Hesper AI
BlogTechnical
TechnicalApril 13, 2026·10 min read·Hesper AI Threat Research

Parallel processing in SIU: how running 15 investigation phases simultaneously changes the claims workflow

Sequential investigation is the hidden reason SIU cases take 14+ days. A technical walk-through of the 15 investigation phases, which can run in parallel, how dependencies are resolved, and why this architecture changes throughput by two orders of magnitude.

15+
Investigation phases per case
Document forensics, OSINT, financial, medical, network
2-4 hrs
End-to-end with full parallelism
Vs. 14+ days sequential
~12
Phases that can run fully in parallel
Bounded only by external API concurrency
~3
Phases with hard dependencies
Analysis, report generation, sign-off

Manual SIU investigation is slow for one reason more than any other: the work is sequential. An investigator pulls the claim file, then reviews prior claims, then queries NICB, then reviews medical records, then runs OSINT, then reviews statements. Each step waits for the previous to finish. Two weeks of calendar time is, in actual work terms, maybe 40-80 hours of focused investigator work stretched across serial handoffs and external dependencies.

Autonomous AI investigation agents do not replace the individual phases. They run them in parallel. A week of sequential work becomes a few hours of parallel work. The compression is architectural, not algorithmic. This piece walks through the 15 investigation phases, which can run in parallel, what dependencies exist, and how the workflow changes at the operations level.

This is a companion to the broader rules-based vs autonomous AI piece. If you want the investigation workflow in its original manual form, start with how insurance companies investigate fraud.

Why sequential investigation is slow

Sequential investigation has three latency sources, all of which compound:

  • Active-work latency - the actual investigator time on each task, typically 30 minutes to 4 hours per task.
  • Wait-state latency - time spent waiting on external parties (database queries with turnaround SLAs, vendor deliverables, statement recordings), typically 24-72 hours per hop.
  • Context-switch latency - the investigator's time cost to pick a case back up after a wait state, typically 15-30 minutes per resumption.

A case that requires six external hops (NICB query, LexisNexis query, medical records request, OSINT scrape, financial records request, witness scheduling) accumulates ~4-5 days of wait-state latency before any investigator work even happens. The investigator's actual focused time on the case might be 30-40 hours; the calendar time stretches to 14+ days.

The 15 investigation phases

A complete autonomous investigation runs 15+ phases per case. The exact phase list varies by claim type (auto, workers' comp, property, liability, medical), but the structure is consistent:

#PhasePurposeTypical data sources
1Claim intake normalisationStandardise claim data structureClaims system APIs (Guidewire, Duck Creek, Majesco)
2Policy and coverage analysisVerify coverage and termsPolicy admin system, product catalogue
3Prior claims historyFlag repeat claimants and patternsISO ClaimSearch, NICB, internal history
4Document forensicsPixel-level analysis of submitted documentsClaim documents, medical records, invoices
5Medical record analysisReview medical records for inconsistenciesMedical records, billing, treatment codes
6Financial analysisLoss calculation, billing patterns, motiveClaim invoices, bank records, tax filings
7Statement cross-referencingRecorded statements and EUOs analysedTranscripts, recordings, written statements
8OSINT and public recordsSocial media, court filings, business recordsPublic web, court databases, business registries
9DMV and government recordsVehicle, license, title verificationState DMV, NMVTIS, government databases
10Provider network analysisCollusion patterns, billing anomaliesProvider billing data, peer benchmarks
11Entity resolutionIdentity verification across sourcesKYC databases, identity graphs
12Timeline reconstructionChronological narrative from all sourcesAll phase outputs with timestamps
13Network graph constructionRelationships between claimants, providers, attorneysAll phase outputs, fraud ring databases
14Findings synthesisCross-reference inconsistencies, scoreAll phase outputs
15Report generationAudit-ready investigation reportAll phase outputs + findings

Phases 1-3 are ingest and context. Phases 4-11 are evidence gathering, each operating on independent data sources and independently parallelisable. Phases 12-15 are synthesis, analysis, and output - these have dependencies on the evidence-gathering phases completing, so they wait on phases 1-11 before starting.

Dependency graph and parallelism

A properly architected autonomous investigation agent resolves phase dependencies automatically. The dependency graph has three tiers:

Phase parallelism breakdown

Ingest tier (phases 1-3) - required first~3 phases
Evidence tier (phases 4-11) - fully parallel~8 phases
Synthesis tier (phases 12-15) - waits on evidence~4 phases

Evidence-tier phases run in parallel bounded only by external API concurrency limits (rate limits from NICB, DMV, medical record providers, etc.). On a typical case, this means ~8 phases completing in 1-2 hours total rather than 1-2 hours each serially (8-16 hours of sequential work).

Synthesis-tier phases have a hard dependency on evidence-tier completion - you cannot synthesise findings until you have the findings. But synthesis is computationally cheap (minutes, not hours) once the evidence is in. The wall-clock time from last evidence phase completing to final report ready is typically 5-15 minutes.

Parallelism is a consumer of external API capacity

Running 8+ phases in parallel against external data sources means running concurrent API calls. Well-designed systems respect rate limits (NICB, LexisNexis, medical record EHRs all have them), use connection pooling, and retry with backoff on failures. This is a standard distributed-systems problem but one that vendors can get wrong. Ask during evaluation how concurrency is managed.

What changes in the workflow

At the SIU operations level, parallel processing changes the workflow in four ways:

1. Investigations complete before the investigator opens the case

The investigator's first view of a case is the completed investigation report with evidence, findings, and recommendation - not an empty case file to start work on. The first action is review, not evidence gathering.

2. Investigator time-per-case drops to review duration

A manual investigator spent 40-80 hours of active work per case across 14+ days. A review-oriented investigator spends 30-60 minutes per case - reading the report, evaluating the findings, making the determination, and signing off. Caseload scales from ~10 to 800+ per month as a direct consequence.

3. Queue dynamics change

Under manual workflows, cases queue because investigators are working on other cases. Under parallel autonomous investigation, there is no 'other cases' blocker - every flagged claim runs through the agent in parallel. The queue shifts from 'waiting for an investigator' to 'waiting for a reviewer', which is a much shorter wait.

4. Coverage rises mechanically

Because the throughput constraint lifts, every flagged claim can receive full investigation. The 75% of flagged claims that currently close without investigation become in-scope. For the economic consequences, see insurance claims leakage: how uninvestigated claims drain profitability.

Error handling and human sign-off

Parallel processing introduces error modes that sequential processing does not have. Three classes matter:

  • Partial-failure errors - one of the 15 phases fails (rate-limited API, data unavailable, format mismatch). The system should surface the missing phase in the report, not silently omit it.
  • Contradictory-finding errors - two phases produce contradictory findings (e.g., OSINT contradicts a statement). These should be flagged, not resolved by the AI - the investigator decides.
  • Confidence-scoring calibration - each finding should have a confidence score; low-confidence findings should be explicitly marked for human review.

The human sign-off step is where these error modes are resolved. An investigator reviewing an AI-generated report is reviewing for: missing phases, contradictory findings, low-confidence items, and context the AI could not have. The sign-off is not a rubber stamp - it is the quality control step that keeps the automated pipeline compliant with the investigator-decides regulatory standard.

Parallel vs sequential benchmarks

MetricSequential manualParallel autonomous
Active work per case40-80 hoursMostly compute-bound
Calendar time per case14-21 days2-4 hours
Concurrent cases / investigator~10 activeBounded by review bandwidth
Cases / investigator / month~10 closed800+ closed
External API calls per caseSequential, retries manualParallel, automatic retries
Wait-state latency~4-5 days / caseMinutes

For the full benchmark set, see benchmarking SIU performance in 2026. For the economics, see from three weeks to four hours.

Key takeaways

  • Manual SIU investigations are slow because the work is sequential - active-work latency, wait-state latency, and context-switch latency compound across the workflow.
  • Autonomous AI investigation agents run 15+ investigation phases per case, with the evidence-gathering tier (~8 phases) fully parallelisable and the synthesis tier (~4 phases) dependent on evidence completion.
  • End-to-end wall clock time compresses from 14+ days to 2-4 hours. The compression is architectural, not algorithmic.
  • Investigator workflow changes: cases arrive complete, review is the first action, time-per-case drops from 40-80 hours to 30-60 minutes, caseload rises from ~10 to 800+ per month.
  • Error modes (partial failure, contradictory findings, low-confidence items) are surfaced for human sign-off, which preserves the regulator-required investigator-decides standard.

Frequently asked questions

Parallel processing means that instead of gathering evidence sequentially (pull claim file, then query NICB, then review medical records, then run OSINT), an autonomous AI investigation agent executes all independent phases simultaneously. For a typical case, this means ~8 evidence-gathering phases (document forensics, medical record analysis, OSINT, database queries, financial analysis, statement cross-referencing, entity resolution, provider network analysis) running concurrently against their respective data sources. The phases complete in whatever time the slowest external API response takes, not in the sum of all phase times. A week of sequential work becomes a few hours of parallel work.

Three tiers: (1) ingest tier (claim intake normalisation, policy analysis, prior claims history) - must run first; (2) evidence tier (document forensics, medical record analysis, financial analysis, statement cross-referencing, OSINT, DMV records, provider network analysis, entity resolution) - fully parallel; (3) synthesis tier (timeline reconstruction, network graph construction, findings synthesis, report generation) - dependent on evidence tier completion. The evidence tier is the largest (~8 phases) and most parallelisable. The synthesis tier is computationally cheap and completes in minutes once evidence is in.

Two ways: first, investigator time-per-case drops from 40-80 hours of active work to 30-60 minutes of review. Second, cases arrive at the investigator already fully investigated - the investigator's first action is reviewing the AI-generated report, not gathering evidence from scratch. Caseload closed per investigator per month scales from ~10 (manual benchmark) to 800+ (AI-augmented benchmark). Investigation coverage of flagged claims rises from 25% to 100% because the throughput constraint lifts.

Well-architected autonomous investigation systems handle partial failures gracefully. If one phase fails (rate-limited external API, data unavailable, format mismatch), the system logs the failure, retries with exponential backoff, and if the failure persists, surfaces it in the investigation report as 'phase incomplete' rather than silently omitting. The investigator reviewing the report sees the incomplete phase and can decide whether to proceed with available evidence, request the phase be re-run, or escalate to a manual investigation. This is different from manual workflows where a missing evidence source might never be revisited.

Only if the system is poorly designed. External data sources (NICB, LexisNexis, state DMVs, medical record EHR systems) all have rate limits and concurrent-connection caps. Properly architected autonomous investigation systems respect these limits through connection pooling, request throttling, and prioritisation queues. Parallelism operates across independent sources simultaneously rather than overloading any single source. Vendors should document their concurrency strategy during evaluation - how they handle rate limits, backoff on failures, and prioritisation when multiple investigations are running at once.

← More articles on the Hesper AI blog

See Hesper AI on your documents

Request a demo and we'll run an analysis on your real document samples.