Agentic Document Extraction: What the Agents of Chaos Paper Gets Wrong About AI Parsing
February 28, 2026
The Headlines Are Dramatic. The Reality Is More Nuanced.
A new paper from Northeastern, Harvard, MIT, and Stanford just dropped. It is called Agents of Chaos and the headlines are dramatic: destroyed servers, DoS attacks, catastrophic failures when AI agents interact.
The research tested OpenClaw agent-to-agent interactions and found that routine glitches can cascade into system-wide failures. ZDNET ran it as breaking news. IEEE Spectrum called it a messy future. WIRED included it in their security roundup.
Here is what they tested: autonomous agents performing tasks like managing emails, scheduling, and file operations. When multiple agents interacted without human oversight, individual errors compounded into what the researchers called qualitatively new failure modes.
Fair enough. Multi-agent orchestration without guardrails is risky. Nobody serious disputes that.
But here is what the paper does not address: agentic document extraction, done correctly, is not multi-agent chaos. It is a single agent with a well-defined task, clear input boundaries, and deterministic validation.
What Agentic Document Extraction Actually Is
Traditional OCR reads text from an image. It outputs raw characters. If the scan is crooked, the text is garbled. If a table has merged cells, the structure breaks. If handwriting varies, accuracy drops.
Agentic extraction is different. Instead of a dumb read-and-dump pipeline, an agentic system:
- Analyzes the document type before attempting extraction
- Selects the appropriate extraction strategy based on document structure
- Validates extracted fields against expected patterns
- Self-corrects when confidence scores are low
- Routes uncertain results for human review instead of guessing
Parseur published a comprehensive guide to agentic extraction this week. LandingAI just launched an entire platform around it. The market is moving fast because the technology actually works when properly constrained.
Why Document Extraction Does Not Cascade Like the Paper Describes
The Agents of Chaos failures happened because agents were modifying shared resources without coordination. One agent deleted a file another agent needed. Agents entered feedback loops. Error handling created new errors.
Document extraction does not have this problem because:
- The input is a static file (PDF, image, scan). The agent reads it. It does not modify it.
- The output is structured data (JSON, CSV). No shared state with other agents.
- Validation is mathematical. Does the invoice total match the line items? Do the dates parse correctly? Is the SSN format valid?
- The failure mode is graceful: low confidence score triggers human review, not cascading system failure.
This is fundamentally different from agents managing each other's email clients or negotiating shared calendar access.
The Real Risk in Document Parsing
The actual risk in production document extraction is not cascading agent failure. It is silent accuracy degradation.
A parser that returns 95% accuracy on invoices sounds good. But if 5% of invoices have wrong totals and nobody catches them, that is real money lost. The fix is not removing agents from the pipeline. It is building verification into every extraction.
Here is what a production pipeline looks like:
- Document ingestion with type classification (is this a W-2, invoice, bank statement, or contract?)
- OCR pass with Mistral OCR or similar high-accuracy engine (costs about $1-2 per 1,000 pages)
- LLM extraction that maps raw text to structured fields using a predefined schema
- Confidence scoring on every extracted field
- Automated validation (math checks, format checks, cross-reference checks)
- Human review queue for anything below threshold
What This Means for Businesses Using Document Parsing
If you are a CPA processing W-2s during tax season, the Agents of Chaos paper is irrelevant to your use case. You are not running multi-agent swarms. You are uploading PDFs and getting structured data back.
The tools that work for this today:
- Dext (formerly Receipt Bank): Good for receipts and invoices. $30-50/month. Limited document types.
- Hubdoc: Decent for bill capture. Free with Xero subscription. Narrow use case.
- Docparser: Template-based extraction. Works well for consistent document formats. $39-299/month.
- Dokyumi: Schema-defined extraction using Mistral OCR plus Claude. Works with any document type. $79/month starting. Full API access.
- Google Document AI: Enterprise pricing. Good accuracy but complex setup.
The cost math is simple. A bookkeeper spending 15-20 minutes per document on manual data entry, processing 200 documents per month, is spending 50-67 hours on extraction. At $25/hour, that is $1,250-1,675/month in labor. Any tool under $200/month pays for itself immediately.
The Takeaway
Multi-agent orchestration has real risks. The Agents of Chaos paper documents them well. But document extraction is a solved problem when the pipeline is properly constrained. The agents parsing your invoices are not the agents destroying servers. Different architecture, different risk profile, different conversation entirely.
Tax season is here. The documents are piling up. The parsing tools work. Use them.
More from Dokyumi
Start extracting in under 2 minutes
100 free extractions every month. No credit card required.