Curious about what keeps experts, CEOs and other decision-makers in the Intelligent Document Processing (IDP) space on their toes? Get food for thought on IDP-related topics from the industry’s leading minds.
In this opinion piece, Dr. Tobias Grüning, Chief Research Officer at Intelligent Document Processing (IDP) vendor PLANET AI, delves into the topic of the growing IDP market, which is attracting opportunistic players (“agent washing”), and the questions technical decision-makers should ask when evaluating IDP solutions.
Intro
A government agency evaluates three AI solutions for its document workflows. All three promise “AI-powered document processing” and “Agentic AI.” During the technical deep dive, the reality is more sobering: each solution is built on the same basic pattern — a prompt template wrapped around a GPT API call, with no feedback loops, no validation, and no ability to learn.
This scenario is more common than most vendors would like to admit. Gartner estimates that of the thousands of vendors claiming “Agentic AI” capabilities, only around 130 actually deliver genuine agent-based functionality. Meanwhile, simple LLM wrappers achieve just 66–77 % accuracy on document data extraction tasks, compared to 93–98 % for specialized IDP systems.
The ability to tell real agentic architecture from API cosmetics is fast becoming a critical competency for technical decision-makers in the IDP space. The difference is not a nuance; it determines whether a solution is production-ready.
Wrapper vs workflow: what’s under the hood
LLM wrappers follow one pattern: input, prompt template, API call, output. There is no planning, no tool use, no memory, and no self-correction. The intelligence lives in the external model, not in the application itself.
Agentic workflows are fundamentally different. Andrew Ng’s widely cited framework identifies four design patterns that define genuine agent-based systems. Reflection means the system critiques and iteratively improves its own output. Tool Use enables access to external resources such as databases, APIs, or code execution environments. Planning breaks complex tasks into sub-steps with dynamic adjustment. Multi-Agent Collaboration coordinates specialized agents working in parallel on different parts of a problem.
Ng’s benchmarks make the performance gap concrete. A smaller, lower-cost model operating in zero-shot mode achieved around 48 % accuracy on the HumanEval coding benchmark. The most capable model available at the time reached 67 %. But the smaller model, running inside an agentic workflow, hit 95.1 %, substantially outperforming the larger model. This finding has since been replicated across numerous benchmarks. Architecture beats model size.
Three failure modes that matter in production
These differences are not academic. In practice, LLM wrapper architectures introduce significant failure modes that compound at scale. The first is hallucination without a safety net. LLMs generate plausible-but-wrong data in 5 to 20 % of complex extraction cases. A 2024 Arxiv study demonstrated mathematically that hallucination cannot be eliminated due to fundamental computational constraints. It can only be intercepted through external validation layers. LLM wrappers have none.
The second is inconsistent output structure. LLMs are non-deterministic. The same document processed twice may return a date as “25 Dec 2024” in one run and “2024-12-25” in the next. Audits of current top models show measurable “data drift” that accumulates into significant errors when processing thousands of documents with dozens of fields. Many AI integration failures trace directly back to this inconsistency.
The third is data sovereignty. Cloud-based LLM APIs route document content to third-party servers. Every prompt can contain personal data — customer names, account details, medical records. For regulated industries in Europe, dependence on US cloud infrastructure is a strategic exposure, not just a compliance checkbox.
“Agent washing” and how to spot it
The IDP market is growing at over 25 % annually and reached approximately $2.3 billion in 2024. This growth is driving a broader shift: from passive data extraction toward proactive document-to-decision automation, enabled by agentic OCR with vision-language models and LLM-powered reasoning pipelines.
Gartner predicts that by 2027, enterprises will deploy small, task-specific AI models three times more frequently than general-purpose LLMs. A clear signal that specialization outperforms the “one model for everything” approach.
But rapid market growth also attracts opportunism. Gartner has flagged the rise of “agent washing”, the rebranding of existing chatbots, RPA tools, and AI assistants as “Agentic AI” without meaningful new capabilities underneath. The forecast is direct: more than 40 % of Agentic AI projects will be abandoned by the end of 2027, as costs escalate and business value fails to materialize.
Technical decision-makers evaluating IDP solutions should ask questions that go beyond marketing claims:
- Can the solution run on-premises, or does it depend on US cloud APIs? Is the architecture model-agnostic?
- How does the system prevent hallucinations? Are there validation layers — business rules, schema enforcement, cross-reference checks — built into the pipeline?
- Are results traceable through source references, confidence scores, and audit trails?
- Does the vendor use proprietary core technology, or are they primarily integrating a third-party API?
Beyond architecture, evaluators should request evidence of autonomous task completion without continuous human oversight, demonstrable reasoning and planning capabilities beyond text generation, and ROI metrics tied to specific business outcomes rather than benchmark scores.
Decision-makers who ask these questions consistently will quickly identify which vendors bring genuine AI substance and which are presenting a new frontend over someone else’s API dashboard.
Conclusion
The availability of powerful foundation models has lowered the barrier to entry for IDP solutions. But that accessibility also creates an illusion of maturity. A polished UI and an API integration can look convincing in a demo. They rarely survive contact with production volumes.
The vendors that will define the IDP market in the next few years share a common architecture: proprietary recognition models as a high-quality data foundation, specialized agents for different processing stages, full data sovereignty for European compliance requirements, and continuous research-driven development rather than pure API integration. The key question for decision-makers is no longer “which LLM is running in the backend.” It is “what intelligence has been built around it.”

About the Author
Dr. Tobias Grüning is Chief Research Officer at PLANET AI. A mathematician by training, Tobias earned his PhD in AI-based handwriting recognition for historical documents. He has been leading the research department at PLANET AI since 2018. The team has been dedicated to AI-driven document processing from the very beginning, and in recent years has increasingly focused on leveraging LLM-based technologies for document analysis.
For more news from PLANET AI, click here.
You may also like:
The AI Special Assistant: Utilizing the Right OCR as a Data Foundation

📨Get IDP industry news, distilled into 5 minutes or less, once a week. Delivered straight to your inbox ↓