Benchmark Methodology

Document Set

We test every platform on a standardized set of 1,000+ documents sourced from real businesses. The set includes invoices (40%), bank statements (15%), tax forms (10%), receipts (10%), contracts (10%), shipping documents (10%), and identity documents (5%).

Documents span three quality levels: digital-native PDFs (40%), high-quality scans (35%), and degraded/low-quality scans (25%). Multi-language documents represent 20% of the set.

Scoring Dimensions

  • Extraction Accuracy (30%) — Field-level accuracy on our standard document set. Blind-tested with ground truth comparison.
  • Ease of Use (20%) — Time to first extraction, learning curve, and whether non-technical users can operate it.
  • Value for Money (15%) — Total cost of ownership at three volume tiers (100, 1,000, 10,000 pages/month), including engineering costs.
  • Integrations (15%) — API quality, pre-built connectors, and ecosystem breadth.
  • Scalability (10%) — Throughput, latency under load, and infrastructure overhead.
  • Support & Docs (10%) — Documentation quality, support responsiveness, and community resources.

Testing Process

Each platform is tested by at least two team members. Accuracy testing is fully blind — the evaluator doesn't know which platform produced which output during scoring.

We re-run benchmarks quarterly to account for model updates and new feature releases. Scores are updated within 30 days of any major platform change.

Affiliate Disclosure

Some links on this site are affiliate links. This means we may earn a small commission if you purchase through our links. This never influences our rankings or recommendations — scores are derived from blind benchmarks, not business relationships.