Stay free if you only need up to 500 pages per month and access to read, layout, prebuilt, and custom models. Upgrade if you need run read, layout, invoice, receipt, id models on-prem and air-gapped/disconnected operation supported. Most solo builders can start free.
Why it matters: Pricing complexity: per-page costs vary by model type (Read $1.50/1K, Prebuilt $10/1K, Custom $50/1K for first 1M pages) and add-ons charge extra
Available from: Standard (S0) Pay-as-you-go
Why it matters: Steeper learning curve than turnkey SaaS â requires Azure subscription, resource provisioning, and key management
Available from: Standard (S0) Pay-as-you-go
Why it matters: No built-in workflow, approval, or human-in-the-loop UI; you must build review interfaces yourself or use Power Platform
Available from: Standard (S0) Pay-as-you-go
Why it matters: Custom model accuracy on highly variable layouts can require iterative retraining and careful sample curation
Available from: Standard (S0) Pay-as-you-go
Why it matters: Documentation sprawl across multiple API versions (v2.1, v3.0, v3.1, v4.0) can confuse new users choosing where to start
Available from: Standard (S0) Pay-as-you-go
Why it matters: Advanced feature not available in free plan.
Available from: Standard (S0) Pay-as-you-go
They are the same service â Microsoft renamed Azure Form Recognizer to Azure AI Document Intelligence in July 2023 to better reflect its expanded capabilities beyond just forms. All existing Form Recognizer APIs, SDKs, and resources continue to work without modification. The rebrand coincided with the v3.1 release and the rollout of new prebuilt models for contracts, US tax documents, and health insurance cards. The current generally available version is v4.0, released in 2024.
There are two tiers: a free F0 tier that processes up to 500 pages per month at no cost, and a pay-as-you-go S0 tier billed per 1,000 pages. As of 2026, the Read OCR model costs $1.50 per 1,000 pages, prebuilt models (invoice, receipt, ID, etc.) cost $10 per 1,000 pages, the Layout model costs $10 per 1,000 pages, and custom models cost $50 per 1,000 pages for the first 1M pages with volume discounts beyond. Add-on capabilities like high-resolution mode and query fields incur additional charges.
Document Intelligence requires a minimum of five sample documents of the same type to train a custom model, which is one of the lowest sample requirements in the document AI category. For best accuracy with variable layouts, Microsoft recommends 50+ samples covering the variations you expect to see in production. The Document Intelligence Studio provides a no-code labeling interface where you draw bounding boxes and assign field names, then trigger training directly from the browser. You can iterate by adding more samples and retraining without losing previous configuration.
Yes, Microsoft offers Document Intelligence as Docker containers that can run on-premises, in your own Azure VNet, or at the edge. Container support covers the Read OCR, Layout, and several prebuilt models including Invoice, Receipt, and ID Document. This makes it suitable for healthcare, government, and financial services scenarios where documents cannot leave a controlled network boundary. Container usage still requires an Azure resource for billing and metering, but document content never leaves your environment.
All three hyperscaler services offer similar core capabilities â OCR, layout extraction, and prebuilt models for common documents â but differ on specifics. Azure has the broadest language coverage (309+ for Read), the lowest custom training sample requirement (5 documents), and the strongest integration with Microsoft 365 and Azure OpenAI for RAG. AWS Textract excels in raw OCR speed and AWS ecosystem integration, while Google Document AI offers strong specialized parsers (lending, procurement) and advanced entity extraction. Choose Azure if you're already on Microsoft's stack or need on-premises container deployment.
Start with the free plan â upgrade when you need more.
Get Started Free âStill not sure? Read our full verdict â
Last verified March 2026