Complete pricing guide for Anthropic Claude on AWS Bedrock. Compare all plans, analyze costs, and find the perfect tier for your needs.
Not sure if free is enough? See our Free vs Paid comparison →
Still deciding? Read our full verdict on whether Anthropic Claude on AWS Bedrock is worth it →
mo
mo
mo
Pricing sourced from Anthropic Claude on AWS Bedrock · Last verified March 2026
Functionally you get the same underlying Claude models, but on Bedrock the traffic is authenticated via AWS IAM and SigV4, data stays within your VPC, and billing consolidates onto your AWS invoice. Bedrock also adds managed services (Knowledge Bases, Agents, Guardrails) that have no direct equivalent on Anthropic's API.
Yes. Amazon Bedrock is a HIPAA-eligible service under an executed BAA with AWS and carries SOC 1/2/3, ISO 27001, and PCI DSS compliance. In GovCloud regions it also holds FedRAMP High authorization.
Bedrock bills Claude on a per-model, pay-as-you-go basis using separate input-token and output-token rates. For example, Haiku 3.5 starts at $0.80 per 1M input tokens, Sonnet 4 at $3.00, and Opus 4 at $15.00. Output tokens are billed at higher rates. Provisioned Throughput plans offer committed capacity at hourly rates, and enterprise agreements can negotiate further discounts.
The Claude model families — including Opus 4, Sonnet 4, and Haiku 3.5 tiers — are available on Bedrock, though availability may vary by region. Check the Bedrock console for current model availability in your target region.
Yes. You can create VPC interface endpoints for Bedrock using AWS PrivateLink so that all InvokeModel traffic stays within your private network and never traverses the public internet.
AI builders and operators use Anthropic Claude on AWS Bedrock to streamline their workflow.
Try Anthropic Claude on AWS Bedrock Now →Google Cloud's unified platform for machine learning and generative AI, offering 180+ foundation models, custom training, and enterprise MLOps tools.
Compare Pricing →Cloud platform for running open-source AI models with serverless inference, fine-tuning, and dedicated GPU infrastructure optimized for production workloads.
Compare Pricing →