Stay free if you only need 10 million free tokens per new api key and access to all apis (embeddings, reranker, reader, search). Upgrade if you need volume discounts on token purchases and dedicated support and slas. Most solo builders can start free.
Why it matters: Token-based pricing can be difficult to predict for variable workloads — costs can spike unexpectedly with high-volume embedding or reading tasks without careful monitoring
Available from: Pay-as-you-go
Why it matters: Reader API struggles with heavily JavaScript-dependent single-page applications and sites behind aggressive anti-bot measures, limiting coverage of modern web apps
Available from: Pay-as-you-go
Why it matters: Documentation is fragmented across multiple product pages without a unified developer portal or comprehensive getting-started guide for the full platform
Available from: Pay-as-you-go
Why it matters: Self-hosted models require significant GPU resources (jina-embeddings-v4 is 3.8B parameters) for production throughput, making local deployment expensive for smaller teams
Available from: Pay-as-you-go
Why it matters: No built-in vector database — Jina provides excellent embeddings and reranking but teams need external storage solutions (Pinecone, Weaviate, Qdrant) for complete search systems
Available from: Pay-as-you-go
Simply prepend r.jina.ai/ to any URL. For example, to read https://example.com/article, visit https://r.jina.ai/https://example.com/article. You can also pass an API key header for higher rate limits and additional features. The response is clean markdown suitable for LLM context windows.
Jina-embeddings-v4 is a 3.8B parameter multimodal model that handles both text and images in the same embedding space, which OpenAI's text-embedding models cannot do natively. It supports 89+ languages with multi-vector (late interaction) outputs for higher precision. On multilingual benchmarks, Jina consistently outperforms OpenAI's offerings.
Yes. Jina publishes models on Hugging Face (jinaai/jina-embeddings-v4, jinaai/jina-reranker-v3) for local deployment. This enables air-gapped environments, data sovereignty compliance, and latency optimization. You'll need GPU infrastructure for production throughput given the 3.8B parameter model size.
DeepSearch is an agentic research tool that iteratively searches the web, reads pages, and reasons about findings until reaching comprehensive answers. Unlike regular search that returns ranked results, DeepSearch autonomously investigates complex questions. It's API-compatible with OpenAI's Chat schema for easy integration.
One API key works for all Jina services — embeddings, reranking, reader, search, and DeepSearch. The token pool is shared across all services, so you manage one balance rather than separate quotas. New accounts get 10M free tokens that work across the entire platform.
Yes. Jina AI is SOC 2 Type I and Type II compliant with the AICPA. They never use customer API requests or data for model training — your data remains strictly yours. This meets enterprise requirements for data privacy and security in regulated industries.
Start with the free plan — upgrade when you need more.
Get Started Free →Still not sure? Read our full verdict →
Last verified March 2026