For AI agents: The documentation index is at https://docs.digitalocean.com/llms.txt. Markdown versions of pages use the same URL with index.html.md in place of the HTML page (for example, append index.html.md to the directory path instead of opening the HTML document).
Inference has a usage-based pricing model, so costs scale with your actual usage.
Bring Your Own Models (BYOM)
BYOM model weights are stored in a service-managed, non-accessible Spaces location, and are billed at $5.00 per month. We do not charge you for browsing or managing imported models in Model Catalog. Costs apply only for storing model weights and for using those models with other paid features, such as dedicated inference deployments.
Model Playground
Usage is charged at the same rate as serverless inference.
Serverless Inference
Serverless inference is billed by DigitalOcean for both open-source and commercial models. Prices align with each provider’s published rates for transparency.
Based on your tier, you have an allocated amount of usage before we charge you. For example, $25 for tier 1. Once you’ve hit that limit, we charge you for that usage. Additional inference usage is capped until you pay that bill.
The following shows pricing for foundation models available through serverless inference.
Anthropic Models
When using Anthropic commercial models with your own model API keys, billing is handled directly by Anthropic at the provider’s rates.
Claude Sonnet 4.6, Sonnet 4.5, and Sonnet 4 support an input context window of up to 1M tokens.
| Model |
Serverless Inference |
| Claude Sonnet 4.6 |
Prompts ≤200K tokens$3.00 per 1M input tokens $15.00 per 1M output tokens Prompts >200K tokens$6.00 per 1M input tokens $22.50 per 1M output tokens Prompt caching$3.75 per 1M cache creation 5m input tokens $6.00 per 1M cache creation 1h input tokens $0.30 per 1M cache read input tokens
|
| Claude Sonnet 4.5 |
Prompts ≤200K tokens$3.00 per 1M input tokens $15.00 per 1M output tokens Prompts >200K tokens$6.00 per 1M input tokens $22.50 per 1M output tokens Prompt caching$3.75 per 1M cache creation 5m input tokens $6.00 per 1M cache creation 1h input tokens $0.30 per 1M cache read input tokens
|
| Claude Sonnet 4 |
Prompts ≤200K tokens$3.00 per 1M input tokens $15.00 per 1M output tokens Prompts >200K tokens$6.00 per 1M input tokens $22.50 per 1M output tokens Prompt caching$3.75 per 1M cache creation 5m input tokens $6.00 per 1M cache creation 1h input tokens $0.30 per 1M cache read input tokens
|
| Claude Haiku 4.5 |
Input/output tokens$1.00 per 1M input tokens $5.00 per 1M output tokens Prompt caching$1.25 per 1M cache creation 5m input tokens $2.00 per 1M cache creation 1h input tokens $1.00 per 1M cache read input tokens
|
| Claude Opus 4.7 |
Input/output tokens$5.00 per 1M input tokens $25.00 per 1M output tokens Prompt caching$6.25 per 1M cache creation 5m input tokens $10.00 per 1M cache creation 1h input tokens $0.50 per 1M cache read input tokens
|
| Claude Opus 4.6 |
Prompts ≤200K tokens$5.00 per 1M input tokens $25.00 per 1M output tokens Prompts >200K tokens$10.00 per 1M input tokens $37.50 per 1M output tokens Prompt caching$6.25 per 1M cache creation 5m input tokens $10.00 per 1M cache creation 1h input tokens $0.50 per 1M cache read input tokens
|
| Claude Opus 4.5 |
Input/output tokens$5.00 per 1M input tokens $25.00 per 1M output tokens Prompt caching$6.25 per 1M cache creation 5m input tokens $10.00 per 1M cache creation 1h input tokens $0.50 per 1M cache read input tokens
|
| Claude Opus 4.1 |
Input/output tokens$15.00 per 1M input tokens $75.00 per 1M output tokens Prompt caching$18.75 per 1M cache creation 5m input tokens $30.00 per 1M cache creation 1h input tokens $1.50 per 1M cache read input tokens
|
| Claude Opus 4 |
Input/output tokens$15.00 per 1M input tokens $75.00 per 1M output tokens Prompt caching$18.75 per 1M cache creation 5m input tokens $30.00 per 1M cache creation 1h input tokens $1.50 per 1M cache read input tokens
|
Arcee Models
| Model |
Serverless Inference |
| Trinity Large |
Input/output tokens$0.25 per 1M input tokens $0.90 per 1M output tokens Prompt caching$0.06 per 1M cache read input tokens
|
fal Models
| Model |
Serverless Inference |
| Fast SDXL |
$0.0011 per compute second |
| Flux Schnell |
$0.0030 per megapixel |
| Stable Audio 2.5 (Text-to-Audio) |
$0.00058 per compute second |
| Multilingual TTS v2 |
$0.10 per 1000 characters |
OpenAI Models
When using OpenAI commercial models with your own model API keys, billing is handled directly by OpenAI at the provider’s rates.
| Model |
Serverless Inference |
| gpt-oss-120b |
Input/output tokens$0.10 per 1M input tokens $0.70 per 1M output tokens
|
| gpt-oss-20b |
Input/output tokens$0.05 per 1M input tokens $0.45 per 1M output tokens
|
| GPT-5.4 |
Input/output tokens$2.50 per 1M input tokens $15.00 per 1M output tokens Prompt caching$0.25 per 1M cache read input tokens
|
| GPT-5.4 mini |
Input/output tokens$0.75 per 1M input tokens $4.50 per 1M output tokens Prompt caching$0.075 per 1M cache read input tokens
|
| GPT-5.4 nano |
Input/output tokens$0.20 per 1M input tokens $1.25 per 1M output tokens Prompt caching$0.02 per 1M cache read input tokens
|
| GPT-5.4 pro |
Input/output tokens$30.00 per 1M input tokens $180.00 per 1M output tokens
|
| GPT-5.3-Codex |
Input/output tokens$1.75 per 1M input tokens $14.00 per 1M output tokens Prompt caching$0.175 per 1M cache read input tokens
|
| GPT-5.2 |
Input/output tokens$1.75 per 1M input tokens $14.00 per 1M output tokens Prompt caching$0.175 per 1M cache read input tokens
|
| GPT-5.2 pro |
Input/output tokens$21.00 per 1M input tokens $168.00 per 1M output tokens
|
| GPT-5.1-Codex-Max |
Input/output tokens$1.25 per 1M input tokens $10.00 per 1M output tokens Prompt caching$0.125 per 1M cache read input tokens
|
| GPT-5 |
Input/output tokens$1.25 per 1M input tokens $10.00 per 1M output tokens Prompt caching$0.125 per 1M cache read input tokens
|
| GPT-5 mini |
Input/output tokens$0.25 per 1M input tokens $2.00 per 1M output tokens Prompt caching$0.025 per 1M cache read input tokens
|
| GPT-5 nano |
Input/output tokens$0.05 per 1M input tokens $0.40 per 1M output tokens Prompt caching$0.005 per 1M cache read input tokens
|
| GPT-4.1 |
Input/output tokens$2.00 per 1M input tokens $8.00 per 1M output tokens Prompt caching$0.50 per 1M cache read input tokens
|
| GPT-4o |
Input/output tokens$2.50 per 1M input tokens $10.00 per 1M output tokens Prompt caching$1.25 per 1M cache read input tokens
|
| GPT-4o mini |
Input/output tokens$0.15 per 1M input tokens $0.60 per 1M output tokens Prompt caching$0.075 per 1M cache read input tokens
|
| o1 |
Input/output tokens$15.00 per 1M input tokens $60.00 per 1M output tokens Prompt caching$7.50 per 1M cache read input tokens
|
| o3 |
Input/output tokens$2.00 per 1M input tokens $8.00 per 1M output tokens Prompt caching$0.50 per 1M cache read input tokens
|
| o3-mini |
Input/output tokens$1.10 per 1M input tokens $4.40 per 1M output tokens Prompt caching$0.55 per 1M cache read input tokens
|
| GPT-image-1 |
Input/output tokens$5.00 per 1M input tokens $40.00 per 1M output tokens Prompt caching$1.25 per 1M cache read input tokens
|
| GPT Image 1.5 |
Input/output tokens$5.00 per 1M input tokens $10.00 per 1M output tokens Prompt caching$1.00 per 1M cache read input tokens
|
| GPT Image 2 |
Text input$5.00 per 1M tokens Text output$0.00 per 1M tokens Text cache read$1.25 per 1M tokens Image input$8.00 per 1M tokens Image output$30.00 per 1M tokens Image cache read$2.00 per 1M tokens
|
DigitalOcean-Hosted Models
The following models are discounted 30% during off-peak hours, 05:00 to 11:00 UTC each day:
See the (off-peak) rows in the following table for off-peak rates.
| Provider |
Model |
Serverless Inference |
| Alibaba |
Qwen3-32B |
Input/output tokens$0.25 per 1M input tokens $0.55 per 1M output tokens
|
| Alibaba |
Qwen3 Coder Flash |
Input/output tokens$0.45 per 1M input tokens $1.70 per 1M output tokens
|
| Alibaba |
Qwen 3.5 397B A17B |
Input/output tokens$0.55 per 1M input tokens $3.50 per 1M output tokens
|
| Alibaba |
Qwen 3 TTS (1.7B) |
$20.00 per 1M character tokens |
| Alibaba |
Wan2.2-T2V-A14B |
$0.60 per 1M video tokens |
| DeepSeek |
DeepSeek R1 Distill Llama 70B |
Input/output tokens$0.99 per 1M input tokens $0.99 per 1M output tokens
|
| DeepSeek |
DeepSeek V4 Pro |
Input/output tokens$1.74 per 1M input tokens $3.48 per 1M output tokens
|
| DeepSeek |
DeepSeek V3.2 |
Input/output tokens$0.50 per 1M input tokens $1.60 per 1M output tokens
|
| Google |
Gemma 4 |
Input/output tokens$0.18 per 1M input tokens $0.50 per 1M output tokens
|
| Intfloat |
E5 Mistral 7B Instruct |
Input/output tokens$0.09 per 1M input tokens $0.00 per 1M output tokens
|
| MiniMax |
MiniMax M2.5 (Public Preview) |
Input/output tokens$0.30 per 1M input tokens $1.20 per 1M output tokens Input/output tokens (off-peak)$0.21 per 1M input tokens $0.84 per 1M output tokens
|
| Moonshot AI |
Kimi K2.5 |
Input/output tokens$0.50 per 1M input tokens $2.70 per 1M output tokens Input/output tokens (off-peak)$0.35 per 1M input tokens $1.89 per 1M output tokens
|
| Moonshot AI |
Kimi K2.6 |
Input/output tokens$0.95 per 1M input tokens $4.00 per 1M output tokens
|
| Meta |
Llama 3.3 Instruct-70B |
Input/output tokens$0.65 per 1M input tokens $0.65 per 1M output tokens
|
| Meta |
Llama 4 Maverick 17B 128E Instruct |
Input/output tokens$0.25 per 1M input tokens $0.87 per 1M output tokens
|
| Mistral AI |
Ministral 3 14B Instruct |
Input/output tokens$0.20 per 1M input tokens $0.20 per 1M output tokens
|
| NVIDIA |
Nemotron-3-Super-120B (Public Preview) |
Input/output tokens$0.30 per 1M input tokens $0.65 per 1M output tokens
|
| NVIDIA |
Nemotron Nano 3 Omni |
Input/output tokens$0.50 per 1M input tokens $0.90 per 1M output tokens
|
| NVIDIA |
Nemotron Nano 12B v2 VL |
Input/output tokens$0.20 per 1M input tokens $0.60 per 1M output tokens
|
| Stability AI |
Stable Diffusion 3.5 Large |
$0.08 per 1M image tokens |
| Z.ai |
GLM 5 |
Input/output tokens$1.00 per 1M input tokens $3.20 per 1M output tokens
|
Web Search Requests
You are charged $10 per 1000 requests for using web search with serverless inference.
Dedicated Inference
Dedicated Inference is billed per GPU-hour based on the GPU you use.
| GPU |
Price |
| AMD MI300X |
$2.59 per hour |
| AMD MI300X (8x) |
$20.70 per hour |
| AMD MI325X |
$2.98 per hour |
| AMD MI325X (8x) |
$23.82 per hour |
| AMD MI350X |
$6.89 per hour |
| NVIDIA B300 |
$10.39 per hour |
| NVIDIA B300 (8x) |
$83.10 per hour |
| NVIDIA H100 |
$4.41 per hour |
| NVIDIA H100 (8x) |
$30.32 per hour |
| NVIDIA H200 |
$4.47 per hour |
| NVIDIA H200 (8x) |
$35.78 per hour |
|
|
Batch Inference
Batch inference is charged at up to a 50% discount on OpenAI and Anthropic models.
You are only charged for completed requests. If a batch job fails, is blocked by guardrails, or expires partway through, requests that were not processed are not charged.
Model Evaluations
Model evaluations for candidate models deployed on Serverless Inference, and for judge models, are charged at the same token rates as serverless inference.
Inference Router
Inference Router is available in public preview and enabled for all users. You can contact support for questions or assistance.
There is no additional cost to using Inference Router during public preview. Using inference routing forwards requests to foundation models for serverless inference and dedicated inference. You are billed for the models that serve each request.
Knowledge Bases
Knowledge base pricing is shown per million tokens, but billing is calculated per thousand tokens.
You’re billed for both indexing and storage:
-
Tokens used for indexing and retrieval query vectorization: We charge for tokens used to generate embeddings during indexing and to vectorize user queries during retrieval. Both use the same embeddings model pricing.
Indexing pricing is the same for manual and auto-indexing. Indexing charges apply only when changes are detected, such as new, updated, or deleted files or URLs. If auto-indexing is paused or no changes are found, there are no indexing charges.
Retrieval requests sent through a MCP server are billed the same as retrieval requests sent directly to the knowledge base retrieve endpoint. This includes the tokens used to vectorize the retrieval query with the selected embeddings model.
For example, a 10 MB dataset is about 3 million tokens, and a 1 GB dataset is about 250 million tokens.
Actual costs depend on the embeddings model:
| Model |
Price |
all-mini-lm-l6-v2 |
$0.009 per 1M input tokens |
multi-qa-mpnet-base-dot-v1 |
$0.009 per 1M input tokens |
gte-large-en-v1.5 |
$0.09 per 1M input tokens |
Qwen3 Embedding 0.6B |
$0.04 per 1,000,000 tokens |
BGE-M3 |
$0.02 per 1,000,000 tokens |
E5 Large V2 |
$0.02 per 1,000,000 tokens |
One token is roughly four characters (approximately 75 words per 100 tokens). Non-Latin scripts, emojis, or binary data may increase token counts.
-
Reranking tokens: If reranking is enabled, tokens used to rerank results are billed based on the selected reranking model. For supported reranking models, see available reranking models.
| Model |
Price |
BGE Reranker v2 m3 |
$0.01 per 1M reranking tokens |
-
Storage: Embeddings are stored in OpenSearch. See OpenSearch pricing.
Chunking has no separate charge. Chunking costs depend on embedding token usage, OpenSearch database, and the selected embeddings model.
Chunking strategy cost depends on how many tokens the strategy embeds and returns:
- Section-based and fixed length chunking are the most cost-efficient because they use simple splitting and have predictable token usage.
- Semantic chunking costs more because it uses the embeddings model to detect semantic boundaries and embed final chunks, often resulting in 1.5 to 3 times more indexing tokens.
- Hierarchical chunking slightly increases indexing cost by creating parent and child embeddings. It can also increase retrieval cost because agents receive both child and parent chunks for each lookup.
Changing your chunking strategy or configuration requires re-indexing the affected data source, which consumes additional tokens. For guidance on chunking configurations and best practices, see our chunking parameters reference and chunking best practices.
If you use RAG Playground, answer generation is billed separately based on the selected serverless inference model. Free tokens for RAG Playground are not separate; they are shared with Model Playground.
Agent creation is free. We charge for model usage and for additional features like knowledge bases, guardrails, and log stream insights. We display prices per million tokens and bill per thousand tokens for accuracy.
Model usage is billed by DigitalOcean. You are charged for all input and output tokens processed by the agent at the same token rates as serverless inference. Token usage depends on factors such as input length, agent instructions, attached knowledge bases, and configuration settings. To optimize usage, test your agents and adjust their parameters.
Agent Guardrails
Charges apply for all tokens processed through agent guardrails:
| Guardrail |
Price |
| Content Moderation |
$0.20 per 1,000,000 tokens |
| Jailbreak Detection |
$0.20 per 1,000,000 tokens |
| Sensitive Data Detection |
$0.34 per 1,000,000 tokens |
Costs are per token. Creating, editing, or duplicating guardrails has no additional cost.
Functions
If you attach DigitalOcean Functions to your agent, you are billed at functions pricing.
Agent Evaluations
Agent evaluations are charged by token usage at the same rates as model usage.
Log Stream Insights
Log Stream Insights uses a third-party model to analyze agent trace data. You are charged per token:
| Tokens |
Price |
| Input |
$1.10 per 1,000,000 tokens |
| Output |
$4.40 per 1,000,000 tokens |
Agent Development Kit
You are not charged for using the Agent Development Kit during public preview. However, you are billed for other DigitalOcean Inference features you use with your agent deployment:
-
We charge for model usage for Agent Development Kit (ADK). If you are using a DigitalOcean-hosted model, you are charged for those model keys.
-
For agent evaluations, token usage is charged to the agent model keys. For example, if your agent uses a serverless inference endpoint key, any token usage is charged to that key. If the agent uses a third-party model key, or a key to a model not hosted on DigitalOcean, you are charged by the hosting provider.
-
If you enable Log Stream Insights for your agent deployment, you are charged for tokens when new insights are generated.
For General Availability, agent deployment hosting, measured in GiB-sec, will be charged. We will also be charging for judge input and output tokens, which are the tokens used for judging the agent inputs and outputs against the test case’s chosen metrics. These costs are waived during public preview.