Available Foundation and Embedding Models for DigitalOcean Gradient™ AI Platform
Validated on 29 Sep 2025 • Last edited on 3 Nov 2025
DigitalOcean Gradient™ AI Platform lets you build fully-managed AI agents with knowledge bases for retrieval-augmented generation, multi-agent routing, guardrails, and more, or use serverless inference to make direct requests to popular foundation models.
The following foundation and embedding models are available for Gradient AI Platform. For pricing, see Gradient AI Platform’s pricing page.
Foundation Models
Gradient AI Platform supports both open source and commercial foundation models. You can use these models for serverless inference, in agents, or to test configurations in the Agent Playground.
Open source models are generally published by research labs, available under open licenses, and offered using DigitalOcean API access keys. Commercial models are proprietary and require the provider’s API keys to access, such as OpenAI API keys and Anthropic API keys.
We offer the following foundation models:
| Model | Model ID | Parameters | Max Output Tokens | Usage Notes |
|---|---|---|---|---|
| Qwen3-32B | alibaba-qwen3-32b |
32 billion | 40,960 | Only for serverless inference. |
| Model | Model ID | Parameters | Max Tokens |
|---|---|---|---|
| Claude Sonnet 4 | anthropic-claude-sonnet-4 |
Not published | 64,000 |
| Claude 3.7 Sonnet | anthropic-claude-3.7-sonnet |
Not published | 128,000 |
| Claude 3.5 Sonnet | anthropic-claude-3.5-sonnet |
Not published | 8,192 |
| Claude 3.5 Haiku | anthropic-claude-3.5-haiku |
Not published | 8,000 |
| Claude Opus 4 | anthropic-claude-opus-4 |
Not published | 32,000 |
| Claude 3 Opus | anthropic-claude-3-opus |
Not published | 4,096 |
| Model | Model ID | Parameters | Max Tokens | Usage Notes |
|---|---|---|---|---|
| DeepSeek R1 Distill Llama 70B | deepseek-r1-distill-llama-70b |
70 billion | 32,768 | When using DeepSeek models in a user-facing agent, we strongly recommend adding all available guardrails for a safer conversational experience. |
| Model | Model ID | Type | Usage Notes |
|---|---|---|---|
| Fast SDXL | fal-ai/fast-sdxl |
Image generation | Multimodal and generative model, only for serverless inference. |
| Flux Schnell | fal-ai/flux/schnell |
Image generation | Multimodal and generative model, only for serverless inference. |
| Stable Audio 2.5 (Text-to-Audio) | fal-ai/stable-audio-25/text-to-audio |
Text-to-audio | Multimodal and generative model, only for serverless inference. |
| Multilingual TTS v2 | fal-ai/elevenlabs/tts/multilingual-v2 |
Text-to-speech | Multimodal and generative model, only for serverless inference. |
| Model | Model ID | Parameters | Max Tokens |
|---|---|---|---|
| Llama 3.3 Instruct-70B | llama3.3-70b-instruct |
70 billion | 128,000 |
| Llama 3.1 Instruct-8B | llama3-8b-instruct |
8 billion | 128,000 |
| Model | Model ID | Parameters | Max Tokens |
|---|---|---|---|
| NeMo | mistral-nemo-instruct-2407 |
12 billion | 128,000 |
| Model | Model ID | Parameters | Max Tokens |
|---|---|---|---|
| gpt-oss-120b | openai-gpt-oss-120b |
117 billion | 131,072 |
| gpt-oss-20b | openai-gpt-oss-20b |
21 billion | 131,072 |
| GPT-5 | openai-gpt-5 |
Not published | Not published |
| GPT-5 mini | openai-gpt-5-mini |
Not published | Not published |
| GPT-5 nano | openai-gpt-5-nano |
Not published | Not published |
| GPT-4.1 | openai-gpt-4.1 |
Not published | Not published |
| GPT-4o | openai-gpt-4o |
Not published | Not published |
| GPT-4o mini | openai-gpt-4o-mini |
Not published | Not published |
| o1 | openai-o1 |
Not published | Not published |
| o3 | openai-o3 |
Not published | Not published |
| o3-mini | openai-o3-mini |
Not published | Not published |
| GPT-image-1 | openai-gpt-image-1 |
Not published | Not published |
Embedding Models
An embedding model converts data into vector embeddings. Gradient AI Platform stores vector embeddings in an OpenSearch database cluster for use with agent knowledge bases. We offer the following embedding models:
| Provider | Type | Model and Version | Parameters |
|---|---|---|---|
| Tongyi Lab, Alibaba | General text embeddings (GTE) | Alibaba-NLP/gte-large-en-v1.5 | 434 million |
| UKP Lab, Technical University of Darmstadt | Sentence Transformer (SBERT) | sentence-transformers/all-MiniLM-L6-v2 | 22.7 million |
| UKP Lab, Technical University of Darmstadt | Sentence Transformer (SBERT) | sentence-transformers/multi-qa-mpnet-base-dot-v1 | 109 million |