Identify the right model for your use case by filtering available foundation models by capabilities and price.
DigitalOcean Gradient™ AI Inference Hub
Validated on 16 Mar 2026 • Last edited on 16 Apr 2026
DigitalOcean Gradient™ AI Inference Hub provides a single control plane for managing inference workflows. It includes a Model Catalog where you can view available foundation models, including both DigitalOcean-hosted and third-party commercial models, compare capabilities and pricing, and run inference using serverless or dedicated deployments. DigitalOcean Gradient AI Inference Hub is in private preview. You can contact support for questions or assistance.
Test and compare foundation models in the Model Playground.
Send API requests directly to foundation models without creating an AI agent or managing infrastructure.
Deploy open-source and commercial LLMs on dedicated GPUs as an inference endpoint.
Latest Updates
16 April 2026
-
The following Anthropic model is now available on DigitalOcean Gradient AI Inference Hub for serverless inference:
For more information, see the Available Models page.
3 April 2026
-
The following models are deprecated from the Model Catalog:
- Meta Llama 3.1 8B-Instruct
- Mistral NeMo
Migrate to Llama 3.3 70B-Instruct (
llama3.3-70b-instruct) and gpt-oss-20b (openai-gpt-oss-20b) models respectively, to avoid service disruption.
2 April 2026
-
The following client libraries for Gradient AI Inference Hub are now available in the official DigitalOcean SDKs. You can use the SDKs to manage serverless and dedicated inference:
-
The Python client library is now available in the official DigitalOcean Python client library PyDo. For more information, see the following reference documentation:
-
The TypeScript client library is now available in the official DigitalOcean TypeScript library DoTs.
-
The official Go client library is available at Gradient Go library.
The Gradient™ SDK will be deprecated in a future release.
-
For more information, see the full release notes.