DigitalOcean Gradient™ AI Inference Hub

Validated on 16 Mar 2026 • Last edited on 16 Apr 2026

DigitalOcean Gradient™ AI Inference Hub provides a single control plane for managing inference workflows. It includes a Model Catalog where you can view available foundation models, including both DigitalOcean-hosted and third-party commercial models, compare capabilities and pricing, and run inference using serverless or dedicated deployments. DigitalOcean Gradient AI Inference Hub is in private preview. You can contact support for questions or assistance.

Browse Models in Model Catalog

Identify the right model for your use case by filtering available foundation models by capabilities and price.

Use Model Playground

Test and compare foundation models in the Model Playground.

Use Serverless Inference

Send API requests directly to foundation models without creating an AI agent or managing infrastructure.

Deploy to Dedicated Inference Endpoints

Deploy open-source and commercial LLMs on dedicated GPUs as an inference endpoint.

Latest Updates

16 April 2026

3 April 2026

  • The following models are deprecated from the Model Catalog:

    • Meta Llama 3.1 8B-Instruct
    • Mistral NeMo

    Migrate to Llama 3.3 70B-Instruct (llama3.3-70b-instruct) and gpt-oss-20b (openai-gpt-oss-20b) models respectively, to avoid service disruption.

2 April 2026

For more information, see the full release notes.

We can't find any results for your search.

Try using different keywords or simplifying your search terms.