DigitalOcean Gradient™ AI Inference Hub Quickstart
Validated on 26 Jun 2018 • Last edited on 16 Mar 2026
DigitalOcean Gradient™ AI Inference Hub provides a single control plane for managing inference workflows. It includes a Model Catalog where you can view available foundation models, including both DigitalOcean-hosted and third-party commercial models, compare capabilities and pricing, and run inference using serverless or dedicated deployments. DigitalOcean Gradient AI Inference Hub is in public preview and enabled for all users. You can contact support for questions or assistance.
Browse the Model Catalog
-
To access Model Catalog, go to the DigitalOcean Control Panel and open the Model Catalog tab in Inference Hub.
-
Browse the available foundation models. For more information about supported models and their capabilities, see our models page.
-
Click a model to open its model card and view details such as capabilities, pricing, and deployment options.
-
To test the model, click Model Playground in the top-right corner of the model card.
To learn more about browsing and filtering models, see Browse Models in Model Catalog.
Test a Model in the Model Playground
-
To access Model Playground, go to the DigitalOcean Control Panel, and then open the Model Playground tab in Inference Hub.
-
Select a foundation model. For more information about supported models and their capabilities, see our models page.
-
Enter a prompt and review the model response.
-
Adjust settings such as temperature and token limits to test different outputs.
Next Steps
Once you’ve browsed models in Inference Hub, you can continue working with the following features:
- Use Serverless Inference: Create model access keys and send API requests to foundation models without managing infrastructure.
- Use Dedicated Inference: Host open-source or commercial LLMs on dedicated GPUs, scale them, and deploy them as inference endpoints.