Inference Quickstart
Validated on 27 Apr 2026 • Last edited on 27 Apr 2026
Inference provides a single control plane for managing inference workflows. It includes a Model Catalog where you can view available foundation models, including both DigitalOcean-hosted and third-party commercial models, compare model capabilities and pricing, use routing to match inference requests to the best-fit model, and run inference using serverless or dedicated deployments.
Browse the Model Catalog
-
To access Model Catalog, go to the DigitalOcean Control Panel and click Model Catalog under INFERENCE.
-
Browse the available foundation models. For more information about supported models and their capabilities, see our models page.
-
Click a model to open its model card and view details such as capabilities, pricing, and deployment options.
-
To test the model, click Model Playground in the top-right corner of the model card.
To learn more about browsing and filtering models, see Browse Models in Model Catalog.
Once you’ve browsed models, you can continue working with the following features:
- Use Serverless Inference: Create model access keys and send API requests to foundation models without managing infrastructure.
- Use Dedicated Inference: Host open-source or commercial LLMs on dedicated GPUs, scale them, and deploy them as inference endpoints.
Test a Model in the Model Playground
-
To access Model Playground, go to the DigitalOcean Control Panel, and click Serverless Inference under INFERENCE. Then, select the Model Playground tab.
-
Select a foundation model. For more information about supported models and their capabilities, see our models page.
-
Enter a prompt and optionally upload images. Then, review the model response.
-
Adjust settings such as temperature and token limits to test different outputs.
Integrate with Coding Agents
Coding agents, such as Codex and Claude Code, can use the inference endpoints https://inference.do-ai.run as a drop-in proxy to run inference requests on DigitalOcean.
- Install the coding agent. Installation steps vary by provider.
- Set up the model access key.
- Configure the agent to use DigitalOcean.
- Run the agent to use DigitalOcean inference endpoints.
For more information, see Use With Coding Agents.