DigitalOcean Gradient™ AI Platform API Reference

Validated on 9 Oct 2024 • Last edited on 9 Apr 2026

DigitalOcean Gradient™ AI Platform lets you build fully-managed AI agents with knowledge bases for retrieval-augmented generation, multi-agent routing, guardrails, and more, or use serverless inference to make direct requests to popular foundation models.

The DigitalOcean Gradient™ AI Platform API endpoints are organized into the following groups:

  • GradientAI Platform (87 endpoints): The API lets you build GPU-powered AI agents with pre-built or custom foundation models, function and agent routes, and RAG pipelines with knowledge bases.
  • Dedicated Inference (13 endpoints): Dedicated Inference delivers scalable production-grade LLM hosting on DigitalOcean. Create, list, get, update, and delete Dedicated Inference instances; manage accelerators, CA certificate, sizes, GPU model config, and access tokens.
  • Agent Inference (1 endpoints): DigitalOcean Gradient™ AI Agentic Cloud allows you to create multi-agent workflows to power your AI applications. This allows developers to integrate agents into your AI applications.
  • Serverless Inference (5 endpoints): DigitalOcean Gradient™ AI Agentic Cloud allows access to serverless inference models. You can access models by providing an inference key.

We can't find any results for your search.

Try using different keywords or simplifying your search terms.