GenAI Platform Limitspublic
Validated on 19 Feb 2025 • Last edited on 17 Jun 2025
The DigitalOcean GenAI Platform lets you work with popular foundation models and build GPU-powered AI agents with fully-managed deployment, or send direct requests using serverless inference. Create agents that incorporate guardrails, functions, agent routing, and retrieval-augmented generation (RAG) pipelines with knowledge bases.
Limits
-
Currently, the GenAI Platform is only available in Toronto.
-
You cannot currently use
doctl
, the official DigitalOcean CLI, to manage your GenAI Platform resources. -
You are not able to bring your own models to use for your Artificial Intelligence (AI) agents. You can see the models we currently offer on our model overview page.
-
You currently cannot change embedding models after creating a knowledge base.
-
The number of agents you can create and the tokens your agents can use daily depends on the limits set for your account. We set dynamic resource limits to protect our platform against bad actors. We allocate a predetermined amount of tokens for each model your team uses. Each agent on your team draws tokens from that model’s amount of tokens. Tokens are replenished on a rolling 24-hour basis. Contact the Support team to request a limit increase.
-
In the Model Playground, teams have a daily token limit for each model. These tokens are separate from agent tokens and replenish every 24 hours. For example, if you use 500 tokens testing the DeepSeek model at 09:05 on Wednesday, those tokens are replenished at 09:05 on Thursday. Token limits are not exposed to users, but if you regularly reach this limit during testing, you can request an increase by contacting our Support team.
-
When creating a knowledge base, you can either select a DigitalOcean Space where your data is located or upload local files as the data source. You must first upload your data in a DigitalOcean Space, then create the knowledge base to attach to your model using the GenAI Platform.
-
Currently, we only support web functions for function routing from agents.
-
If you have a public agent that calls a private function, the private function can be called by anyone having the function’s URL. We recommend setting your function to Secure Web Function which enables authentication. For more information on configuring the function’s access and security, see Access and Security in the Functions documentation.
-
To manage compute resources and ensure fair resource distribution, GenAI Platform has limits on resource creation and model usage. If you require a limit increase, contact support.
-
For web crawling data sources, the crawler indexes up to 5500 pages and skips inaccessible or disallowed links to prevent excessively large indexing jobs.
-
You cannot currently re-index a previously crawled seed URL. To re-index the content, delete the seed URL, and then add it again to start a new crawl.
-
Currently, you cannot customize detection rules for guardrails.
-
You cannot view the sources used to generate a response in the chatbot interface.
-
Bringing Your Own Models (BYOM) is not currently supported with serverless inference. To view the models we offer, see the available model offerings.
-
Serverless inference supports the two to three most recent stable versions of each model to ensure consistent performance and reliable maintenance. For the list of supported models and versions, see the available model offerings.
-
Serverless inference model endpoints support OpenAI-compatible request formats but may not be compatible with all OpenAI tools and plugins.
-
Serverless inference provides access to commercial models, but not all model-specific features are currently supported. For example, features like Anthropic’s extended thinking are not available.
-
OpenAI models accessed through serverless inference do not support zero data retention. If your use case requires strict data privacy or compliance, consider using a different model or contact support for guidance.
-
Currently, you cannot access the OpenAI o1 model through serverless inference. All other supported models are available as model endpoints.
-
Currently, you cannot re-index specific data sources within a knowledge base. To re-index any changed data sources, you need to re-index all the data sources.
-
Currently, you cannot change the OpenSearch database for an existing knowledge base. To use a different database, create a new knowledge base with the same data sources and choose the new OpenSearch database during setup.
-
Currently, you cannot edit attributes of an knowledge using the DigitalOcean API. Alternatively, edit your knowledge base using the DigitalOcean Control Panel.
-
You cannot view the sources used to generate a response in the chatbot interface.