Batch Inference
Generated on 28 Apr 2026
This content is automatically generated from https://github.com/digitalocean-labs/mcp-digitalocean/blob/main/pkg/registry/genai-batchinference/README.md.
GenAI Batch Inference (genai-batchinference)
What this service does
This package registers Batch Inference tools for the DigitalOcean MCP server. Batch Inference allows you to submit large batches of inference requests (JSONL) for asynchronous processing via OpenAI-compatible or Anthropic models. The tools call the DigitalOcean inference proxy at inference.do-ai.run through godo.
Enable these tools by including the service key when you configure the server: service name genai-batchinference.
Code layout
| Path | Purpose |
|---|---|
batch_inference_tools.go |
Tool handlers and MCP tool definitions |
batch_inference_tools_test.go |
Unit tests (mocked API client) |
generate.go |
mockgen directive |
mocks.go |
Generated mock for godo.BatchInferenceService |
Tools exposed
| Tool | What it does |
|---|---|
genai-batch-inference-create-file |
Create a presigned URL for uploading a JSONL input file. |
genai-batch-inference-create |
Create a new batch inference job (OpenAI or Anthropic). |
genai-batch-inference-get |
Get a batch inference job’s status and metadata. |
genai-batch-inference-get-results |
Get the presigned download URL for completed job results. |
genai-batch-inference-cancel |
Request cancellation of a running batch inference job. |
genai-batch-inference-list |
List batch inference jobs with cursor-based pagination. |
Arguments (summary)
genai-batch-inference-create-file
| Argument | Required | Description |
|---|---|---|
FileName |
yes | Name of the JSONL file (must end in .jsonl) |
genai-batch-inference-create
| Argument | Required | Description |
|---|---|---|
Provider |
yes | openai or anthropic |
FileID |
yes | UUID of a previously uploaded .jsonl file |
CompletionWindow |
yes | e.g. 24h |
RequestID |
no | Client-supplied idempotency key |
Endpoint |
no | OpenAI batch API path (required for OpenAI, e.g. /v1/chat/completions) |
genai-batch-inference-get / genai-batch-inference-get-results / genai-batch-inference-cancel
| Argument | Required | Description |
|---|---|---|
BatchID |
yes | UUID of the batch inference job |
genai-batch-inference-list
| Argument | Required | Description |
|---|---|---|
Status |
no | Filter by status (e.g. completed, in_progress) |
Limit |
no | Max jobs per page |
After |
no | Cursor from previous page’s endCursor |
Responses
- create-file: Returns
file_id,upload_url, andexpires_at. - create: Returns batch object with
batch_id,status,provider,request_counts, timestamps. - get: Returns batch object (same shape as create response).
- get-results: Returns
output_file_idand nesteddownloadwithpresigned_urlandexpires_at. - cancel: Returns the batch object with updated status.
- list: Returns Relay-style
edges(each withnodeandcursor) andpage_info(hasNextPage,endCursor).
Auth
Callers need a DigitalOcean API token with access to GenAI Batch Inference endpoints. The feature may be gated behind FlipperGenAIBatchInference per account/team.