How to View Serverless Inference Metrics
Validated on 27 Apr 2026 • Last edited on 27 Apr 2026
Inference provides a single control plane for managing inference workflows. It includes a Model Catalog where you can view available foundation models, including both DigitalOcean-hosted and third-party commercial models, compare model capabilities and pricing, use routing to match inference requests to the best-fit model, and run inference using serverless or dedicated deployments.
Inference observability provides real-time and historical metrics across latency, throughput, error rates, token consumption, cost attribution, and rate limiting. To view the metrics, in the DigitalOcean Control Panel, click INFERENCE in the left menu, and then click Serverless Inference. Then, select the Analyze tab to view the metrics.