← Back
NVIDIA
NVIDIA Dynamo 1.0 reaches production with 7x inference throughput boost on Blackwell
· releasefeatureperformanceplatformintegration · developer.nvidia.com ↗

Production-Ready Distributed Inference

NVIDIA Dynamo 1.0 is now available as a mature, production-grade framework for distributed inference of large reasoning and generative AI models across multiple GPU nodes. The framework addresses the challenge of orchestrating complex AI workloads—particularly agentic AI workflows that integrate multiple models and external tools—in large-scale environments requiring careful coordination across GPUs.

Performance Achievements and Benchmarks

Dynamo delivers significant performance improvements, with benchmarks demonstrating 7x throughput gains on NVIDIA Blackwell hardware when using disaggregated serving. The framework has achieved strong results in trusted third-party benchmarks including MLPerf and SemiAnalysis InferenceX, validating its position as a production-grade inference platform. Performance improvements are particularly notable with wide expert parallelism on NVIDIA GB200 NVL72 configurations.

Key Features and Optimizations

Recent enhancements include:

  • Agentic inference optimizations: Priority-based routing and cache pinning for multi-model agentic workflows
  • Multimodal acceleration: Disaggregated encode/prefill/decode stages, embedding cache management, and multimodal KV routing
  • Startup performance: ModelExpress enables 7x faster startup through checkpoint restore and weight streaming via NVIDIA NVLink
  • Kubernetes integration: Grove API for topology-aware scheduling on NVIDIA GB300 NVL72; zero-config deployment with DGDR
  • Resilience features: Layered fault detection, request cancellation/migration, and pip-installable KV Block Manager with object storage integration

Engine Support and Ecosystem Integration

Dynamo supports leading open-source inference engines including SGLang, NVIDIA TensorRT LLM, and vLLM. Early adopters including AstraZeneca, Baseten, ByteDance, CoreWeave, DigitalOcean, Pinterest, Tencent Cloud, Together AI, and Vultr have deployed Dynamo in production. Major cloud providers—Alibaba Cloud, AWS, Google Cloud, Microsoft Azure, and Oracle Cloud Infrastructure—have built native integrations with their managed Kubernetes environments.

Developer Action Items

Developers and organizations running large-scale inference deployments should evaluate Dynamo 1.0 for distributed multi-node inference scenarios. The framework's support for multiple inference engines and cloud platforms enables flexible integration into existing infrastructure. Documentation and deployment guides are available through major cloud provider offerings and NVIDIA's developer resources.