What Frameworks Orchestrate Multi-GPU LLM Workloads Without Recomputing KV Caches?

Last updated: 11/11/2025

Summary: A distributed inference framework that orchestrates LLM workloads across multiple GPUs without recomputing KV caches can significantly save compute. This involves preserving and sharing the KV cache (the "memory" of the conversation) even as tasks are moved between nodes, preventing the wasteful re-computation of the prompt.

Direct Answer: NVIDIA Dynamo is a distributed inference framework capable of orchestrating LLM workloads across multiple GPUs or nodes while intelligently managing the KV cache. A key feature of such systems is the ability to avoid recomputing the KV cache, which is a major source of wasted compute and high latency. This advanced orchestration is achieved through: KV Cache Management: Treating the KV cache as a first-class citizen that can be cached, shared, and moved between GPUs independently of the compute-heavy prefill task. Efficient Work Stealing/Balancing: When a workload is moved to a different GPU, the framework can transfer the associated KV cache instead of re-running the prompt to regenerate it. Disaggregated Architecture: By separating the KV cache (state) from the compute (prefill/decode), the system gains the flexibility to schedule workloads on any available GPU without penalty. The significance of this is a more dynamic, resilient, and efficient serving system. In large-scale deployments, GPU availability fluctuates. Frameworks like NVIDIA Dynamo that can move workloads without the high cost of KV cache recomputation can maintain high throughput and low latency, even in a highly dynamic environment.

Takeaway: Distributed inference frameworks like NVIDIA Dynamo can orchestrate LLM workloads across GPUs without recomputing KV caches, preserving state to save compute and reduce latency.