At Ray Summit 2025, Robert Shaw from Red Hat shares how vLLM is evolving to support p/d disaggregation and wide expert parallelism, two critical techniques for serving frontier-scale models like DeepSeek-R1 across large GPU clusters. He begins by breaking down how prefill/decode (p/d) disaggregation enables more efficient resource utilization by separating the heavy, bursty prefill phase from the latency-sensitive decode phase—allowing clusters to scale each stage independently for maximum throughput and cost efficiency. Robert then dives into the implementation of wide expert parallelism (EP) in vLLM, explaining how it allows MoE-based models to leverage large numbers of experts across multi-node environments. He details the orchestration, scheduling, and memory-management challenges of EP at cluster scale, and the design decisions that make these deployments practical. Finally, he explores the tradeoffs and system-level considerations that arise when serving massive models such as DeepSeek-R1—including GPU topology, communication overhead, batching behavior, and cluster elasticity. Attendees will gain a deep understanding of how vLLM implements next-generation parallelism strategies and what it takes to run cluster-scale MoE and non-MoE LLMs efficiently in real production environments. Subscribe to our YouTube channel to stay up-to-date on the future of AI! 🔗 Connect with us: LinkedIn: X: Website:











