DualPath: Breaking the Storage Wall
Episode Summary
A deep dive into DualPath, a system that solves the storage bandwidth bottleneck in agentic LLM inference — then a scale-by-scale walkthrough of how the same bottleneck affects everyone from Raspberry Pi clusters to DGX SuperPods. As AI agents run multi-turn sessions (100+ turns, 95%+ KV-cache reuse), the bottleneck shifts from compute to storage I/O. DualPath exploits idle decode-engine storage NICs to load KV-cache and transfer it via RDMA to prefill engines, achieving 1.87x offline throughput and 1.96x online serving improvements. We break down the architecture, then walk from RPi5 to Mac mini to DGX Spark to production, showing where the diagnosis applies universally and where the specific cure requires datacenter hardware.
Paper Discussed
DualPath: Breaking the Storage Bandwidth Bottleneck in Agentic LLM InferencearXiv:2602.21548 — HTML versionAuthors: Yongtong Wu, Shaoyuan Chen, Yinmin Zhong, Rilin Huang, Yixuan Tan, Wentao Zhang, Liyue Zhang, Shangyan Zhou, Yuxuan Liu, Shunfeng Zhou, Mingxing Zhang, Xin Jin, Panpan HuangAffiliations: Peking University, Tsinghua University, DeepSeek-AIHardware Scale Walkthrough
Raspberry Pi 5 Cluster
~30 TOPS NPU, Gigabit Ethernet, USB 3.0 storageSame I/O bottleneck physics, no RDMA or traffic isolation availableDiagnosis applies; cure doesn'tMac mini M4 / Mac Studio
16-32GB unified memory, Thunderbolt 4 (40Gbps bidirectional)Single bus carries all traffic — no compute/storage network separationThunderbolt 5 at 120Gbps starts to change the equationDGX Spark Cluster
8x Sparks: 128GB each, 1TB total, ConnectX-7 with real RDMATwo MikroTik switches: one compute network, one storage network4 prefill + 4 decode engines (1:1 P/D ratio — middle of bottleneck-free range)~$30K all-in (8 × $3K Sparks + ~$2,600 switches + cables)DGX Spark home cluster build video — 6,367 tok/s on Qwen 34B BF16This is where DualPath's architecture becomes directly feasibleQSFP28 vs QSFP56 cable differences matter for bandwidthProduction Scale (Paper's Target)
DGX SuperPOD: 8 GPUs/node, 8x 400Gbps CNICs, 1x 400Gbps SNICPhysically isolated compute and storage networksFull DualPath: 1.87x offline, 1.96x online throughputKey Concepts
Prefill-Decode Disaggregation — Separating prompt processing from token generation onto dedicated engine pools. See DistServe.KV-Cache — Cached attention keys and values, stored to avoid recomputation on subsequent turns.Cache-Compute Ratio — GB of KV-cache to load per PFLOP of compute. The universal diagnostic for whether you're I/O-bound or compute-bound.RDMA — Remote Direct Memory Access. Direct memory-to-memory transfer without CPU involvement.Layerwise Prefill — Per-layer KV-cache loading to overcome HBM limits. See LayerKV.3FS — DeepSeek's distributed file system. GitHub.InfiniBand Virtual Lanes — Hardware QoS for traffic isolation.Key Numbers
Metric
Value
Avg agent turns (production traces)
157
Avg append tokens per turn
429
KV-cache hit rate
98.7%
Cache-compute ratio (DeepSeek V3.2)
13–36 GB/PFLOP
Cache-compute ratio (Qwen 32B, FP16)
117–267 GB/PFLOP
Offline throughput improvement
up to 1.87x
Online serving throughput improvement
1.96x average
I/O-compute ratio degradation (Ampere→Blackwell)
14.4x
Bottleneck-free P/D ratio range
1/7 to 7/2
Scale tested
up to 1,152 GPUs
Related Work
Mooncake: KV Cache-Centric LLM Serving — DRAM-based caching approachDistServe: Disaggregating Prefill and Decoding — PD disaggregationDeepSeek-V3 Technical Report — Model architectureFlashMLA — Efficient attention kernelDeepEP — Expert parallel communicationSGLang — Structured LLM servingModels Evaluated
DeepSeek V3.2 660B (MoE with sparse attention)DS 27B (downscaled V3.2)Qwen2.5-32B (dense, GQA)Author Profiles
Yinmin Zhong — Peking UniversityXin Jin — Peking UniversityMingxing Zhang — Tsinghua UniversityThis episode was generated with AI assistance.