pith. machine review for the scientific record. sign in

arxiv: 2604.27085 · v1 · submitted 2026-04-29 · 💻 cs.DC · cs.AI· cs.LG

Recognition: unknown

Efficient Training on Multiple Consumer GPUs with RoundPipe

Authors on Pith no claims yet

Pith reviewed 2026-05-07 10:33 UTC · model grok-4.3

classification 💻 cs.DC cs.AIcs.LG
keywords pipeline parallelismLLM fine-tuningconsumer GPUsRoundPipeweight bindingLoRAdistributed training
0
0 comments X

The pith

RoundPipe dynamically dispatches model stages round-robin across stateless GPUs to eliminate the weight binding bottleneck in pipeline parallelism.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces RoundPipe to address the weight binding issue in pipeline parallelism for fine-tuning large language models on consumer GPUs. Existing methods tie unevenly sized stages, such as large LM heads, to specific GPUs, limiting overall throughput to the slowest device and creating pipeline bubbles. RoundPipe instead treats the GPUs as a pool of stateless workers and dispatches stages dynamically in round-robin fashion, supported by priority-aware transfers, event-based synchronization, and automated partitioning. A sympathetic reader would care because this could make fine-tuning of models up to hundreds of billions of parameters practical on affordable multi-GPU servers without high-end networking.

Core claim

RoundPipe breaks the weight binding constraint on consumer GPU servers by treating GPUs as a pool of stateless execution workers and dynamically dispatching computation stages across devices in a round-robin manner to achieve a near-zero-bubble pipeline, while integrating a priority-aware transfer scheduling engine, a fine-grained distributed event-based synchronization protocol, and an automated layer partitioning algorithm to ensure training correctness and efficiency.

What carries the argument

Round-robin dynamic dispatching of computation stages to stateless GPU workers, which decouples stage assignment from fixed device binding.

If this is right

  • RoundPipe delivers 1.48 to 2.16 times speedups compared to state-of-the-art baselines for fine-tuning models ranging from 1.7 billion to 32 billion parameters.
  • It enables LoRA fine-tuning of the Qwen3-235B model using sequences of 31K length on a single consumer-grade server with eight RTX 4090 GPUs.
  • The approach supports training correctness and convergence for arbitrary model architectures and sequence lengths.
  • Pipeline bubbles are reduced to near zero through the combination of dynamic dispatch and supporting mechanisms.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar dynamic dispatching logic could be adapted for distributed inference to improve utilization on heterogeneous hardware.
  • Researchers might experiment with combining RoundPipe with other memory optimization techniques like quantization to further scale model sizes.
  • The open-source library could serve as a base for testing the method on non-LLM workloads such as vision transformers or reinforcement learning agents.

Load-bearing premise

The dynamic dispatching of stages combined with the priority-aware transfer engine and event-based synchronization protocol introduces negligible overhead while preserving training correctness and convergence for arbitrary model architectures and sequence lengths.

What would settle it

Running the same training task on a 32B model both with RoundPipe and with a baseline pipeline schedule, then comparing the achieved tokens per second and the final model loss to check for slowdowns or divergence.

Figures

Figures reproduced from arXiv: 2604.27085 by Huichuan Zheng, Jiwu Shu, Shiwei Gao, Yibin Luo, Youyou Lu.

Figure 1
Figure 1. Figure 1: Looped BFS schedule and RoundPipe schedule when training a 12-layer model with language model head (layer 13) on 4 GPUs. F/B denotes forward/backward, and numbers indicate the index of layers contained in the stage. Looped BFS schedule processes 8 microbatches at a time while RoundPipe processes them in two rounds. GPU. We term this the weight binding issue. No matter how the pipeline is partitioned (stand… view at source ↗
Figure 2
Figure 2. Figure 2: Theoretical time of recomputing and reloading activations of a transformer layer. Calculation details in Ap￾pendix B.2. layer. This reduces the intermediate activations’ size to 2𝑠𝑏ℎ bytes per layer [22]. While the extra computation may seem expensive, it is often more efficient than swapping activations to host memory over PCIe view at source ↗
Figure 3
Figure 3. Figure 3: Bubble ratio of Looped schedules under ideal balanced partition and real-world imbalanced partition on 8 GPUs. Real-world bubble ratios are collected in §5.6.1 across all 𝑁 GPUs. This reduces the bubble ratio to approxi￾mately 𝑁 · (𝑁 −1) 𝑆·𝑀+𝑁 · (𝑁 −1) , and decreases as the number of stages grows. Imbalance Bubbles. Theoretical analyses of pipeline bub￾bles assume uniform stage execution times, which are … view at source ↗
Figure 4
Figure 4. Figure 4: Asynchronous optimizer update in RoundPipe. executes all 𝑀𝑅 micro-batches of the current round for its assigned stage before the next stage slot is dispatched. Between rounds, the dispatch seamlessly resumes where the previous round left off. That is, the starting index updates to 𝑔0 ← (𝑔0 + 𝑆) mod 𝑁, assigning the first stage of the new round to the GPU next in line. This dispatch logic ensures a continuo… view at source ↗
Figure 5
Figure 5. Figure 5: RoundPipe system overview. 4.1 RoundPipe Overview RoundPipe adopts a single-controller architecture inspired by Ray [30] and veRL/HybridFlow [43], separating the control plane, which manages task scheduling and ordering, from the data plane, which handles hardware-level execution and device data transfers. As shown in view at source ↗
Figure 6
Figure 6. Figure 6: Illustrated example of simple compute-transfer overlap on two consecutive stages with two microbatches (MB) each. Each color represents one stage. Stage 1 Stage 2 Stage 3 MB1 MB2 MB1 MB2 MB1 MB2 Time Chunked Parameter Upload Activation Upload Compute Activation Download Chunked Gradient Download view at source ↗
Figure 7
Figure 7. Figure 7: RoundPipe multi-stream workflow over three con￾secutive stages with two microbatches (MB) each. Each color represents one stage. and GPUs continuously. As shown in view at source ↗
Figure 8
Figure 8. Figure 8: (a) The blocking approach copies weights (P cp) and gradients (G cp) on the main thread. (b) RoundPipe’s event-based protocol offloads copies to the optimizer worker (O) and uses per-layer events. The four arrows from left to right correspond to ordering constraints (1)-(4), respectively. (1) Protects weight integrity. P copy must wait until the GPU has finished uploading parameters for the previous iterat… view at source ↗
Figure 9
Figure 9. Figure 9: Training Throughput on 8×RTX 4090. Qwen3 1.7B LLaMA-3.1 8B GPT-OSS 20B Qwen3 32B Qwen3 235B-LoRA 0K 20K 40K 60K Max Sequence Length OOM OOM OOM OOM OOM OOM OOM OOM OOM OOM OOM OOM N/A OOM OOM Max Sequence Length on 8×RTX 4090 ZeRO-2 FSDP ZeRO-Infinity Megatron-TP Megatron-PP Mobius RoundPipe view at source ↗
Figure 10
Figure 10. Figure 10: Maximum trainable sequence length on 8×RTX 4090. 5.2 End-to-End Performance on 4090 Figures 9 and 10 present end-to-end throughput and maximum sequence length results on the 4090 server. 4 Throughput. BothRoundPipe andRoundPipe-sync achieve the highest throughput across all five models. RoundPipe out￾performs the fastest existing systems across 1.7-32B models by 1.48 ∼ 2.16× on training throughput, and Ro… view at source ↗
Figure 11
Figure 11. Figure 11: Training Throughput on 8×A800. Qwen3 1.7B LLaMA-3.1 8B GPT-OSS 20B Qwen3 32B Qwen3 235B-LoRA 0K 50K 100K 150K 200K 250K 300K Max Sequence Length OOM OOM OOM OOM OOM N/A Max Sequence Length on 8×A800 SXM ZeRO-2 FSDP ZeRO-Infinity Megatron-TP Megatron-PP RoundPipe view at source ↗
Figure 12
Figure 12. Figure 12: Maximum trainable sequence length on 8×A800. RoundPipe achieves it by storing stage-boundary activations in host memory and recomputing layer-internal activations on demand. RoundPipe avoids TP’s heavy communication over￾head, thereby achieving higher training throughput than TP. As model size scales up, non-offloading systems are bottlenecked by model footprint and fail to scale sequence lengths, whereas… view at source ↗
Figure 14
Figure 14. Figure 14: Throughput vs. sequence length for Qwen3-1.7B view at source ↗
Figure 17
Figure 17. Figure 17: Operational intensity vs. batch size for representa￾tive dense and MoE models at 𝑠=2048. Horizontal lines mark the ridge-point OI for GPUs in view at source ↗
read the original abstract

Fine-tuning Large Language Models (LLMs) on consumer-grade GPUs is highly cost-effective, yet constrained by limited GPU memory and slow PCIe interconnects. Pipeline parallelism combined with CPU offloading mitigates these hardware bottlenecks by reducing communication overhead. However, existing PP schedules suffer from an inherent limitation termed the weight binding issue. Binding uneven model stages (e.g., the LM head is large) to GPUs limits the pipeline's throughput to that of the GPU with the heaviest load, leading to severe pipeline bubbles. In this paper, we propose RoundPipe, a novel pipeline schedule that breaks the weight binding constraint on consumer GPU servers. RoundPipe treats GPUs as a pool of stateless execution workers and dynamically dispatches computation stages across devices in a round-robin manner, achieving a near-zero-bubble pipeline. To ensure training correctness and system efficiency, RoundPipe integrates a priority-aware transfer scheduling engine, a fine-grained distributed event-based synchronization protocol, and an automated layer partitioning algorithm. Evaluations on an 8$\times$ RTX 4090 server demonstrate that RoundPipe achieves 1.48--2.16$\times$ speedups over state-of-the-art baselines when fine-tuning 1.7B to 32B models. Remarkably, RoundPipe enables LoRA fine-tuning of the Qwen3-235B model with 31K sequence length on a single server. RoundPipe is publicly available as an open-source Python library with comprehensive documentation.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper proposes RoundPipe, a pipeline parallelism schedule for fine-tuning LLMs on consumer GPUs that treats GPUs as stateless workers and dispatches model stages dynamically in round-robin fashion to break the weight-binding bottleneck of prior PP methods. It augments this with a priority-aware transfer engine, a fine-grained distributed event-based synchronization protocol, and an automated layer partitioning algorithm. On an 8× RTX 4090 server, it reports 1.48–2.16× speedups versus state-of-the-art baselines for 1.7B–32B models and demonstrates LoRA fine-tuning of the 235B Qwen3 model at 31K sequence length.

Significance. If the correctness and overhead claims hold, the result would be significant for enabling large-model fine-tuning on affordable single-server hardware. The open-source Python library release with documentation strengthens reproducibility. Concrete speedups across a range of model sizes and the 235B capability are notable empirical contributions to distributed training systems for consumer GPUs.

major comments (3)
  1. [§4.2] §4.2 (event-based synchronization protocol): the description does not supply a formal invariant, proof sketch, or exhaustive dependency analysis showing that round-robin dynamic dispatch plus event ordering preserves correct per-microbatch weight versions, activation tensors, and gradient accumulation across forward/backward passes for arbitrary model graphs.
  2. [§3.3] §3.3 (automated partitioning algorithm): no formal load-balance guarantee or worst-case analysis is given that the algorithm produces balanced stages for every architecture and sequence length; evaluations cover only the reported 1.7B–32B models plus one 235B case.
  3. [§5] §5 (evaluation): speedups are presented without a quantitative overhead breakdown (dispatch, priority scheduling, event synchronization) or micro-benchmark isolating the cost of dynamic dispatching, leaving the “negligible overhead” claim unsupported beyond aggregate throughput numbers.
minor comments (2)
  1. [Figure 3] Figure 3 and Table 2: axis labels and legend entries use inconsistent abbreviations for the baselines; adding a short caption footnote would improve readability.
  2. [Abstract] The abstract claims “near-zero-bubble” but the text never quantifies bubble fraction or compares it directly to the theoretical minimum; a short definition or measurement method should be added.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback and the recommendation for major revision. We address each major comment point by point below, indicating revisions where we agree the manuscript can be strengthened.

read point-by-point responses
  1. Referee: [§4.2] §4.2 (event-based synchronization protocol): the description does not supply a formal invariant, proof sketch, or exhaustive dependency analysis showing that round-robin dynamic dispatch plus event ordering preserves correct per-microbatch weight versions, activation tensors, and gradient accumulation across forward/backward passes for arbitrary model graphs.

    Authors: We acknowledge that the manuscript presents the event-based synchronization protocol primarily through its design and operational examples rather than a formal proof. The protocol uses unique event identifiers tied to microbatches and stages to enforce ordering. We agree a proof sketch would improve rigor. In the revised version we will add a concise invariant and proof sketch showing that round-robin dispatch with event ordering preserves per-microbatch weight versions and tensor dependencies for the transformer graphs considered. revision: yes

  2. Referee: [§3.3] §3.3 (automated partitioning algorithm): no formal load-balance guarantee or worst-case analysis is given that the algorithm produces balanced stages for every architecture and sequence length; evaluations cover only the reported 1.7B–32B models plus one 235B case.

    Authors: The partitioning algorithm is a heuristic that estimates per-layer compute and communication costs to produce balanced stages for transformer models. It does not claim a formal worst-case guarantee for arbitrary graphs. Evaluations span 1.7B–32B models plus the 235B case, which covers the target use case. We will revise §3.3 to explicitly state the algorithm’s assumptions and limitations and note that exhaustive analysis for non-transformer architectures lies outside the paper’s scope. revision: partial

  3. Referee: [§5] §5 (evaluation): speedups are presented without a quantitative overhead breakdown (dispatch, priority scheduling, event synchronization) or micro-benchmark isolating the cost of dynamic dispatching, leaving the “negligible overhead” claim unsupported beyond aggregate throughput numbers.

    Authors: We agree that an explicit overhead breakdown would better support the negligible-overhead claim. The current evaluation relies on end-to-end speedups. In the revision we will add micro-benchmarks that isolate the latency of dynamic dispatch, priority scheduling, and event synchronization on the 8× RTX 4090 platform to quantify these costs. revision: yes

Circularity Check

0 steps flagged

RoundPipe presents a new dynamic pipeline design and empirical results with no circular reductions

full rationale

The paper introduces RoundPipe as a novel pipeline schedule that treats GPUs as stateless workers with round-robin dispatching, priority-aware transfers, event-based synchronization, and an automated partitioning algorithm. All central claims rest on this system design plus direct performance measurements on 1.7B–32B models and one 235B case. No equations, fitted parameters, predictions, or first-principles results are defined in terms of themselves or prior self-citations; the work is self-contained as a systems contribution evaluated empirically.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 3 invented entities

The paper's contributions center on newly designed scheduling and synchronization components whose effectiveness is demonstrated empirically but whose foundational assumptions about hardware behavior and overheads are not fully detailed in the abstract.

axioms (1)
  • domain assumption Dynamic dispatching of pipeline stages maintains model training correctness and convergence
    This is implicitly assumed for the system to be usable.
invented entities (3)
  • RoundPipe schedule no independent evidence
    purpose: Novel pipeline schedule for near-zero-bubble execution
    Introduced as the core contribution of the paper.
  • priority-aware transfer scheduling engine no independent evidence
    purpose: Manages data transfers efficiently between CPU and GPUs
    New component integrated into the system.
  • fine-grained distributed event-based synchronization protocol no independent evidence
    purpose: Ensures correct ordering of computations across devices
    Proposed to support the dynamic scheduling.

pith-pipeline@v0.9.0 · 5573 in / 1352 out tokens · 80712 ms · 2026-05-07T10:33:21.674252+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

61 extracted references · 38 canonical work pages · 13 internal anchors

  1. [1]

    Chenxin An, Shansan Gong, Ming Zhong, Xingjian Zhao, Mukai Li, Jun Zhang, Lingpeng Kong, and Xipeng Qiu. 2023. L-Eval: InstitutingStandardizedEvaluationforLongContextLanguageModels. arXiv:2307.11088 [cs.CL]https://arxiv.org/abs/2307.11088

  2. [2]

    {FlashNeuron}:{SSD- Enabled} {Large-Batch} training of very deep neural networks

    Jonghyun Bae, Jongsung Lee, Yunho Jin, Sam Son, Shine Kim, Hak- beomJang,TaeJunHam,andJaeWLee.2021. {FlashNeuron}:{SSD- Enabled} {Large-Batch} training of very deep neural networks. In 19th USENIX conference on file and storage technologies (FAST 21). 387–401

  3. [3]

    Chi-Chung Chen, Chia-Lin Yang, and Hsiang-Yun Cheng. 2018. Ef- ficient and robust parallel dnn training through model parallelism on multi-gpu platform.arXiv preprint arXiv:1809.02839(2018)

  4. [4]

    Training Deep Nets with Sublinear Memory Cost

    TianqiChen,BingXu,ChiyuanZhang,andCarlosGuestrin.2016.Train- ingDeepNetswithSublinearMemoryCost. arXiv:1604.06174[cs.LG] https://arxiv.org/abs/1604.06174

  5. [5]

    Zihao Chen, Chen Xu, Weining Qian, and Aoying Zhou. 2023. Elastic averaging for efficient pipelined DNN training. InProceedings of the 28th ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming. 380–391

  6. [6]

    Tri Dao. 2023. FlashAttention-2: Faster Attention with Better Par- allelism and Work Partitioning. arXiv:2307.08691 [cs.LG]https: //arxiv.org/abs/2307.08691

  7. [7]

    Fu, Stefano Ermon, Atri Rudra, and Christopher Ré

    Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, and Christopher Ré

  8. [8]

    FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness

    FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness. arXiv:2205.14135 [cs.LG]https://arxiv.org/abs/2205. 14135

  9. [9]

    DeepSeek-AI. 2025. DeepSeek-R1 incentivizes reasoning in LLMs through reinforcement learning.Nature645, 8081 (Sept. 2025), 633–638. doi:10.1038/s41586-025-09422-z

  10. [10]

    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova

  11. [11]

    InProceedings of the 2019 conference of the North American chapter of the association for computational linguistics (NAACL)

    Bert:Pre-trainingofdeepbidirectionaltransformersforlanguage understanding. InProceedings of the 2019 conference of the North American chapter of the association for computational linguistics (NAACL). 4171–4186

  12. [12]

    Shiqing Fan, Yi Rong, Chen Meng, Zongyan Cao, Siyu Wang, Zhen Zheng, Chuan Wu, Guoping Long, Jun Yang, Lixue Xia, et al. 2021. DAPPLE: A pipelined data parallel approach for training large models. InProceedings of the 26th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 431–445

  13. [13]

    Paralleltrainingofpre-trainedmodelsviachunk-based dynamic memory management.IEEE Transactions on Parallel and Distributed Systems34, 1 (2022), 304–315

    Jiarui Fang, Zilin Zhu, Shenggui Li, Hui Su, Yang Yu, Jie Zhou, and YangYou.2022. Paralleltrainingofpre-trainedmodelsviachunk-based dynamic memory management.IEEE Transactions on Parallel and Distributed Systems34, 1 (2022), 304–315

  14. [14]

    Yangyang Feng, Minhui Xie, Zijie Tian, Shuo Wang, Youyou Lu, and Jiwu Shu. 2023. Mobius: Fine Tuning Large-Scale Models on CommodityGPUServers.InProceedingsofthe28thACMInternational Conference on Architectural Support for Programming Languages and OperatingSystems,Volume2(Vancouver,BC,Canada)(ASPLOS2023). Association for Computing Machinery, New York, NY,...

  15. [15]

    R. L. Graham. 1969. Bounds on Multiprocessing Timing Anomalies.SIAM J. Appl. Math.17, 2 (1969), 416–429. arXiv:https://doi.org/10.1137/0117039 doi:10.1137/0117039

  16. [16]

    Lei Guan, Wotao Yin, Dongsheng Li, and Xicheng Lu. 2019. XPipe: Efficientpipelinemodelparallelismformulti-GPUDNNtraining.arXiv preprint arXiv:1911.04610(2019)

  17. [17]

    Aaron Harlap, Deepak Narayanan, Amar Phanishayee, Vivek Se- shadri, Nikhil Devanur, Greg Ganger, and Phil Gibbons. 2018. PipeDream: Fast and Efficient Pipeline Parallel DNN Training. arXiv:1806.03377 [cs.DC]https://arxiv.org/abs/1806.03377

  18. [18]

    GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

    Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Mia Xu Chen, Dehao Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V. Le, YonghuiWu,andZhifengChen.2019.GPipe:EfficientTrainingofGiant NeuralNetworksusingPipelineParallelism. arXiv:1811.06965[cs.CV] https://arxiv.org/abs/1811.06965

  19. [19]

    Arpan Jain, Ammar Ahmad Awan, Asmaa M Aljuhani, Jahanzeb Maq- bool Hashmi, Quentin G Anthony, Hari Subramoni, Dhableswar K Panda, Raghu Machiraju, and Anil Parwani. 2020. Gems: Gpu-enabled memory-aware model-parallelism system for distributed dnn training. InSC20: international conference for high performance computing, networking, storage and analysis. I...

  20. [20]

    Mixtral of Experts

    AlbertQ.Jiang,AlexandreSablayrolles,AntoineRoux,ArthurMensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas,EmmaBouHanna,FlorianBressand,GiannaLengyel,Guillaume Bour,GuillaumeLample,LélioRenardLavaud,LucileSaulnier,Marie- Anne Lachaux, Pierre Stock, Sandeep Subramanian, Sophia Yang, Szymon Antoniak, Teven Le Scao, Théophile Ger...

  21. [21]

    MegaScale-MoE:Large-ScaleCommunication- Efficient Training of Mixture-of-Experts Models in Production

    ChaoJin,ZihengJiang,ZhihaoBai,ZhengZhong,JuncaiLiu,XiangLi, NingxinZheng,XiWang,CongXie,QiHuang,WenHeng,YiyuanMa, Wenlei Bao, Size Zheng, Yanghua Peng, Haibin Lin, Xuanzhe Liu, Xin Jin,andXinLiu.2025. MegaScale-MoE:Large-ScaleCommunication- Efficient Training of Mixture-of-Experts Models in Production. arXiv:2505.11432 [cs.LG]https://arxiv.org/abs/2505.11432

  22. [22]

    Scaling Laws for Neural Language Models

    Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Ben- jamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. 2020. Scaling Laws for Neural Language Models. arXiv:2001.08361 [cs.LG]https://arxiv.org/abs/2001.08361

  23. [23]

    Heehoon Kim, Junyeol Ryu, and Jaejin Lee. 2024. TCCL: Discovering Better Communication Paths for PCIe GPU Clusters. InProceedings of the 29th ACM International Conference on Architectural Support for 12 Programming Languages and Operating Systems, Volume 3(La Jolla, CA, USA)(ASPLOS ’24). Association for Computing Machinery, New York, NY, USA, 999–1015. do...

  24. [24]

    Vijay Korthikanti, Jared Casper, Sangkug Lym, Lawrence McAfee, Michael Andersch, Mohammad Shoeybi, and Bryan Catanzaro. 2022. Reducing Activation Recomputation in Large Transformer Models. arXiv:2205.05198 [cs.LG]https://arxiv.org/abs/2205.05198

  25. [25]

    Joel Lamy-Poirier. 2023. Breadth-First Pipeline Parallelism. arXiv:2211.05953 [cs.DC]https://arxiv.org/abs/2211.05953

  26. [26]

    TingfengLan,YusenWu,BinMa,ZhaoyuanSu,RuiYang,TekinBicer, Masahiro Tanaka, Olatunji Ruwase, Dong Li, and Yue Cheng. 2025. ZenFlow: Enabling Stall-Free Offloading Training via Asynchronous Updates. arXiv:2505.12242 [cs.DC]https://arxiv.org/abs/2505.12242

  27. [27]

    Shigang Li and Torsten Hoefler. 2021. Chimera: efficiently training large-scaleneuralnetworkswithbidirectionalpipelines.InProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis. 1–14

  28. [28]

    Changyue Liao, Mo Sun, Zihan Yang, Jun Xie, Kaiqi Chen, Bin- hang Yuan, Fei Wu, and Zeke Wang. 2024. LoHan: Low-Cost High- Performance Framework to Fine-Tune 100B Model on a Consumer GPU. arXiv:2403.06504 [cs.DC]https://arxiv.org/abs/2403.06504

  29. [29]

    Xinzhang Liu, Chao Wang, Zhihao Yang, Zhuo Jiang, Xuncheng Zhao, Haoran Wang, Lei Li, Dongdong He, Luobin Liu, Kaizhe Yuan, Han Gao, Zihan Wang, Yitong Yao, Sishi Xiong, Wenmin Deng, Haowei He, Kaidong Yu, Yu Zhao, Ruiyu Fang, Yuhao Jiang, Yingyan Li, XiaohuiHu,XiYu,JingqiLi,YanweiLiu,QingliLi,XinyuShi,Junhao Niu, Chengnuo Huang, Yao Xiao, Ruiwen Wang, Fe...

  30. [30]

    arXiv:2512.24157 [cs.CL] https://arxiv.org/abs/2512.24157

    Training Report of TeleChat3-MoE. arXiv:2512.24157 [cs.CL] https://arxiv.org/abs/2512.24157

  31. [31]

    AI @ Meta Llama Team. 2024. The Llama 3 Herd of Models. arXiv:2407.21783 [cs.AI]https://arxiv.org/abs/2407.21783

  32. [32]

    Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, and Hao Wu. 2018. Mixed Precision Training. arXiv:1710.03740 [cs.AI]https://arxiv.org/abs/1710.03740

  33. [33]

    Jordan, and Ion Stoica

    Philipp Moritz, Robert Nishihara, Stephanie Wang, Alexey Tumanov, Richard Liaw, Eric Liang, Melih Elibol, Zongheng Yang, William Paul, Michael I. Jordan, and Ion Stoica. 2018. Ray: A Distributed Framework for Emerging AI Applications. arXiv:1712.05889 [cs.DC] https://arxiv.org/abs/1712.05889

  34. [34]

    DeepakNarayanan,AmarPhanishayee,KaiyuShi,XieChen,andMatei Zaharia. 2021. Memory-Efficient Pipeline-Parallel DNN Training. arXiv:2006.09503 [cs.LG]https://arxiv.org/abs/2006.09503

  35. [35]

    Deepak Narayanan, Mohammad Shoeybi, Jared Casper, Patrick LeGresley, Mostofa Patwary, Vijay Anand Korthikanti, Dmitri Vain- brand, Prethvi Kashinkunti, Julie Bernauer, Bryan Catanzaro, Amar Phanishayee, and Matei Zaharia. 2021. Efficient Large-Scale Language Model Training on GPU Clusters Using Megatron-LM. arXiv:2104.04473 [cs.CL]https://arxiv.org/abs/2104.04473

  36. [36]

    OpenAI. 2025. gpt-oss-120b & gpt-oss-20b Model Card. arXiv:2508.10925 [cs.CL]https://arxiv.org/abs/2508.10925

  37. [37]

    Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wain- wright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. 2022. Training language models to follow instructions with human feedback.Advances in neural information processing systems (NIPS)35 (2022), 27730–27744

  38. [38]

    Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Köpf, Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chil- amkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala

  39. [39]

    PyTorch: An Imperative Style, High-Performance Deep Learning Library

    PyTorch:AnImperativeStyle,High-PerformanceDeepLearning Library. arXiv:1912.01703 [cs.LG]https://arxiv.org/abs/1912.01703

  40. [40]

    Penghui Qi, Xinyi Wan, Guangxing Huang, and Min Lin. 2023. Zero Bubble Pipeline Parallelism. arXiv:2401.10241 [cs.DC]https://arxiv. org/abs/2401.10241

  41. [41]

    Samyam Rajbhandari, Jeff Rasley, Olatunji Ruwase, and Yuxiong He. 2020. ZeRO: Memory Optimizations Toward Training Trillion Parameter Models. arXiv:1910.02054 [cs.LG]https://arxiv.org/abs/ 1910.02054

  42. [42]

    Samyam Rajbhandari, Olatunji Ruwase, Jeff Rasley, Shaden Smith, and Yuxiong He. 2021. ZeRO-Infinity: Breaking the GPU Memory Wall for Extreme Scale Deep Learning. arXiv:2104.07857 [cs.DC] https://arxiv.org/abs/2104.07857

  43. [43]

    JieRen,JiaolinLuo,KaiWu,MinjiaZhang,HyeranJeon,andDongLi

  44. [44]

    In2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA)

    Sentinel: Efficient tensor migration and allocation on heteroge- neous memory systems for deep learning. In2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA). IEEE, 598–611

  45. [45]

    Jie Ren, Samyam Rajbhandari, Reza Yazdani Aminabadi, Olatunji Ruwase, Shuangyan Yang, Minjia Zhang, Dong Li, and Yuxiong He

  46. [46]

    arXiv:2101.06840 [cs.DC]https://arxiv.org/abs/2101.06840

    ZeRO-Offload: Democratizing Billion-Scale Model Training. arXiv:2101.06840 [cs.DC]https://arxiv.org/abs/2101.06840

  47. [47]

    vDNN:Virtualizeddeepneuralnetworksfor scalable, memory-efficient neural network design

    Minsoo Rhu, Natalia Gimelshein, Jason Clemons, Arslan Zulfiqar, and StephenWKeckler.2016. vDNN:Virtualizeddeepneuralnetworksfor scalable, memory-efficient neural network design. In2016 49th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO). IEEE, 1–13

  48. [48]

    Baptiste Roziere, Jonas Gehring, Fabian Gloeckle, Sten Sootla, Itai Gat, Xiaoqing Ellen Tan, Yossi Adi, Jingyu Liu, Romain Sauvestre, Tal Remez, et al. 2023. Code llama: Open foundation models for code. arXiv preprint arXiv:2308.12950(2023)

  49. [49]

    Guangming Sheng, Chi Zhang, Zilingfeng Ye, Xibin Wu, Wang Zhang, Ru Zhang, Yanghua Peng, Haibin Lin, and Chuan Wu. 2025. Hybrid- Flow: A Flexible and Efficient RLHF Framework. InProceedings of the Twentieth European Conference on Computer Systems (EuroSys ’25). ACM, 1279–1297. doi:10.1145/3689031.3696075

  50. [50]

    Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2020. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism. arXiv:1909.08053 [cs.CL]https://arxiv.org/abs/1909.08053

  51. [51]

    StrongHold:fastandaffordablebillion- scale deep learning model training

    XiaoyangSun,WeiWang,ShenghaoQiu,RenyuYang,SongfangHuang, JieXu,andZhengWang.2022. StrongHold:fastandaffordablebillion- scale deep learning model training. InProceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis(Dallas, Texas)(SC ’22). IEEE Press, Article 71, 17 pages

  52. [52]

    QwenTeam.2025.Qwen3TechnicalReport.arXiv:2505.09388[cs.CL] https://arxiv.org/abs/2505.09388

  53. [53]

    Team Wan. 2025. Wan: Open and Advanced Large-Scale Video Generative Models. arXiv:2503.20314 [cs.CV]https://arxiv.org/abs/ 2503.20314

  54. [54]

    Linnan Wang, Jinmian Ye, Yiyang Zhao, Wei Wu, Ang Li, Shuai- wen Leon Song, Zenglin Xu, and Tim Kraska. 2018. Superneurons: Dynamic GPU memory management for training deep neural networks. InProceedings of the 23rd ACM SIGPLAN symposium on principles and practice of parallel programming. 41–53

  55. [55]

    Samuel Williams, Andrew Waterman, and David Patterson. 2009. Roofline:aninsightfulvisualperformancemodelformulticorearchitec- tures.Commun. ACM52, 4 (April 2009), 65–76. doi:10.1145/1498765. 1498785

  56. [56]

    Bowen Yang, Jian Zhang, Jonathan Li, Christopher Ré, Christopher Aberger, and Christopher De Sa. 2021. Pipemare: Asynchronous pipeline parallel dnn training.Proceedings of Machine Learning and 13 Table 1.Summary of notations. Notation Meaning 𝑠sequence length 𝑏micro-batch size ℎhidden dimension 𝑚intermediate dimension in MLP 𝑎number of attention heads 𝑘nu...

  57. [57]

    PengCheng Yang, Xiaoming Zhang, Wenpeng Zhang, Ming Yang, and Hong Wei. 2022. Group-based interleaved pipeline parallelism for large-scale DNN training. InInternational Conference on Learning Representations

  58. [58]

    Haoyang Zhang, Yirui Zhou, Yuqi Xue, Yiqi Liu, and Jian Huang

  59. [59]

    InProceedings of the 56th Annual IEEE/ACM International Symposium on Microarchitecture

    G10: Enabling an efficient unified gpu memory and storage architecture with smart tensor migrations. InProceedings of the 56th Annual IEEE/ACM International Symposium on Microarchitecture. 395–410

  60. [60]

    Pinxue Zhao, Hailin Zhang, Fangcheng Fu, Xiaonan Nie, Qibin Liu, FangYang,YuanboPeng,DianJiao,ShuaipengLi,JinbaoXue,Yangyu Tao, and Bin Cui. 2025. MEMO: Fine-grained Tensor Management For Ultra-long Context LLM Training.Proceedings of the ACM on Management of Data3, 1 (Feb. 2025), 1–28. doi:10.1145/3709703

  61. [61]

    Zan Zong, Li Lin, Leilei Lin, Lijie Wen, and Yu Sun. 2023. Str: Hybrid tensor re-generation to break memory wall for dnn training. IEEE Transactions on Parallel and Distributed Systems34, 8 (2023), 2403–2418. A Summary of notations Table 1 defines the notation used in the appendix. Table 2 shows the hardware specs of GPUs used in the following analysis. T...