pith. machine review for the scientific record. sign in

arxiv: 2605.04842 · v1 · submitted 2026-05-06 · 💻 cs.DC

Recognition: unknown

Communication Offloading on SmartNIC DPUs: A Quantitative Approach

Andong Hu, Ivy Peng, Jacob Wahlgren, Maya Gokhale, Roger Pearce

Authors on Pith no claims yet

Pith reviewed 2026-05-08 17:20 UTC · model grok-4.3

classification 💻 cs.DC
keywords SmartNIC DPUcommunication offloadingfire-and-forget modelBuddy enginememory-to-communication ratioperformance predictionDRAM trafficDirect Cache Access
0
0 comments X

The pith

Offloading communication tasks to SmartNIC DPUs speeds up host-dominated workloads by up to 1.55x when the memory-to-communication ratio is high.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper tests whether SmartNIC DPUs can handle the message-routing part of an asynchronous fire-and-forget communication model without hurting overall performance. It introduces the Buddy engine to move those tasks off the host CPU and onto the DPU or a spare CPU core. Evaluation across five applications shows that the ratio of memory work to communication work predicts whether offloading helps, with clear gains in host-heavy codes such as Quicksilver and sparse-matrix transpose. The same experiments also expose a hardware shortcoming: without direct cache access, DRAM traffic rises sharply. Readers should care because the work supplies a simple, measurable rule for deciding when DPU offload is worthwhile and flags a concrete design change needed in future SmartNIC hardware.

Core claim

The memory-to-communication ratio reliably indicates when the Buddy offloading engine, running on a SmartNIC DPU such as the Nvidia BlueField-3, will deliver net speedup by decoupling message routing from the application process. In host-dominated workloads the speedup reaches 1.55x; across the tested set the absence of Direct Cache Access support on the DPU produces a 625x rise in DRAM traffic, demonstrating both the practical value of the offload strategy and the hardware feature required for its full effectiveness.

What carries the argument

The Buddy engine, which decouples the fire-and-forget message-routing service from the main application thread and executes it on either a DPU or a generic x86 core.

If this is right

  • Workloads whose memory operations greatly exceed communication volume obtain measurable CPU relief by moving routing to the DPU.
  • The memory-to-communication ratio supplies a lightweight decision metric for choosing when to activate offload.
  • SmartNIC designs that add Direct Cache Access support would eliminate the large DRAM-traffic penalty observed here.
  • Buddy's portable implementation allows the same offload logic to run on spare host cores when a DPU is unavailable.
  • Asynchronous fire-and-forget communication models become more attractive once the ratio-based predictor is applied.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If the ratio predictor proves stable, runtime systems could insert automatic offload decisions without programmer intervention.
  • The observed DRAM-traffic explosion suggests that cache-coherence features will become a gating item for wider DPU adoption in memory-intensive codes.
  • Repeating the experiments on additional DPUs or on applications drawn from other domains would test whether the ratio rule generalizes beyond the five codes studied.
  • The same decoupling approach could be applied to other network services such as collective operations or RDMA request handling.

Load-bearing premise

The memory-to-communication ratio accurately forecasts offloading benefit and the overhead of running Buddy on the DPU stays low enough to produce net gains in real applications.

What would settle it

Measuring the same five applications after offloading communication to the DPU and finding either no speedup in the high-ratio cases or a DRAM-traffic increase far below 625x would falsify the central claims.

Figures

Figures reproduced from arXiv: 2605.04842 by Andong Hu, Ivy Peng, Jacob Wahlgren, Maya Gokhale, Roger Pearce.

Figure 1
Figure 1. Figure 1: An overview of Nvidia BlueField-3 DPU architecture. NIC by integrating programmable logic (e.g. ASIC/FPGA) on the data path, enabling host-side workloads to be offloaded to the NIC. Data Processing Units (DPUs) go one step further, featuring general-purpose CPUs and a standalone Linux software stack. Several vendors provide commercial DPU solutions with ARM cores, including AMD Pensando, Intel IPU, and Nvi… view at source ↗
Figure 2
Figure 2. Figure 2: An overview of the Buddy design, where the routing agent interfaces between local processes and remote nodes to aggregate individual messages. Overview. Frequent irregular communication is a challenge in HPC systems since it leads to many small inefficient messages across the network. Message aggregation combined with multi-hop routing, a core component for enabling the “fire-and-forget” model, provides hi… view at source ↗
Figure 3
Figure 3. Figure 3: Three offloading scenarios with arrows representing RDMA data transfer. We use a set of five applications to evaluate communication offloading. They come from a variety of domains and feature different communication patterns as shown in view at source ↗
Figure 4
Figure 4. Figure 4: Speedup compared to no offloading. Histogram Quicksilver SSSP Transpose Tri. Count. 0 25 50 75 Network util. (%) DPU x86 None view at source ↗
Figure 7
Figure 7. Figure 7: Impact of tuning and optimizations on application performance. scenarios, although it deviates slightly for None in Histogram. Quicksilver has by far the highest ratio around 72, explaining the low network utilization we observed, followed by Sparse Transpose with 2.3 and Histogram with 1.5. SSSP and Triangle Counting have more communication than memory traffic with ratios of 0.56 and 0.53, respectively. W… view at source ↗
Figure 8
Figure 8. Figure 8: Performance of scaling Histogram and Quicksilver with Dpu offloading up to 8 nodes and the average transfer sizes. around 8 buffers with 8.0x and 12.4x, respectively, while other applications were not significantly affected.Utilizing multiple threads in the routing agent improves performance for most applications, with up to 6.2x speedup at 7 or 8 threads in Histogram. Triangle Counting peaks at 5 threads … view at source ↗
Figure 10
Figure 10. Figure 10: Amount of data loaded from DRAM in the routing agent (log scale). Intel Gold BlueField-3 Grace 0 20 40 Memory loads (% of comm.) view at source ↗
read the original abstract

SmartNIC Data Processing Units (DPUs) offer a promising solution for saving high-end CPU resources by offloading tasks to programmable cores near the network interface. In this work, we explore the feasibility of SmartNIC DPUs in supporting an asynchronous communication model called "fire-and-forget", particularly its core message routing service. We design a communication offloading engine called Buddy that decouples communication tasks from the application process. Buddy runs flexibly on SmartNIC DPUs such as the Nvidia BlueField-3 DPU and generic x86 CPUs. Our evaluation results in five applications identify the memory-to-communication ratio as a key predictor of the offloading performance. Host-dominated workloads, such as Quicksilver and Sparse Matrix Transpose, achieved up to 1.55x speedup with communication offloaded to the DPU. We further identify a 625x increase in DRAM traffic due to the absence of Direct Cache Access support on the DPU, highlighting a critical need in future SmartNIC designs.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces Buddy, a flexible communication offloading engine for SmartNIC DPUs (e.g., Nvidia BlueField-3) that decouples asynchronous 'fire-and-forget' message routing from the host application process. Evaluation across five applications identifies the memory-to-communication ratio as the key predictor of offloading benefit; host-dominated workloads (Quicksilver, Sparse Matrix Transpose) achieve up to 1.55x speedup when offloaded to the DPU, while the absence of Direct Cache Access on the DPU produces a 625x increase in DRAM traffic.

Significance. If the empirical results hold after improved controls, the work supplies concrete, workload-characteristic guidance for deciding when communication offloading to DPUs is profitable and flags a concrete hardware limitation (missing DCA) that future SmartNIC designs must address. The quantitative framing and real-hardware measurements on a production DPU are strengths that could influence both systems software and hardware roadmaps in distributed and high-performance computing.

major comments (2)
  1. [Evaluation] Evaluation section: the central claim that the memory-to-communication ratio reliably predicts offloading benefit rests on results from only five applications without reported isolation of DPU kernel execution time versus host savings, without a CPU-based offload baseline, and without controls for message-size distribution or cache behavior. Consequently the 1.55x speedup figures for Quicksilver and Sparse Matrix Transpose cannot yet be attributed causally to the ratio rather than to unmeasured confounding factors.
  2. [Evaluation] Evaluation section: no error bars, standard deviations, or run counts are supplied for any speedup or traffic number (including the 625x DRAM-traffic increase), and baseline details plus exclusion criteria are absent. This directly weakens the soundness of the quantitative claims that the abstract presents as concrete evidence.
minor comments (2)
  1. [Abstract] Abstract: the phrase 'five applications' should be followed by the explicit list of workloads (Quicksilver, Sparse Matrix Transpose, …) so readers immediately see the sampling scope.
  2. [Evaluation] The manuscript would benefit from a short table summarizing per-application memory-to-communication ratios together with the observed speedups; this would make the predictive claim visually testable.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback on our evaluation. We address each major comment below and indicate the changes we will make to strengthen the manuscript.

read point-by-point responses
  1. Referee: Evaluation section: the central claim that the memory-to-communication ratio reliably predicts offloading benefit rests on results from only five applications without reported isolation of DPU kernel execution time versus host savings, without a CPU-based offload baseline, and without controls for message-size distribution or cache behavior. Consequently the 1.55x speedup figures for Quicksilver and Sparse Matrix Transpose cannot yet be attributed causally to the ratio rather than to unmeasured confounding factors.

    Authors: We agree that isolating DPU kernel time from host savings and adding a CPU offload baseline would improve causal attribution. Buddy supports execution on both DPU and CPU; we will add a direct CPU baseline comparison in the revision. We will also include per-workload breakdowns of DPU versus host execution time for Quicksilver and Sparse Matrix Transpose. The five applications were chosen to cover a spectrum of memory-to-communication ratios, and we will expand the text to detail how message-size distributions and cache behavior were controlled or measured. We will qualify the predictive claim to note that it is supported by observed trends across the tested workloads rather than proven in isolation. revision: partial

  2. Referee: Evaluation section: no error bars, standard deviations, or run counts are supplied for any speedup or traffic number (including the 625x DRAM-traffic increase), and baseline details plus exclusion criteria are absent. This directly weakens the soundness of the quantitative claims that the abstract presents as concrete evidence.

    Authors: We acknowledge the omission. All reported speedups and traffic figures are averages over at least five runs per configuration; we will add error bars, standard deviations, and explicit run counts to the revised figures and tables. We will also expand the evaluation section with full baseline hardware/software configurations and any data exclusion criteria to improve reproducibility. revision: yes

Circularity Check

0 steps flagged

No circularity: results from direct hardware measurements

full rationale

The paper reports an empirical quantitative study of the Buddy offloading engine on real SmartNIC DPUs (Nvidia BlueField-3) and x86 CPUs across five applications. The memory-to-communication ratio is identified post hoc as a predictor solely from observed performance data and speedups (e.g., 1.55x on Quicksilver and Sparse Matrix Transpose) measured on hardware; no equations, fitted parameters, or first-principles derivations are presented that could reduce to inputs by construction. The 625x DRAM traffic observation is likewise a direct hardware measurement. No self-citations, uniqueness theorems, or ansatzes are invoked to justify core claims. The evaluation is self-contained against external benchmarks via real-system testing rather than any closed derivation loop.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The work rests on standard hardware assumptions about DPU programmability and application communication patterns; no free parameters or invented entities are introduced.

axioms (1)
  • domain assumption The DPU can execute the Buddy communication engine with low enough overhead to produce net performance gains in host-dominated workloads.
    Invoked when claiming 1.55x speedups from offloading.

pith-pipeline@v0.9.0 · 5476 in / 1260 out tokens · 23818 ms · 2026-05-08T17:20:32.066893+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Post-Moore Technologies for Plasma Simulation: A Community Roadmap

    cs.ET 2026-05 unverdicted novelty 4.0

    No single post-Moore technology replaces current HPC for plasma simulations, but FPGA-class accelerators offer near-term kernel offload, non-von Neumann architectures medium-term operator acceleration, and quantum com...

Reference graph

Works this paper leans on

18 extracted references · cited by 1 Pith paper

  1. [1]

    In: 2020 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS) (2020)

    Alian, M., Yuan, Y., Zhang, J., Wang, R., Jung, M., Kim, N.S.: Data direct I/O characterization for future I/O system exploration. In: 2020 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS) (2020)

  2. [2]

    In: International Conference on High Performance Computing (2021)

    Bayatpour, M., Sarkauskas, N., Subramoni, H., Maqbool Hashmi, J., Panda, D.K.: BluesMPI: Efficient MPI non-blocking Alltoall offloading designs on modern Blue- Field smart NICs. In: International Conference on High Performance Computing (2021)

  3. [3]

    In: Proceedings of the 48th International Conference on Parallel Processing (2019)

    Brock, B., Buluç, A., Yelick, K.: BCL: A cross-platform distributed data structures library. In: Proceedings of the 48th International Conference on Parallel Processing (2019)

  4. [4]

    In: 2020 USENIX Annual Technical Conference (USENIX ATC 20) (2020)

    Farshin, A., Roozbeh, A., Maguire Jr, G.Q., Kostić, D.: Reexamining direct cache access to optimize I/O intensive applications for multi-hundred-gigabit networks. In: 2020 USENIX Annual Technical Conference (USENIX ATC 20) (2020)

  5. [5]

    In: Proceedings of the 2006 ACM/IEEE conference on Supercomputing (2006)

    Garg, R., Sabharwal, Y.: Software routing and aggregation of messages to optimize the performance of HPCC Randomaccess benchmark. In: Proceedings of the 2006 ACM/IEEE conference on Supercomputing (2006)

  6. [6]

    In: Proceedings of the 2024 SIGCOMM Workshop on Networks for AI Computing (2024)

    Gu, T., Fei, J., Canini, M.: OmNICCL: Zero-cost sparse AllReduce with direct cache access and SmartNICs. In: Proceedings of the 2024 SIGCOMM Workshop on Networks for AI Computing (2024)

  7. [7]

    In: 2022 IEEE International Parallel and Distributed Processing Symposium (IPDPS) (2022)

    Karamati, S., Hughes, C., Hemmert, K.S., Grant, R.E., Schonbein, W.W., Levy, S., Conte, T.M., Young, J., Vuduc, R.W.: “Smarter” NICs for faster molecular dynam- ics: a case study. In: 2022 IEEE International Parallel and Distributed Processing Symposium (IPDPS) (2022)

  8. [8]

    In: 15th Annual IEEE Symposium on High- Performance Interconnects (HOTI 2007) (2007)

    León, E.A., Ferreira, K.B., Maccabe, A.B.: Reducing the impact of the memory wall for I/O using cache injection. In: 15th Annual IEEE Symposium on High- Performance Interconnects (HOTI 2007) (2007)

  9. [9]

    In: 2024 IEEE International Parallel and Distributed Processing Symposium (IPDPS)

    Li, Y., Kashyap, A., Chen, W., Guo, Y., Lu, X.: Accelerating lossy and lossless compression on emerging bluefield dpu architectures. In: 2024 IEEE International Parallel and Distributed Processing Symposium (IPDPS). pp. 373–385 (2024)

  10. [10]

    In: 2019 IEEE/ACM 9th Workshop on Irregular Applications: Architectures and Algorithms (IA3) (2019)

    Maley, F.M., DeVinney, J.G.: Conveyors for streaming many-to-many communica- tion. In: 2019 IEEE/ACM 9th Workshop on Irregular Applications: Architectures and Algorithms (IA3) (2019)

  11. [11]

    IEEE Computer Architecture Letters (2025)

    Mamandipoor, A., Tran, H.D., Alian, M.: SDT: Cutting datacenter tax through simultaneous data-delivery threads. IEEE Computer Architecture Letters (2025)

  12. [12]

    In: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (2023)

    Steil, T., Reza, T., Priest, B., Pearce, R.: Embracing irregular parallelism in HPC with YGM. In: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (2023)

  13. [13]

    In: 2023 IEEE International Parallel and Distributed Processing Symposium (IPDPS) (2023)

    Suresh, K.K., Michalowicz, B., Ramesh, B., Contini, N., Yao, J., Xu, S., Shafi, A., Subramoni, H., Panda, D.: A novel framework for efficient offloading of commu- nication operations to BlueField SmartNICs. In: 2023 IEEE International Parallel and Distributed Processing Symposium (IPDPS) (2023)

  14. [14]

    Future Generation Computer Sys- tems (2025)

    Tibbetts, N., Ibtisum, S., Puri, S.: A survey on heterogeneous computing using SmartNICs and emerging data processing units. Future Generation Computer Sys- tems (2025)

  15. [15]

    In: International Conference on High Performance Com- puting (2021) Communication Offloading on SmartNIC DPUs: A Quantitative Approach 15

    Tithi, J.J., Petrini, F., Richards, D.F.: Lessons learned from accelerating Quicksil- ver on programmable integrated unified memory architecture (PIUMA) and how that’s different from CPU. In: International Conference on High Performance Com- puting (2021) Communication Offloading on SmartNIC DPUs: A Quantitative Approach 15

  16. [16]

    In: Proceedings of the In- ternational Conference for High Performance Computing, Networking, Storage and Analysis (2025)

    Usman, M., Benito, M., Iserte, S., Peña, A.J.: ODOS-MPI: HPC-friendly Smart- NIC offloading of computation/communication kernels. In: Proceedings of the In- ternational Conference for High Performance Computing, Networking, Storage and Analysis (2025)

  17. [17]

    In: Proceedings of the SC’23 Workshops of The International Con- ference on High Performance Computing, Network, Storage, and Analysis (2023)

    Usman,M.,Iserte,S.,Ferrer,R.,Peña,A.J.:DPUoffloadingprogrammingwiththe OpenMP API. In: Proceedings of the SC’23 Workshops of The International Con- ference on High Performance Computing, Network, Storage, and Analysis (2023)

  18. [18]

    Proceedings of the ACM on Measurement and Analysis of Computing Systems6(1) (2022)

    Wang, M., Xu, M., Wu, J.: Understanding I/O direct cache access performance for end host networking. Proceedings of the ACM on Measurement and Analysis of Computing Systems6(1) (2022)