pith. machine review for the scientific record. sign in

arxiv: 2605.07599 · v1 · submitted 2026-05-08 · 💻 cs.DC · cs.ET

Recognition: no theorem link

Stencil Computations on Tenstorrent Wormhole

Lorenzo Piarulli , Daniele De Sensi

Authors on Pith no claims yet

Pith reviewed 2026-05-11 02:07 UTC · model grok-4.3

classification 💻 cs.DC cs.ET
keywords stencil computationAI acceleratorTenstorrent WormholeHPC kernelsenergy efficiencyRISC-Vdataflow architectureperformance profiling
0
0 comments X

The pith

Stencil computations map to the Tenstorrent Wormhole with kernel times matching CPU but full runs slowed by transfers and setup.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper tests whether AI dataflow accelerators can run traditional scientific kernels by adapting a 2D five-point stencil to the Tenstorrent Wormhole. It creates two versions: Axpy breaks the stencil into element-wise submatrix steps while MatMul recasts the whole operation as a matrix multiplication. End-to-end the CPU stays three times faster, yet the accelerator's isolated kernel time is comparable once profiling separates out PCIe moves, device startup, and host preprocessing. For large inputs the Axpy version also consumes less energy than the CPU baseline. The work uses these measurements to name concrete hardware and software limits that currently block wider use of such accelerators for HPC tasks.

Core claim

We investigate the mapping of 2D 5-point stencil computations onto the Tenstorrent Wormhole, a RISC-V AI dataflow accelerator. We develop two heterogeneous implementations: Axpy, which decomposes the stencil into element-wise submatrix operations, and MatMul, which reformulates it as a matrix multiplication. While the CPU baseline remains 3x faster end-to-end, profiling reveals that the isolated Wormhole kernel is competitive with CPU execution, with the gap driven by PCIe transfers, device initialization, and host-side preprocessing. Despite slower runtime, Axpy achieves lower energy consumption than the CPU baseline for large inputs.

What carries the argument

The Axpy and MatMul heterogeneous implementations that adapt the five-point stencil to the Wormhole RISC-V dataflow architecture.

If this is right

  • Reducing PCIe, initialization, and host preprocessing costs would make the Wormhole kernel competitive end-to-end for stencil workloads.
  • Energy advantages in the Axpy mapping could favor the accelerator for large problems even if runtime is not yet superior.
  • Detailed profiling of data-movement and setup overheads supplies concrete targets for hardware and software changes.
  • The same decomposition techniques could extend to other grid-based scientific kernels on similar accelerators.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The overhead pattern observed here would likely appear in other AI accelerators unless they integrate memory more tightly with the host.
  • Testing the same mappings on larger grids or three-dimensional stencils would show whether the competitiveness holds at scale.
  • If the main limiter is data movement, on-device memory or direct host integration would close most of the end-to-end gap.

Load-bearing premise

That the Axpy and MatMul versions are efficient and fair mappings of the stencil and that the chosen CPU baseline and measurement method permit direct comparison.

What would settle it

A side-by-side run of the identical stencil on Wormhole and CPU where data already resides in accelerator memory with no PCIe transfers or host preprocessing, measuring only kernel time and energy.

Figures

Figures reproduced from arXiv: 2605.07599 by Daniele De Sensi, Lorenzo Piarulli.

Figure 1
Figure 1. Figure 1: Tenstorrent Wormhole architecture. The 10×12 grid contains 64 Tensix cores (T), DRAM controllers (D), Eth￾ernet interfaces (E), PCIe controller, and ARC management core (image from [8]). the stencil accesses neighboring elements that may span different tiles. 3.2 Tenstorrent Wormhole Architecture Tenstorrent Wormhole is a RISC-V AI accelerator available as a PCIe card (n150d/n300d variants). The architectu… view at source ↗
Figure 2
Figure 2. Figure 2: Tensix core architecture. Five baby RISC-V cores coordinate NoC data movement (Router 0/1), computation (Compute block with matrix and vector engines), and local L1 memory access (image from [15]) [PITH_FULL_IMAGE:figures/full_fig_p003_2.png] view at source ↗
Figure 5
Figure 5. Figure 5: Execution time comparison between Axpy and MatMul heterogeneous implementations (log scale). Axpy is approximately 75× faster across all configurations. matrices must then be aligned to 32×32 tiles by padding: the column vector is padded to 32 × 1 and replicated 32 times to form a 32 × 32 tile; the input matrix rows are padded from 9 to 32 columns. Finally, tilize_nfaces() converts the row-major data to Wo… view at source ↗
Figure 4
Figure 4. Figure 4: MatMul pipeline. The padded input is converted to stencil-to-row format, the stencil kernel is flattened to a column vector, both are aligned to 32×32 tiles via padding, and the tilized data is sent to Wormhole for matrix multipli￾cation. the total number of tiles is 𝑇 = (𝑁/32) 2 , and each core re￾ceives ⌈𝑇 /64⌉ tiles. For each tile index 𝑘, the compute kernel executes: out32×32 𝑘 = 0.2532×32 ⊙ [PITH_FUL… view at source ↗
Figure 6
Figure 6. Figure 6: Execution time breakdown by phase. Top: Axpy shows balanced distribution across CPU, memcpy, and Wormhole. Bottom: MatMul is dominated by CPU-side tiling conversions (≈90%) [PITH_FULL_IMAGE:figures/full_fig_p005_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Execution time comparison between Axpy and CPU baseline implementations. consistent with the quadratic growth of the stencil-to-row expansion in MatMul. The phase breakdown in [PITH_FULL_IMAGE:figures/full_fig_p005_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Theoretical performance under UVM (top, 900 GB/s NVLink-C2C bandwidth with 450 GB/s per￾direction) and UPM (bottom, zero transfer overhead) sce￾narios compared to the CPU baseline. UPM model. We model complete memory unification, as in AMD MI300A [3], where CPU and accelerator share coher￾ent physical memory. This eliminates transfers entirely and also removes the need for tiling functions (tilize/untilize… view at source ↗
read the original abstract

As investment in AI-focused accelerators grows and their deployment in supercomputing facilities expands, understanding whether these architectures can efficiently support traditional scientific kernels is critical for the future of High-Performance Computing. We investigate the mapping of 2D 5-point stencil computations onto the Tenstorrent Wormhole, a RISC-V AI dataflow accelerator. We develop two heterogeneous implementations: Axpy, which decomposes the stencil into element-wise submatrix operations, and MatMul, which reformulates it as a matrix multiplication. While the CPU baseline remains 3x faster end-to-end, profiling reveals that the isolated Wormhole kernel is competitive with CPU execution, with the gap driven by PCIe transfers, device initialization, and host-side preprocessing. Despite slower runtime, Axpy achieves lower energy consumption than the CPU baseline for large inputs. Through detailed profiling and theoretical analysis, we identify key architectural and software limitations of the current platform and outline concrete hardware and software directions that could make AI accelerators competitive for HPC workloads.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper maps 2D 5-point stencil computations to the Tenstorrent Wormhole RISC-V AI accelerator via two heterogeneous implementations: Axpy (decomposing into element-wise submatrix operations) and MatMul (reformulating as matrix multiplication). It reports that a CPU baseline remains 3x faster end-to-end, but the isolated Wormhole kernel is competitive, with the gap driven by PCIe transfers, device initialization, and host-side preprocessing. Axpy achieves lower energy consumption than the CPU baseline for large inputs, supported by profiling and theoretical analysis that identify architectural and software limitations and suggest hardware/software directions for competitiveness in HPC workloads.

Significance. If the performance attribution and energy results hold, the work contributes empirical evidence on adapting AI dataflow accelerators to traditional scientific kernels, with the separation of kernel vs. overhead times and the theoretical analysis providing concrete guidance for future platform improvements. The reproducible profiling approach and focus on energy efficiency for large inputs are strengths that could inform supercomputing deployments.

major comments (2)
  1. [Abstract] Abstract: The central claim that 'Axpy achieves lower energy consumption than the CPU baseline for large inputs' is load-bearing for the efficiency argument, yet the text does not specify the energy measurement methodology (e.g., device-only sensors/APIs for Wormhole vs. full-system or package-level tools such as RAPL for CPU) or confirm that host-side preprocessing and PCIe costs are accounted equivalently on both platforms. This risks the reported advantage being an artifact of inconsistent accounting rather than architectural merit.
  2. [Abstract] The manuscript's soundness assessment notes the absence of full methods, error bars, or raw data, which directly impacts verification of the performance gap attribution (PCIe, initialization, host preprocessing) and the competitiveness of the isolated kernel. Without these, the claim that the Wormhole kernel is competitive cannot be fully evaluated against the CPU baseline.
minor comments (2)
  1. The abstract and profiling description would benefit from explicit cross-references to the sections detailing the Axpy and MatMul implementations, including any pseudocode or kernel launch parameters.
  2. Clarify the stencil size, grid dimensions, and input scales used for the 'large inputs' energy comparison to allow direct reproduction.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the careful and constructive review. The comments correctly identify areas where greater methodological transparency is needed to support the claims. We address each point below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim that 'Axpy achieves lower energy consumption than the CPU baseline for large inputs' is load-bearing for the efficiency argument, yet the text does not specify the energy measurement methodology (e.g., device-only sensors/APIs for Wormhole vs. full-system or package-level tools such as RAPL for CPU) or confirm that host-side preprocessing and PCIe costs are accounted equivalently on both platforms. This risks the reported advantage being an artifact of inconsistent accounting rather than architectural merit.

    Authors: We agree that the abstract does not explicitly describe the energy measurement methodology, which is necessary for proper interpretation. In the full manuscript, Wormhole energy is obtained from on-device power sensors and APIs, while CPU energy is measured at the package level using RAPL. Both platforms incorporate host-side preprocessing, PCIe transfers, and initialization costs into the reported energy figures by capturing the full end-to-end execution. To eliminate any ambiguity, we will revise the abstract to state the measurement approach concisely and add a dedicated methods subsection that details the tools, equivalence of accounting, and any platform-specific adjustments. revision: yes

  2. Referee: [Abstract] The manuscript's soundness assessment notes the absence of full methods, error bars, or raw data, which directly impacts verification of the performance gap attribution (PCIe, initialization, host preprocessing) and the competitiveness of the isolated kernel. Without these, the claim that the Wormhole kernel is competitive cannot be fully evaluated against the CPU baseline.

    Authors: We accept that the current manuscript lacks sufficient detail on experimental methods, error bars, and public raw data, which hinders independent verification of the overhead attributions and kernel competitiveness. The profiling breakdowns and theoretical analysis are present, but we will expand the methods section with complete hardware/software configurations, add error bars (standard deviation across repeated runs) to all performance and energy plots, and release the raw measurement data together with analysis scripts in a public repository upon acceptance. These changes will enable full evaluation of the reported results. revision: yes

Circularity Check

0 steps flagged

Purely empirical performance study with no derivations or self-referential claims

full rationale

The paper reports direct measurements of runtime, energy consumption, and profiling data for two heterogeneous stencil implementations (Axpy and MatMul) on the Wormhole accelerator, compared against an external CPU baseline. No equations, predictions, fitted parameters, or first-principles derivations are presented that could reduce to their own inputs. The abstract and description reference theoretical analysis only in the context of identifying architectural limitations from observed data, not as load-bearing self-referential logic. Energy claims rest on reported measurements rather than any constructed equivalence. This is a standard empirical HPC benchmarking study with no circularity.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

No free parameters, new axioms, or invented entities are introduced; the work rests on standard definitions of 5-point stencil and hardware performance counters.

pith-pipeline@v0.9.0 · 5466 in / 1121 out tokens · 30012 ms · 2026-05-11T02:07:46.272880+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

17 extracted references · 17 canonical work pages

  1. [1]

    Dissecting the Tenstorrent Blackhole Architecture via Mi- crobenchmarking.https://asplos.dev/wordpress/wp-content/uploads/ 2025/09/TT_bench-1.pdf

    2025. Dissecting the Tenstorrent Blackhole Architecture via Mi- crobenchmarking.https://asplos.dev/wordpress/wp-content/uploads/ 2025/09/TT_bench-1.pdf. Accessed: 2025-11-01

  2. [2]

    Giorgio Amati, Matteo Turisini, Andrea Monterubbiano, Mattia Pal- adino, Elisabetta Boella, Daniele Gregori, and Danilo Croce. 2025. Ac- celerating Gravitational N-Body Simulations Using the RISC-V-Based Tenstorrent Wormhole. InProceedings of the SC’25 Workshops of the International Conference for High Performance Computing, Networking, Storage and Analys...

  3. [3]

    AMD Corporation. 2023. AMD Instinct MI300A APU Product Overview.https://www.amd.com/en/products/accelerators/instinct/ mi300a.html. Accessed: 2025-10-05

  4. [4]

    Nick Brown and Ryan Barton. 2024. Accelerating Stencils on the Tenstorrent Grayskull RISC-V Accelerator. InSC24-W: Workshops of the International Conference for High Performance Computing, Networking, Storage and Analysis. IEEE, 1690–1700. doi:10.1109/SCW63240.2024. 00211

  5. [5]

    Nick Brown, Jake Davies, and Felix Le Clair. 2025. Exploring Fast Fourier Transforms on the Tenstorrent Wormhole. InISC 2025 Work- shops (Lecture Notes in Computer Science). Springer. arXiv:2506.15437

  6. [6]

    Z. Cai, R. Giordano, et al. 2023. Assessing Tenstorrent’s RISC-V Matrix Multiply Acceleration.arXiv preprint arXiv:2305.10314(2023)

  7. [7]

    Yuetao Chen, Kun Li, Yuhao Wang, Donglin Bai, Lei Wang, Lingxiao Ma, Liang Yuan, Yunquan Zhang, Ting Cao, and Mao Yang. 2024. ConvStencil: Transform Stencil Computation to Matrix Multiplication on Tensor Cores. InProceedings of the 29th ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming (PPoPP ’24). ACM, 333–347

  8. [8]

    corsix. 2024. Community Highlight: Tenstorrent Wormhole Series Part 2: Which disabled rows? Tenstorrent News- room.https://tenstorrent.com/newsroom/community-highlight- tenstorrent-wormhole-series-part-2-which-disabled-rowsAccessed: 2024-11-18

  9. [9]

    Giannozzi, O

    P. Giannozzi, O. Baseggio, P. Bonfà, D. Brunato, R. Car, I. Carnimeo, C. Cavazzoni, S. De Gironcoli, P. Delugas, F. Ferrari Ruffino, et al. 2020. Quantum ESPRESSO Toward the Exascale.The Journal of Chemical Physics152, 15 (2020)

  10. [10]

    NVIDIA Corporation. 2023. NVIDIA GH200 Grace Hopper Su- perchip.https://www.nvidia.com/en-us/data-center/grace-hopper- superchip/. Accessed: 2025-10-05

  11. [11]

    Schmidhuber

    J. Schmidhuber. 2015. Deep Learning in Neural Networks: An Overview.Neural Networks61 (2015), 85–117

  12. [12]

    Bartosz Taudul. 2024. Tracy Profiler.https://github.com/wolfpld/ tracy

  13. [13]

    Tenstorrent. 2025. Blackhole Accelerator Specifications.https://docs. tenstorrent.com/aibs/blackhole/specifications.html. Accessed: 2025- 11-01

  14. [14]

    Tenstorrent. 2025. Wormhole Accelerator Specifications.https:// docs.tenstorrent.com/aibs/wormhole/specifications.html. Accessed: 2025-09-11

  15. [15]

    2024.Tensix Neo: RISC-V-based AI IP For Extraordinary AI Performance

    Tenstorrent Inc. 2024.Tensix Neo: RISC-V-based AI IP For Extraordinary AI Performance. Tenstorrent.https://tenstorrent.com/ip/tensix-neo Accessed: 2024-11-18

  16. [16]

    Jasmina Vasiljevic and Davor Capalija. 2024. Blackhole & TT- Metalium: The Standalone AI Computer and Its Programming Model. InHot Chips 36

  17. [17]

    A. Zeni, G. Guidi, M. Ellis, N. Ding, M. D. Santambrogio, S. Hofmeyr, A. Buluç, L. Oliker, and K. Yelick. 2020. Logan: High-Performance GPU-Based X-Drop Long-Read Alignment. In2020 IEEE International Parallel and Distributed Processing Symposium (IPDPS). IEEE, 462–471