pith. machine review for the scientific record. sign in

arxiv: 2605.00319 · v1 · submitted 2026-05-01 · 💻 cs.AR

Recognition: unknown

A PVT-Resilient Subthreshold SRAM-Based In-Memory Computing Accelerator with In-Situ Regulation for Energy-Efficient Spiking Neural Networks

Bing-Han Liu, Chien-Nan Liu, Hung-Ming Chen, I-Wen Wang, Shih-Hang Kao, Shyh-Jye Jou, Tian-Sheuan Chang, Wei-Zen Chen, Yang-Chan Hung, Yu-Chia Chen

Authors on Pith no claims yet

Pith reviewed 2026-05-09 19:07 UTC · model grok-4.3

classification 💻 cs.AR
keywords subthreshold SRAMcomputing-in-memoryspiking neural networksPVT resilienceenergy efficiencyin-situ regulationedge AI28-nm CMOS
0
0 comments X

The pith

Subthreshold SRAM CIM with in-situ sensors and regulators enables PVT-resilient SNN acceleration at over 1000 TOPS/W.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

This paper presents a computing-in-memory macro built from subthreshold SRAM cells for spiking neural networks, adding on-chip current sensors and distributed voltage regulators to counteract process, voltage, and temperature variations that normally destabilize low-voltage operation. Programmable neuron thresholds implemented with memory cells further stabilize firing behavior, while a stride-tick batching schedule reuses input data to shrink buffer size. A 28-nm prototype chip with a 1024-by-1304 array reaches 93.64 percent accuracy on keyword spotting and reports peak figures of 1181.42 TOPS/W and 7.24 TOPS/mm². Readers focused on edge devices would care because the work shows a concrete route to reliable, high-efficiency inference on sparse event-driven networks without moving to higher supply voltages.

Core claim

The authors claim that integrating in-situ current sensors, distributed voltage regulators, and memory-cell-based programmable neuron thresholds inside a large-scale subthreshold current-mode SRAM CIM array allows robust operation for spiking neural networks; when paired with stride-tick batching that exploits SNN sparsity and data reuse, the fabricated 28-nm design delivers 93.64% keyword-spotting accuracy, up to 1181.42 TOPS/W, and 7.24 TOPS/mm².

What carries the argument

In-situ current sensors and distributed voltage regulators combined with programmable memory-cell-based neuron firing thresholds that stabilize subthreshold current-mode computation across a 1024 wordline by 1304 bitline CIM array.

If this is right

  • Exploitation of SNN sparsity yields additional energy savings on top of the baseline CIM design.
  • Stride-tick batching measurably cuts buffer overhead and improves input data reuse.
  • Programmable neuron thresholds increase robustness to PVT variations without external calibration.
  • The overall architecture provides a practical path to high-performance edge SNN hardware.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same in-situ regulation approach could be ported to other low-voltage memory-based accelerators facing similar variation problems.
  • The efficiency numbers imply the design might support always-on inference on battery-constrained sensors if the macro is integrated into a full system.
  • Scaling the array further would reveal whether regulation overhead grows linearly or sub-linearly with size.

Load-bearing premise

The added sensors and regulators can offset PVT variations in subthreshold current-mode arrays without area or power costs large enough to erase the reported efficiency advantage.

What would settle it

Direct measurement of the fabricated chip across a wide temperature or voltage range that shows accuracy falling below 90% or efficiency below 500 TOPS/W would disprove the claimed resilience and efficiency.

Figures

Figures reproduced from arXiv: 2605.00319 by Bing-Han Liu, Chien-Nan Liu, Hung-Ming Chen, I-Wen Wang, Shih-Hang Kao, Shyh-Jye Jou, Tian-Sheuan Chang, Wei-Zen Chen, Yang-Chan Hung, Yu-Chia Chen.

Figure 1
Figure 1. Figure 1: The challenges of current CIM designs. notably current-mode instability and pronounced sensitivity to process–voltage–temperature (PVT) variations. Prior SNN￾CIM prototypes span RRAM [1] and eDRAM [2] technologies as well as SRAM-based designs [3]–[5]. While these efforts pursue high energy efficiency and ADC-less operation [4], achieving robust, large-scale subthreshold operation with con￾current high acc… view at source ↗
Figure 2
Figure 2. Figure 2: (a) Left: 8T SRAM output current under nominal view at source ↗
Figure 3
Figure 3. Figure 3: CIM macro with monitor sensors and distributed view at source ↗
Figure 4
Figure 4. Figure 4: Simulated RBL current variation at different tempera view at source ↗
Figure 5
Figure 5. Figure 5: Current variation of 8T SRAM cell (a) IDAC driven view at source ↗
Figure 6
Figure 6. Figure 6: The proposed voltage regulator and its timing diagram. view at source ↗
Figure 8
Figure 8. Figure 8: Transimpedance error amplifier. implementing transimpedance gain that should be adjusted across PVT variations. In this design, the ITH corresponds to the output current of five unity cells. The proposed circuit consumes only 0.9% of the total chip power (12.39 mW), where each sense amplifier and the ITH generator consume 25.2 µW and 0.9 µW, respectively. The total area overhead is 364 µm2 for 128 instance… view at source ↗
Figure 9
Figure 9. Figure 9: (a) The spike generation architecture. (b) SNN neuron. view at source ↗
Figure 10
Figure 10. Figure 10: The proposed example SNN model for keyword view at source ↗
Figure 11
Figure 11. Figure 11: SNN model training flow. TABLE I: Results of applying variations and variation-aware training on the proposed SNN model. Condition Accuracy (%) Ideal Model 96.58 With Variations (No Adjustment) 59.64 With Variation-Aware Training 93.64 B. Hardware Architecture view at source ↗
Figure 12
Figure 12. Figure 12: Proposed CIM macro and stride-tick batching view at source ↗
Figure 13
Figure 13. Figure 13: Typical SNN and stride-tick batching dataflow. view at source ↗
Figure 15
Figure 15. Figure 15: Chip micrograph view at source ↗
Figure 16
Figure 16. Figure 16: Shmoo plot. ranges from 0.28 V to 0.3 V during the CIM mode operation. The test chip was fabricated via an educational multi-project wafer (MPW) shuttle, where a limited number of dies are available. Experimental characterization of three sample chips yielded consistent performance across all units. Furthermore, the design’s robustness against process, voltage, and temper￾ature (PVT) variations was verifi… view at source ↗
Figure 14
Figure 14. Figure 14: Pooling write-back (PWB) timing diagram. view at source ↗
Figure 17
Figure 17. Figure 17: Experimental setup view at source ↗
read the original abstract

This paper presents a PVT-resilient, subthreshold SRAM-based computing-in-memory (CIM) macro tailored for energy-efficient spiking neural networks (SNNs). The macro integrates in-situ current sensors and distributed voltage regulators to enable robust large-scale (1024 wordlines, 1304 bitlines and 128 shared neuron cells) subthreshold current-mode CIM, mitigating energy overheads and process-voltage-temperature (PVT) sensitivity. The neuron cells adopt a programmable, memory cell-based firing threshold to enhance neuron robustness against PVT variations. The architecture uses a stride-tick batching schedule to significantly reduce buffer overhead with enhanced input data reuse. Exploiting the high sparsity of SNNs, the proposed system demonstrates significant improvements in energy efficiency and variation tolerance. Fabricated in 28-nm CMOS, the prototype attains 93.64\% accuracy on keyword spotting, delivers up to 1181.42 TOPS/W, and achieves 7.24 TOPS/mm^2, demonstrating a viable and efficient solution for high-performance edge SNN processing.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript presents a fabricated 28-nm CMOS subthreshold SRAM-based CIM macro for energy-efficient SNNs. It integrates in-situ current sensors and distributed voltage regulators to mitigate PVT variations in a 1024x1304 current-mode array, uses programmable memory-cell-based neuron thresholds for robustness, and employs stride-tick batching for data reuse and sparsity exploitation. The prototype reports 93.64% accuracy on keyword spotting, peak efficiency of 1181.42 TOPS/W, and 7.24 TOPS/mm².

Significance. If the efficiency and resilience claims hold after isolating overheads and providing full measurement details, this would be a meaningful contribution to edge SNN hardware by demonstrating a viable subthreshold CIM solution with integrated PVT mitigation. The use of a fabricated prototype with concrete metrics is a strength, though the current presentation limits assessment of net gains.

major comments (2)
  1. [Results] Results section: The headline figures of 1181.42 TOPS/W and 7.24 TOPS/mm² are reported for the overall macro without a power or area breakdown that isolates the static and dynamic overheads of the in-situ current sensors, distributed voltage regulators, and programmable neuron thresholds. These blocks are central to the PVT-resilience claim; without separation, it is impossible to confirm that the core CIM efficiency is preserved rather than offset by auxiliary circuitry.
  2. [Experimental validation] Experimental validation: The abstract and results claim 93.64% accuracy and the efficiency metrics from silicon, yet provide no details on measurement conditions (supply voltage, temperature, clocking), number of dies characterized, error bars, or performance across PVT corners. This omission directly undermines verification of the variation-tolerance claims for the large subthreshold array.
minor comments (1)
  1. [Abstract] The abstract introduces the array size (1024 wordlines, 1304 bitlines, 128 neurons) late; moving this detail to the opening sentence would improve readability.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments, which help strengthen the presentation of our results and experimental validation. We address each major comment below and will revise the manuscript to incorporate the suggested improvements where possible.

read point-by-point responses
  1. Referee: [Results] Results section: The headline figures of 1181.42 TOPS/W and 7.24 TOPS/mm² are reported for the overall macro without a power or area breakdown that isolates the static and dynamic overheads of the in-situ current sensors, distributed voltage regulators, and programmable neuron thresholds. These blocks are central to the PVT-resilience claim; without separation, it is impossible to confirm that the core CIM efficiency is preserved rather than offset by auxiliary circuitry.

    Authors: We agree that isolating the overheads of the auxiliary circuits is valuable for rigorously validating the net benefits of our PVT-resilience techniques. The reported figures represent the complete fabricated macro, including all supporting blocks, as this is the relevant system-level performance for edge SNN deployment. In the revised manuscript, we will add a detailed power and area breakdown (both estimated and measured where available) that separates the contributions of the in-situ current sensors, distributed voltage regulators, and programmable neuron thresholds from the core 1024x1304 current-mode array. This will enable readers to assess whether the core CIM efficiency is preserved. revision: yes

  2. Referee: [Experimental validation] Experimental validation: The abstract and results claim 93.64% accuracy and the efficiency metrics from silicon, yet provide no details on measurement conditions (supply voltage, temperature, clocking), number of dies characterized, error bars, or performance across PVT corners. This omission directly undermines verification of the variation-tolerance claims for the large subthreshold array.

    Authors: We acknowledge that additional experimental details are needed to fully substantiate the variation-tolerance claims. In the revised version, we will expand the results section with the available measurement conditions, including the specific supply voltages, temperature ranges, clocking schemes, number of dies characterized, and any statistical metrics such as error bars from repeated measurements. For PVT corner performance, we will report all data obtained during silicon characterization; where full corner coverage was limited by testing resources, we will explicitly note the scope and any observed trends in resilience. revision: partial

Circularity Check

0 steps flagged

No significant circularity; experimental hardware prototype with measured results

full rationale

This is a fabricated 28-nm CMOS prototype paper reporting silicon measurements for accuracy (93.64% on keyword spotting), energy efficiency (up to 1181.42 TOPS/W), and area efficiency (7.24 TOPS/mm²). No mathematical derivations, equations, predictive models, or fitted parameters are present that could reduce to inputs by construction. Claims rest on direct experimental benchmarks rather than analytical chains, self-citations, or ansatzes. The report is self-contained against external benchmarks with no load-bearing self-citation or definitional circularity.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract-only review yields no explicit free parameters, axioms, or invented entities; the design implicitly relies on standard CMOS process assumptions and the general property of SNN sparsity.

pith-pipeline@v0.9.0 · 5533 in / 1150 out tokens · 75660 ms · 2026-05-09T19:07:38.802147+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

21 extracted references · 1 canonical work pages

  1. [1]

    NIMBLE: A neuromorphic learning scheme and memristor based computing-in- memory engine for EMG based hand gesture recognition,

    F. Tian, J. Jiang, J. Liang, Z. Zhang, J. Shi, C. Fanget al., “NIMBLE: A neuromorphic learning scheme and memristor based computing-in- memory engine for EMG based hand gesture recognition,” inIEEE International Symposium on Circuits and Systems (ISCAS), 2022, pp. 2695–2699

  2. [2]

    A reconfigurable 1T1C eDRAM-based spiking neural network computing-in-memory processor for high system-level efficiency,

    S. Kim, S. Kim, S. Um, S. Kim, Z. Li, S. Kimet al., “A reconfigurable 1T1C eDRAM-based spiking neural network computing-in-memory processor for high system-level efficiency,” inIEEE International Sym- posium on Circuits and Systems (ISCAS), 2023, pp. 1–5

  3. [3]

    An 82nw 0.53pJ/SOP clock-free spiking neural network with 40µs latency for AloT wake-up functions using ultimate-event-driven bionic architecture and computing-in-memory technique,

    Y . Liu, Z. Wang, W. He, L. Shen, Y . Zhang, P. Chenet al., “An 82nw 0.53pJ/SOP clock-free spiking neural network with 40µs latency for AloT wake-up functions using ultimate-event-driven bionic architecture and computing-in-memory technique,” inIEEE International Solid-State Circuits Conference (ISSCC), 2022, pp. 372–374

  4. [4]

    Neuro-CIM: Adc-less neuromorphic computing-in-memory processor with operation gating/stopping and digital-analog networks,

    S. Kim, S. Kim, S. Um, S. Kim, K. Kim, and H.-J. Yoo, “Neuro-CIM: Adc-less neuromorphic computing-in-memory processor with operation gating/stopping and digital-analog networks,”IEEE Journal of Solid- State Circuits, vol. 58, no. 10, pp. 2931–2945, 2023

  5. [5]

    30.2 a 22nm 0.26nw/synapse spike-driven spiking neural network processing unit using time-step-first dataflow and sparsity-adaptive in-memory com- puting,

    Y . Liu, Y . Ma, N. Shang, T. Zhao, P. Chen, M. Wuet al., “30.2 a 22nm 0.26nw/synapse spike-driven spiking neural network processing unit using time-step-first dataflow and sparsity-adaptive in-memory com- puting,” inIEEE International Solid-State Circuits Conference (ISSCC), 2024, pp. 484–486

  6. [6]

    Spinalflow: An architecture and dataflow tailored for spiking neural networks,

    S. Narayanan, K. Taht, R. Balasubramonian, E. Giacomin, and P.- E. Gaillardon, “Spinalflow: An architecture and dataflow tailored for spiking neural networks,” in2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA). IEEE, 2020, pp. 349– 362

  7. [7]

    Spikesim: An end-to-end compute-in-memory hardware eval- uation tool for benchmarking spiking neural networks,

    A. Moitra, A. Bhattacharjee, R. Kuang, G. Krishnan, Y . Cao, and P. Panda, “Spikesim: An end-to-end compute-in-memory hardware eval- uation tool for benchmarking spiking neural networks,”IEEE Transac- tions on Computer-Aided Design of Integrated Circuits and Systems, vol. 42, no. 11, pp. 3815–3828, 2023

  8. [8]

    A 65-nm 8T SRAM compute-in-memory macro with column ADCs for processing neural networks,

    C. Yu, T. Yoo, K. T. C. Chai, T. T.-H. Kim, and B. Kim, “A 65-nm 8T SRAM compute-in-memory macro with column ADCs for processing neural networks,”IEEE Journal of Solid-State Circuits, vol. 57, no. 11, pp. 3466–3476, 2022

  9. [9]

    A 65 nm 73 Kb SRAM-based computing-in-memory macro with dynamic-sparsity controlling,

    X. Qiao, J. Song, X. Tang, H. Luo, N. Pan, X. Cuiet al., “A 65 nm 73 Kb SRAM-based computing-in-memory macro with dynamic-sparsity controlling,”IEEE Transactions on Circuits and Systems II: Express Briefs, vol. 69, no. 6, pp. 2977–2981, 2022

  10. [10]

    4b/4b/8b precision charge-domain 8T-SRAM based cim for CNN processing,

    Q. Zang, W. L. Goh, Y . S. Chong, and A. T. Do, “4b/4b/8b precision charge-domain 8T-SRAM based cim for CNN processing,” inIEEE 5th International Conference on Artificial Intelligence Circuits and Systems (AICAS), 2023, pp. 1–5

  11. [11]

    A charge domain SRAM compute-in-memory macro with C-2C ladder- based 8-bit MAC unit in 22-nm FinFET process for edge inference,

    H. Wang, R. Liu, R. Dorrance, D. Dasalukunte, D. Lake, and B. Carlton, “A charge domain SRAM compute-in-memory macro with C-2C ladder- based 8-bit MAC unit in 22-nm FinFET process for edge inference,” IEEE Journal of Solid-State Circuits, vol. 58, no. 4, pp. 1037–1050, 2023

  12. [12]

    A 8-b-precision 6T SRAM computing-in-memory macro using segmented- bitline charge-sharing scheme for AI edge chips,

    J.-W. Su, Y .-C. Chou, R. Liu, T.-W. Liu, P.-J. Lu, P.-C. Wuet al., “A 8-b-precision 6T SRAM computing-in-memory macro using segmented- bitline charge-sharing scheme for AI edge chips,”IEEE Journal of Solid- State Circuits, vol. 58, no. 3, pp. 877–892, 2023

  13. [13]

    C3SRAM: An in- memory-computing SRAM macro based on robust capacitive coupling computing mechanism,

    Z. Jiang, S. G. Yin, J. sun Seo, M. Seoket al., “C3SRAM: An in- memory-computing SRAM macro based on robust capacitive coupling computing mechanism,”IEEE Journal of Solid-State Circuits, vol. 55, no. 7, pp. 1888–1897, 2020

  14. [14]

    A 22nm 832Kb hybrid-domain floating-point SRAM in-memory- compute macro with 16.2–70.2 TFLOPS/W for high-accuracy AI-edge devices,

    P.-C. Wu, J.-W. Su, L.-Y . Hong, J.-S. Ren, C.-H. Chien, H.-Y . Chen et al., “A 22nm 832Kb hybrid-domain floating-point SRAM in-memory- compute macro with 16.2–70.2 TFLOPS/W for high-accuracy AI-edge devices,” inIEEE International Solid-State Circuits Conference (ISSCC), 2023, pp. 126–128

  15. [15]

    34.3 a 22nm 64kb lightning-like hybrid computing-in-memory macro with a compressed adder tree and analog-storage quantizers for transformer and CNNs,

    A. Guo, F. Dong, J. Chen, Z. Yuan, X. Huet al., “34.3 a 22nm 64kb lightning-like hybrid computing-in-memory macro with a compressed adder tree and analog-storage quantizers for transformer and CNNs,” in IEEE International Solid-State Circuits Conference (ISSCC), 2024, pp. 570–572

  16. [16]

    A machine-learning classifier implemented in a standard 6T SRAM array,

    J. Zhang, Z. Wang, N. Vermaet al., “A machine-learning classifier implemented in a standard 6T SRAM array,” inProceedings of VLSI Symposium on Technology and Circuits, 2016, pp. 1–2

  17. [17]

    A compilation framework for sram computing-in-memory systems with optimized weight mapping and error correction,

    Y . Bai, Y . Li, H. Zhang, A. Jiang, Y . Du, and L. Du, “A compilation framework for sram computing-in-memory systems with optimized weight mapping and error correction,”IEEE Transactions on Computer- Aided Design of Integrated Circuits and Systems, vol. 43, no. 8, pp. 2379–2392, 2024

  18. [18]

    A 28-nm 50.1-tops/w p-8t sram compute- in-memory macro design with bl charge-sharing-based in-sram dac/adc operations,

    K. Lee, J. Kim, and J. Park, “A 28-nm 50.1-tops/w p-8t sram compute- in-memory macro design with bl charge-sharing-based in-sram dac/adc operations,”IEEE Journal of Solid-State Circuits, vol. 59, no. 6, pp. 1926–1937–1667, 2024

  19. [19]

    One timestep is all you need: Training spiking neural networks with ultra low latency,

    S. S. Chowdhury, N. Rathi, K. Royet al., “One timestep is all you need: Training spiking neural networks with ultra low latency,”arXiv preprint arXiv:2110.05929, 2021. Shih-Hang Kaoreceived the B. S. and M. S. de- grees in electronics engineering from National Yang Ming Chiao Tung University, Hsinchu, Taiwan, in 2021 and 2024. He is currently working at R...

  20. [20]

    Since 2004, he has been a Professor with the Electronics Engineering Department, National Chiao Tung University, and served as the Chairperson from 2006 to 2009. From January 2016 to December 2017, he was appointed as the Director General of the Science Education and International Cooperation Department, Ministry of Science and Technology, Taiwan. He was ...

  21. [21]

    He served as the Institute director from 2015-2018

    He is a professor at the Institute of Electronics and Department of Electronics and Electrical Engi- neering, National Yang Ming Chiao Tung University (NYCU). He served as the Institute director from 2015-2018. He was the Deputy Executive Director of the National SoC Program and the Principal Investigator of the National Project on Intelligent Electronics...