pith. machine review for the scientific record. sign in

arxiv: 2605.13699 · v1 · submitted 2026-05-13 · 💻 cs.AR

Recognition: 2 theorem links

· Lean Theorem

Memristor Technologies for Dynamic Vision Sensors: A Critical Assessment and Research Roadmap

Authors on Pith no claims yet

Pith reviewed 2026-05-14 17:42 UTC · model grok-4.3

classification 💻 cs.AR
keywords memristordynamic vision sensorDVSedge AIneuromorphic computingin-memory computeevent-driven visiontechnology readiness level
0
0 comments X

The pith

An end-to-end integrated DVS-memristor system is the field's open challenge, with concrete accuracy and power targets.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper reviews memristor technologies for use with dynamic vision sensors to enable low-energy event-driven processing at the edge. It shows that the two technologies are mature on their own but lack a unifying framework that separates what has been built from what is only projected. Across six application areas, half depend entirely on forecasts, while existing hardware remains at technology readiness levels 2-5. A three-paradigm taxonomy is used to classify architectures and compare them with current digital neuromorphic options. The central conclusion is that the next required step is fabrication of a complete integrated system that meets explicit performance benchmarks.

Core claim

The paper identifies an end-to-end integrated dynamic vision sensor and memristor system as the open challenge, with testable targets for accuracy and power that would overcome data-movement energy limits in edge AI deployments.

What carries the argument

A three-paradigm architectural taxonomy combined with an evidence-grading method that distinguishes fabricated demonstrations from projected systems.

If this is right

  • Research efforts should shift from separate component development to full-system integration and fabrication.
  • Application domains such as robotics and surveillance can now use the identified power and accuracy targets to guide hardware specifications.
  • Direct comparisons with digital neuromorphic processors will become possible once integrated prototypes exist.
  • Technology readiness levels for these systems must advance from the current 2-5 range to enable practical edge deployment.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Success would allow always-on vision processing in power-constrained devices such as wearables and remote sensors.
  • The same integration approach could later be applied to other event-driven sensors beyond vision.
  • The concrete targets provide measurable milestones that funding programs and development teams can track.

Load-bearing premise

That the projected orders-of-magnitude energy savings from analog in-memory computation will actually appear once systems move beyond today's early laboratory demonstrations.

What would settle it

A working prototype of a complete DVS-memristor chip that meets or misses the stated accuracy and power targets when tested in one of the surveyed application domains.

Figures

Figures reproduced from arXiv: 2605.13699 by Edris Zaman Farsa, Mohamad Yazan Sadoun, Sarah Sharif, Yaser Mike Banad.

Figure 1
Figure 1. Figure 1: Structured PRISMA-style screening pipeline using the PRISMA reporting guide [PITH_FULL_IMAGE:figures/full_fig_p005_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Memristor and photomemristor device physics. [PITH_FULL_IMAGE:figures/full_fig_p006_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Comparison of frame-based and event-based camera outputs. (a) A rotating-disk [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Biological versus neuromorphic vision systems. Top: biological neural networks process [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Timeline of key milestones in memristor-based vision systems (2008–2025), colour [PITH_FULL_IMAGE:figures/full_fig_p011_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Three computing paradigms for vision, mapping onto the R1/R2/R3 taxonomy. [PITH_FULL_IMAGE:figures/full_fig_p012_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Taxonomy of memristor integration roles in DVS systems, refining Figure [PITH_FULL_IMAGE:figures/full_fig_p012_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Material-system compatibility matrix for memristor-based DVS integration. Rows are [PITH_FULL_IMAGE:figures/full_fig_p017_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: End-to-end sensor-to-crossbar pipeline illustrating how an event-driven front-end [PITH_FULL_IMAGE:figures/full_fig_p020_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Technology Readiness Level (TRL) assessment. Horizontal bars indicate the assessed [PITH_FULL_IMAGE:figures/full_fig_p025_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Research roadmap for memristor-DVS integration. Three phases with falsifiable [PITH_FULL_IMAGE:figures/full_fig_p031_11.png] view at source ↗
read the original abstract

Edge-AI deployment is bottlenecked by data-movement energy; pairing event-driven vision sensors with in-memory analog compute could lift that ceiling by orders of magnitude. Both technologies are individually mature; the framework distinguishing fabricated demonstrations from projected systems is missing. Of six application domains surveyed (robotics, autonomous vehicles, AR/VR, surveillance, medical imaging, IoT), half rest entirely on projection, and existing hardware sits at Technology Readiness Levels 2-5. This evidence-graded review applies a three-paradigm architectural taxonomy and benchmarks the gap against current digital neuromorphic alternatives. It identifies an end-to-end integrated DVS-memristor system as the field's open challenge, with testable accuracy and power targets.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript surveys memristor technologies for dynamic vision sensors across six application domains (robotics, autonomous vehicles, AR/VR, surveillance, medical imaging, IoT). It distinguishes fabricated demonstrations (TRL 2-5) from projections, applies a three-paradigm architectural taxonomy, benchmarks against digital neuromorphic alternatives, and concludes that an end-to-end integrated DVS-memristor system remains the open challenge, with specific testable accuracy and power targets.

Significance. If the evidence grading and taxonomy hold, the work supplies a needed organizing framework for the field by clarifying the integration gap and setting concrete, falsifiable targets. The explicit separation of demonstrated hardware from projections, plus direct comparison to digital baselines, strengthens the roadmap value.

major comments (2)
  1. [Application Domains Survey] In the survey of the six domains, the statement that half rest entirely on projection requires explicit listing of which three domains and the exact projected metrics (e.g., energy per event or accuracy) used to reach that count; without this, the TRL assessment cannot be independently verified.
  2. [Architectural Taxonomy] The three-paradigm taxonomy is presented as an organizing tool, yet the manuscript does not provide a side-by-side quantitative comparison table of the paradigms against the digital neuromorphic baselines on the same accuracy/power metrics; this weakens the benchmarking claim.
minor comments (2)
  1. [Figures] Figure captions for the TRL progression diagrams should include the exact reference numbers from the surveyed literature for each data point.
  2. [Introduction] The abstract states 'orders of magnitude' energy gains; the main text should cite the specific projected ratios and the source papers for those ratios.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments and the recommendation for minor revision. The feedback highlights opportunities to improve verifiability and quantitative support, which we have addressed in the revised manuscript.

read point-by-point responses
  1. Referee: [Application Domains Survey] In the survey of the six domains, the statement that half rest entirely on projection requires explicit listing of which three domains and the exact projected metrics (e.g., energy per event or accuracy) used to reach that count; without this, the TRL assessment cannot be independently verified.

    Authors: We agree that explicit identification strengthens the claim. The revised manuscript now includes a dedicated subsection and accompanying table that lists the three domains relying entirely on projections (AR/VR, medical imaging, and IoT). For each domain we specify the exact projected metrics drawn from the cited works, including energy per event (e.g., sub-pJ targets) and accuracy thresholds (e.g., >90% event-based classification), together with the source references. This change allows direct verification of the TRL 2-5 grading. revision: yes

  2. Referee: [Architectural Taxonomy] The three-paradigm taxonomy is presented as an organizing tool, yet the manuscript does not provide a side-by-side quantitative comparison table of the paradigms against the digital neuromorphic baselines on the same accuracy/power metrics; this weakens the benchmarking claim.

    Authors: We accept that a consolidated quantitative table would reinforce the benchmarking section. The revised manuscript adds a new comparison table that places the three paradigms (hybrid, fully analog in-memory, and event-driven memristor) alongside digital baselines (Loihi, TrueNorth) using the most comparable reported figures for accuracy on event-based datasets and energy per event. Where metrics are not directly reported we note the normalization assumptions and data gaps, preserving transparency while providing the requested side-by-side view. revision: yes

Circularity Check

0 steps flagged

No significant circularity

full rationale

The paper is a literature survey and roadmap that grades existing demonstrations by Technology Readiness Level, applies an external three-paradigm taxonomy, and identifies an integration gap from surveyed sources. No internal equations, fitted parameters, or self-referential derivations appear; all claims about energy projections, accuracy targets, and the open challenge of an end-to-end DVS-memristor system rest on citations to independent prior work rather than reducing to the paper's own inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The review rests on the domain assumption that event-driven sensing plus in-memory analog compute can deliver large energy reductions, drawn from neuromorphic hardware literature rather than derived within the paper.

axioms (1)
  • domain assumption Pairing event-driven vision sensors with in-memory analog compute can lift the data-movement energy bottleneck by orders of magnitude.
    Core premise stated in the abstract and used to frame the entire assessment.

pith-pipeline@v0.9.0 · 5427 in / 1256 out tokens · 44851 ms · 2026-05-14T17:42:37.543630+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

151 extracted references · 151 canonical work pages

  1. [1]

    A 128×128 120 dB 15µs latency asynchronous temporal contrast vision sensor,

    P. Lichtsteiner, C. Posch, and T. Delbrück, “A 128×128 120 dB 15µs latency asynchronous temporal contrast vision sensor,”IEEE Journal of Solid-State Circuits, vol. 43, no. 2, pp. 566–576, 2008

  2. [2]

    In-memory computing with resistive switching devices,

    D. Ielmini and H.-S. P. Wong, “In-memory computing with resistive switching devices,” Nature Electronics, vol. 1, no. 6, pp. 333–343, 2018

  3. [3]

    Memory devices and applications for in-memory computing,

    A. Sebastian, M. Le Gallo, R. Khaddam-Aljameh, and E. Eleftheriou, “Memory devices and applications for in-memory computing,”Nature Nanotechnology, vol. 15, no. 7, pp. 529–544, 2020

  4. [4]

    The missing memristor found,

    D. B. Strukov, G. S. Snider, D. R. Stewart, and R. S. Williams, “The missing memristor found,”Nature, vol. 453, no. 7191, pp. 80–83, 2008

  5. [5]

    Training and operation of an integrated neuromorphic network based on metal- oxide memristors,

    M. Prezioso, F. Merrikh-Bayat, B. D. Hoskins, G. C. Adam, K. K. Likharev, and D. B. Strukov, “Training and operation of an integrated neuromorphic network based on metal- oxide memristors,”Nature, vol. 521, no. 7550, pp. 61–64, 2015

  6. [6]

    Reconfigurable in-sensor processing based on a multi-phototransistor–one-memristor array,

    B. Dang, T. Zhang, X. Wu, K. Liu, R. Huang, and Y. Yang, “Reconfigurable in-sensor processing based on a multi-phototransistor–one-memristor array,”Nature Electronics, vol. 7, pp. 991–1003, 2024. 32

  7. [7]

    Memristor–CMOS hybrid circuits implementing event-driven neural networks for dynamic vision sensor camera,

    R. Yoon, S. Oh, S. Cho, and K.-S. Min, “Memristor–CMOS hybrid circuits implementing event-driven neural networks for dynamic vision sensor camera,”Micromachines, vol. 15, no. 4, p. 426, 2024

  8. [8]

    Event-based vision: A survey,

    G. Gallego, T. Delbrück, G. Orchard, C. Bartolozzi, B. Taba, A. Censi, S. Leutenegger, A. J. Davison, J. Conradt, K. Daniilidis, and D. Scaramuzza, “Event-based vision: A survey,”IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 1, pp. 154–180, 2022

  9. [9]

    Recent event camera innovations: A survey,

    B. Chakravarthi, A. A. Verma, K. Daniilidis, C. Fermüller, and Y. Yang, “Recent event camera innovations: A survey,” inComputer Vision – ECCV 2024 Workshops. Springer, 2024, pp. 316–332

  10. [10]

    An application-driven survey on event-based neuromorphic computer vision,

    D. Cazzato and F. Bono, “An application-driven survey on event-based neuromorphic computer vision,”Information, vol. 15, no. 8, p. 472, 2024

  11. [11]

    Technology and integration roadmap for optoelectronic memristor,

    J. Wang, N. Ilyas, Y. Ren, Y. Ji, S. Li, C. Li, F. Liu, D. Gu, and K.-W. Ang, “Technology and integration roadmap for optoelectronic memristor,”Advanced Materials, vol. 36, no. 9, p. 2307393, 2024

  12. [12]

    Progress of materials and devices for neuromorphic vision sensors,

    S. W. Cho, C. Jo, Y.-H. Kim, and S.-K. Park, “Progress of materials and devices for neuromorphic vision sensors,”Nano-Micro Letters, vol. 14, no. 1, p. 203, 2022

  13. [13]

    In-sensor computing: Materials, devices, and integration technologies,

    T. Wan, B. Shao, S. Ma, Y. Zhou, Q. Li, and Y. Chai, “In-sensor computing: Materials, devices, and integration technologies,”Advanced Materials, vol. 35, p. 2203830, 2023

  14. [14]

    On non-von Neumann flexible neuromorphic vision sensors,

    H. Wang, B. Sun, S. S. Ge, J. Su, and M. L. Jin, “On non-von Neumann flexible neuromorphic vision sensors,”npj Flexible Electronics, vol. 8, p. 30, 2024

  15. [15]

    Emerging photoelectric devices for neuromorphic vision applications: Principles, developments, and outlooks,

    Y. Zhang, Z. Huang, and J. Jiang, “Emerging photoelectric devices for neuromorphic vision applications: Principles, developments, and outlooks,”Science and Technology of Advanced Materials, vol. 24, no. 1, p. 2186689, 2023

  16. [16]

    The building blocks of a brain-inspired computer,

    J. D. Kendall and S. Kumar, “The building blocks of a brain-inspired computer,”Applied Physics Reviews, vol. 7, no. 1, p. 011305, 2020

  17. [17]

    Physics for neuromorphic comput- ing,

    D. Marković, A. Mizrahi, D. Querlioz, and J. Grollier, “Physics for neuromorphic comput- ing,”Nature Reviews Physics, vol. 2, pp. 499–510, 2020

  18. [18]

    2022 roadmap on neuromorphic computing and engineering,

    D. V. Christensen, R. Dittmann, B. Linares-Barranco, A. Sebastian, M. Le Gallo, A. Redaelli, S. Slesazeck, T. Mikolajick, S. Spiga, S. Menzelet al., “2022 roadmap on neuromorphic computing and engineering,”Neuromorphic Computing and Engineering, vol. 2, no. 2, p. 022501, 2022

  19. [19]

    Opportu- nities for neuromorphic computing algorithms and applications,

    C. D. Schuman, S. R. Kulkarni, M. Parsa, J. P. Mitchell, P. Date, and B. Kay, “Opportu- nities for neuromorphic computing algorithms and applications,”Nature Computational Science, vol. 2, no. 1, pp. 10–19, 2022

  20. [20]

    Neuromorphic computing at scale,

    D. Kudithipudi, C. Schuman, C. M. Vineyard, T. Pandit, C. Merkel, R. Kubendran, J. B. Aimone, G. Orchard, C. Mayr, R. Benosman, J. Hays, C. Young, C. Bartolozzi, A. Majumdar, S. G. Cardwell, M. Payvand, S. Buckley, S. Kulkarni, H. A. Gonzalez, G. Cauwenberghs, C. S. Thakur, A. Subramoney, and S. Furber, “Neuromorphic computing at scale,”Nature, vol. 637, ...

  21. [21]

    Bioinspired artificial vision system based on photoelectric memristors,

    S. Mao, Y. Zhao, Z. Cao, S. Zhu, G. Zhou, and B. Sun, “Bioinspired artificial vision system based on photoelectric memristors,”Materials Science and Engineering: R: Reports, vol. 167, p. 101137, 2026. 33

  22. [22]

    Devices, functions, and applications of artificial neuromorphic visual systems,

    J. Liu, B. Li, C. Liu, D. Sun, and H. Cheng, “Devices, functions, and applications of artificial neuromorphic visual systems,”Advanced Science, vol. 12, no. 46, p. e16379, 2025

  23. [23]

    Toward energy-efficient machine vision: Advances in optoelectronic memristors,

    S. Pan, S. Wu, J. Ming, and H. Ling, “Toward energy-efficient machine vision: Advances in optoelectronic memristors,”Advanced Optical Materials, vol. 13, no. 30, p. 2501992, 2025

  24. [24]

    Optoelectronic devices for in-sensor computing,

    Q. Ren, S. Zhu, G. Ma, Y. Li, M. Zhu, Z. Zeng, and Y. Chai, “Optoelectronic devices for in-sensor computing,”Advanced Materials, vol. 37, no. 10, p. 2407476, 2024

  25. [25]

    Review of memristors for in-memory computing and spiking neural networks,

    M. Shooshtari, T. Serrano-Gotarredona, and B. Linares-Barranco, “Review of memristors for in-memory computing and spiking neural networks,”Advanced Intelligent Systems, vol. 8, p. 2500806, 2025

  26. [26]

    Hardware, algo- rithms, and applications of the neuromorphic vision sensor: A review,

    C. Cimarelli, J. A. Millan-Romera, H. Voos, and J. L. Sanchez-Lopez, “Hardware, algo- rithms, and applications of the neuromorphic vision sensor: A review,”Sensors, vol. 25, no. 19, p. 6208, 2025

  27. [27]

    Deep learning for event-based vision: A comprehensive survey and benchmarks,

    X. Zheng, Y. Liu, Y. Lu, T. Hua, T. Pan, W. Zhang, D. Tao, and L. Wang, “Deep learning for event-based vision: A comprehensive survey and benchmarks,”arXiv preprint arXiv:2302.08890v3, 2024

  28. [28]

    Emerging trends and applications of neuromorphic dynamic vision sensors: A survey,

    H. Aliakbarpour, A. Moori, J. Khorramdel, E. Blasch, and O. Tahri, “Emerging trends and applications of neuromorphic dynamic vision sensors: A survey,”IEEE Sensors Reviews, vol. 1, pp. 14–64, 2024

  29. [29]

    Event-based spiking neural networks for object detection: A review of datasets, architectures, learning rules, and implementation,

    C. Iaboni, Y. Wu, and P. Abichandani, “Event-based spiking neural networks for object detection: A review of datasets, architectures, learning rules, and implementation,”IEEE Access, 2024

  30. [30]

    When in-memory computing meets spiking neural networks—a perspective on device-circuit-system-and-algorithm co-design,

    A. Moitra, A. Bhattacharjee, Y. Li, Y. Kim, and P. Panda, “When in-memory computing meets spiking neural networks—a perspective on device-circuit-system-and-algorithm co-design,”Applied Physics Reviews, vol. 11, p. 031325, 2024

  31. [31]

    Finateu, A

    T. Finateu, A. Niwa, D. Matolin, K. Tsuchimoto, A. Mascheroni, E. Reynaud, P. Mostafalu, F. Brady, L. Chotard, F. LeGoff, H. Takahashi, H. Hagekatsu, T. Endo, Y. Lievre, Y. Nishimura, D. Renshaw, and C. Posch, “A 1280×720 back-illuminated stacked temporal contrast event-based vision sensor with 4.86µm pixels, 1.066GEPS readout, programmable event-rate con...

  32. [32]

    A 1280×960 dynamic vision sensor with a 4.95-µm pixel pitch and motion artifact minimization,

    Y. Suh, S. Choi, M. Ito, J. Kim, Y. Lee, J. Seo, H. Jung, D.-H. Yeo, S. Namgung, J. Bong et al., “A 1280×960 dynamic vision sensor with a 4.95-µm pixel pitch and motion artifact minimization,” inProc. IEEE International Symposium on Circuits and Systems (ISCAS), 2020

  33. [33]

    Thousands of conductance levels in memristors integrated on CMOS,

    M. Rao, H. Tang, J. Wu, W. Song, M. Zhang, W. Yin, Y. Zhuo, F. Kiani, B. Chen, X. Jianget al., “Thousands of conductance levels in memristors integrated on CMOS,” Nature, vol. 615, pp. 823–829, 2023

  34. [34]

    A compute-in- memory chip based on resistive random-access memory,

    W. Wan, R. Kubendran, C. Schaefer, S. B. Eryilmaz, W. Zhang, D. Wu, S. Deiss, P. Raina, H. Qian, B. Gao, S. Joshi, H. Wu, H.-S. P. Wong, and G. Cauwenberghs, “A compute-in- memory chip based on resistive random-access memory,”Nature, vol. 608, no. 7923, pp. 504–512, 2022

  35. [35]

    Fully hardware-implemented memristor convolutional neural network,

    P. Yao, H. Wu, B. Gao, J. Tang, Q. Zhang, W. Zhang, J. J. Yang, and H. Qian, “Fully hardware-implemented memristor convolutional neural network,”Nature, vol. 577, no. 7792, pp. 641–646, 2020. 34

  36. [36]

    Equivalent-accuracy accelerated neural-network training using analogue memory,

    S. Ambrogio, P. Narayanan, H. Tsai, R. M. Shelby, I. Boybat, C. di Nolfo, S. Sidler, M. Giordano, M. Bodini, N. C. P. Farinha, B. Killeen, C. Cheng, Y. Jaoudi, and G. W. Burr, “Equivalent-accuracy accelerated neural-network training using analogue memory,” Nature, vol. 558, pp. 60–67, 2018

  37. [37]

    Programming memristor arrays with arbitrarily high precision for analog computing,

    W. Song, M. Rao, Y. Li, C. Li, Y. Zhuo, F. Cai, M. Wu, W. Yin, Z. Li, Q. Wei, S. Lee, H. Zhu, L. Gong, M. Barnell, Q. Wu, P. A. Beerel, M. S.-W. Chen, N. Ge, M. Hu, Q. Xia, and J. J. Yang, “Programming memristor arrays with arbitrarily high precision for analog computing,”Science, vol. 383, no. 6685, pp. 903–910, 2024

  38. [38]

    An analog-AI chip for energy-efficient speech recognition and transcription,

    S. Ambrogio, P. Narayanan, A. Okazaki, A. Fasoli, C. Mackin, K. Hosokawa, A. Nomura, T. Yasuda, A. Chen, A. Frizet al., “An analog-AI chip for energy-efficient speech recognition and transcription,”Nature, vol. 620, pp. 768–775, 2023

  39. [39]

    A 64-core mixed- signal in-memory compute chip based on phase-change memory for deep neural network inference,

    M. Le Gallo, R. Khaddam-Aljameh, M. Stanisavljevic, A. Vasilopoulos, B. Kersting, M. Dazzi, G. Karunaratne, M. Bräendli, A. Singh, S. M. Muelleret al., “A 64-core mixed- signal in-memory compute chip based on phase-change memory for deep neural network inference,”Nature Electronics, vol. 6, pp. 680–693, 2023

  40. [40]

    The PRISMA 2020 statement: An updated guideline for reporting systematic reviews,

    M. J. Page, J. E. McKenzie, P. M. Bossuytet al., “The PRISMA 2020 statement: An updated guideline for reporting systematic reviews,”BMJ, vol. 372, p. n71, 2021

  41. [41]

    In-memory-computing realization with a photodiode/memristor based vision sensor,

    N. Vasileiadis, V. Ntinas, G. C. Sirakoulis, and P. Dimitrakis, “In-memory-computing realization with a photodiode/memristor based vision sensor,”Materials, vol. 14, no. 18, p. 5223, 2021

  42. [42]

    Memristor-based pixel for event-detection vision sensor,

    M. Laiho and E. Lehtonen, “Memristor-based pixel for event-detection vision sensor,” in IEEE International Symposium on Circuits and Systems (ISCAS), 2015, pp. 2313–2316

  43. [43]

    A dual-functional perovskite-based photodetector and memristor for visual memory,

    F. Cao, Z. Hu, T. Yan, E. Hong, X. Deng, L. Wu, and X. Fang, “A dual-functional perovskite-based photodetector and memristor for visual memory,”Advanced Materials, vol. 35, no. 44, p. 2304550, 2023

  44. [44]

    In-situ artificial retina with all-in-one reconfigurable photomemristor networks,

    Y. Cai, Y. Jiang, C. Sheng, Z. Wu, L. Chen, B. Tian, C. Duan, S. Xiong, Y. Zhan, C. Cong, Z. Qiu, Y. Qin, R. Liu, and L. Hu, “In-situ artificial retina with all-in-one reconfigurable photomemristor networks,”npj Flexible Electronics, vol. 7, p. 29, 2023

  45. [45]

    Optical–electrical coordinately modulated memristor based on 2D ferroelectric RP perovskite for artificial vision applications,

    H. Wang, J. Yang, Z. Yang, G. Liu, Y. Tang, Y. Shao, and X. Yan, “Optical–electrical coordinately modulated memristor based on 2D ferroelectric RP perovskite for artificial vision applications,”Advanced Science, vol. 11, no. 33, p. 2403150, 2024

  46. [46]

    Fully integrated multi-mode optoelectronic memristor array for diversified in-sensor computing,

    H. Huang, X. Liang, Y. Wang, J. Tang, Y. Li, Y. Du, W. Sun, J. Zhang, P. Yao, X. Mou, F.Xu, J.Zhang, Y.Lu, Z.Liu, J.Wang, Z.Jiang, R.Hu, Z.Wang, Q.Zhang, B.Gao, X.Bai, L. Fang, Q. Dai, H. Yin, H. Qian, and H. Wu, “Fully integrated multi-mode optoelectronic memristor array for diversified in-sensor computing,”Nature Nanotechnology, vol. 20, no. 1, pp. 93–103, 2025

  47. [47]

    Dynamic machine vision with retinomorphic photomemristor- reservoir computing,

    H. Tan and S. van Dijken, “Dynamic machine vision with retinomorphic photomemristor- reservoir computing,”Nature Communications, vol. 14, p. 2169, 2023

  48. [48]

    Analogue signal and image processing with large memristor crossbars,

    C. Li, M. Hu, Y. Li, H. Jiang, N. Ge, E. Montgomery, J. Zhang, W. Song, N. Dávila, C. E. Graves, Z. Li, J. P. Strachan, P. Lin, Z. Wang, M. Barnell, Q. Wu, R. S. Williams, J. J. Yang, and Q. Xia, “Analogue signal and image processing with large memristor crossbars,” Nature Electronics, vol. 1, no. 1, pp. 52–59, 2018. 35

  49. [49]

    Full hardware implementation of neuromorphic visual system based on multimodal optoelectronic resistive memory arrays for versatile image processing,

    G. Zhou, J. Li, Q. Song, L. Wang, Z. Ren, B. Sun, X. Hu, W. Wang, G. Xu, X. Chen, L. Cheng, F. Zhou, and S. Duan, “Full hardware implementation of neuromorphic visual system based on multimodal optoelectronic resistive memory arrays for versatile image processing,”Nature Communications, vol. 14, p. 8489, 2023

  50. [50]

    Memristor—the missing circuit element,

    L. O. Chua, “Memristor—the missing circuit element,”IEEE Transactions on Circuit Theory, vol. 18, no. 5, pp. 507–519, 1971

  51. [51]

    Memristive devices for computing,

    J. J. Yang, D. B. Strukov, and D. R. Stewart, “Memristive devices for computing,”Nature Nanotechnology, vol. 8, no. 1, pp. 13–24, 2013

  52. [52]

    Role of electrode materials in resistive switching mechanisms of oxide-based memristors for enhanced neuromorphic computing: A comprehensive study,

    A. Gooran-Shoorakchaly, S. S. Sharif, and Y. M. Banad, “Role of electrode materials in resistive switching mechanisms of oxide-based memristors for enhanced neuromorphic computing: A comprehensive study,”Memories—Materials, Devices, Circuits and Systems, vol. 11, p. 100133, 2025

  53. [53]

    Memristor synapse—a device- level critical review,

    S. Chandrasekaran, Y.-F. Chang, and F. M. Simanjuntak, “Memristor synapse—a device- level critical review,”Nanomaterials, vol. 16, no. 3, p. 179, 2026

  54. [54]

    Current opinions on memristor-accelerated machine learning hardware,

    M. Jiang, Y. Xu, Z. Li, and C. Li, “Current opinions on memristor-accelerated machine learning hardware,”Current Opinion in Solid State and Materials Science, 2025

  55. [55]

    Investigating the effect of electrical and thermal transport properties on oxide-based memristors performance and reliability,

    A. Gooran-Shoorakchaly, S. S. Sharif, and Y. M. Banad, “Investigating the effect of electrical and thermal transport properties on oxide-based memristors performance and reliability,”Scientific Reports, vol. 15, p. 18646, 2025

  56. [56]

    TEAM: ThrEshold Adaptive Memristor model,

    S. Kvatinsky, E. G. Friedman, A. Kolodny, and U. C. Weiser, “TEAM: ThrEshold Adaptive Memristor model,”IEEE Transactions on Circuits and Systems I: Regular Papers, vol. 60, no. 1, pp. 211–221, 2013

  57. [57]

    Computational model of Ta2O5/TaOx memristors: Predicting resistive switching behavior and filament growth dynamics,

    A. Gooran-Shoorakchaly, S. S. Sharif, and Y. M. Banad, “Computational model of Ta2O5/TaOx memristors: Predicting resistive switching behavior and filament growth dynamics,” inProc. IEEE 67th International Midwest Symposium on Circuits and Systems (MWSCAS), 2024, pp. 552–556

  58. [58]

    A QVGA 143 dB dynamic range frame-free PWM image sensor with lossless pixel-level video compression and time-domain CDS,

    C. Posch, D. Matolin, and R. Wohlgenannt, “A QVGA 143 dB dynamic range frame-free PWM image sensor with lossless pixel-level video compression and time-domain CDS,” IEEE Journal of Solid-State Circuits, vol. 46, no. 1, pp. 259–275, 2011

  59. [59]

    A 240×180 130 dB 3µs latency global shutter spatiotemporal vision sensor,

    C. Brandli, R. Berner, M. Yang, S.-C. Liu, and T. Delbrück, “A 240×180 130 dB 3µs latency global shutter spatiotemporal vision sensor,”IEEE Journal of Solid-State Circuits, vol. 49, no. 10, pp. 2333–2341, 2014

  60. [60]

    GenX320 event-based metavision sensor — product brief,

    Prophesee, “GenX320 event-based metavision sensor — product brief,” Prophesee SA product documentation, https://www.prophesee.ai/event-based-sensor-genx320/, 2024

  61. [61]

    A 640×480 dynamic vision sensor with a 9µm pixel and 300Meps address-event representation,

    B. Son, Y. Suh, S. Kim, H. Jung, J.-S. Kim, C. Shin, K. Park, K. Lee, J. Park, J. Woo, Y. Roh, H. Lee, Y. Wang, I. Ovsiannikov, and H. Ryu, “A 640×480 dynamic vision sensor with a 9µm pixel and 300Meps address-event representation,” inIEEE International Solid-State Circuits Conference (ISSCC), 2017, pp. 66–67

  62. [62]

    Event-based vision on FPGAs – a survey,

    T. Kryjak, “Event-based vision on FPGAs – a survey,” inProc. 27th Euromicro Conference on Digital System Design (DSD), 2024

  63. [63]

    Spiking neural networks on FPGA: A survey of methodologies and recent advancements,

    M. Karamimanesh, E. Abiri, M. Shahsavari, K. Hassanli, A. van Schaik, and J. Eshraghian, “Spiking neural networks on FPGA: A survey of methodologies and recent advancements,” Neural Networks, vol. 186, p. 107256, 2025. 36

  64. [64]

    Loihi: A neuromorphic manycore processor with on-chip learning,

    M. Davies, N. Srinivasa, T.-H. Lin, G. Chinya, Y. Cao, S. H. Choday, G. Dimou, P. Joshi, N. Imam, S. Jain, Y. Liao, C.-K. Lin, A. Lines, R. Liu, D. Mathaikutty, S. McCoy, A. Paul, J. Tse, G. Venkataramanan, Y.-H. Weng, A. Wild, Y. Yang, and H. Wang, “Loihi: A neuromorphic manycore processor with on-chip learning,”IEEE Micro, vol. 38, no. 1, pp. 82–99, 2018

  65. [65]

    A million spiking-neuron integrated circuit with a scalable communication network and interface,

    P. A. Merolla, J. V. Arthur, R. Alvarez-Icaza, A. S. Cassidy, J. Sawada, F. Akopyan, B. L. Jackson, N. Imam, C. Guo, Y. Nakamura, B. Brezzo, I. Vo, S. K. Esser, R. Appuswamy, B. Taba, A. Amir, M. D. Flickner, W. P. Risk, R. Manohar, and D. S. Modha, “A million spiking-neuron integrated circuit with a scalable communication network and interface,” Science,...

  66. [66]

    The SpiNNaker project,

    S. B. Furber, F. Galluppi, S. Temple, and L. A. Plana, “The SpiNNaker project,”Proceed- ings of the IEEE, vol. 102, no. 5, pp. 652–665, 2014

  67. [67]

    Speck: A smart event-based vision sensor with a low latency 327K neuron convolutional neural network processing pipeline,

    O. Richter, Y. Xing, M. De Marchi, C. Nielsen, M. Katsimpris, R. Cattaneo, Y. Ren, Y. Hu, Q. Liu, S. Sheik, T. Demirci, and N. Qiao, “Speck: A smart event-based vision sensor with a low latency 327K neuron convolutional neural network processing pipeline,” 2023

  68. [68]

    Dot-product engine as computing memory to accelerate machine learning algorithms,

    M. Hu, J. P. Strachan, Z. Li, and R. S. Williams, “Dot-product engine as computing memory to accelerate machine learning algorithms,” inProc. 17th IEEE International Symposium on Quality Electronic Design (ISQED), 2016, pp. 374–379

  69. [69]

    NeuroSim: A circuit-level macro model for benchmarking neuro-inspired architectures in online learning,

    P.-Y. Chen, X. Peng, and S. Yu, “NeuroSim: A circuit-level macro model for benchmarking neuro-inspired architectures in online learning,”IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, vol. 37, no. 12, pp. 3067–3080, 2018

  70. [70]

    Complementary resistive switches for passive nanocrossbar memories,

    E. Linn, R. Rosezin, C. Kügeler, and R. Waser, “Complementary resistive switches for passive nanocrossbar memories,”Nature Materials, vol. 9, no. 5, pp. 403–406, 2010

  71. [71]

    Wafer-scale fabrication of memristive passive crossbar circuits for brain-scale neuromorphic computing,

    S. Choi, S. S. Bezugam, T. Bhattacharya, D. Kwon, and D. B. Strukov, “Wafer-scale fabrication of memristive passive crossbar circuits for brain-scale neuromorphic computing,” Nature Communications, 2025

  72. [72]

    A mixed-precision memristor and SRAM compute-in-memory AI processor,

    W.-S. Khwa, T.-H. Wen, H.-H. Hsu, W.-H. Huanget al., “A mixed-precision memristor and SRAM compute-in-memory AI processor,”Nature, vol. 639, no. 8055, pp. 617–623, 2025

  73. [73]

    A 28-nm 135.19 TOPS/W bootstrapped-SRAM compute-in-memory accelerator with layer-wise precision and sparsity,

    W. Mao, D. Liu, H. Zhou, K. Li, Q. Wu, J. Yang, L. Zhang, and H. Yu, “A 28-nm 135.19 TOPS/W bootstrapped-SRAM compute-in-memory accelerator with layer-wise precision and sparsity,”IEEE Transactions on Circuits and Systems I: Regular Papers, 2025

  74. [74]

    Efficient neuromorphic signal processing with Loihi 2,

    G. Orchard, E. P. Frady, D. B.-D. Rubin, S. Sanborn, S. B. Shrestha, F. T. Sommer, and M. Davies, “Efficient neuromorphic signal processing with Loihi 2,” inIEEE Workshop on Signal Processing Systems (SiPS), 2021, pp. 254–259

  75. [75]

    Effect of edge roughness on static characteristics of graphene nanoribbon field effect transistor,

    Y. M. Banadaki and A. Srivastava, “Effect of edge roughness on static characteristics of graphene nanoribbon field effect transistor,”Electronics, vol. 5, no. 1, p. 11, 2016

  76. [76]

    Y. M. Banadaki and S. Sharifi,Graphene Nanostructures: Modeling, Simulation, and Applications in Electronics and Photonics. Jenny Stanford Publishing, 2019

  77. [77]

    Physical modeling of graphene nanoribbon field effect transistor using non-equilibrium green function approach for integrated circuit design,

    Y. M. Banadaki, “Physical modeling of graphene nanoribbon field effect transistor using non-equilibrium green function approach for integrated circuit design,” Ph.D. dissertation, Louisiana State University and Agricultural and Mechanical College, 2016. 37

  78. [78]

    A novel graphene nanoribbon field effect transistor for integrated circuit design,

    Y. M. Banadaki and A. Srivastava, “A novel graphene nanoribbon field effect transistor for integrated circuit design,” in2013 IEEE 56th International Midwest Symposium on Circuits and Systems (MWSCAS). IEEE, 2013, pp. 924–927

  79. [79]

    Two-dimensional material-based devices for in-sensor computing,

    Chen, Wang, and Yan, “Two-dimensional material-based devices for in-sensor computing,” npj Unconventional Computing, vol. 2, p. 19, 2025

  80. [80]

    Light- enhanced molecular polarity enabling multispectral color-cognitive memristor for neuro- morphic visual system,

    J.-M. Lee, B. H. Jeong, E. Kamaraj, D. Kim, H. Kim, S. Park, and H. J. Park, “Light- enhanced molecular polarity enabling multispectral color-cognitive memristor for neuro- morphic visual system,”Nature Communications, 2023

Showing first 80 references.