pith. machine review for the scientific record. sign in

arxiv: 2604.19993 · v1 · submitted 2026-04-21 · 💻 cs.AR · cs.LG

Recognition: unknown

Algorithm and Hardware Co-Design for Efficient Complex-Valued Uncertainty Estimation

Authors on Pith no claims yet

Pith reviewed 2026-05-10 00:55 UTC · model grok-4.3

classification 💻 cs.AR cs.LG
keywords complex-valued neural networksBayesian inferencedropoutuncertainty quantificationFPGA acceleratorsneural architecture searchhardware-software co-designalgorithm-hardware co-design
0
0 comments X

The pith

Bayesian dropout enables uncertainty quantification directly in complex-valued neural networks with hardware-efficient implementations.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows how to bring uncertainty estimation to complex-valued neural networks by extending dropout-based Bayesian methods to handle complex arithmetic. This is important for applications like signal processing where complex numbers are natural but confidence in predictions has been hard to obtain. The dual nature of complex values creates new opportunities for mixing real and imaginary components across layers, which an automated search exploits to find configurations that are both accurate and hardware-friendly. A framework then generates optimized FPGA accelerators from these designs, delivering large speedups and low power use compared to GPU baselines. If successful, this makes complex-valued models practical for embedded or real-time systems that require both accuracy and reliability estimates.

Core claim

We propose, for the first time, dropout-based Bayesian Complex-Valued Neural Networks (BayesCVNNs) to enable uncertainty quantification for complex-valued applications, exhibiting broad applicability and efficiency for hardware implementation due to modularity. As the dual-part nature of complex values broadens the design space with novel layer-mixing and part-mixing configurations, an automated search identifies optimal setups for real and imaginary parts. A framework generates customized FPGA-based accelerators using optimized building blocks, with experiments showing the best configurations attain higher performance with lower hardware costs.

What carries the argument

dropout-based variational inference extended to complex-valued layers combined with automated search over layer-mixing and part-mixing configurations

Load-bearing premise

That dropout remains a valid variational approximation to Bayesian inference when the network weights and operations are complex-valued rather than real-valued.

What would settle it

Observing whether the uncertainty estimates from BayesCVNNs are well-calibrated on a complex-valued test set, or whether the automatically searched configurations use fewer hardware resources than the best manually designed ones while maintaining or improving accuracy.

Figures

Figures reproduced from arXiv: 2604.19993 by He Li, Mark Chen, Wayne Luk, Zehuan Zhang.

Figure 1
Figure 1. Figure 1: Bayesian Layer Configurations. By contrast, in real domains, the Bayesian layer configuration is fixed. Therefore, BayesCVNNs exhibit a 3 𝑁 times larger design space than BayesNNs in real domains, offering flexibility to explore trade-offs between accuracy, uncertainty, and hardware costs. 4 Framework for Accelerator Generation A framework is developed to produce the optimal model configura￾tion and genera… view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the Framework. 4.3 Hardware Mapping Existing Qkeras libraries do not support complex layers. To enable the construction of BayesCVNNs, we develop a set of building blocks with hardware mapping schemes. According to operational characteristics, complex layers can be categorized into three classes. The first class includes complex convolutional and complex fully connected layers, both requiring f… view at source ↗
Figure 3
Figure 3. Figure 3: Pseudocode of Bernoulli Dropout. 5 Experiments and Evaluation The proposed BayesCVNNs are implemented using PyTorch 1.9.0. Vivado-HLS 2019.2 is used for hardware implementation. QKeras 0.9.0 is used for quantization. The produced C-synthesis reports provide the performance of latency and resource consumption. The place-and-route optimizations are implemented by Vivado 2019.2 for the final designs. The FPGA… view at source ↗
Figure 4
Figure 4. Figure 4: Search Results of Complex-Valued Tasks. 5.2 Analysis of Mapping Schemes The performance difference between the two mapping schemes comes from the exploited reusability. To quantitatively investigate the effects, we take complex fully connected layers as an exam￾ple. We generate and synthesize the core operations, i.e. four sub￾operations, with varying input and output dimensions. The input dimension is set… view at source ↗
Figure 5
Figure 5. Figure 5: Resource and Latency of Mapping Schemes. [PITH_FULL_IMAGE:figures/full_fig_p005_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Resource Utilization of Different Models. [PITH_FULL_IMAGE:figures/full_fig_p006_6.png] view at source ↗
read the original abstract

Complex-Valued Neural Networks (CVNNs) have significant advantages in handling tasks that involve complex numbers. However, existing CVNNs are unable to quantify predictive uncertainty. We propose, for the first time, dropout-based Bayesian Complex-Valued Neural Networks (BayesCVNNs) to enable uncertainty quantification for complex-valued applications, exhibiting broad applicability and efficiency for hardware implementation due to modularity. Furthermore, as the dual-part nature of complex values significantly broadens the design space and enables novel configurations based on layer-mixing and part-mixing, we introduce an automated search approach to effectively identify optimal configurations for both real and imaginary components. To facilitate deployment, we present a framework that generates customized FPGA-based accelerators for BayesCVNNs, leveraging a set of optimized building blocks. Experiments demonstrate the best configuration can be effectively found via the automated search, attaining higher performance with lower hardware costs compared with manually crafted models. The optimized accelerators achieve approximately 4.5x and 13x speedups on different models with less than 10% power consumption compared to GPU implementations, and outperform existing work in both algorithm and hardware aspects. Our code is publicly available at: https://github.com/zehuanzhang/BayesCVNN.git.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes the first dropout-based Bayesian Complex-Valued Neural Networks (BayesCVNNs) for predictive uncertainty quantification in CVNNs. It introduces an automated search over layer-mixing and part-mixing configurations to optimize accuracy and hardware cost, together with a modular FPGA accelerator generation framework that produces custom accelerators from optimized building blocks. Experiments are said to show that the search finds superior configurations and that the resulting accelerators deliver ~4.5× and 13× speedups versus GPU baselines at <10% power.

Significance. If the central claims are substantiated, the work would provide a practical route to uncertainty-aware complex-valued models together with hardware co-design, which is relevant for signal-processing and communications applications. Public release of the code is a clear positive.

major comments (2)
  1. [Abstract and §4] Abstract and §4 (Experiments): the claim that the automated search attains “higher performance with lower hardware costs” and that accelerators achieve 4.5×/13× speedups rests on unspecified datasets, baselines, uncertainty calibration metrics (e.g., ECE, NLL), and synthesis results. Without these details the experimental support for the co-design claim cannot be evaluated.
  2. [§3] §3 (BayesCVNN construction): the paper applies real-valued dropout (layer-wise or part-wise) inside complex multiplication and addition to obtain a variational posterior, yet provides no calibration study, comparison against exact inference, or analysis of whether the induced variational family respects the circular geometry of complex posteriors. This assumption is load-bearing for the entire uncertainty-quantification claim.
minor comments (2)
  1. [§3] Notation for real/imaginary part mixing and the precise definition of the search space (layer-mixing vs. part-mixing) should be formalized with a small table or diagram early in §3 to improve readability.
  2. [Abstract] The abstract states “outperform existing work in both algorithm and hardware aspects” without naming the prior CVNN or Bayesian hardware papers being compared; a short related-work paragraph or table would clarify the novelty.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the thorough and constructive review. The comments highlight important areas for clarification and strengthening, particularly around experimental details and theoretical justification of the variational approximation. We address each point below and commit to revisions that improve the manuscript without misrepresenting our contributions.

read point-by-point responses
  1. Referee: [Abstract and §4] Abstract and §4 (Experiments): the claim that the automated search attains “higher performance with lower hardware costs” and that accelerators achieve 4.5×/13× speedups rests on unspecified datasets, baselines, uncertainty calibration metrics (e.g., ECE, NLL), and synthesis results. Without these details the experimental support for the co-design claim cannot be evaluated.

    Authors: We agree that the abstract and experimental section would benefit from greater specificity to allow full evaluation of the claims. In the revised version, we will explicitly name the datasets (complex-valued signal processing and communications benchmarks used in §4), detail the GPU baselines (including hardware platforms and software implementations), report uncertainty calibration metrics such as ECE and NLL alongside accuracy, and expand the synthesis results with concrete FPGA metrics (resource utilization, latency, power). These additions will directly support the co-design performance claims. revision: yes

  2. Referee: [§3] §3 (BayesCVNN construction): the paper applies real-valued dropout (layer-wise or part-wise) inside complex multiplication and addition to obtain a variational posterior, yet provides no calibration study, comparison against exact inference, or analysis of whether the induced variational family respects the circular geometry of complex posteriors. This assumption is load-bearing for the entire uncertainty-quantification claim.

    Authors: We acknowledge the importance of validating the variational approximation. Our method extends the established real-valued dropout variational inference to CVNNs via layer-wise and part-mixing strategies applied to complex operations. Exact inference is intractable for networks of this size (as is standard in the BNN literature), precluding direct comparison. In the revision we will add: (i) a discussion of how the part-mixing and layer-mixing configurations preserve key complex properties, (ii) empirical calibration results using ECE and NLL, and (iii) analysis of the induced posterior relative to circular symmetry. This strengthens the uncertainty-quantification foundation while preserving the paper's focus on the novel co-design. revision: partial

Circularity Check

0 steps flagged

No circularity: novel BayesCVNN construction and search evaluated empirically

full rationale

The paper introduces dropout-based variational inference for complex-valued networks as a new construction, along with an automated search over layer-mixing and part-mixing configurations whose outputs are measured against external GPU baselines and manual designs. No derivation step reduces to a self-definition, fitted parameter renamed as prediction, or load-bearing self-citation chain; the central claims rest on empirical validation rather than tautological equivalence to inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

The central claim rests on the unstated premise that dropout variational inference transfers to complex arithmetic without invalidating the uncertainty estimates and that the expanded mixing search space contains practically superior points. No free parameters, axioms, or invented physical entities are explicitly introduced in the abstract.

pith-pipeline@v0.9.0 · 5521 in / 1243 out tokens · 56924 ms · 2026-05-10T00:55:36.855274+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

51 extracted references · 5 canonical work pages

  1. [1]

    Maruf Ahmad, Lei Zhang, and Muhammad EH Chowdhury. 2024. FPGA im- plementation of complex-valued neural network for polar-represented image classification.Sensors, 24, 3, 897

  2. [2]

    H. N. Reza Mohammadi Asiyabi, M. Datcu, and A. Anghel. 2023. S1slc_cvdl: a complex-valued annotated single look complex sentinel-1 sar dataset for complex-valued deep networks.IEEE Dataport. Early access. (Apr. 2023). doi:1 0.1109/TGRS.2023.3267185

  3. [3]

    Hiromitsu Awano and Masanori Hashimoto. 2020. BYNQnet: Bayesian neural network with quadratic activations for sampling-free uncertainty estimation on FPGA. In2020 Design, Automation & Test in Europe Conference & Exhibition (DATE). IEEE, 1402–1407

  4. [4]

    Jose Agustin Barrachina, Chenfang Ren, Christele Morisseau, Gilles Vieillard, and J-P Ovarlez. 2021. Complex-valued vs. real-valued neural networks for clas- sification perspectives: an example on non-circular data. InIEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2990–2994

  5. [5]

    David M Blei, Alp Kucukelbir, and Jon D McAuliffe. 2017. Variational inference: a review for statisticians.Journal of the American Statistical Association, 112, 518, 859–877

  6. [6]

    Charles Blundell, Julien Cornebise, Koray Kavukcuoglu, and Daan Wierstra

  7. [7]

    InInternational Conference on Machine Learning

    Weight uncertainty in neural network. InInternational Conference on Machine Learning. PMLR, 1613–1622

  8. [8]

    Ruizhe Cai, Ao Ren, Ning Liu, Caiwen Ding, Luhao Wang, Xuehai Qian, Massoud Pedram, and Yanzhi Wang. 2018. VIBNN: hardware acceleration of Bayesian neural networks.ACM SIGPLAN Notices, 53, 2, 476–488

  9. [9]

    Yice Cao, Yan Wu, Peng Zhang, Wenkai Liang, and Ming Li. 2019. Pixel-wise polsar image classification via a novel complex-valued deep fully convolutional network.Remote Sensing, 11, 22, 2653

  10. [10]

    Elizabeth Cole, Joseph Cheng, John Pauly, and Shreyas Vasanawala. 2021. Analysis of deep complex-valued convolutional neural networks for MRI re- construction and phase-focused applications.Magnetic Resonance in Medicine, 86, 2, 1093–1109

  11. [11]

    Hongxiang Fan, Mark Chen, Liam Castelli, Zhiqiang Que, He Li, Kenneth Long, and Wayne Luk. 2023. When Monte-Carlo dropout meets multi-exit: op- timizing Bayesian neural networks on FPGA. InACM/IEEE Design Automation Conference. IEEE, 1–6

  12. [12]

    Hongxiang Fan, Martin Ferianc, Zhiqiang Que, Shuanglong Liu, Xinyu Niu, Miguel RD Rodrigues, and Wayne Luk. 2022. FPGA-based acceleration for Bayesian convolutional neural networks.IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41, 12, 5343–5356

  13. [13]

    Hongxiang Fan, Martin Ferianc, Zhiqiang Que, Xinyu Niu, Miguel Rodrigues, and Wayne Luk. 2022. Accelerating bayesian neural networks via algorithmic and hardware optimizations.IEEE Transactions on Parallel and Distributed Systems, 33, 12, 3387–3399

  14. [14]

    Hongxiang Fan, Martin Ferianc, Miguel Rodrigues, Hongyu Zhou, Xinyu Niu, and Wayne Luk. 2021. High-performance FPGA-based accelerator for Bayesian neural networks. In2021 58th ACM/IEEE Design Automation Conference (DAC). IEEE, 1063–1068

  15. [15]

    Martin Ferianc, Divyansh Manocha, Hongxiang Fan, and Miguel Rodrigues

  16. [16]

    InInternational Conference on Artificial Neural Networks

    Combinet: compact convolutional bayesian neural network for image seg- mentation. InInternational Conference on Artificial Neural Networks. Springer, 483–494

  17. [17]

    Martin Ferianc, Zhiqiang Que, Hongxiang Fan, Wayne Luk, and Miguel Ro- drigues. 2021. Optimizing Bayesian recurrent neural networks on an FPGA- based accelerator. In2021 International Conference on Field-Programmable Tech- nology (ICFPT). IEEE, 1–10

  18. [18]

    Yoshiki Fujiwara and Shinya Takamaeda-Yamazaki. 2021. ASBNN: acceleration of Bayesian convolutional neural networks by algorithm-hardware co-design. In2021 IEEE 32nd International Conference on Application-specific Systems, Architectures and Processors (ASAP). IEEE, 226–233

  19. [19]

    Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a Bayesian approximation: representing model uncertainty in deep learning. InInternational Conference on Machine Learning. PMLR, 1050–1059

  20. [20]

    Jingkun Gao, Bin Deng, Yuliang Qin, Hongqiang Wang, and Xiang Li. 2018. Enhanced radar imaging using a complex-valued convolutional neural network. IEEE Geoscience and Remote Sensing Letters, 16, 1, 35–39

  21. [21]

    Ethan Goan and Clinton Fookes. 2020. Bayesian neural networks: an introduc- tion and survey.Case Studies in Applied Bayesian Data Science: CIRM Jean-Morlet Chair, Fall 2018, 45–87

  22. [22]

    Su Lee Goh, Mo Chen, DH Popović, K Aihara, D Obradovic, and DP Mandic

  23. [23]

    Complex-valued forecasting of wind profile.Renewable Energy, 31, 11, 1733–1750

  24. [24]

    Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. 2017. On calibration of modern neural networks. InInternational conference on machine learning. PMLR, 1321–1330

  25. [25]

    Laurent Valentin Jospin, Hamid Laga, Farid Boussaid, Wray Buntine, and Mohammed Bennamoun. 2022. Hands-on Bayesian neural networks—a tutorial for deep learning users.IEEE Computational Intelligence Magazine, 17, 2, 29–48

  26. [26]

    Alex Kendall, Vijay Badrinarayanan, and Roberto Cipolla. 2015. Bayesian seg- net: model uncertainty in deep convolutional encoder-decoder architectures for scene understanding.arXiv preprint arXiv:1511.02680

  27. [27]

    Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. 2002. Gradient- based learning applied to document recognition.Proceedings of the IEEE, 86, 11, 2278–2324

  28. [28]

    Hyunwuk Lee, Hyungjun Jang, Sungbin Kim, Sungwoo Kim, Wonho Cho, and Won Woo Ro. 2023. Exploiting inherent properties of complex numbers for accelerating complex valued neural networks. InProceedings of the 56th Annual IEEE/ACM International Symposium on Microarchitecture, 1121–1134

  29. [29]

    Ruochen Li, Jiazhen Pan, Youxiang Zhu, Juncheng Ni, and Daniel Rueckert

  30. [30]

    InInternational Workshop on Machine Learning in Medical Imaging

    Classification, regression and segmentation directly from k-space in cardiac mri. InInternational Workshop on Machine Learning in Medical Imaging. Springer, 31–41

  31. [31]

    Mingqi Lv, Wei Xu, and Tieming Chen. 2019. A hybrid deep convolutional and recurrent neural network for complex activity recognition using multimodal sensors.Neurocomputing, 362, 33–40

  32. [32]

    Huilin Mu, Yun Zhang, Yicheng Jiang, and Chang Ding. 2021. CV-GMTINet: GMTI using a deep complex-valued convolutional neural network for multi- channel SAR-GMTI system.IEEE Transactions on Geoscience and Remote Sensing, 60, 1–15

  33. [33]

    Radford Neal. 1992. Bayesian learning via stochastic dynamics.Advances in Neural Information Processing Systems, 5

  34. [34]

    Radford M Neal et al. 2011. Mcmc using hamiltonian dynamics.Handbook of Markov Chain Monte Carlo, 2, 11, 2

  35. [35]

    Othmane-Latif Ouabi, Radmila Pribić, and Sorin Olaru. 2021. Stochastic complex- valued neural networks for radar. In28th European Signal Processing Conference, 1442–1446. doi:10.23919/Eusipco47968.2020.9287425

  36. [36]

    Hongwu Peng et al. 2021. Binary complex neural network acceleration on FPGA. In2021 IEEE 32nd International Conference on Application-specific Systems, Architectures and Processors (ASAP). IEEE, 85–92

  37. [37]

    Călin-Adrian Popa. 2017. Complex-valued convolutional neural networks for real-valued image classification. InInternational Joint Conference on Neural Networks. IEEE, 816–822

  38. [38]

    Zhiqiang Que, Hongxiang Fan, Gabriel Figueiredo, Ce Guo, Wayne Luk, Ryota Yasudo, and Masato Motomura. 2025. Trustworthy deep learning accelera- tion with customizable design flow automation. InProceedings of the 15th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies, 1–13

  39. [39]

    Koosha Sharifani and Mahyar Amini. 2023. Machine learning and deep learn- ing: a review of methods and applications.World Information Technology and Engineering Journal, 10, 07, 3897–3904

  40. [40]

    Yuki Sunaga, Ryo Natsuaki, and Akira Hirose. 2019. Land form classification and similar land-shape discovery by using complex-valued convolutional neural networks.IEEE Transactions on Geoscience and Remote Sensing, 57, 10, 7907– 7917

  41. [41]

    Chiheb Trabelsi et al. 2017. Deep complex networks.arXiv preprint arXiv: 1705.09792

  42. [42]

    Shanshan Wang, Huitao Cheng, Leslie Ying, Taohui Xiao, Ziwen Ke, Hairong Zheng, and Dong Liang. 2020. Deepcomplexmri: exploiting deep residual net- work for fast parallel mr imaging with complex convolution.Magnetic resonance imaging, 68, 136–147

  43. [43]

    Max Welling and Yee W Teh. 2011. Bayesian learning via stochastic gradi- ent langevin dynamics. InProceedings of the 28th international conference on machine learning (ICML-11). Citeseer, 681–688

  44. [44]

    Linfang Xiao, Yilong Liu, Zheyuan Yi, Yujiao Zhao, Linshan Xie, Peibei Cao, Alex TL Leong, and Ed X Wu. 2022. Partial fourier reconstruction of complex mr images using complex-valued convolutional neural networks.Magnetic Resonance in Medicine, 87, 2, 999–1014

  45. [45]

    Ximei Yang, Ronny G Guendel, Alexander Yarovoy, and Francesco Fioranelli

  46. [46]

    In2022 IEEE radar conference (radarConf22)

    Radar-based human activities classification with complex-valued neural networks. In2022 IEEE radar conference (radarConf22). IEEE, 1–6

  47. [47]

    Zehuan Zhang, Hongxiang Fan, Hao Chen, Lukasz Dudziak, and Wayne Luk

  48. [48]

    InProceedings of the 61st ACM/IEEE Design Automation Conference (DAC), 1–6

    Hardware-aware neural dropout search for reliable uncertainty prediction on FPGA. InProceedings of the 61st ACM/IEEE Design Automation Conference (DAC), 1–6

  49. [49]

    Zehuan Zhang, Matej Genci, Hongxiang Fan, Andreas Wetscherek, and Wayne Luk. 2024. Accelerating MRI uncertainty estimation with mask-based Bayesian neural network. InIEEE 35th International Conference on Application-specific Systems, Architectures and Processors. IEEE, 107–115

  50. [50]

    Zhilu Zhang, Adrian V Dalca, and Mert R Sabuncu. 2019. Confidence calibration for convolutional neural networks using structured dropout.arXiv preprint arXiv:1906.09551. DAC ’26, July 26–29, 2026, Long Beach, CA, USA Zhang et al

  51. [51]

    Ke Zou, Zhihao Chen, Xuedong Yuan, Xiaojing Shen, Meng Wang, and Huazhu Fu. 2023. A review of uncertainty estimation and its application in medical imaging.Meta-Radiology, 100003