Recognition: no theorem link
The Luna Bound Propagator for Formal Analysis of Neural Networks
Pith reviewed 2026-05-15 00:54 UTC · model grok-4.3
The pith
A new C++ bound propagator called Luna computes tighter and faster bounds for neural network verification than the existing Python alpha-CROWN.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Luna is a C++ bound propagator that implements Interval Bound Propagation, DeepPoly/CROWN, and alpha-CROWN analysis on general computational graphs, and it outperforms the state-of-the-art Python alpha-CROWN implementation in both bound tightness and computational efficiency on supported VNN-COMP 2025 benchmarks.
What carries the argument
The Luna bound propagator, a C++ reimplementation of the parameterized CROWN (alpha-CROWN) method that computes certified bounds for neural network outputs across multiple analysis modes.
If this is right
- Existing DNN verifiers gain easier integration through the C++ interface.
- Verification runs on supported benchmarks finish with tighter certified bounds and shorter computation times.
- A single framework now covers interval, DeepPoly/CROWN, and alpha-CROWN analyses on general graphs.
- Public release allows direct incorporation into production verification pipelines.
Where Pith is reading between the lines
- Safety-critical applications such as autonomous systems could incorporate tighter formal guarantees at lower overhead.
- Extensions to additional network layers or new bound-propagation variants become feasible once the C++ base is in place.
- Re-implementations in other performance-oriented languages could be compared directly against Luna to isolate language-specific gains.
Load-bearing premise
The C++ implementation correctly and faithfully reproduces the mathematical behavior of the original Python alpha-CROWN without introducing numerical or algorithmic discrepancies that affect the reported tightness or speed gains.
What would settle it
Running the identical VNN-COMP 2025 benchmark networks through both Luna and the reference Python alpha-CROWN and checking whether Luna indeed returns strictly tighter bounds at lower runtimes.
Figures
read the original abstract
The parameterized CROWN analysis, a.k.a., alpha-CROWN has emerged as a practically successful abstract interpretation method for neural network verification. However, existing implementations of alpha-CROWN are limited to Python, which complicates integration into existing DNN verifiers and long-term production-level systems. We introduce Luna, a new abstract-interpretation-based bound propagator implemented in C++. Luna supports Interval Bound Propagation, the DeepPoly/CROWN analysis, and the alpha-CROWN analysis over a general computational graph. We describe the architecture of Luna and show that it outperforms the state-of-the-art alpha-CROWN implementation in terms of both bound tightness and computational efficiency on supported benchmarks from VNN-COMP 2025. Luna is publicly available at https://github.com/ai-ar-research/luna.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces Luna, a C++ implementation of abstract interpretation-based bound propagation supporting Interval Bound Propagation, DeepPoly/CROWN, and alpha-CROWN over general computational graphs for neural network verification. It claims that this implementation outperforms the state-of-the-art Python alpha-CROWN in both bound tightness and computational efficiency on supported VNN-COMP 2025 benchmarks, with the code released publicly at a GitHub repository.
Significance. If the C++ reimplementation faithfully reproduces the mathematical behavior of alpha-CROWN without introducing discrepancies, Luna could enable easier integration of advanced verification techniques into production systems and long-term maintenance, addressing a practical limitation of existing Python tools. The public code release supports reproducibility and potential community adoption.
major comments (2)
- [Experimental Evaluation] Experimental section (results on VNN-COMP 2025): the headline claim of tighter bounds and faster runtime for Luna's alpha-CROWN mode rests on direct comparison to an external Python baseline, but no equivalence verification (e.g., matching intermediate alpha parameters, relaxation matrices, or final bounds on identical inputs) is reported; without this, observed differences could stem from unverified algorithmic or numerical deviations rather than genuine improvements.
- [Implementation Architecture] Implementation architecture (Section 3): the description of how alpha parameter optimization and linear relaxations are realized in C++ does not include explicit checks or invariants ensuring mathematical equivalence to the reference Python alpha-CROWN, which is load-bearing for interpreting the tightness results as faithful reimplementation gains.
minor comments (2)
- [Abstract] Abstract: the phrase 'supported benchmarks from VNN-COMP 2025' is vague; listing the specific networks or properties used would improve clarity without lengthening the text.
- Notation: the paper uses 'alpha-CROWN' interchangeably with 'parameterized CROWN' but does not define the exact parameterization (e.g., how alpha variables are initialized or optimized) in a single location, which could confuse readers unfamiliar with the baseline.
Simulated Author's Rebuttal
We thank the referee for the constructive comments, which highlight important aspects of reproducibility and faithfulness in our comparison. We address each major comment below and will incorporate revisions to strengthen the manuscript.
read point-by-point responses
-
Referee: [Experimental Evaluation] Experimental section (results on VNN-COMP 2025): the headline claim of tighter bounds and faster runtime for Luna's alpha-CROWN mode rests on direct comparison to an external Python baseline, but no equivalence verification (e.g., matching intermediate alpha parameters, relaxation matrices, or final bounds on identical inputs) is reported; without this, observed differences could stem from unverified algorithmic or numerical deviations rather than genuine improvements.
Authors: We agree that explicit equivalence verification is necessary to support the claim that performance differences arise from implementation rather than algorithmic deviations. In the revised manuscript we will add a new subsection in the experimental evaluation that reports direct comparisons on small networks and selected VNN-COMP instances, confirming that Luna produces identical alpha parameters, relaxation matrices, and final bounds (within floating-point tolerance) to the reference Python alpha-CROWN implementation when given identical inputs and initialization. This verification will be performed using the same random seeds and optimization settings. revision: yes
-
Referee: [Implementation Architecture] Implementation architecture (Section 3): the description of how alpha parameter optimization and linear relaxations are realized in C++ does not include explicit checks or invariants ensuring mathematical equivalence to the reference Python alpha-CROWN, which is load-bearing for interpreting the tightness results as faithful reimplementation gains.
Authors: We acknowledge that the current description of the C++ realization lacks explicit statements of the mathematical invariants used. In the revised Section 3 we will add a dedicated paragraph (and accompanying pseudocode) that enumerates the key invariants preserved by our implementation: identical linear relaxation forms for ReLU, identical gradient-based alpha optimization procedure, and identical bound propagation order over the computational graph. We will also describe the unit tests we maintain that assert numerical agreement on these quantities against the reference Python code. revision: yes
Circularity Check
No circularity in Luna's implementation claims or benchmark comparisons
full rationale
The paper presents a C++ reimplementation of established bound-propagation techniques (IBP, DeepPoly/CROWN, alpha-CROWN) and reports empirical tightness and speed results on external VNN-COMP 2025 benchmarks. No derivation chain, first-principles result, or fitted parameter is introduced that reduces to a self-defined quantity. All load-bearing claims rest on direct runtime comparisons against an independent Python baseline rather than on quantities defined in terms of Luna's own outputs or prior self-citations. The architecture description and public code release further separate the contribution from any self-referential construction.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
Bai, J., Lu, F., Zhang, K., et al.: Onnx: Open neural network exchange.https: //github.com/onnx/onnx(2019)
work page 2019
-
[2]
In: International Conference on Computer Aided Verification
Bak, S., Tran, H.D., Hobbs, K., Johnson, T.T.: Improved Geometric Path Enu- meration for Verifying ReLU Neural Networks. In: International Conference on Computer Aided Verification. pp. 66–96. Springer (2020)
work page 2020
-
[3]
arXiv preprint arXiv:2412.19985 (2024)
Brix, C., Bak, S., Johnson, T.T., Wu, H.: The fifth international verification of neu- ral networks competition (vnn-comp 2024): Summary and results. arXiv preprint arXiv:2412.19985 (2024)
-
[4]
In: International Conference on Computer Aided Verification
Duong, H., Nguyen, T., Dwyer, M.B.: Neuralsat: a high-performance verification tool for deep neural networks. In: International Conference on Computer Aided Verification. pp. 409–423. Springer (2025)
work page 2025
-
[5]
arXiv preprint arXiv:1810.12715 (2018)
Gowal, S., Dvijotham, K., Stanforth, R., Bunel, R., Qin, C., Uesato, J., Arand- jelovic, R.,Mann, T.,Kohli, P.: OntheEffectivenessofIntervalBoundPropagation for Training Verifiably Robust Models. arXiv preprint arXiv:1810.12715 (2018)
-
[6]
In: International Joint Con- ference on Artificial Intelligence
Henriksen, P., Lomuscio, A.: DEEPSPLIT: An Efficient Splitting Method for Neu- ral Network Verification via Indirect Effect Analysis. In: International Joint Con- ference on Artificial Intelligence. pp. 2549–2555. ijcai.org (8 2021)
work page 2021
-
[7]
In: Giacomo, G.D., Catalá, A., Dilkina, B., Milano, M., Barro, S., Bugarín, A., Lang, J
Henriksen, P., Lomuscio, A.R.: Efficient Neural Network Verification via Adap- tive Refinement and Adversarial Search. In: Giacomo, G.D., Catalá, A., Dilkina, B., Milano, M., Barro, S., Bugarín, A., Lang, J. (eds.) European Conference on Artificial Intelligence. vol. 325, pp. 2513–2520. IOS Press (2020)
work page 2020
-
[8]
Katz, G., Huang, D., Ibeling, D., Julian, K., Lazarus, C., Lim, R., Shah, P., Thakoor, S., Wu, H., Zeljić, A., Dill, D., Kochenderfer, M., Barrett, C.: The Marabou Framework for Verification and Analysis of Deep Neural Networks. In: Proc. 31st Int. Conf. on Computer Aided Verification (CAV). pp. 443–452 (2019)
work page 2019
-
[9]
arXiv preprint arXiv:2512.19007 (2025)
Kaulen, K., Ladner, T., Bak, S., Brix, C., Duong, H., Flinkow, T., Johnson, T.T., Koller, L., Manino, E., Nguyen, T.H., et al.: The 6th international verification of neural networks competition (vnn-comp 2025): Summary and results. arXiv preprint arXiv:2512.19007 (2025)
-
[10]
Lopez, D.M., Choi, S.W., Tran, H., Johnson, T.T.: NNV 2.0: The Neural Network VerificationTool.In:Enea,C.,Lal,A.(eds.)InternationalConferenceonComputer Aided Verification. vol. 13965, pp. 397–412. Springer (2023)
work page 2023
-
[11]
Towards Deep Learning Models Resistant to Adversarial Attacks
Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083 (2017)
work page internal anchor Pith review Pith/arXiv arXiv 2017
-
[12]
Proceedings of the ACM on Programming Languages6(POPL), 1–33 (2022)
Müller, M.N., Makarchuk, G., Singh, G., Püschel, M., Vechev, M.: Prima: general and precise neural network certification via scalable convex hull approximations. Proceedings of the ACM on Programming Languages6(POPL), 1–33 (2022)
work page 2022
-
[13]
Advances in neural information processing sys- tems32(2019)
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., et al.: Pytorch: An imperative style, high- performance deep learning library. Advances in neural information processing sys- tems32(2019)
work page 2019
-
[14]
Advances in Neural Information Process- ing Systems32, 15098–15109 (2019)
Singh, G., Ganvir, R., Püschel, M., Vechev, M.: Beyond the Single Neuron Convex Barrier for Neural Network Certification. Advances in Neural Information Process- ing Systems32, 15098–15109 (2019)
work page 2019
-
[15]
Singh, G., Gehr, T., Püschel, M., Vechev, M.: An Abstract Domain for Certifying Neural Networks. Proceedings of the ACM on Programming Languages3(POPL), 1–30 (2019) TheLunaBound Propagator for Formal Analysis of Neural Networks 9
work page 2019
-
[16]
In: International Conference on Learning Representations (2019)
Singh, G., Gehr, T., Püschel, M., Vechev, M.: Boosting Robustness Certification of Neural Networks. In: International Conference on Learning Representations (2019)
work page 2019
-
[17]
In: International Conference on Learning Represen- tations (2019)
Tjeng, V., Xiao, K.Y., Tedrake, R.: Evaluating robustness of neural networks with mixed integer programming. In: International Conference on Learning Represen- tations (2019)
work page 2019
-
[18]
In: International Confer- ence on Computer Aided Verification
Tran, H.D., Yang, X., Manzanas Lopez, D., Musau, P., Nguyen, L.V., Xiang, W., Bak,S.,Johnson,T.T.:NNV:theNeuralNetworkVerificationToolforDeepNeural Networks and Learning-enabled Cyber-Physical Systems. In: International Confer- ence on Computer Aided Verification. pp. 3–17. Springer (2020)
work page 2020
-
[19]
Advances in Neural Information Processing Systems 31, 6369–6379 (2018)
Wang, S., Pei, K., Whitehouse, J., Yang, J., Jana, S.: Efficient Formal Safety Analysis of Neural Networks. Advances in Neural Information Processing Systems 31, 6369–6379 (2018)
work page 2018
-
[20]
Advances in Neural Information Processing Sys- tems34, 29909–29921 (2021)
Wang, S., Zhang, H., Xu, K., Lin, X., Jana, S., Hsieh, C.J., Kolter, J.Z.: Beta- crown: Efficient Bound Propagation with Per-neuron Split Constraints for Neural Network Robustness Verification. Advances in Neural Information Processing Sys- tems34, 29909–29921 (2021)
work page 2021
-
[21]
arXiv preprint arXiv:2401.14461 (2024)
Wu, H., Isac, O., Zeljić, A., Tagomori, T., Daggitt, M., Kokke, W., Refaeli, I., Amir, G., Julian, K., Bassan, S., et al.: Marabou 2.0: A Versatile Formal Analyzer of Neural Networks. arXiv preprint arXiv:2401.14461 (2024)
-
[22]
Advances in Neural Information Processing Systems33, 1129–1141 (2020)
Xu, K., Shi, Z., Zhang, H., Wang, Y., Chang, K.W., Huang, M., Kailkhura, B., Lin, X., Hsieh, C.J.: Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond. Advances in Neural Information Processing Systems33, 1129–1141 (2020)
work page 2020
- [23]
-
[24]
Advances in Neural Information Processing Systems35, 1656–1670 (2022)
Zhang, H., Wang, S., Xu, K., Li, L., Li, B., Jana, S., Hsieh, C.J., Kolter, J.Z.: General Cutting Planes for Bound-propagation-based Neural Network Verification. Advances in Neural Information Processing Systems35, 1656–1670 (2022)
work page 2022
-
[25]
Zhang, H., Weng, T.W., Chen, P.Y., Hsieh, C.J., Daniel, L.: Efficient Neural Net- work Robustness Certification with General Activation Functions. Advances in Neural Information Processing Systems31, 4944–4953 (2018) 10 Henry LeCates and Haoze Wu A Architecture and Core Components This appendix records implementation-level details that complement Section ...
work page 2018
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.