Recognition: no theorem link
Minimal Filling Architectures of Polynomial Neural Networks: Counterexamples, Frontier Search, and Defects
Pith reviewed 2026-05-12 02:15 UTC · model grok-4.3
The pith
A minimal filling architecture for polynomial neural networks need not have unimodal hidden-layer widths.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
We exhibit a counterexample to the minimal unimodal conjecture for polynomial neural networks with power activation functions. With input and output widths held fixed, there exists a minimal filling architecture whose sequence of hidden widths is not unimodal. The architecture was discovered by frontier search and certified by recursive dimension bounds combined with symbolic computation. Several of its subarchitectures display large defect, unlike the predominantly small-defect behavior of previously examined examples.
What carries the argument
Frontier search over candidate width sequences, certified by recursive dimension bounds on the represented function space and by direct symbolic computation.
If this is right
- Minimal filling architectures for polynomial networks can have non-unimodal hidden widths.
- The unimodal conjecture does not hold in general for power activations.
- Subarchitectures of a minimal filling network can exhibit large defect.
- Frontier search combined with dimension bounds can locate counterexamples that symbolic checks then confirm.
Where Pith is reading between the lines
- Network design rules that assume unimodal widths may miss some of the smallest architectures.
- Large-defect subnetworks inside a minimal architecture may point to regions where further width reductions are still possible.
- The same search-and-certify method could be applied to other activation families or to deeper networks to test related conjectures.
Load-bearing premise
The architecture located by the search is genuinely minimal, fills the target function space, and truly fails to be unimodal.
What would settle it
An explicit computation showing either that the dimension of functions realized by this width sequence falls short of the target dimension or that some strictly smaller architecture with unimodal widths already fills the space.
Figures
read the original abstract
We provide a counterexample to the minimal unimodal conjecture for polynomial neural networks (PNNs) with power activation functions. Fixing the input and output widths, the conjecture states that any minimal filling architecture has unimodal widths for the hidden layers. We found a counterexample via a frontier search and certified it using recursive dimension bounds and symbolic computation. Notably, several subarchitectures of this example exhibit large defect, in contrast with the predominantly small-defect behavior observed in prior examples.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript provides a counterexample to the minimal unimodal conjecture for polynomial neural networks (PNNs) with power activation functions. For fixed input and output widths, the conjecture claims that every minimal filling architecture has unimodal hidden-layer widths. The authors locate a counterexample by frontier search and certify that it is minimal and non-unimodal by means of recursive dimension bounds together with symbolic computation. They further report that several of its subarchitectures display large defects, in contrast to the small-defect pattern seen in earlier examples.
Significance. A verified counterexample would refute the minimal unimodal conjecture and thereby sharpen the understanding of which width sequences can realize minimal filling architectures in PNNs. The combination of automated frontier search with rigorous, machine-checkable certification constitutes a methodological contribution that could be reused for related conjectures. The observation of large defects in subarchitectures supplies new empirical material on defect behavior.
major comments (1)
- [Certification procedure] Certification procedure (the section describing recursive dimension bounds and symbolic computation): the manuscript states that the discovered architecture is minimal and non-unimodal, yet the explicit recursive bounds, the symbolic expressions that were checked, and the termination criteria of the computation are not reproduced in sufficient detail for an independent reader to reconstruct the verification. Because this certification is the sole warrant for the central claim, the omission is load-bearing.
minor comments (2)
- [Frontier search] The description of the frontier-search algorithm would benefit from a concise pseudocode block or a reference to the exact implementation parameters (depth limit, branching factor, pruning rule) used to locate the counterexample.
- [Defect analysis] Table or figure that lists the widths of the reported counterexample and its subarchitectures should include the corresponding defect values so that the “large defect” claim can be inspected at a glance.
Simulated Author's Rebuttal
We thank the referee for the positive evaluation of the significance of our counterexample and the methodological aspects of the work. We address the single major comment below and will incorporate the requested clarifications in a revised manuscript.
read point-by-point responses
-
Referee: [Certification procedure] Certification procedure (the section describing recursive dimension bounds and symbolic computation): the manuscript states that the discovered architecture is minimal and non-unimodal, yet the explicit recursive bounds, the symbolic expressions that were checked, and the termination criteria of the computation are not reproduced in sufficient detail for an independent reader to reconstruct the verification. Because this certification is the sole warrant for the central claim, the omission is load-bearing.
Authors: We agree that the current manuscript does not supply the certification details at a level sufficient for independent reconstruction. In the revised version we will expand the relevant section to include: (i) the complete set of recursive dimension bounds applied at each layer, (ii) the precise symbolic expressions that were verified by the computer-algebra system, and (iii) the termination criteria used both for the frontier search and for the subsequent minimality check. We will also add a short algorithmic outline and, if space permits, a supplementary file containing the raw computation log. revision: yes
Circularity Check
No significant circularity detected
full rationale
The paper reports a computational search for a counterexample to the minimal unimodal conjecture for PNNs, certified by recursive dimension bounds and symbolic computation. This process is a direct enumeration and verification procedure rather than a derivation that reduces by construction to fitted inputs, self-definitional relations, or load-bearing self-citations. No equations or steps in the described argument equate a claimed prediction or uniqueness result to its own inputs; the certification step supplies independent evidence external to any ansatz or prior fit within the paper itself.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[2]
M.-C. Brandenburg, G. Loho, and G. Montufar , The real tropical geometry of neural networks for binary classification , Transactions on Machine Learning Research, (2024). https://openreview.net/forum?id=I7JWf8XA2w URL
work page 2024
-
[4]
B. Finkel, J. I. Rodriguez, C. Wu, and T. Yahl , A ctivation degree thresholds and expressiveness of polynomial neural networks , Algebraic Statistics, 16 (2025), pp. 113--130. https://doi.org/10.2140/astat.2025.16.113 DOI
-
[5]
O’Connor, and Kevin McGuinness
M. Goyal, R. Goyal, and B. Lall , I mproved P olynomial N eural N etworks with N ormalised A ctivations , in 2020 International Joint Conference on Neural Networks (IJCNN), 2020, pp. 1--8. https://doi.org/10.1109/IJCNN48605.2020.9207535 DOI
-
[6]
N. Henry, G. L. Marchetti, and K. Kohn , Geometry of lightning self-attention: Identifiability and dimension , in International Conference on Learning Representations, Y. Yue, A. Garg, N. Peng, F. Sha, and R. Yu, eds., vol. 2025, 2025, pp. 14400--14416. https://proceedings.iclr.cc/paper_files/paper/2025/file/259e59fe23ebd09252647fed42949182-Paper-Conferen...
work page 2025
- [7]
- [8]
-
[10]
K. Kubjas, J. Li, and M. Wiesmann , G eometry of P olynomial N eural N etworks , Algebraic Statistics, 15 (2024), pp. 295--328. https://doi.org/10.2140/astat.2024.15.295 DOI
-
[11]
G. L. Marchetti, V. Shahverdi, S. Mereta, M. Trager, and K. Kohn , P osition: A lgebra U nveils D eep L earning -- A n I nvitation to N euroalgebraic G eometry , in Proceedings of the 42nd International Conference on Machine Learning, PMLR, 2025. https://doi.org/10.48550/arXiv.2501.18915 DOI
-
[13]
K. Usevich, R. Borsoi, C. D\' e rand, and M. Clausel , Identifiability of deep polynomial neural networks , in Advances in Neural Information Processing Systems, vol. 38, Curran Associates, Inc., 2025, pp. 81809--81858. https://openreview.net/forum?id=MrUsZfQ9pC URL
work page 2025
-
[14]
L. Zhang, G. Naitzat, and L.-H. Lim , Tropical geometry of deep neural networks , in Proceedings of the 35th International Conference on Machine Learning, J. Dy and A. Krause, eds., vol. 80 of Proceedings of Machine Learning Research, PMLR, 10--15 Jul 2018, pp. 5824--5832. https://proceedings.mlr.press/v80/zhang18i.html URL
work page 2018
-
[15]
J. Zhou, H. Qian, X. Lu, Z. Duan, H. Huang, and Z. Shao , Polynomial activation neural networks: M odeling, stability analysis and coverage BP -training , Neurocomputing, 359 (2019), pp. 227--240. https://doi.org/10.1016/j.neucom.2019.06.004 DOI
-
[16]
Geometry of Lightning Self-Attention: Identifiability and Dimension , url =
Henry, Nathan and Marchetti, Giovanni Luca and Kohn, Kathl\'. Geometry of Lightning Self-Attention: Identifiability and Dimension , url =. International Conference on Learning Representations , editor =
-
[17]
Lim, Lek-Heng and Nelson, Bradley J. , TITLE =. Notices Amer. Math. Soc. , FJOURNAL =. 2023 , NUMBER =. doi:10.1090/noti2666 , url =
-
[18]
Transactions on Machine Learning Research , issn=
The Real Tropical Geometry of Neural Networks for Binary Classification , author=. Transactions on Machine Learning Research , issn=. 2024 , url=
work page 2024
-
[19]
Proceedings of the 35th International Conference on Machine Learning , pages =
Tropical Geometry of Deep Neural Networks , author =. Proceedings of the 35th International Conference on Machine Learning , pages =. 2018 , editor =
work page 2018
-
[21]
Kohn, Kathl\'en and Merkh, Thomas and Mont\'ufar, Guido and Trager, Matthew , TITLE =. SIAM J. Appl. Algebra Geom. , FJOURNAL =. 2022 , NUMBER =. doi:10.1137/21M1441183 , url =
-
[22]
Kohn, Kathl\'en and Mont\'ufar, Guido and Shahverdi, Vahid and Trager, Matthew , TITLE =. SIAM J. Appl. Algebra Geom. , FJOURNAL =. 2024 , NUMBER =. doi:10.1137/23M1565504 , url =
-
[23]
Chrysos, Grigorios G. and Moschoglou, Stylianos and Bouritsas, Giorgos and Deng, Jiankang and Panagakis, Yannis and Zafeiriou, Stefanos , title =. IEEE Trans. Pattern Anal. Mach. Intell. , month1 = aug, pages =. 2022 , issue_date1 =. doi:10.1109/TPAMI.2021.3058891 , abstract =
-
[24]
Identifiability of Deep Polynomial Neural Networks , url =
Usevich, Konstantin and Borsoi, Ricardo and D\'. Identifiability of Deep Polynomial Neural Networks , url =. Advances in Neural Information Processing Systems , editor2 =
-
[25]
SIAM Journal on Applied Algebra and Geometry , volume =
Arjevani, Yossi and Bruna, Joan and Kileel, Joe and Polak, Elzbieta and Trager, Matthew , title =. SIAM Journal on Applied Algebra and Geometry , volume =. 2026 , doi =. https://doi.org/10.1137/25M1732994 , abstract =
-
[26]
P. Langley , title =. Proceedings of the 17th International Conference on Machine Learning (ICML 2000) , address =. 2000 , pages =
work page 2000
-
[27]
T. M. Mitchell. The Need for Biases in Learning Generalizations. 1980
work page 1980
-
[28]
M. J. Kearns , title =
-
[29]
Machine Learning: An Artificial Intelligence Approach, Vol. I. 1983
work page 1983
-
[30]
R. O. Duda and P. E. Hart and D. G. Stork. Pattern Classification. 2000
work page 2000
-
[31]
Suppressed for Anonymity , author=
-
[32]
A. Newell and P. S. Rosenbloom. Mechanisms of Skill Acquisition and the Law of Practice. Cognitive Skills and Their Acquisition. 1981
work page 1981
-
[33]
A. L. Samuel. Some Studies in Machine Learning Using the Game of Checkers. IBM Journal of Research and Development. 1959
work page 1959
-
[34]
Kileel, J. and Trager, M. and Bruna, J. , title =. Advances in Neural Information Processing Systems , volume =. 2019 , url =
work page 2019
- [35]
-
[36]
Finkel, Bella and Rodriguez, Jose Israel and Wu, Chenxi and Yahl, Thomas , title =. Algebraic Statistics , year =
-
[37]
arXiv preprint arXiv:2511.19703 , year =
Massarenti, Alessio and Mella, Massimiliano , title=. arXiv preprint arXiv:2511.19703 , year =. doi:10.48550/arXiv.2511.19703 , url =
-
[38]
Proceedings of the 42nd International Conference on Machine Learning , year =
Marchetti, Giovanni Luca and Shahverdi, Vahid and Mereta, Stefano and Trager, Matthew and Kohn, Kathl\'. Proceedings of the 42nd International Conference on Machine Learning , year =
-
[39]
Zhou, Jun and Qian, Huimin and Lu, Xinbiao and Duan, Zhaoxia and Huang, Haoqian and Shao, Zhen , title =. Neurocomputing , volume =. 2019 , issn =
work page 2019
-
[40]
Goyal, Mohit and Goyal, Rajan and Lall, Brejesh , booktitle=. 2020 , volume=
work page 2020
-
[41]
Geometry and Optimization of Shallow Polynomial Networks , author=. 2025 , eprint=
work page 2025
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.