Recognition: unknown
GESR: A Genetic Programming-Based Symbolic Regression Method with Gene Editing
Pith reviewed 2026-05-14 21:15 UTC · model grok-4.3
The pith
Two BERT models replace random mutations and crossovers in genetic programming to raise efficiency on symbolic regression tasks.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
GESR trains two BERT models to serve as gene editors inside a genetic programming loop. The first model guides mutation of individual expression symbols through its masked language modeling capability. The second model guides crossover by predicting the most useful cut point between two parent individuals. By substituting these directed operations for the random ones used in conventional GP, the method produces fewer low-fitness offspring and reaches target expressions in fewer generations.
What carries the argument
Two BERT models that predict beneficial mutations of symbols and crossover points between expression trees.
If this is right
- Guided rather than random genetic operators can measurably lower the computational cost of evolving symbolic expressions.
- The two-BERT editing scheme delivers competitive accuracy while cutting the number of generations needed on standard regression benchmarks.
- Replacing stochastic variation steps with learned predictors inside evolutionary loops reduces the production of detrimental individuals.
- The overall workflow remains a population-based search but executes each generation more productively.
Where Pith is reading between the lines
- The same predictor-guided editing pattern could be tested on other tree-structured evolutionary problems such as program synthesis or circuit design.
- Larger or domain-specific pretraining corpora of mathematical expressions might further improve the accuracy of the mutation and crossover predictors.
- Hybrid systems that interleave deep-learning guidance with evolutionary search may scale to higher-dimensional or noisier data sets where pure random GP currently struggles.
Load-bearing premise
The BERT predictions actually steer the population toward higher-fitness expressions rather than merely reproducing patterns seen during their training.
What would settle it
A head-to-head run on the same symbolic regression benchmarks in which GESR requires at least as many fitness evaluations as ordinary GP to reach the same accuracy level.
read the original abstract
Mathematical formulas serve as a language through which humans communicate with nature. Discovering mathematical laws from scientific data to describe natural phenomena has been a long-standing pursuit of humanity for centuries. In the field of artificial intelligence, this challenge is known as the symbolic regression problem. Among existing symbolic regression approaches, Genetic Programming (GP) based on evolutionary algorithms remains one of the most classical and widely adopted methods. GP simulates the evolutionary process across generations through genetic mutation and crossover. However, mutations and crossovers in GP are entirely random. While this randomness effectively mimics natural evolution, it inevitably produces both beneficial and detrimental variations. If there existed a metaphorical `God` capable of foreseeing which genetic mutations or crossovers would yield superior outcomes and performing targeted gene editing accordingly, the efficiency of evolution could be substantially improved. Motivated by this idea, we propose in this paper a symbolic regression approach based on gene editing, termed GESR. In GESR, we trained two "hands of God" (two BERT models). Among them, the first leverages the BERT's masked language modeling capability to guide the mutation of genes (expression symbols). The other BERT model guides the crossover of individual genes by predicting the crossover point. Experimental results demonstrate that GESR significantly improves computational efficiency compared with traditional GP algorithms and achieves strong overall performance across multiple symbolic regression tasks.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces GESR, a genetic programming (GP) approach to symbolic regression in which two BERT models replace random mutation and crossover operators. One BERT uses masked language modeling to guide mutations of expression symbols; the second predicts crossover points. The central claim is that this learned 'gene editing' yields substantially higher computational efficiency than standard GP while maintaining strong performance across multiple symbolic regression benchmarks.
Significance. If the empirical claims are substantiated, the work would demonstrate a practical way to inject statistical priors from expression corpora into evolutionary search, addressing a long-standing inefficiency in GP-based symbolic regression. The hybrid use of language-model guidance is novel in this domain and, if shown to transfer beyond training distributions, could influence subsequent research on learned operators in evolutionary algorithms.
major comments (3)
- [Abstract] Abstract: The claims that GESR 'significantly improves computational efficiency' and 'achieves strong overall performance' are presented without any quantitative metrics (runtime, RMSE, R², success rate), baseline algorithms, number of benchmarks, or statistical tests. The experimental section must supply these details with tables or figures to allow assessment of the central claim.
- [Method] Method (BERT training): Both BERT models are trained exclusively with masked-language-modeling objectives on expression symbols and contain no regression loss or fitness signal from the target dataset. This design choice means any efficiency gain rests on an untested assumption that higher-likelihood edits also produce lower regression error; the manuscript should either add a fitness-aware fine-tuning stage or provide an ablation showing that MLM guidance outperforms random valid edits on held-out regression tasks.
- [Experiments] Experimental setup: No information is given on the corpus used to train the BERT models (source, size, diversity, overlap with evaluation benchmarks), GP hyperparameters, baseline implementations (standard GP, other SR methods), or evaluation protocol (number of runs, statistical significance). These omissions prevent verification that reported gains are not artifacts of distribution matching.
minor comments (2)
- [Abstract/Introduction] The informal metaphor of 'hands of God' and 'God' in the abstract and introduction should be replaced with precise technical language for a formal journal.
- [Method] Notation for the two BERT models and their inputs/outputs should be defined consistently in a single table or figure to improve readability.
Simulated Author's Rebuttal
We thank the referee for the constructive and detailed review. The comments highlight important areas for clarification and strengthening. We address each major comment below and indicate the revisions planned for the next version of the manuscript.
read point-by-point responses
-
Referee: [Abstract] The claims that GESR 'significantly improves computational efficiency' and 'achieves strong overall performance' are presented without any quantitative metrics (runtime, RMSE, R², success rate), baseline algorithms, number of benchmarks, or statistical tests. The experimental section must supply these details with tables or figures to allow assessment of the central claim.
Authors: We agree that the abstract would be strengthened by including concrete quantitative support. The experiments section already presents tables and figures with runtime comparisons, RMSE and R² values, success rates, baseline comparisons (including standard GP), and statistical tests across the benchmark suite. In the revised manuscript we will update the abstract to summarize key metrics, such as average runtime reduction and the fraction of benchmarks on which GESR matches or exceeds baseline performance. revision: yes
-
Referee: [Method] Both BERT models are trained exclusively with masked-language-modeling objectives on expression symbols and contain no regression loss or fitness signal from the target dataset. This design choice means any efficiency gain rests on an untested assumption that higher-likelihood edits also produce lower regression error; the manuscript should either add a fitness-aware fine-tuning stage or provide an ablation showing that MLM guidance outperforms random valid edits on held-out regression tasks.
Authors: The MLM-only training is intentional: it learns general syntactic priors over mathematical expressions from a broad corpus, supporting transfer across different regression problems. A full fitness-aware fine-tuning stage would risk overfitting to particular datasets and reduce generality. To directly test the value of the learned guidance, we will add an ablation study in the revised manuscript that compares MLM-guided edits against random but syntactically valid edits on held-out benchmarks, reporting the resulting differences in efficiency and regression accuracy. revision: partial
-
Referee: [Experiments] No information is given on the corpus used to train the BERT models (source, size, diversity, overlap with evaluation benchmarks), GP hyperparameters, baseline implementations (standard GP, other SR methods), or evaluation protocol (number of runs, statistical significance). These omissions prevent verification that reported gains are not artifacts of distribution matching.
Authors: We apologize for these descriptive omissions in the initial submission. The revised manuscript will contain a dedicated experimental-setup subsection that specifies the source, size, and diversity of the expression corpus used for BERT training, the steps taken to avoid overlap with evaluation benchmarks, all GP hyperparameters, the concrete implementations of the baseline algorithms, and the full evaluation protocol (number of independent runs per benchmark together with the statistical significance tests employed). revision: yes
Circularity Check
No significant circularity in derivation chain
full rationale
The paper introduces GESR as a GP variant that augments mutation and crossover with two BERT models trained via masked language modeling on expression symbols. The central claims rest on comparative experimental results across symbolic regression benchmarks rather than any closed-form derivation or self-referential definition. No equations are presented that reduce a prediction to a fitted input by construction, and no uniqueness theorem or ansatz is imported via self-citation to force the architecture. The method is therefore self-contained against external benchmarks.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption BERT masked-language-model pre-training transfers useful priors for predicting beneficial edits to symbolic expressions.
- domain assumption Predicting a single crossover point is sufficient to produce fitter offspring on average.
Reference graph
Works this paper leans on
-
[1]
arXiv preprint arXiv:2111.00053 (2021)
Mundhenk, T.N., Landajuela, M., Glatt, R., Santiago, C.P., Faissol, D.M., Petersen, B.K.: Symbolic regression via neural-guided genetic programming population seeding. arXiv preprint arXiv:2111.00053 (2021)
-
[2]
IEEE Computational Intelligence Magazine20(2), 58–79 (2025)
Han, X., Zhong, J., Ma, Z., Mu, X., Gligorovski, N.: Transformer-assisted genetic programming for symbolic regression [research frontier]. IEEE Computational Intelligence Magazine20(2), 58–79 (2025)
2025
-
[3]
Evolutionary Algorithms and Neural Networks: Theory and Applications, 43–55 (2019)
Mirjalili, S., Mirjalili, S.: Genetic algorithm. Evolutionary Algorithms and Neural Networks: Theory and Applications, 43–55 (2019)
2019
-
[4]
Multimedia tools and applications80, 8091–8126 (2021)
Katoch, S., Chauhan, S.S., Kumar, V.: A review on genetic algorithm: past, present, and future. Multimedia tools and applications80, 8091–8126 (2021)
2021
-
[5]
IEEE Transactions on Systems, Man, and Cyber- netics, Part C (Applications and Reviews)40(2), 121–144 (2009)
Espejo, P.G., Ventura, S., Herrera, F.: A survey on the application of genetic programming to classification. IEEE Transactions on Systems, Man, and Cyber- netics, Part C (Applications and Reviews)40(2), 121–144 (2009)
2009
-
[6]
The Journal of Machine Learning Research13(1), 2171–2175 (2012)
Fortin, F.-A., De Rainville, F.-M., Gardner, M.-A.G., Parizeau, M., Gagné, C.: Deap: Evolutionary algorithms made easy. The Journal of Machine Learning Research13(1), 2171–2175 (2012)
2012
-
[7]
In: Proceedings
Augusto, D.A., Barbosa, H.J.: Symbolic regression via genetic programming. In: Proceedings. Vol. 1. Sixth Brazilian Symposium on Neural Networks, pp. 173–178 (2000). IEEE
2000
-
[8]
Cranmer, M.: Interpretable machine learning for science with pysr and symboli- cregression. jl. arXiv preprint arXiv:2305.01582 (2023) 16
work page internal anchor Pith review Pith/arXiv arXiv 2023
-
[9]
arXiv preprint arXiv:1912.04871 (2019)
Petersen,B.K.,Landajuela,M.,Mundhenk,T.N.,Santiago,C.P.,Kim,S.K.,Kim, J.T.: Deep symbolic regression: Recovering mathematical expressions from data via risk-seeking policy gradients. arXiv preprint arXiv:1912.04871 (2019)
-
[10]
In: Advances in Neural Information Processing Systems, vol
Landajuela, M., Lee, C.S., Yang, J., Glatt, R., Santiago, C., Mundhenk, T.N., Aravena, I., Mulcahy, G., Petersen, B.: A unified framework for deep symbolic regression. In: Advances in Neural Information Processing Systems, vol. 35, pp. 33985–33998 (2022)
2022
-
[11]
arXiv preprint arXiv:2205.13134 (2022)
Sun, F., Liu, Y., Wang, J.-X., Sun, H.: Symbolic physics learner: Discovering governing equations via monte carlo tree search. arXiv preprint arXiv:2205.13134 (2022)
-
[12]
Expert Systems with Applications, 127591 (2025)
Li, Y., Li, W., Yu, L., Wu, M., Liu, J., Li, W., Hao, M.: Discovering mathematical formulas from data via gpt-guided monte carlo tree search. Expert Systems with Applications, 127591 (2025)
2025
-
[13]
arXiv preprint arXiv:2309.13705 (2023)
Li, W., Li, W., Yu, L., Wu, M., Liu, J., Li, Y.: A neural-guided dynamic sym- bolic network for exploring mathematical expressions from data. arXiv preprint arXiv:2309.13705 (2023)
-
[14]
arXiv preprint arXiv:2305.14656 (2023)
Xu, Y., Liu, Y., Sun, H.: Rsrm: Reinforcement symbolic regression machine. arXiv preprint arXiv:2305.14656 (2023)
-
[15]
Nature Computational Science, 1–14 (2025)
Ruan, K., Xu, Y., Gao, Z.-F., Liu, Y., Guo, Y., Wen, J.-R., Sun, H.: Discovering physical laws with parallel symbolic enumeration. Nature Computational Science, 1–14 (2025)
2025
-
[16]
Science Advances6(16), 2631 (2020)
Udrescu, S.-M., Tegmark, M.: Ai feynman: A physics-inspired method for symbolic regression. Science Advances6(16), 2631 (2020)
2020
-
[17]
Advances in Neural Information Processing Systems33, 4860–4871 (2020)
Udrescu, S.-M., Tan, A., Feng, J., Neto, O., Wu, T., Tegmark, M.: Ai feynman 2.0: Pareto-optimal symbolic regression exploiting graph modularity. Advances in Neural Information Processing Systems33, 4860–4871 (2020)
2020
-
[18]
Extrapolation and learning equations
Martius, G., Lampert, C.H.: Extrapolation and learning equations. arXiv preprint arXiv:1610.02995 (2016)
work page internal anchor Pith review Pith/arXiv arXiv 2016
-
[19]
IEEE transactions on neural networks and learning systems 32(9), 4166–4177 (2020)
Kim, S., Lu, P.Y., Mukherjee, S., Gilbert, M., Jing, L., Čeperić, V., Soljačić, M.: Integration of neural network-based symbolic regression in deep learning for scientific discovery. IEEE transactions on neural networks and learning systems 32(9), 4166–4177 (2020)
2020
-
[20]
arXiv preprint arXiv:2304.08915 (2023) 17
Zeng, P., Song, X., Lensen, A., Ou, Y., Sun, Y., Zhang, M., Lv, J.: Differentiable genetic programming for high-dimensional symbolic regression. arXiv preprint arXiv:2304.08915 (2023) 17
-
[21]
In: Proceedings of the AAAI Conference on Artificial Intelligence, vol
Li, Y., Li, W., Yu, L., Wu, M., Liu, J., Wei, S., Deng, Y., Hao, M.: Metasymnet: A tree-like symbol network with adaptive architecture and activation functions. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, pp. 27081–27089 (2025)
2025
-
[22]
In: International Conference on Machine Learning, pp
Biggio,L.,Bendinelli,T.,Neitz,A.,Lucchi,A.,Parascandolo,G.:Neuralsymbolic regression that scales. In: International Conference on Machine Learning, pp. 936–945 (2021). PMLR
2021
-
[23]
arXiv preprint arXiv:2303.06833 (2023)
Shojaee, P., Meidani, K., Farimani, A.B., Reddy, C.K.: Transformer-based planning for symbolic regression. arXiv preprint arXiv:2303.06833 (2023)
-
[24]
arXiv preprint arXiv:2302.11223 (2023)
Kamienny, P.-A., Lample, G., Lamprier, S., Virgolin, M.: Deep generative sym- bolic regression with monte-carlo-tree-search. arXiv preprint arXiv:2302.11223 (2023)
-
[25]
arXiv preprint arXiv:2404.06330 (2024)
Li, Y., Li, W., Yu, L., Wu, M., Liu, J., Li, W., Hao, M., Wei, S., Deng, Y.: Genera- tive pre-trained transformer for symbolic regression base in-context reinforcement learning. arXiv preprint arXiv:2404.06330 (2024)
-
[26]
arXiv preprint arXiv:2310.02227 (2023)
Meidani, K., Shojaee, P., Reddy, C.K., Farimani, A.B.: Snip: Bridging mathe- matical symbolic and numeric realms with unified pre-training. arXiv preprint arXiv:2310.02227 (2023)
-
[27]
Information Fusion114, 102681 (2025)
Li,Y.,Liu,J.,Wu,M.,Yu,L.,Li,W.,Ning,X.,Li,W.,Hao,M.,Deng,Y.,Wei,S.: Mmsr: symbolic regression is a multi-modal information fusion task. Information Fusion114, 102681 (2025)
2025
-
[28]
ChatSR: Multimodal Large Language Models for Scientific Formula Discovery
Li, Y., Li, W., Yu, L., Wu, M., Liu, J., Li, W., Wei, S., Deng, Y.: Mllm-sr: Con- versational symbolic regression base multi-modal large language models. arXiv preprint arXiv:2406.05410 (2024)
work page internal anchor Pith review Pith/arXiv arXiv 2024
-
[29]
In: Pro- ceedings of the 36th International Conference on Machine Learning
Lee, J., Lee, Y., Kim, J., Kosiorek, A., Choi, S., Teh, Y.W.: Set transformer: A framework for attention-based permutation-invariant neural networks. In: Pro- ceedings of the 36th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 97, pp. 3744–3753. PMLR, ??? (2019). https://proceedings.mlr.press/v97/lee19d.html
2019
-
[30]
biometrika78(3), 691–692 (1991)
Nagelkerke, N.J.,et al.: A note on a general definition of the coefficient of determination. biometrika78(3), 691–692 (1991)
1991
-
[31]
Ozer,D.J.:Correlationandthecoefficientofdetermination.Psychologicalbulletin 97(2), 307 (1985)
1985
-
[32]
MIT Press, Cambridge, MA, USA (1992) 18
Koza, J.R.: Genetic Programming: On the Programming of Computers by Means of Natural Selection. MIT Press, Cambridge, MA, USA (1992) 18
1992
-
[33]
Evolutionary Computation6(3), 231–252 (1998) https://doi.org/10.1162/evco.1998.6.3.231
Poli, R., Langdon, W.B.: Schema theory for genetic programming with one-point crossover and point mutation. Evolutionary Computation6(3), 231–252 (1998) https://doi.org/10.1162/evco.1998.6.3.231
-
[34]
BERT: Pre-training of deep bidi- rectional transformers for language understanding
Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Com- putational Linguistics: Human Language Technologies (NAACL-HLT 2019), pp. 4171–4186. Association for Computational Linguistics, Min...
-
[35]
Advances in Neural Information Processing Systems 35, 10269–10281 (2022)
Kamienny, P.-A., d’Ascoli, S., Lample, G., Charton, F.: End-to-end symbolic regression with transformers. Advances in Neural Information Processing Systems 35, 10269–10281 (2022)
2022
-
[36]
Advances in Neural Information Processing Systems 36(2024)
Shojaee,P.,Meidani,K.,BaratiFarimani,A.,Reddy,C.:Transformer-basedplan- ning for symbolic regression. Advances in Neural Information Processing Systems 36(2024)
2024
-
[37]
arXiv preprint arXiv:2206.10540 (2022)
Matsubara, Y., Chiba, N., Igarashi, R., Ushiku, Y.: Rethinking symbolic regression datasets and benchmarks for scientific discovery. arXiv preprint arXiv:2206.10540 (2022)
-
[38]
The Journal of the Acoustical Society of America99(5), 2985–3002 (1996)
Berglund, B., Hassmen, P., Job, R.S.: Sources and effects of low-frequency noise. The Journal of the Acoustical Society of America99(5), 2985–3002 (1996)
1996
-
[39]
Journal of Fluid Mechanics615, 253–292 (2008)
Tam, C.K., Viswanathan, K., Ahuja, K., Panda, J.: The sources of jet noise: experimental evidence. Journal of Fluid Mechanics615, 253–292 (2008)
2008
-
[40]
Neuroimage37(4), 1286–1300 (2007)
Beall, E.B., Lowe, M.J.: Isolating physiologic noise sources with independently determined spatial measures. Neuroimage37(4), 1286–1300 (2007)
2007
-
[41]
Sensors22(3), 1241 (2022)
Ziyadinov, V., Tereshonok, M.: Noise immunity and robustness study of image recognition using a convolutional neural network. Sensors22(3), 1241 (2022)
2022
-
[42]
In: Proceedings of the Acm/ieee 42nd International Conference on Software Engineering, pp
Gao, X., Saha, R.K., Prasad, M.R., Roychoudhury, A.: Fuzz testing based data augmentation to improve robustness of deep neural networks. In: Proceedings of the Acm/ieee 42nd International Conference on Software Engineering, pp. 1147– 1158 (2020)
2020
-
[43]
SIAM Journal on Applied Mathematics28(4), 831–838 (1975)
Newton, T.A., Liepnik, J.S.: Chaos in nonlinear oscillators. SIAM Journal on Applied Mathematics28(4), 831–838 (1975)
1975
-
[44]
Lorenz,E.N.:Deterministicnonperiodicflow.JournaloftheAtmosphericSciences 20(2), 130–141 (1963)
1963
-
[45]
Physics Letters A71(2–3), 155–157 19 (1979)
Rössler, O.E.: An equation for hyperchaos. Physics Letters A71(2–3), 155–157 19 (1979)
1979
-
[46]
Chaos, Solitons & Fractals13(9), 1965–1971 (2002)
Jha, R.: Hyperchaos generated from a modified lorenz system. Chaos, Solitons & Fractals13(9), 1965–1971 (2002)
1965
-
[47]
Chaos, Solitons & Fractals25(3), 579–586 (2005)
Pang, S., Liu, Y.: A new hyperchaotic system and its control. Chaos, Solitons & Fractals25(3), 579–586 (2005)
2005
-
[48]
Physics Letters A76(3), 201–204 (1980)
Shimizu, T., Morioka, N.: On the bifurcation of a symmetric limit cycle to chaos in a simple model. Physics Letters A76(3), 201–204 (1980)
1980
-
[49]
Automatica28(2), 347–354 (1992)
Genesio, R., Tesi, A.: A harmonic balance approach to chaos in nonlinear systems. Automatica28(2), 347–354 (1992)
1992
-
[50]
Physics Letters A53(1), 77–78 (1975)
Haken, H.: Analogy between higher instabilities in fluids and lasers. Physics Letters A53(1), 77–78 (1975)
1975
-
[51]
Journal of Chemical Physics48(4), 1695–1700 (1968)
Prigogine, I., Lefever, R.: Symmetry breaking instabilities in dissipative systems. Journal of Chemical Physics48(4), 1695–1700 (1968)
1968
-
[52]
Journal of Fluid Mechanics237, 209–229 (1992)
Rucklidge, A.M.: Chaos in a low-order model of magnetoconvection. Journal of Fluid Mechanics237, 209–229 (1992)
1992
-
[53]
Biophysical Journal1(6), 445–466 (1961)
FitzHugh, R.: Impulses and physiological states in theoretical models of nerve membrane. Biophysical Journal1(6), 445–466 (1961)
1961
-
[54]
Proceedings of the IRE50(10), 2061–2070 (1962)
Nagumo, J.-I., Arimoto, S., Yoshizawa, S.: An active pulse transmission line simulating nerve axon. Proceedings of the IRE50(10), 2061–2070 (1962)
2061
-
[55]
Chaos, Solitons & Fractals36(5), 1305–1314 (2008)
Chen, W.-C.: Chaotic dynamics in nonlinear finance systems. Chaos, Solitons & Fractals36(5), 1305–1314 (2008)
2008
-
[56]
Chaos, Solitons & Fractals23(4), 1307–1317 (2005)
Li, D., Chen, G.: A new chaotic system and its synchronization. Chaos, Solitons & Fractals23(4), 1307–1317 (2005)
2005
-
[57]
Tellus A 36(2), 98–110 (1984)
Lorenz, E.N.: Irregularity: A fundamental property of the atmosphere. Tellus A 36(2), 98–110 (1984)
1984
-
[58]
Physical Review E50(2), 647–650 (1994)
Sprott, J.C.: Some simple chaotic flows. Physical Review E50(2), 647–650 (1994)
1994
-
[59]
Springer, ??? (1983)
Guckenheimer, J., Holmes, P.: Nonlinear Oscillations, Dynamical Systems, and Bifurcations of Vector Fields. Springer, ??? (1983)
1983
-
[60]
Strizhak, P.E., Kawczynski, A.L.: Chaos in catalytic reactions. Chemical Engi- neering Science55(24), 5837–5848 (2000) 20 Appendix For GESR: A Genetic Programming–Based Symbolic Regression Method with Gene Editing A pseudocode for GESR 23 A.1 GESR: Learning-guided Genetic Programming Framework. . . . . . . 23 A.2 Training pair construction pipeline for mu...
-
[61]
Newton–Liepnik System [43] The Newton–Liepnik system is defined by the following three-dimensional autonomous ordinary differential equations: ˙x=−ax+y+ 10yz, ˙y=−x−ay+ 5xz, ˙z=bz−5xy, a= 0.4, b= 0.175.(7) 49 Table 13: Chaotic dynamics 13 to 16. Chaotic Dynamic Governing Equations Parameters Finance ˙x= (1/b−a)x+z+xy ˙y=−by−x 2 ˙z=−x−cz a: 0.001, ...
-
[62]
Under suitable parameters, it typically exhibits two or more positive Lyapunov exponents, manifesting hyperchaotic behavior
Hyper-Lorenz System (Four-Dimensional Hyperchaos) [44, 45] The Hyper-Lorenz system is a four-dimensional extension of the classical Lorenz system, governed by: ˙x=a(y−x) +w, ˙y=cx−y−xz, ˙z=xy−bz, ˙w=dw−xz, a= 10, b= 2.667, c= 28, d= 1.1.(8) By retaining the core nonlinear convection termsxzandxyfrom the Lorenz system while introducing an addit...
-
[63]
Hyper-Jha System (Four-Dimensional Hyperchaos) [46] The Hyper-Jha system is another Lorenz-type four-dimensional hyperchaotic system described by: ˙x=a(y−x) +w, ˙y=cx−y−xz, ˙z=xy−bz, ˙w=−dy, (9) Unlike the Hyper-Lorenz system, Hyper-Jha introduces a linear decay term−dyin the auxiliary state equation, altering energy feedback pathways among st...
-
[64]
This results in phase- space projections that differ markedly from those of Hyper-Lorenz and Hyper-Jha, while preserving the Lorenz-type nonlinear backbone
Hyper-Pang System (Four-Dimensional Hyperchaos) [47] The Hyper-Pang system is another Lorenz-type hyperchaotic model governed by: ˙x=a(y−x), ˙y=cx−y−xz+w, ˙z=xy−bz, ˙w=−d(x+y), (10) A distinguishing feature of this system is that the auxiliary variablewcou- ples simultaneously to multiple original state variables through linear combinations, r...
-
[65]
The nonlinear termsxzandx 2 introduce modulation effects in fast and slow chan- nels, enabling complex chaotic behavior even in low dimensions
Shimizu–Morioka System [48] The Shimizu–Morioka system is described by the following three-dimensional autonomous ODE: ˙x=y, ˙y=x−ay−xz, ˙z=−bz+x 2, a= 0.85, b= 0.5.(11) Originating from low-dimensional laser dynamics modeling, this system can be interpreted as a state-dependent feedback oscillator coupled with a slow variable. The nonlinear terms...
-
[66]
Its simple yet explicit nonlinear source makes it an important benchmark for evaluating the trade-off between nonlinear identification accuracy and model compactness
Genesio–Tesi System [49] The Genesio–Tesi system is a polynomial chaotic model governed by: ˙x=y, ˙y=z, ˙z=−cx−by−az+x 2, a= 1.2, b= 2.92, c= 6.(12) This system exhibits an integration-chain structure, with chaos driven solely by the quadratic termx2. Its simple yet explicit nonlinear source makes it an important benchmark for evaluating the trade...
-
[67]
The bilinear termsxzandxydescribe nonlinear interactions between light and matter
Laser System [50] The Laser system originates from single-mode laser rate equations and is described by: ˙x=a(y−x), ˙y=bx−y−xz, ˙z=−cz+xy, (13) Here, state variables typically correspond to electric field intensity, polarization, and population inversion. The bilinear termsxzandxydescribe nonlinear interactions between light and matter. Under suit...
-
[68]
Although structurally simple, the system can exhibit complex phase-space behavior under appropriate damping parameters, making it a foundational prototype in nonlinear dynamics
Duffing Autonomous System [59] The Duffing system is a classical model of nonlinear vibration, expressed in autonomous form as: ˙x=y, ˙y=x−x 3 −ay, ˙z=−bz, (14) Here,xdenotes displacement andyvelocity, whilex 3 captures nonlinear stiffness effects. Although structurally simple, the system can exhibit complex phase-space behavior under appropriate ...
-
[69]
This system is a canonical benchmark in nonequilibrium thermodynamics and chemical kinetics
Brusselator System [51] The Brusselator system models autocatalytic chemical reactions: ˙x=A+x 2y−(B+ 1)x, ˙y=Bx−x 2y, ˙z=−z+x, (15) The nonlinear termx 2yintroduces positive feedback, enabling oscillatory and chaotic behavior far from equilibrium. This system is a canonical benchmark in nonequilibrium thermodynamics and chemical kinetics
-
[70]
Kawczynski–Strizhak System [60] The Kawczynski–Strizhak system is a third-order polynomial chaotic model: ˙x=y, ˙y=z, ˙z=−az−by−x+x 2, (16) 53 Its chaos is driven entirely by the quadratic termx 2, making it suitable for analyzing how nonlinear forcing influences stability and bifurcation behavior
-
[71]
Rucklidge System [52] The Rucklidge system, originally developed for thermal convection instability, is defined as: ˙x=−kx+ay−yz, ˙y=x, ˙z=y 2 −z, (17) It can be viewed as a Lorenz-type variant, with nonlinear coupling termsyzand y2 shaping its stretching-and-folding phase-space geometry
-
[72]
FitzHugh–Nagumo System [53, 54] The FitzHugh–Nagumo system is a reduced model of Hodgkin–Huxley neuron dynamics: ˙x=c x− x3 3 +y , ˙y=− 1 c (x−a+by), ˙z=−z+x, (18) The cubic nonlinearity governs neuronal excitation and recovery dynamics and can produce oscillatory and chaotic firing patterns
-
[73]
Finance System [55] The Finance system models nonlinear macroeconomic interactions: ˙x=z+ (y−a)x, ˙y= 1−by−x 2, ˙z=−x−cz, (19) The quadratic termx 2 captures saturation effects and economic feedback loops, producing chaotic market-like fluctuations
-
[74]
Dequan–Li System [56] The Dequan–Li system is a Lorenz-type chaotic system: ˙x=a(y−x) +yz, ˙y=cx−xz+dy, ˙z=xy−bz, (20) Additional bilinear couplings reshape attractor geometry, making it suitable for testing modeling under mixed linear–nonlinear structures. 54
-
[75]
Hadley Circulation Model [57] The Hadley circulation model describes large-scale atmospheric circulation: ˙x=−y 2 −z 2 −ax+ac, ˙y=xy−bxz−y+d, ˙z=bxy+xz−z, (21) Its nonlinear energy-exchange terms allow complex low-dimensional climate dynamics
-
[76]
(22) Chaos arises from the nonsmooth nonlinearitysign(x), making this system a challenging benchmark for symbolic regression under nonsmooth dynamics
Sprott–Jerk System [58] The Sprott–Jerk system is a piecewise nonsmooth chaotic model: ˙x=y, ˙y=z, ˙z=−az−y+ sign(x). (22) Chaos arises from the nonsmooth nonlinearitysign(x), making this system a challenging benchmark for symbolic regression under nonsmooth dynamics. J.2 Comprehensive Evaluation Protocol Based on Per-DimensionR 2 and Short-Horizo...
-
[77]
cos(x)−1U(−1,1,20) Nguyen-6sin(x 1) + sin(x1 +x 2 1)U(−1,1,20) Nguyen-7log(x 1 + 1) + log(x2 1 + 1)U(0,2,20) Nguyen-8 √xU(0,4,20) Nguyen-9sin(x) + sin(x 2 2)U(0,1,20) Nguyen-102 sin(x) cos(x 2)U(0,1,20) Nguyen-11x x2 1 U(0,1,20) Nguyen-12x 4 1 −x 3 1 + 1 2 x2 2 −x 2 U(0,1,20) Nguyen-2′ 4x4 1 + 3x3 1 + 2x2 1 +xU(−1,1,20) Nguyen-5′ sin(x2
-
[78]
cos(x)−2U(−1,1,20) Nguyen-8′ 3√xU(0,4,20) Nguyen-8′′ 3 q x2 1 U(0,4,20) Nguyen-1c 3.39x3 1 + 2.12x2 1 + 1.78xU(−1,1,20) Nguyen-5c sin(x2
-
[79]
65 Name Expression Dataset Neat-1x 4 1 +x 3 1 +x 2 1 +xU(−1,1,20) Neat-2x 5 1 +x 4 1 +x 3 1 +x 2 1 +xU(−1,1,20) Neat-3sin(x 2
cos(x)−0.75U(−1,1,20) Nguyen-7c log(x+ 1.4) + log(x 2 1 + 1.3)U(0,2,20) Nguyen-8c √ 1.23xU(0,4,20) Nguyen-10c sin(1.5x) cos(0.5x2)U(0,1,20) Korns-11.57 + 24.3∗x 4 1 U(−1,1,20) Korns-20.23 + 14.2 (x4+x1) (3x2) U(−1,1,20) Korns-34.9 (x2−x1+ x1 x3 (3x3)) −5.41U(−1,1,20) Korns-40.13sin(x 1)−2.3U(−1,1,20) Korns-53 + 2.13log(|x 5|)U(−1,1,20) Korns-61.3 + 0.13 p...
-
[80]
66 Name Expression Dataset Livermore-174sin(x 1)cos(x2)U(−3,3,100) Livermore-18sin(x 2 1)∗cos(x 1)−5U(−3,3,100) Livermore-19x 5 1 +x 4 1 +x 2 1 +x 1 U(−3,3,100) Livermore-20e (−x2
cos(x)−1U(−1,1,20) Neat-4log(x+ 1) + log(x 2 1 + 1)U(0,2,20) Neat-52 sin(x) cos(x 2)U(−1,1,100) Neat-6 Px k=1 1 k E(1,50,50) Neat-72−2.1 cos(9.8x 1) sin(1.3x2)E(−50,50,10 5) Neat-8 e−(x1 )2 1.2+(x2−2.5)2 U(0.3,4,100) Neat-9 1 1+x−4 1 + 1 1+x−4 2 E(−5,5,21) Keijzer-10.3x 1sin(2πx1)U(−1,1,20) Keijzer-22.0x 1sin(0.5πx1)U(−1,1,20) Keijzer-30.92x 1sin(2.41πx1)...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.