pith. machine review for the scientific record. sign in

arxiv: 2605.02933 · v1 · submitted 2026-04-30 · 💻 cs.NE

Recognition: unknown

Relation Reasoning with LLMs in Expensive Optimization

Authors on Pith no claims yet

Pith reviewed 2026-05-09 20:36 UTC · model grok-4.3

classification 💻 cs.NE
keywords expensive optimizationsurrogate-assisted evolutionary algorithmslarge language modelspairwise relation reasoningreinforcement learningblack-box optimizationevolutionary algorithmszero-shot surrogate
0
0 comments X

The pith

A large language model fine-tuned to reason about solution relations can guide evolutionary optimization of expensive problems without retraining.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Expensive optimization problems restrict how many times the true objective function can be evaluated because each evaluation costs a great deal. Surrogate-assisted evolutionary algorithms therefore use predictive models to choose promising candidates, yet conventional surrogates typically require retraining at every generation as the population shifts. This paper shows that a large language model, trained via reinforcement learning on evolutionary trajectories to judge pairwise relations between solutions, can serve as a stable surrogate. An anchor-based prompting method keeps the input size linear rather than quadratic in population size, while a voting step converts the model's relation outputs into selection scores. Experiments on standard single- and multi-objective benchmarks report higher optimization performance than prior surrogate methods and general large language models, together with support for quantization and zero-shot use.

Core claim

The paper establishes that relation-based surrogate modeling, cast as an in-context pairwise reasoning task inside a large language model, enables effective guidance within evolutionary algorithms for expensive black-box problems. Training the model on trajectories with GRPO, combined with anchor-based iterative context construction and voting-based aggregation of predicted relations into scores, produces improved relation prediction and state-of-the-art optimization results on single- and multi-objective benchmarks while allowing a zero-shot surrogate paradigm without per-generation retraining.

What carries the argument

The anchor-based iterative context construction strategy that reduces prompt complexity from quadratic to linear in population size, together with a voting-based aggregation scheme that converts predicted relations into scores for offspring selection.

Load-bearing premise

That the LLM's predictions of relations between candidate solutions, learned from evolutionary trajectories, will generalize reliably to new problems and supply accurate guidance for selection without any retraining during the search.

What would settle it

A test on held-out optimization benchmarks in which the R2SAEA algorithm produces final solution quality no better than strong SAEA baselines after the same evaluation budget, or in which the model's pairwise relation accuracy falls sharply on populations not seen during training.

Figures

Figures reproduced from arXiv: 2605.02933 by Aimin Zhou, Bingdong Li, Hao Hao, Ye Lu.

Figure 1
Figure 1. Figure 1: Overview of the R2SAEA framework. It comprises three components: an SAEA module, an LLM-driven relation model, [PITH_FULL_IMAGE:figures/full_fig_p004_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Prompt template for relation reasoning. A naive construction would enumerate all pairwise relations within Dctx ∪ Dquery. Let |Dctx| = n and |Dquery| = q. This yields O [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Median runtime performance of LSEA and R2SAEA [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Offline evaluation of ReLLM-C1 and general LLMs [PITH_FULL_IMAGE:figures/full_fig_p009_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Offline evaluation of ReLLM-C2 and general LLMs [PITH_FULL_IMAGE:figures/full_fig_p009_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Element Acc (accele) on the offline test set under different decision dimensions (D) and subsample sizes (N) [PITH_FULL_IMAGE:figures/full_fig_p011_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Binary Acc (accbin) on the offline test set under different decision dimensions (D) and subsample sizes (n). E. The Impact of Normalization Strategies Normalization is required in our RL fine-tuning and infer￾ence pipeline, since decision variables originate from different benchmark functions and different evolutionary stages, and therefore may have heterogeneous value ranges. We denote the context set as … view at source ↗
Figure 9
Figure 9. Figure 9: Completion tokens per second (CTs) by Model Size [PITH_FULL_IMAGE:figures/full_fig_p012_9.png] view at source ↗
Figure 8
Figure 8. Figure 8: Total tokens per second (TTs) by Model Size and [PITH_FULL_IMAGE:figures/full_fig_p012_8.png] view at source ↗
read the original abstract

Expensive optimization problems (EOPs) are black-box tasks with costly objective evaluations and no gradient access, making the evaluation budget the key bottleneck. Surrogate-assisted evolutionary algorithms (SAEAs) reduce evaluations via surrogate predictions, but conventional surrogates often require frequent retraining as populations evolve, incurring overhead. This paper proposes R2SAEA, a reinforcement-trained relation-based large language model (LLM) surrogate assisted evolutionary algorithm. We cast relation-based surrogate modeling as an in-context pairwise reasoning task. To enable efficient inference in evolutionary loops, we develop an anchor-based iterative context construction strategy that reduces prompt complexity from quadratic to linear in population size, and a voting-based aggregation scheme that converts predicted relations into scores for offspring selection. We further build an RL pipeline from evolutionary trajectories and fine-tune Qwen2.5 with GRPO. Experiments on single- and multi-objective benchmarks show improved relation prediction and state-of-the-art optimization performance over strong SAEA baselines and general LLMs. Quantization also enables efficient edge deployment, supporting a zero-shot surrogate paradigm without per-generation retraining. Code and models are available at https://github.com/Septend9/R2SAEA.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes R2SAEA, a surrogate-assisted evolutionary algorithm (SAEA) for expensive optimization problems (EOPs) that replaces conventional surrogates with a relation-reasoning LLM. The LLM (Qwen2.5) is fine-tuned via GRPO on evolutionary trajectories to predict pairwise relations between candidate solutions. An anchor-based iterative context construction reduces prompt length from quadratic to linear in population size, and a voting aggregation converts relation predictions into scalar scores for offspring selection. The method is positioned as enabling a zero-shot surrogate paradigm without per-generation retraining. Experiments on single- and multi-objective benchmarks are reported to show improved relation prediction and state-of-the-art optimization performance relative to strong SAEA baselines and general LLMs; quantization for edge deployment is also demonstrated.

Significance. If the generalization claims hold, the work would be significant for the SAEA community by showing that relation-based LLM reasoning can serve as a reusable, low-overhead surrogate that avoids repeated retraining. The open release of code and models is a clear strength that supports reproducibility and follow-on research. The approach also opens a new direction for leveraging LLM reasoning capabilities in black-box optimization rather than direct value prediction.

major comments (2)
  1. [§4 (Training Pipeline)] §4 (Training Pipeline): The paper does not characterize the distribution of evolutionary trajectories used for GRPO fine-tuning (problem classes, dimensions, objective landscape features, or diversity metrics). This is load-bearing for the zero-shot generalization claim, because without such characterization it is impossible to determine whether reported gains on the test benchmarks reflect true out-of-distribution robustness or partial overlap with the training distribution.
  2. [§5 (Experimental Results)] §5 (Experimental Results): The optimization performance claims (SOTA over SAEA baselines) are presented without reported statistical tests, number of independent runs, or variance measures. In addition, the precise configuration of the “strong SAEA baselines” is not detailed enough to allow direct replication or assessment of whether the comparison is fair on evaluation budget and hyperparameter tuning.
minor comments (2)
  1. [Abstract] The abstract states “improved relation prediction” without naming the metric or reporting numerical gains; adding a brief quantitative statement would strengthen the summary.
  2. [§5] Figure captions and axis labels in the experimental plots should explicitly state the number of function evaluations used and whether shaded regions represent standard deviation or standard error.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments, which help strengthen the presentation of our work on R2SAEA. We address each major comment below and will incorporate revisions to improve clarity and rigor.

read point-by-point responses
  1. Referee: [§4 (Training Pipeline)] The paper does not characterize the distribution of evolutionary trajectories used for GRPO fine-tuning (problem classes, dimensions, objective landscape features, or diversity metrics). This is load-bearing for the zero-shot generalization claim, because without such characterization it is impossible to determine whether reported gains on the test benchmarks reflect true out-of-distribution robustness or partial overlap with the training distribution.

    Authors: We agree that a detailed characterization of the training trajectories is important for supporting the zero-shot generalization claims. In the revised manuscript, we will add a new subsection in §4 describing the distribution of evolutionary trajectories collected for GRPO fine-tuning. This will include the specific problem classes (e.g., CEC 2017/2020 single- and multi-objective benchmarks), dimension ranges (10D to 50D), objective landscape features (unimodal/multimodal, separable/non-separable), and diversity metrics such as average population variance and convergence statistics across the collected trajectories. revision: yes

  2. Referee: [§5 (Experimental Results)] The optimization performance claims (SOTA over SAEA baselines) are presented without reported statistical tests, number of independent runs, or variance measures. In addition, the precise configuration of the “strong SAEA baselines” is not detailed enough to allow direct replication or assessment of whether the comparison is fair on evaluation budget and hyperparameter tuning.

    Authors: We acknowledge that the current experimental reporting lacks sufficient statistical detail and baseline transparency. In the revised §5, we will explicitly state the number of independent runs (20 per benchmark instance), report mean performance with standard deviations, and include statistical significance tests (Wilcoxon rank-sum test with p-values) comparing R2SAEA against each baseline. We will also expand the baseline descriptions with a dedicated table listing exact configurations, including population sizes, surrogate hyperparameters, evaluation budgets per generation, and any tuning procedures used, to enable direct replication and confirm fairness of the comparisons. revision: yes

Circularity Check

0 steps flagged

No circularity; claims rest on independent benchmark evaluation after training on trajectories

full rationale

The paper trains the LLM surrogate via GRPO on evolutionary trajectories, then deploys the resulting model for relation prediction and offspring selection on separate single- and multi-objective benchmarks. The anchor-based context construction and voting aggregation are procedural engineering steps that convert LLM outputs into selection scores without defining the performance metric in terms of the training data itself. No equations, self-citations, or uniqueness theorems are invoked that reduce the reported SOTA gains to a fitted quantity or prior result by construction. Generalization to new EOPs is presented as an empirical outcome of the experiments rather than a definitional necessity.

Axiom & Free-Parameter Ledger

1 free parameters · 2 axioms · 0 invented entities

Since only the abstract is available, the full set of assumptions cannot be audited. The central claim rests on the LLM learning generalizable pairwise relations from optimization trajectories and on the anchor-based and voting mechanisms preserving useful information.

free parameters (1)
  • GRPO and fine-tuning hyperparameters
    The reinforcement learning pipeline and model fine-tuning necessarily involve multiple hyperparameters whose values are chosen or fitted during training.
axioms (2)
  • domain assumption LLM pairwise relation predictions can be aggregated via voting into reliable scalar scores that improve evolutionary selection
    This conversion step is required for the surrogate to guide offspring selection.
  • domain assumption Anchor-based iterative context construction retains sufficient relational information while reducing prompt complexity
    The linear-complexity strategy is presented as preserving accuracy.

pith-pipeline@v0.9.0 · 5505 in / 1566 out tokens · 48214 ms · 2026-05-09T20:36:35.303222+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

70 extracted references · 18 canonical work pages · 6 internal anchors

  1. [1]

    Evolutionary computation for ex- pensive optimization: A survey,

    J.-Y . Li, Z.-H. Zhan, and J. Zhang, “Evolutionary computation for ex- pensive optimization: A survey,”Machine Intelligence Research, vol. 19, no. 1, pp. 3–23, 2022

  2. [2]

    A survey on evolutionary neural architecture search,

    Y . Liu, Y . Sun, B. Xue, M. Zhang, G. G. Yen, and K. C. Tan, “A survey on evolutionary neural architecture search,”IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 2, pp. 550–570, 2023

  3. [3]

    Derivative-free optimization via classifi- cation,

    Y . Yu, H. Qian, and Y .-Q. Hu, “Derivative-free optimization via classifi- cation,” inProceedings of the AAAI Conference on Artificial Intelligence, vol. 30, no. 1, 2016

  4. [4]

    Surrogate-assisted evolutionary computation: Recent advances and future challenges,

    Y . Jin, “Surrogate-assisted evolutionary computation: Recent advances and future challenges,”Swarm and Evolutionary Computation, vol. 1, no. 2, pp. 61–70, 2011

  5. [5]

    Dual relational surrogate-assisted evolution for expensive constrained multiobjective optimization,

    S. Liu, P. Chen, and Q. Lin, “Dual relational surrogate-assisted evolution for expensive constrained multiobjective optimization,”IEEE Transac- tions on Evolutionary Computation, 2025. 13 TABLE IV: Performance of C1-Trained Models across Varying Parameter Sizes and Quantization Precisions on the Offline Test Set. 3B 7B Metric Qwen2.5-instruct ReLLM-C1 ReL...

  6. [6]

    Comparison-based opti- mizers need comparison-based surrogates,

    I. Loshchilov, M. Schoenauer, and M. Sebag, “Comparison-based opti- mizers need comparison-based surrogates,” inInternational conference on parallel problem solving from nature. Springer, 2010, pp. 364–373

  7. [7]

    Large language models as surrogate models in evolutionary algorithms: A preliminary study,

    H. Hao, X. Zhang, and A. Zhou, “Large language models as surrogate models in evolutionary algorithms: A preliminary study,”

  8. [8]

    Available: https://arxiv.org/abs/2406.10675

    [Online]. Available: https://arxiv.org/abs/2406.10675

  9. [9]

    Expensive optimization via relation,

    ——, “Expensive optimization via relation,”IEEE Transactions on Evolutionary Computation, 2025

  10. [10]

    DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models

    Z. Shao, P. Wang, Q. Zhu, R. Xu, J. Song, X. Bi, H. Zhang, M. Zhang, Y . Li, Y . Wuet al., “Deepseekmath: Pushing the limits of mathematical reasoning in open language models,”arXiv preprint arXiv:2402.03300, 2024

  11. [11]

    A review of surrogate-assisted evolutionary multi-objective optimization,

    Q. Wang, Y . Zhang, D. Gong, X. Song, C. He, X. Ji, and F. Chu, “A review of surrogate-assisted evolutionary multi-objective optimization,” IEEE Transactions on Emerging Topics in Computational Intelligence, 2026

  12. [12]

    Multiobjective evolutionary algorithms: A survey of the state of the art,

    A. Zhou, B.-Y . Qu, H. Li, S.-Z. Zhao, P. N. Suganthan, and Q. Zhang, “Multiobjective evolutionary algorithms: A survey of the state of the art,”Swarm and evolutionary computation, vol. 1, no. 1, pp. 32–49, 2011

  13. [13]

    Multiobjective optimization using coupled response surface model and evolutionary algorithm

    Y . Lian and M.-S. Liou, “Multiobjective optimization using coupled response surface model and evolutionary algorithm.”AIAA journal, vol. 43, no. 6, pp. 1316–1325, 2005

  14. [14]

    Surrogate-assisted co- operative swarm optimization of high-dimensional expensive problems,

    C. Sun, Y . Jin, R. Cheng, J. Ding, and J. Zeng, “Surrogate-assisted co- operative swarm optimization of high-dimensional expensive problems,” IEEE Transactions on Evolutionary Computation, vol. 21, no. 4, pp. 644–660, 2017

  15. [15]

    A gaussian process surrogate model assisted evolutionary algorithm for medium scale expensive optimization problems,

    B. Liu, Q. Zhang, and G. G. Gielen, “A gaussian process surrogate model assisted evolutionary algorithm for medium scale expensive optimization problems,”IEEE Transactions on Evolutionary Computation, vol. 18, no. 2, pp. 180–192, 2013

  16. [16]

    An approximated domination re- lationship based on binary classifiers for evolutionary multiobjective optimization,

    H. Hao, A. Zhou, and H. Zhang, “An approximated domination re- lationship based on binary classifiers for evolutionary multiobjective optimization,” in2021 IEEE Congress on Evolutionary Computation (CEC). IEEE, 2021, pp. 2427–2434

  17. [17]

    A classification- based surrogate-assisted evolutionary algorithm for expensive many- objective optimization,

    L. Pan, C. He, Y . Tian, H. Wang, X. Zhang, and Y . Jin, “A classification- based surrogate-assisted evolutionary algorithm for expensive many- objective optimization,”IEEE Transactions on Evolutionary Computa- tion, vol. 23, no. 1, pp. 74–88, 2018

  18. [18]

    Fuzzy-classification assisted solution preselection in evolutionary optimization,

    A. Zhou, J. Zhang, J. Sun, and G. Zhang, “Fuzzy-classification assisted solution preselection in evolutionary optimization,” inProceedings of the AAAI conference on artificial intelligence, vol. 33, no. 01, 2019, pp. 2403–2410

  19. [19]

    Binary relation learning and classifying for preselection in evolutionary algorithms,

    H. Hao, J. Zhang, X. Lu, and A. Zhou, “Binary relation learning and classifying for preselection in evolutionary algorithms,”IEEE Transac- tions on Evolutionary Computation, vol. 24, no. 6, pp. 1125–1139, 2020

  20. [20]

    Enhancing saeas with unevaluated solutions: a case study of relation model for expensive optimization,

    H. Hao, X. Zhang, and A. Zhou, “Enhancing saeas with unevaluated solutions: a case study of relation model for expensive optimization,” Science China Information Sciences, vol. 67, no. 2, p. 120103, 2024

  21. [21]

    Expensive multiobjective evolutionary op- timization assisted by dominance prediction,

    Y . Yuan and W. Banzhaf, “Expensive multiobjective evolutionary op- timization assisted by dominance prediction,”IEEE Transactions on Evolutionary Computation, vol. 26, no. 1, pp. 159–173, 2021

  22. [22]

    Expensive multiobjective optimization by relation learning and prediction,

    H. Hao, A. Zhou, H. Qian, and H. Zhang, “Expensive multiobjective optimization by relation learning and prediction,”IEEE Transactions on Evolutionary Computation, vol. 26, no. 5, pp. 1157–1170, 2022

  23. [23]

    Platemo: A matlab platform for evolutionary multi-objective optimization [educational forum],

    Y . Tian, R. Cheng, X. Zhang, and Y . Jin, “Platemo: A matlab platform for evolutionary multi-objective optimization [educational forum],”IEEE Computational Intelligence Magazine, vol. 12, no. 4, pp. 73–87, 2017

  24. [24]

    Evolutionary computation in the era of large language model: Survey and roadmap,

    X. Wu, S. hao Wu, J. Wu, L. Feng, and K. C. Tan, “Evolutionary computation in the era of large language model: Survey and roadmap,” 2024

  25. [25]

    Large language models as optimizers,

    C. Yang, X. Wang, Y . Lu, H. Liu, Q. V . Le, D. Zhou, and X. Chen, “Large language models as optimizers,” 2024

  26. [26]

    Evolution of heuristics: Towards efficient automatic algorithm design using large language mode,

    F. Liu, X. Tong, M. Yuan, X. Lin, F. Luo, Z. Wang, Z. Lu, and Q. Zhang, “Evolution of heuristics: Towards efficient automatic algorithm design using large language mode,” 2024

  27. [27]

    Algorithm evolution using large language model,

    F. Liu, X. Tong, M. Yuan, and Q. Zhang, “Algorithm evolution using large language model,” 2023

  28. [28]

    Large Language Models to Enhance Bayesian Optimization

    T. Liu, N. Astorga, N. Seedat, and M. van der Schaar, “Large language models to enhance bayesian optimization,”arXiv preprint arXiv:2402.03921, 2024

  29. [29]

    Language model crossover: Variation through few-shot prompting

    E. Meyerson, M. J. Nelson, H. Bradley, A. Moradi, A. K. Hoover, and J. Lehman, “Language model crossover: Variation through few-shot prompting,”arXiv preprint arXiv:2302.12170, 2023

  30. [30]

    Large language models as evolution strategies,

    R. Lange, Y . Tian, and Y . Tang, “Large language models as evolution strategies,” inProceedings of the Genetic and Evolutionary Computation Conference Companion, 2024, pp. 579–582

  31. [31]

    As-llm: When algorithm se- lection meets large language model,

    X. Wu, Y . Zhong, J. Wu, and K. C. Tan, “As-llm: When algorithm se- lection meets large language model,”arXiv preprint arXiv:2311.13184, 2023

  32. [32]

    L-autoda: Leveraging large language models for automated decision-based adversarial attacks,

    P. Guo, F. Liu, X. Lin, Q. Zhao, and Q. Zhang, “L-autoda: Leveraging large language models for automated decision-based adversarial attacks,” arXiv preprint arXiv:2401.15335, 2024

  33. [33]

    Optimus: Optimization modeling using MIP solvers and large language models

    A. AhmadiTeshnizi, W. Gao, and M. Udell, “Optimus: Optimization modeling using mip solvers and large language models,”arXiv preprint arXiv:2310.06116, 2023

  34. [34]

    The openelm library: Leveraging progress in language models for novel evolutionary algorithms,

    H. Bradley, H. Fan, T. Galanos, R. Zhou, D. Scott, and J. Lehman, “The openelm library: Leveraging progress in language models for novel evolutionary algorithms,” inGenetic Programming Theory and Practice XX. Springer, 2024, pp. 177–201

  35. [35]

    Large language models as surrogate models in evolutionary algorithms: A preliminary study,

    H. Hao, X. Zhang, and A. Zhou, “Large language models as surrogate models in evolutionary algorithms: A preliminary study,”Swarm and Evolutionary Computation, vol. 91, p. 101741, 2024

  36. [36]

    Deep reinforcement learning from human preferences,

    P. F. Christiano, J. Leike, T. Brown, M. Martic, S. Legg, and D. Amodei, “Deep reinforcement learning from human preferences,”Advances in neural information processing systems, 2017

  37. [37]

    Learning to summarize with human feedback,

    N. Stiennon, L. Ouyang, J. Wu, D. Ziegler, R. Lowe, C. V oss, A. Rad- ford, D. Amodei, and P. F. Christiano, “Learning to summarize with human feedback,”Advances in neural information processing systems, vol. 33, pp. 3008–3021, 2020

  38. [38]

    Proximal Policy Optimization Algorithms

    J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Prox- imal policy optimization algorithms,”arXiv preprint arXiv:1707.06347, 2017

  39. [39]

    Online estimation and inference for robust policy evaluation in reinforcement learning.The Annals of Statistics, 53(5):2128–2152, 2025c

    Z. Liu, X. Guo, Z. Yang, F. Lou, L. Zeng, M. Li, Q. Qi, Z. Liu, Y . Han, D. Cheng, X. Feng, H. J. Wang, C. Shi, and L. Zhang, “Fin-r1: A large language model for financial reasoning through reinforcement learning,” 2025. [Online]. Available: https://arxiv.org/abs/2503.16252

  40. [40]

    Legalδ: Enhancing legal reasoning in llms via reinforcement learning with chain-of-thought guided information gain,

    X. Dai, B. Xu, Z. Liu, Y . Yan, H. Xie, X. Yi, S. Wang, and G. Yu, “Legalδ: Enhancing legal reasoning in llms via reinforcement learning with chain-of-thought guided information gain,” 2025. [Online]. Available: https://arxiv.org/abs/2508.12281

  41. [41]

    Evaluating Test-Time Scaling LLMs for Legal Reasoning: OpenAI o1, DeepSeek-R1, and Beyond , url=

    Y . Hu, Y . Yu, L. Gan, B. Wei, K. Kuang, and F. Wu, “Evaluating test-time scaling llms for legal reasoning: Openai o1, deepseek-r1, and beyond,” inFindings of the Association for Computational Linguistics: EMNLP 2025. Association for Computational Linguistics, 2025, p. 13759–13781. [Online]. Available: http://dx.doi.org/10.18653/v1/2025.findings-emnlp.742

  42. [42]

    Legal mathematical reasoning with llms: Procedural alignment through two-stage reinforcement learning,

    K. Zhang, G. Xie, W. Yu, M. Xu, X. Tang, Y . Li, and J. Xu, “Legal mathematical reasoning with llms: Procedural alignment through two-stage reinforcement learning,” 2025. [Online]. Available: https://arxiv.org/abs/2504.02590 14

  43. [43]

    Yu, and Qingsong Wen

    Z. Chu, S. Wang, J. Xie, T. Zhu, Y . Yan, J. Ye, A. Zhong, X. Hu, J. Liang, P. S. Yu, and Q. Wen, “Llm agents for education: Advances and applications,” 2025. [Online]. Available: https://arxiv.org/abs/2503.11733

  44. [44]

    Cultivating helpful, personalized, and creative ai tutors: A framework for pedagogical alignment using reinforcement learning,

    S. Song, W. Liu, Y . Lu, R. Zhang, T. Liu, J. Lv, X. Wang, A. Zhou, F. Tan, B. Jiang, and H. Hao, “Cultivating helpful, personalized, and creative ai tutors: A framework for pedagogical alignment using reinforcement learning,” 2025. [Online]. Available: https://arxiv.org/abs/2507.20335

  45. [45]

    Benchmark functions for the cec’2008 special session and competition on high-dimensional real-parameter optimization,

    K. Tang, X. Yao, P. Suganthan, C. MacNish, Y . Chen, C. Chen, and Z. Yang, “Benchmark functions for the cec’2008 special session and competition on high-dimensional real-parameter optimization,”Na- ture Inspired Computation and Applications Laboratory, USTC, Hefei, China, 2007

  46. [46]

    Problem definitions and evaluation criteria for the cec 2010 competition on constrained real-parameter optimization,

    R. Mallipeddi and P. N. Suganthan, “Problem definitions and evaluation criteria for the cec 2010 competition on constrained real-parameter optimization,”Nanyang Technological University, Singapore, vol. 24, p. 910, 2010

  47. [47]

    A large population size can be unhelpful in evolutionary algorithms,

    T. Chen, K. Tang, G. Chen, and X. Yao, “A large population size can be unhelpful in evolutionary algorithms,”Theoretical Computer Science, vol. 436, pp. 54–70, 2012

  48. [48]

    Evolutionary algorithm with dynamic population size for constrained multiobjective optimiza- tion,

    B.-C. Wang, Z.-Y . Shui, Y . Feng, and Z. Ma, “Evolutionary algorithm with dynamic population size for constrained multiobjective optimiza- tion,”Swarm and Evolutionary Computation, vol. 73, p. 101104, 2022

  49. [49]

    The impact of population sizes and diversity on the adaptability of evolution strategies in dynamic environments,

    L. Schonemann, “The impact of population sizes and diversity on the adaptability of evolution strategies in dynamic environments,” in Proceedings of the 2004 Congress on Evolutionary Computation (IEEE Cat. No. 04TH8753), vol. 2. IEEE, 2004, pp. 1270–1277

  50. [50]

    An estimation of distribution algorithm with cheap and expensive local search methods,

    A. Zhou, J. Sun, and Q. Zhang, “An estimation of distribution algorithm with cheap and expensive local search methods,”IEEE Transactions on Evolutionary Computation, vol. 19, no. 6, pp. 807–822, 2015

  51. [51]

    On the robustness of a simple domain reduction scheme for simulation-based optimization,

    N. Stander and K. J. Craig, “On the robustness of a simple domain reduction scheme for simulation-based optimization,”Engineering Com- putations, vol. 19, no. 4, pp. 431–450, 2002

  52. [52]

    Efficient global optimiza- tion of expensive black-box functions,

    D. R. Jones, M. Schonlau, and W. J. Welch, “Efficient global optimiza- tion of expensive black-box functions,”Journal of Global optimization, vol. 13, no. 4, pp. 455–492, 1998

  53. [53]

    A surrogate-assisted multiswarm optimization algorithm for high-dimensional computationally expensive problems,

    F. Li, X. Cai, L. Gao, and W. Shen, “A surrogate-assisted multiswarm optimization algorithm for high-dimensional computationally expensive problems,”IEEE transactions on cybernetics, vol. 51, no. 3, pp. 1390– 1402, 2020

  54. [54]

    Hollander, D

    M. Hollander, D. A. Wolfe, and E. Chicken,Nonparametric statistical methods. John Wiley & Sons, 2013

  55. [55]

    Evolutionary programming made faster,

    X. Yao, Y . Liu, and G. Lin, “Evolutionary programming made faster,” IEEE Transactions on Evolutionary computation, vol. 3, no. 2, pp. 82– 102, 1999

  56. [56]

    A classification and pareto dom- ination based multiobjective evolutionary algorithm,

    J. Zhang, A. Zhou, and G. Zhang, “A classification and pareto dom- ination based multiobjective evolutionary algorithm,” in2015 IEEE congress on evolutionary computation (CEC). IEEE, 2015, pp. 2883– 2890

  57. [57]

    A surrogate-assisted reference vector guided evolutionary algorithm for computationally expensive many-objective optimization,

    T. Chugh, Y . Jin, K. Miettinen, J. Hakanen, and K. Sindhya, “A surrogate-assisted reference vector guided evolutionary algorithm for computationally expensive many-objective optimization,”IEEE Trans- actions on Evolutionary Computation, vol. 22, no. 1, pp. 129–142, 2016

  58. [58]

    A pairwise com- parison based surrogate-assisted evolutionary algorithm for expensive multi-objective optimization,

    Y . Tian, J. Hu, C. He, H. Ma, L. Zhang, and X. Zhang, “A pairwise com- parison based surrogate-assisted evolutionary algorithm for expensive multi-objective optimization,”Swarm and Evolutionary Computation, vol. 80, p. 101323, 2023

  59. [59]

    Solving multiobjective optimization problems using an artificial immune system,

    C. A. C. Coello and N. C. Cort ´es, “Solving multiobjective optimization problems using an artificial immune system,”Genetic programming and evolvable machines, vol. 6, no. 2, pp. 163–190, 2005

  60. [60]

    Scalable test problems for evolutionary multiobjective optimization,

    K. Deb, L. Thiele, M. Laumanns, and E. Zitzler, “Scalable test problems for evolutionary multiobjective optimization,” inEvolutionary multiob- jective optimization: theoretical advances and applications. Springer, 2005, pp. 105–145

  61. [61]

    A review of multiobjective test problems and a scalable test problem toolkit,

    S. Huband, P. Hingston, L. Barone, and L. While, “A review of multiobjective test problems and a scalable test problem toolkit,”IEEE Transactions on Evolutionary Computation, vol. 10, no. 5, pp. 477–506, 2006

  62. [62]

    A benchmark test suite for evolutionary many-objective optimization,

    R. Cheng, M. Li, Y . Tian, X. Zhang, S. Yang, Y . Jin, and X. Yao, “A benchmark test suite for evolutionary many-objective optimization,” Complex & Intelligent Systems, vol. 3, no. 1, pp. 67–81, 2017

  63. [63]

    J. H. Holland,Adaptation in natural and artificial systems: an intro- ductory analysis with applications to biology, control, and artificial intelligence. MIT press, 1992

  64. [64]

    GPT-4o System Card

    A. Hurst, A. Lerer, A. P. Goucher, A. Perelman, A. Ramesh, A. Clark, A. Ostrow, A. Welihinda, A. Hayes, A. Radfordet al., “Gpt-4o system card,”arXiv preprint arXiv:2410.21276, 2024

  65. [65]

    Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality, Long Context, and Next Generation Agentic Capabilities

    G. Comanici, E. Bieber, M. Schaekermann, I. Pasupat, N. Sachdeva, I. Dhillon, M. Blistein, O. Ram, D. Zhang, E. Rosenet al., “Gem- ini 2.5: Pushing the frontier with advanced reasoning, multimodality, long context, and next generation agentic capabilities,”arXiv preprint arXiv:2507.06261, 2025

  66. [66]

    Introducing Claude 4,

    Anthropic, “Introducing Claude 4,” May 2025. [Online]. Available: https://www.anthropic.com/news/claude-4

  67. [67]

    Qwen3 Technical Report

    A. Yang, A. Li, B. Yang, B. Zhang, B. Hui, B. Zheng, B. Yu, C. Gao, C. Huang, C. Lvet al., “Qwen3 technical report,”arXiv preprint arXiv:2505.09388, 2025

  68. [68]

    Qwen2.5 technical report,

    Qwen, :, A. Yang, B. Yang, B. Zhang, B. Hui, B. Zheng, B. Yu, C. Li, D. Liu, F. Huang, H. Wei, H. Lin, J. Yang, J. Tu, J. Zhang, J. Yang, J. Yang, J. Zhou, J. Lin, K. Dang, K. Lu, K. Bao, K. Yang, L. Yu, M. Li, M. Xue, P. Zhang, Q. Zhu, R. Men, R. Lin, T. Li, T. Tang, T. Xia, X. Ren, X. Ren, Y . Fan, Y . Su, Y . Zhang, Y . Wan, Y . Liu, Z. Cui, Z. Zhang, ...

  69. [69]

    Qwen2.5 Technical Report

    [Online]. Available: https://arxiv.org/abs/2412.15115

  70. [70]

    llama.cpp,

    G. Gerganov, I. Kawrakow, and Contributors, “llama.cpp,” https://github.com/ggerganov/llama.cpp, 2023