pith. machine review for the scientific record. sign in

arxiv: 2604.09095 · v2 · submitted 2026-04-10 · 💻 cs.LG · math.OC

Recognition: unknown

GeoPAS: Geometric Probing for Algorithm Selection in Continuous Black-Box Optimisation

Authors on Pith no claims yet

Pith reviewed 2026-05-10 17:00 UTC · model grok-4.3

classification 💻 cs.LG math.OC
keywords algorithm selectionblack-box optimizationgeometric probingcontinuous optimizationlandscape descriptorsconvolutional networksportfolio optimizationrisk-aware selection
0
0 comments X

The pith

Coarse geometric slices of problems allow better selection among optimization algorithms than always using the strongest single solver.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper proposes representing optimization problem instances through several coarse two-dimensional geometric slices taken at varied locations, orientations, and logarithmic scales. A convolutional network processes these slices while respecting their validity, incorporates scale and amplitude information, and combines the embeddings in a way that does not depend on order. This combined representation then drives a selection model that predicts solver performance on a log scale and adds a penalty for poor tail behavior. When tested on a standard set of benchmark problems in dimensions from 2 to 10 using a portfolio of 12 solvers, the approach yields better average results than picking the single best solver, and the advantage holds even when the test problems are completely unseen or grouped differently.

Core claim

GeoPAS represents a problem instance by multiple coarse two-dimensional slices sampled across locations, orientations, and logarithmic scales. A shared validity-aware convolutional encoder maps each slice to an embedding, conditions it on slice-scale and amplitude statistics, and aggregates the resulting features permutation-invariantly for risk-aware solver selection via log-scale performance prediction with an explicit penalty on tail failures. On the COCO/BBOB benchmark with a 12-solver portfolio in dimensions 2--10, this method improves over the single best solver under leave-instance-out, grouped random, and leave-problem-out evaluation.

What carries the argument

Multi-scale geometric slicing of the search space, processed by a validity-aware convolutional encoder with permutation-invariant aggregation to produce embeddings for solver performance prediction.

If this is right

  • The approach delivers higher average performance than any fixed solver choice across multiple evaluation protocols.
  • Transferable signals from the slices support selection even when problems are split by type or instance.
  • Penalizing tail failures in the prediction reduces the chance of selecting solvers that fail badly on certain instances.
  • Results in dimensions 2 to 10 indicate the method applies to moderate-dimensional continuous problems.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If geometric slices capture key landscape properties, the same probing strategy could support selection in other optimization domains such as noisy or constrained problems.
  • Heavy-tail regimes that still dominate suggest combining the initial static selection with limited online adaptation during the run.
  • Applying similar slice-based representations to higher dimensions would reveal whether logarithmic scaling continues to provide useful signals.
  • The encoder architecture could be adapted for direct landscape feature extraction beyond algorithm selection.

Load-bearing premise

That sparse sampling of coarse two-dimensional slices at different scales and orientations extracts sufficient information about the full problem landscape to predict solver suitability across instances.

What would settle it

Running the selection on a held-out set of problems where the geometric slices fail to correlate with actual solver performance differences, resulting in no improvement or degradation relative to the best single solver.

Figures

Figures reproduced from arXiv: 2604.09095 by Jiabao Brad Wang, Mustafa Misir, Xiang Shi, Yiliang Yuan.

Figure 1
Figure 1. Figure 1: Illustration of multi-scale geometric probing. Oriented square slices with varying relative side length [PITH_FULL_IMAGE:figures/full_fig_p003_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the GeoPAS visual encoder. Each sampled slice contributes a value map [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Histogram of relERT distribution of GeoPAS vs. SBS under LPO. The relERT axis is displayed in log-scale [PITH_FULL_IMAGE:figures/full_fig_p008_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Selection frequencies of GeoPAS vs. VBS under LPO on (left) all datapoints, (middle) datapoints from SBS-non [PITH_FULL_IMAGE:figures/full_fig_p009_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Heatmaps of mean (top), median (middle), and 90th [PITH_FULL_IMAGE:figures/full_fig_p011_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Twenty sampled, normalised slices (value channel) of resolution [PITH_FULL_IMAGE:figures/full_fig_p014_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Selection distributions and pick frequencies (GeoPAS vs. VBS) under LIO, Random, and LPO. [PITH_FULL_IMAGE:figures/full_fig_p015_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Survival curves P(relERT > 𝑡) for GeoPAS and SBS under LIO, Random, and LPO [PITH_FULL_IMAGE:figures/full_fig_p016_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Heatmaps of mean (top), median (middle), and 90th percentile (bottom) [PITH_FULL_IMAGE:figures/full_fig_p017_9.png] view at source ↗
read the original abstract

Automated algorithm selection in continuous black-box optimisation typically relies on fixed landscape descriptors computed under a limited probing budget, yet such descriptors can degrade under problem-split or cross-benchmark evaluation. We propose GeoPAS, a geometric probing approach that represents a problem instance by multiple coarse two-dimensional slices sampled across locations, orientations, and logarithmic scales. A shared validity-aware convolutional encoder maps each slice to an embedding, conditions it on slice-scale and amplitude statistics, and aggregates the resulting features permutation-invariantly for risk-aware solver selection via log-scale performance prediction with an explicit penalty on tail failures. On COCO/BBOB with a 12-solver portfolio in dimensions 2--10, GeoPAS improves over the single best solver under leave-instance-out, grouped random, and leave-problem-out evaluation. These results suggest that multi-scale geometric slices provide a useful transferable static signal for algorithm selection, although a small number of heavy-tail regimes remain and continue to dominate the mean. Our code is available at https://github.com/BradWangW/GeoPAS.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes GeoPAS for automated algorithm selection in continuous black-box optimization. It represents problem instances via multiple coarse 2D geometric slices sampled across locations, orientations, and logarithmic scales; a validity-aware CNN encoder produces embeddings conditioned on scale and amplitude statistics, which are aggregated permutation-invariantly. These features are used for risk-aware log-scale performance prediction with an explicit tail penalty to select from a 12-solver portfolio. On COCO/BBOB benchmarks in dimensions 2-10, the method is reported to improve over the single-best solver under leave-instance-out, grouped-random, and leave-problem-out protocols, though a small number of heavy-tail regimes continue to dominate the mean.

Significance. If the reported gains hold under detailed scrutiny, GeoPAS would demonstrate that multi-scale geometric probing can yield transferable static signals for solver selection that generalize better than fixed landscape descriptors across problem splits. The open-source code strengthens reproducibility. However, the abstract's own caveat on heavy-tail regimes dominating the mean limits the practical significance for overall mean performance, and the lack of reported error bars, statistical tests, or tail-specific breakdowns reduces confidence in the transferability claim.

major comments (2)
  1. [Abstract] Abstract: The central claim of consistent improvement over the single-best solver under leave-problem-out evaluation is load-bearing for the transferability argument, yet no quantitative effect sizes, confidence intervals, or statistical significance tests are provided; without these, it is impossible to determine whether gains are robust or driven by the bulk of easier instances.
  2. [Abstract] Abstract and experimental evaluation: The manuscript acknowledges that 'a small number of heavy-tail regimes remain and continue to dominate the mean' even after the risk-aware log-scale prediction. No breakdown of selection accuracy or regret on tail versus bulk instances is described, so it remains unclear whether the multi-scale geometric slices (plus CNN encoder and conditioning) actually improve performance on the regimes that matter most for the mean.
minor comments (2)
  1. [Method] The probing budget, exact sampling procedure for the 2D slices (locations, orientations, scales), and validity-aware CNN architecture details should be expanded with pseudocode or equations for reproducibility.
  2. [Experiments] Figure and table captions should explicitly state the number of runs, random seeds, and whether error bars represent standard deviation or standard error.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback emphasizing the need for statistical rigor and targeted analysis of heavy-tail regimes. We will revise the manuscript to incorporate effect sizes, confidence intervals, significance tests, and performance breakdowns, which will better support the transferability claims while maintaining the honest acknowledgment of limitations.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim of consistent improvement over the single-best solver under leave-problem-out evaluation is load-bearing for the transferability argument, yet no quantitative effect sizes, confidence intervals, or statistical significance tests are provided; without these, it is impossible to determine whether gains are robust or driven by the bulk of easier instances.

    Authors: We agree that the absence of effect sizes, confidence intervals, and statistical tests weakens the ability to assess robustness. In the revised manuscript we will report relative improvements with standard errors or confidence intervals for the leave-problem-out protocol, together with results from paired statistical tests (e.g., Wilcoxon signed-rank) comparing GeoPAS against the single-best solver. These additions will clarify whether observed gains hold across the distribution of instances rather than being driven by easier cases. revision: yes

  2. Referee: [Abstract] Abstract and experimental evaluation: The manuscript acknowledges that 'a small number of heavy-tail regimes remain and continue to dominate the mean' even after the risk-aware log-scale prediction. No breakdown of selection accuracy or regret on tail versus bulk instances is described, so it remains unclear whether the multi-scale geometric slices (plus CNN encoder and conditioning) actually improve performance on the regimes that matter most for the mean.

    Authors: We accept that a disaggregated analysis is required to evaluate effectiveness on the regimes that dominate the mean. The revised version will include explicit breakdowns of selection accuracy, regret, and solver performance metrics on tail versus bulk instances under all three evaluation protocols. This will be presented in additional tables or figures that isolate the contribution of the multi-scale geometric representation and risk-aware objective on the most challenging cases. revision: yes

Circularity Check

0 steps flagged

No circularity in derivation chain; performance gains evaluated on held-out splits

full rationale

The GeoPAS pipeline extracts geometric slices, encodes them via a CNN, conditions on scale/amplitude, aggregates invariantly, and trains a log-scale performance predictor with tail penalty. All steps are trained on observed solver runs from COCO/BBOB and evaluated under explicit leave-instance-out, grouped-random, and leave-problem-out protocols. No equation reduces a claimed prediction to a fitted input by construction, no self-citation supplies a load-bearing uniqueness result, and no ansatz is smuggled via prior work. The heavy-tail caveat is an acknowledged empirical limitation rather than a definitional collapse.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on the domain assumption that coarse 2D geometric slices capture transferable landscape information sufficient for solver selection; no explicit free parameters or invented entities are stated in the abstract.

axioms (1)
  • domain assumption Multi-scale 2D geometric slices provide a useful transferable static signal about problem difficulty for algorithm selection.
    Invoked as the basis for the probing approach and its claimed generalization across evaluation splits.

pith-pipeline@v0.9.0 · 5491 in / 1368 out tokens · 68508 ms · 2026-05-10T17:00:19.111556+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

44 extracted references · 4 canonical work pages · 1 internal anchor

  1. [1]

    Mohamad Alissa, Kevin Sim, and Emma Hart. 2023. Automated algorithm selection: from feature-based to feature-free approaches.Journal of Heuristics 29, 1 (2023), 1–38

  2. [2]

    Asma Atamna. 2015. Benchmarking IPOP-CMA-ES-TPA and IPOP-CMA-ES- MSR on the BBOB noiseless testbed. InProceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1135–1142

  3. [3]

    Anne Auger, Dimo Brockhoff, and Nikolaus Hansen. 2013. Benchmarking the local metamodel CMA-ES on the noiseless BBOB’2013 test bed. InProceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1225–1232

  4. [4]

    Petr Baudiš and Petr Pošík. 2015. Global line search algorithm hybridized with quadratic interpolation and its extension to separable functions. InProceedings of the 2015 annual conference on genetic and evolutionary computation. 257–264

  5. [5]

    Albert S Berahas, Richard H Byrd, and Jorge Nocedal. 2019. Derivative-free optimization of noisy functions via quasi-Newton methods.SIAM Journal on Optimization29, 2 (2019), 965–993

  6. [6]

    Bernd Bischl, Olaf Mersmann, Heike Trautmann, and Mike Preuß. 2012. Al- gorithm selection based on exploratory landscape analysis and cost-sensitive learning. InProceedings of the 14th annual conference on Genetic and evolutionary computation. 313–320

  7. [7]

    Gjorgjina Cenikj, Ana Nikolikj, and Tome Eftimov. 2025. Recent Advances in Meta-features Used for Representing Black-box Single-objective Continuous Op- timization. InProceedings of the Genetic and Evolutionary Computation Conference Companion. 1471–1494

  8. [8]

    Gjorgjina Cenikj, Ana Nikolikj, Gašper Petelin, Niki Van Stein, Carola Doerr, and Tome Eftimov. 2026. A survey of features used for representing black-box single-objective continuous optimization.Swarm and Evolutionary Computation 101 (2026), 102288

  9. [9]

    Gjorgjina Cenikj, Gašper Petelin, and Tome Eftimov. 2024. Transoptas: Transformer-based algorithm selection for single-objective optimization. InPro- ceedings of the Genetic and Evolutionary Computation Conference Companion. 403–406

  10. [10]

    Gjorgjina Cenikj, Gašper Petelin, Moritz Seiler, Nikola Cenikj, and Tome Eftimov

  11. [11]

    Landscape features in single-objective continuous optimization: Have we hit a wall in algorithm selection generalization?Swarm and Evolutionary Computation94 (2025), 101894

  12. [12]

    Konstantin Dietrich, Diederick Vermetten, Carola Doerr, and Pascal Kerschke

  13. [13]

    InProceedings of the Genetic and Evolutionary Computation Conference

    Impact of training instance selection on automated algorithm selection models for numerical black-box optimization. InProceedings of the Genetic and Evolutionary Computation Conference. 1007–1016

  14. [14]

    Nikolaus Hansen. 2016. The CMA evolution strategy: A tutorial.arXiv preprint arXiv:1604.00772(2016)

  15. [15]

    Nikolaus Hansen, Anne Auger, Raymond Ros, Olaf Mersmann, Tea Tušar, and Dimo Brockhoff. 2021. COCO: A platform for comparing continuous optimizers in a black-box setting.Optimization Methods and Software36, 1 (2021), 114–144

  16. [16]

    2009.Real- Parameter Black-Box Optimization Benchmarking 2009: Noiseless Functions Def- initions

    Nikolaus Hansen, Steffen Finck, Raymond Ros, and Anne Auger. 2009.Real- Parameter Black-Box Optimization Benchmarking 2009: Noiseless Functions Def- initions. Research Report RR-6829. INRIA. https://inria.hal.science/inria- 00362633v2 Version 2, HAL Id: inria-00362633

  17. [17]

    Frank Hutter, Holger Hoos, and Kevin Leyton-Brown. 2013. An evaluation of sequential model-based optimization for expensive blackbox functions. In Proceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1209–1216

  18. [18]

    Waltraud Huyer and Arnold Neumaier. 2009. Benchmarking of MCS on the noiseless function testbed.Online, 2009c. URL http://www. mat. univie. ac. at/˜ neum/papers. html989 (2009)

  19. [19]

    Stefan Ivić, Siniša Družeta, and Luka Grbčić. 2025. Randomness as Refer- ence: Benchmark Metric for Optimization in Engineering.arXiv preprint arXiv:2511.17226(2025)

  20. [20]

    Pascal Kerschke, Holger H Hoos, Frank Neumann, and Heike Trautmann. 2019. Automated algorithm selection: Survey and perspectives.Evolutionary computa- tion27, 1 (2019), 3–45

  21. [21]

    Pascal Kerschke and Heike Trautmann. 2019. Automated algorithm selection on continuous black-box problems by combining exploratory landscape analysis and machine learning.Evolutionary computation27, 1 (2019), 99–127

  22. [22]

    Ana Kostovska, Carola Doerr, Sašo Džeroski, Panče Panov, and Tome Eftimov

  23. [23]

    InProceedings of the Genetic and Evolutionary Wang et al

    Geometric Learning in Black-Box Optimization: A GNN Framework for Algorithm Performance Prediction. InProceedings of the Genetic and Evolutionary Wang et al. Computation Conference Companion. 487–490

  24. [24]

    Manuel López-Ibáñez, Diederick Vermetten, Johann Dreo, and Carola Doerr. 2024. Using the empirical attainment function for analyzing single-objective black-box optimization algorithms.IEEE Transactions on Evolutionary Computation(2024)

  25. [25]

    Ilya Loshchilov, Marc Schoenauer, and Michèle Sebag. 2013. Bi-population CMA-ES agorithms with surrogate models and line searches. InProceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1177–1184

  26. [26]

    Zeyuan Ma, Hongshu Guo, Yue-Jiao Gong, Jun Zhang, and Kay Chen Tan. 2025. Toward automated algorithm design: A survey and practical guide to meta-black- box-optimization.IEEE Transactions on Evolutionary Computation(2025)

  27. [27]

    Francesco Mezzadri. 2006. How to generate random matrices from the classical compact groups.arXiv preprint math-ph/0609050(2006)

  28. [28]

    Ana Nikolikj, Ana Kostovska, Gjorgjina Cenikj, Carola Doerr, and Tome Eftimov

  29. [29]

    In2024 IEEE Congress on Evolutionary Computation (CEC)

    Generalization Ability of Feature-Based Performance Prediction Models: A Statistical Analysis Across Benchmarks. In2024 IEEE Congress on Evolutionary Computation (CEC). IEEE, 1–8

  30. [30]

    Art B Owen. 1998. Scrambling Sobol’and Niederreiter–Xing Points.Journal of complexity14, 4 (1998), 466–489

  31. [31]

    László Pál. 2013. Benchmarking a hybrid multi level single linkage algorithm on the BBOB noiseless testbed. InProceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1145–1152

  32. [32]

    László Pál. 2013. Comparison of multistart global optimization algorithms on the BBOB noiseless testbed. InProceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1153–1160

  33. [33]

    Gašper Petelin and Gjorgjina Cenikj. 2024. On Generalization of ELA Feature Groups. InProceedings of the Genetic and Evolutionary Computation Conference Companion. 419–422

  34. [34]

    Gašper Petelin and Gjorgjina Cenikj. 2025. The Pitfalls of Benchmarking in Algorithm Selection: What We Are Getting Wrong. InProceedings of the Genetic and Evolutionary Computation Conference. 1181–1189

  35. [35]

    Gašper Petelin, Gjorgjina Cenikj, and Tome Eftimov. 2024. Tinytla: Topological landscape analysis for optimization problem classification in a limited sample setting.Swarm and Evolutionary Computation84 (2024), 101448

  36. [36]

    Petr Pošík and Petr Baudiš. 2015. Dimension selection in axis-parallel brent- step method for black-box optimization of separable continuous functions. In Proceedings of the Annual Conference on Genetic and Evolutionary Computation (GECCO) Companion. 1151–1158

  37. [37]

    Raphael Patrick Prager, Moritz Vinzent Seiler, Heike Trautmann, and Pascal Kerschke. 2022. Automated algorithm selection in single-objective continuous optimization: a comparative study of deep learning and landscape analysis meth- ods. InInternational Conference on Parallel Problem Solving from Nature. Springer, 3–17

  38. [38]

    Iván Olarte Rodríguez, Maria Laura Santoni, Fabian Duddeck, Carola Doerr, Thomas Bäck, and Elena Raponi. 2025. MECHBench: A Set of Black-Box Op- timization Benchmarks originated from Structural Mechanics.arXiv preprint arXiv:2511.10821(2025)

  39. [39]

    Moritz Seiler, Urban Škvorc, Gjorgjina Cenikj, Carola Doerr, and Heike Traut- mann. 2024. Learned features vs. Classical ELA on affine BBOB functions. In International Conference on Parallel Problem Solving from Nature. Springer, 137– 153

  40. [40]

    Moritz Seiler, Urban Škvorc, Carola Doerr, and Heike Trautmann. 2024. Synergies of deep and classical exploratory landscape features for automated algorithm selection. InInternational Conference on Learning and Intelligent Optimization. Springer, 361–376

  41. [41]

    Moritz Vinzent Seiler, Pascal Kerschke, and Heike Trautmann. 2025. Deep-ela: Deep exploratory landscape analysis with self-supervised pretrained transform- ers for single-and multi-objective continuous optimization problems.Evolution- ary Computation(2025), 1–27

  42. [42]

    Urban Škvorc, Tome Eftimov, and Peter Korošec. 2021. The effect of sampling methods on the invariance to function transformations when using exploratory landscape analysis. In2021 IEEE Congress on Evolutionary Computation (CEC). IEEE, 1139–1146

  43. [43]

    Bas van Stein, Fu Xing Long, Moritz Frenzel, Peter Krause, Markus Gitterle, and Thomas Bäck. 2023. Doe2vec: Deep-learning based features for exploratory landscape analysis. InProceedings of the Companion Conference on Genetic and Evolutionary Computation. 515–518

  44. [44]

    Diederick Vermetten, Furong Ye, Thomas Bäck, and Carola Doerr. 2025. MA- BBOB: A problem generator for black-box optimization using affine combinations and shifts.ACM Transactions on Evolutionary Learning5, 1 (2025), 1–19. Geometric Probing for Algorithm Selection in Continuous Black-Box Optimisation Table 5: Average wall-clock time (s) to generate the in...