pith. machine review for the scientific record. sign in

arxiv: 2604.25371 · v1 · submitted 2026-04-28 · 🧬 q-bio.QM · cs.CV

Recognition: unknown

PhyloSDF: Phylogenetically-Conditioned Neural Generation of 3D Skull Morphology via Residual Flow Matching

Gary P. T. Choi, Kaikwan Lau

Authors on Pith no claims yet

Pith reviewed 2026-05-07 14:10 UTC · model grok-4.3

classification 🧬 q-bio.QM cs.CV
keywords 3D skull morphologyphylogenetic conditioningresidual flow matchinggenerative modelsmorphological variationDarwin's finchesevolutionary biologyDeepSDF
0
0 comments X

The pith

A phylogenetically conditioned neural model generates novel 3D skull shapes from as few as four specimens per species while matching real variation levels.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper presents a generative model for three-dimensional skull shapes that incorporates known evolutionary relationships among species even when only a handful of specimens are available for each one. It structures a shape representation space so that distances between species codes align closely with their phylogenetic distances and uses a specialized flow matching process that directly supplies each species' average shape while learning only the remaining deviations. This produces new skull models that display variation levels comparable to those observed in real specimens and that can be extended to species left out of training. A sympathetic reader would care because data scarcity has long blocked computational exploration of morphological evolution in groups like birds or mammals, where complete specimen sets are rare.

Core claim

PhyloSDF integrates a DeepSDF auto-decoder regularized by a Phylogenetic Consistency Loss that structures the latent space to correlate with evolutionary distances at Pearson r=0.993, together with a Residual Conditional Flow Matching architecture that factorizes generation into analytic species-centroid lookup and learned residual prediction. Evaluated on 100 micro-CT skulls across 24 species of Darwin's finches and relatives, the model produces 180 novel meshes verified as non-memorized that achieve 88-129 percent of real intra-species variation at the code level, surpasses denoising diffusion, standard flow matching, and Gaussian mixture baselines in Chamfer distance and morphometric Fré{

What carries the argument

The Residual Conditional Flow Matching architecture that performs analytic lookup of species centroids and predicts only residual shape deviations, paired with the Phylogenetic Consistency Loss that enforces correlation between latent codes and phylogenetic distances.

If this is right

  • Smooth interpolation in the structured latent space produces biologically plausible ancestral skull reconstructions.
  • The residual factorization allows generation to succeed where full denoising diffusion fails and where standard flow matching collapses to minimal variation.
  • Leave-one-species-out tests demonstrate extrapolation to unseen species, supporting use for filling gaps in morphological phylogenies.
  • All generated meshes achieve variation levels between 88 and 129 percent of real intra-species ranges while remaining distinct from training examples.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same residual-plus-phylogenetic structure could be tested on other anatomical elements such as limb bones or teeth to check generality beyond skulls.
  • If the latent alignment captures evolutionary signal, the model could be used to simulate morphological responses to different rates of evolutionary change.
  • Connecting the generator to explicit models of selection or drift would allow forward simulation of diversification patterns across a tree.
  • Application to fossil taxa with even fewer specimens would test whether the few-sample regime extends to deep-time reconstruction.

Load-bearing premise

That a high correlation between latent codes and evolutionary distances plus quantitative mesh metrics and non-memorization checks are enough to ensure generated shapes are both novel and biologically plausible rather than artifacts of the loss or preprocessing.

What would settle it

Expert morphometric comparison or direct measurement showing that shapes generated for a held-out species systematically deviate from independent fossil or museum data in ways not predicted by the phylogenetic distances used during training.

Figures

Figures reproduced from arXiv: 2604.25371 by Gary P. T. Choi, Kaikwan Lau.

Figure 1
Figure 1. Figure 1: Overview of the proposed pipeline for phylogenetically-conditioned 3D skull generation. (a) Data preprocessing and representation. Raw STL skull meshes from museum specimens of Darwin’s Finches and their relatives are converted to watertight meshes and normalized, after which signed distance function (SDF) samples are generated for DeepSDF auto-decoder training. A Phylogenetic Consistency Loss augments the… view at source ↗
Figure 2
Figure 2. Figure 2: Overview of our PhyloSDF pipeline. (a) DeepSDF auto-decoder learns zi with phylogenetic consistency loss. (b) Latent augmentation expands codes. (c) Residual CFM learns noise, then residual mapping. (d) Generation: znew = µs + CFM(ϵ), decoded via Marching Cubes. 4.1 Stage 1: Phylogenetically-Regularized DeepSDF We adopt the DeepSDF auto-decoder formulation from Section 3, in which each of the N = 100 speci… view at source ↗
Figure 3
Figure 3. Figure 3: The phylogenetic structure that drives the phylogenetic consistency loss in our formulation. (a) Normalized pairwise phylogenetic distance matrix dphylo for all species in the dataset, with genus-level color coding and hierarchical clustering dendrogram. Colored boxes highlight intra-genus blocks. (b) The phylogenetic consistency loss Lphylo maps genus-level tree distances into the regularized latent space… view at source ↗
Figure 4
Figure 4. Figure 4: Phylogenetic tree and dataset composition for the 24 species and N = 100 specimens used in all experiments. Species are organized into eight biologically defined clades: Geospiza (ground finches, 7 species, 29 specimens), Camarhynchus (tree finches, 3 species, 13 specimens), Certhidea/Platyspiza (warbler and vegetarian finches, 3 species, 9 specimens), Pinaroloxias (Cocos finch, 1 species, 2 specimens), Co… view at source ↗
Figure 5
Figure 5. Figure 5: Species-conditioned skull meshes generated by Residual CFM for the four focal species. Each row shows five independently sampled outputs from the same species conditioning. Row 1 (G. fortis): compact skulls with short, deep beaks characteristic of a seed-cracking ground finch. Row 2 (G. magnirostris): larger cranial vaults and broader beaks consistent with this species’ role as the large ground finch. Row … view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of G. fuliginosa skulls generated by Residual CFM (row a, n = 5) and the GMM baseline (row b, n = 5). CFM samples are geometrically consistent with the G. fuliginosa morphotype across all five outputs, beak shape, cranial vault curvature, and overall skull proportions remain stable, reflecting its superior fidelity (NN-CD 0.00181 vs. 0.00190). GMM samples exhibit greater within-row shape variati… view at source ↗
Figure 7
Figure 7. Figure 7: Method comparison: intra-species diversity. Only Residual CFM recovers biological-level variation. DDPM fails entirely; standard CFM mode-collapses to centroids; GMM achieves partial diversity. visually distinct but geometrically similar meshes. The code-level metric captures the full variation learned by the generative model; the mesh-level metric is more conservative and influenced by the SDF decoder’s s… view at source ↗
Figure 8
Figure 8. Figure 8: Latent interpolation smoothness. Vertex counts (a proxy for skull size) change monotonically along interpolation paths between species centroids, confirming a smooth, well-structured latent space. 61% improvement, confirming that the augmented training set provides the sample volume needed for the flow model to learn a broader residual distribution. Adding phylogenetic loss alone without augmentation (Conf… view at source ↗
Figure 9
Figure 9. Figure 9: Leave-one-species-out generalization. Blue: centroid prediction error (predicted vs. true centroid L2 distance). Red: true within-species spread. The centroid error is consistently below within-species variation, confirming that phylogenetic neighbors provide useful positional information. residual distribution conditioned on this predicted centroid. No model retraining occurs; the experiment tests latent … view at source ↗
Figure 10
Figure 10. Figure 10: Latent space interpolation between three species pairs, each showing 11 uniformly spaced steps from the source centroid (left) to the target centroid (right). (a) From G. fortis to G. magnirostris (congeneric, dphylo = 0.125); beak depth increases monotonically, and cranial vault enlarges progressively. (b) From G. fortis to G. fuliginosa (congeneric, dphylo = 0.125); overall skull size decreases while be… view at source ↗
read the original abstract

Generating novel, biologically plausible three-dimensional morphological structures is a fundamental challenge in computational evolutionary biology, hampered by extreme data scarcity and the requirement that generated shapes respect phylogenetic relationships among species. In this work, we present PhyloSDF, a phylogenetically-conditioned neural generative model for 3D biological morphology that integrates two innovations: (1) a DeepSDF auto-decoder regularized by a novel Phylogenetic Consistency Loss that structures the latent space to correlate with evolutionary distances (Pearson $r=0.993$); (2) a Residual Conditional Flow Matching (Residual CFM) architecture that factorizes generation into analytic species-centroid lookup and learned residual prediction, enabling generation from as few as ~4 specimens per species. We evaluate PhyloSDF on 100 micro-CT-scanned skulls of Darwin's Finches and their relatives across 24 species. The model generates novel meshes achieving 88-129% of real intra-species variation at the code level, with all 180 generated meshes verified as non-memorized. Residual CFM surpasses denoising diffusion (which fails entirely at this scale), standard flow matching (which mode-collapses to 3-6% variation), and a Gaussian mixture baseline in both fidelity (Chamfer Distance 0.00181 vs. 0.00190) and morphometric Fr\'{e}chet distance (10,641 vs. 13,322). Leave-one-species-out experiments across 18 species demonstrate phylogenetic extrapolation capability, and smooth latent interpolations produce biologically plausible ancestral skull reconstructions.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript presents PhyloSDF, a phylogenetically-conditioned generative model for 3D skull shapes using a DeepSDF auto-decoder regularized by a Phylogenetic Consistency Loss (Pearson r = 0.993 with evolutionary distances) and a Residual Conditional Flow Matching (Residual CFM) architecture. On a dataset of 100 micro-CT skulls from 24 species, it generates 180 novel meshes from as few as ~4 specimens per species, achieving 88-129% of real intra-species variation, non-memorization verified, and superior performance over baselines in Chamfer Distance (0.00181) and morphometric Fréchet distance (10,641). Leave-one-species-out experiments across 18 species show phylogenetic extrapolation, with smooth interpolations for ancestral reconstructions.

Significance. If the central claims hold, this work offers a meaningful advance for computational evolutionary biology by tackling extreme data scarcity in 3D morphological generation while enforcing phylogenetic structure. The residual CFM factorization is a practical strength for low-sample regimes (~4 specimens/species), and the explicit non-memorization verification plus quantitative baselines provide a reproducible foundation. Successful phylogenetic extrapolation and ancestral interpolation could support new studies of evolutionary trajectories. The integration of flow matching with a consistency loss on latent codes is a clear technical contribution.

major comments (3)
  1. [Phylogenetic Consistency Loss definition and §4.2] The Phylogenetic Consistency Loss is constructed to directly increase correlation between pairwise latent distances and evolutionary distances; the reported Pearson r=0.993 is therefore the expected outcome of the loss rather than independent evidence of biologically meaningful structure. This global code-level correlation does not automatically translate to per-shape morphological plausibility in the decoded SDF outputs (e.g., respect for functional or developmental constraints). A concrete test—such as expert morphological scoring or comparison against known biomechanical features—would be required to support the claim that generated shapes are evolutionarily plausible rather than loss-induced artifacts.
  2. [§4.3 and §4.4 (evaluation and leave-one-species-out)] With only 100 total specimens across 24 species, the reported success (88-129% intra-species code variation, all 180 meshes non-memorized, Chamfer 0.00181) is quantitatively strong but rests on metrics that can be satisfied by faithful interpolations. The leave-one-species-out experiments measure distribution match but supply no independent biological ground truth for the extrapolated species. The non-memorization verification procedure should be specified in detail (e.g., exact distance threshold or reconstruction error cutoff used).
  3. [Residual CFM architecture description and Table 2] The Residual CFM factorization (analytic centroid lookup plus learned residual) reduces the fitting burden and enables small-sample generation, yet the manuscript does not demonstrate that the residual prediction preserves phylogenetic consistency at the level of the final 3D mesh geometry rather than only at the latent-code level. An ablation removing the phylogenetic loss while keeping Residual CFM would clarify the contribution of each component.
minor comments (2)
  1. [§4.1] Define 'variation at the code level' and 'morphometric Fréchet distance' explicitly in the main text with equations or pseudocode; the abstract numbers are clear but the precise computation should be reproducible from the methods.
  2. [Figures 3-5] Figure captions for generated meshes and interpolations should include direct side-by-side comparisons with real specimens and scale information to allow readers to assess visual plausibility.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their thoughtful and constructive comments, which highlight important aspects of validation and experimental design. We address each major comment below and will revise the manuscript accordingly to improve clarity and strengthen the evidence presented.

read point-by-point responses
  1. Referee: The Phylogenetic Consistency Loss is constructed to directly increase correlation between pairwise latent distances and evolutionary distances; the reported Pearson r=0.993 is therefore the expected outcome of the loss rather than independent evidence of biologically meaningful structure. This global code-level correlation does not automatically translate to per-shape morphological plausibility in the decoded SDF outputs (e.g., respect for functional or developmental constraints). A concrete test—such as expert morphological scoring or comparison against known biomechanical features—would be required to support the claim that generated shapes are evolutionarily plausible rather than loss-induced artifacts.

    Authors: We agree that the Pearson r=0.993 is a direct result of the Phylogenetic Consistency Loss design, which explicitly regularizes latent pairwise distances to align with evolutionary distances. This is intentional to embed phylogenetic structure. Evidence that this structure produces plausible decoded morphologies comes from the leave-one-species-out experiments, where extrapolated shapes align with phylogenetic expectations in morphospace, and from the ancestral interpolation results showing smooth, biologically consistent transitions. Quantitative support includes the achieved intra-species variation (88-129%) and superior Chamfer/Fréchet distances relative to baselines. While expert morphological scoring or biomechanical validation would be valuable, it is outside the scope of the current computational study; we will revise the discussion section to explicitly distinguish latent-space regularization from decoded-shape plausibility and add qualitative comparisons to known Darwin's finch morphological traits. revision: partial

  2. Referee: With only 100 total specimens across 24 species, the reported success (88-129% intra-species code variation, all 180 meshes non-memorized, Chamfer 0.00181) is quantitatively strong but rests on metrics that can be satisfied by faithful interpolations. The leave-one-species-out experiments measure distribution match but supply no independent biological ground truth for the extrapolated species. The non-memorization verification procedure should be specified in detail (e.g., exact distance threshold or reconstruction error cutoff used).

    Authors: We will revise the methods and supplementary sections to fully specify the non-memorization verification procedure, including the exact Chamfer distance threshold and reconstruction error cutoff used to confirm all 180 meshes are novel. On the concern that metrics could be met by interpolations, the Residual CFM generates variation by learning residuals around species centroids rather than simple averaging, enabling the reported 88-129% of real intra-species variation. For leave-one-species-out, we acknowledge the lack of independent ground truth for the precise morphology of held-out species; the experiments instead validate phylogenetic consistency by showing that generated shapes match expected evolutionary positioning and intra-species distribution statistics. We will expand the evaluation discussion to address these limitations and the implications for extrapolation claims. revision: yes

  3. Referee: The Residual CFM factorization (analytic centroid lookup plus learned residual) reduces the fitting burden and enables small-sample generation, yet the manuscript does not demonstrate that the residual prediction preserves phylogenetic consistency at the level of the final 3D mesh geometry rather than only at the latent-code level. An ablation removing the phylogenetic loss while keeping Residual CFM would clarify the contribution of each component.

    Authors: We agree this ablation would strengthen the analysis. In the revised manuscript we will add results from a variant trained with Residual CFM but without the Phylogenetic Consistency Loss, comparing effects on both latent-space correlation and final mesh-level metrics (Chamfer Distance, Fréchet distance, and variation statistics). This will quantify whether phylogenetic structure is preserved in the decoded 3D geometry and clarify the individual contributions of each component. revision: yes

Circularity Check

1 steps flagged

Phylogenetic Consistency Loss enforces latent-evolutionary distance correlation by design, making r=0.993 a constructed outcome

specific steps
  1. fitted input called prediction [Abstract]
    "a DeepSDF auto-decoder regularized by a novel Phylogenetic Consistency Loss that structures the latent space to correlate with evolutionary distances (Pearson r=0.993)"

    The loss is formulated to enforce correlation between latent codes and evolutionary distances; therefore the high reported Pearson correlation (r=0.993) is achieved by construction as the direct outcome of the regularization rather than emerging independently from the data or architecture.

full rationale

The paper's core claim of phylogenetically-conditioned generation rests on a DeepSDF auto-decoder regularized by a novel Phylogenetic Consistency Loss that explicitly structures the latent space to correlate with evolutionary distances. The reported Pearson r=0.993 is the direct quantitative target and result of this loss term rather than an independent validation. While the Residual CFM factorization and non-memorization checks provide separate support for generation from limited data, the phylogenetic structuring step reduces to the loss formulation. This constitutes moderate circularity (fitted input called prediction) without invalidating the overall architecture. No self-citation chains, ansatz smuggling, or other load-bearing reductions appear in the abstract or described claims.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

Review is based solely on the abstract; the central claims rest on standard assumptions of neural generative models plus the domain assumption that evolutionary distances provide a meaningful supervisory signal for 3D morphology. No invented physical entities are introduced.

axioms (1)
  • domain assumption Evolutionary distances between species can be used as a supervisory signal to structure a latent shape space in a biologically meaningful way
    Invoked by the Phylogenetic Consistency Loss that targets Pearson correlation with these distances.

pith-pipeline@v0.9.0 · 5591 in / 1500 out tokens · 96540 ms · 2026-05-07T14:10:31.422659+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

52 extracted references · 3 canonical work pages · 3 internal anchors

  1. [1]

    A virtual world of paleontology,

    J. A. Cunningham, I. A. Rahman, S. Lautenschlager, E. J. Rayfield, and P. C. Donoghue, “A virtual world of paleontology,”Trends in Ecology & Evolution, vol. 29, no. 6, pp. 347–357, 2014

  2. [2]

    Open data and digital morphology,

    T. G. Davies, I. A. Rahman, S. Lautenschlager, J. A. Cunningham, R. J. Asher, P. M. Barrett, K. T. Bates, S. Bengtson, R. B. J. Benson, D. M. Boyer, J. Braga, J. A. Bright, L. P. A. M. Claessens, P. G. Cox, X.-P. Dong, A. R. Evans, P. L. Falkingham, M. Friedman, R. J. Garwood, A. Goswami, J. R. Hutchinson, N. S. Jeffery, Z. Johanson, R. Lebrun, C. Martíne...

  3. [3]

    Evolution and development of shape: integrating quantitative approaches,

    C. P. Klingenberg, “Evolution and development of shape: integrating quantitative approaches,”Nature Reviews Genetics, vol. 11, no. 9, pp. 623–635, 2010

  4. [4]

    Micro-computed tomography for natural history specimens: a handbook of best practice protocols,

    K. Keklikoglou, S. Faulwetter, E. Chatzinikolaou, P. Wils, J. Brecko, J. Kvaček, B. Metscher, and C. Arvanitidis, “Micro-computed tomography for natural history specimens: a handbook of best practice protocols,”European Journal of Taxonomy, no. 522, pp. 1–55, 2019. 20

  5. [5]

    A practical guide to sliding and surface semilandmarks in morphometric analyses,

    C. Bardua, R. N. Felice, A. Watanabe, A. C. Fabre, and A. Goswami, “A practical guide to sliding and surface semilandmarks in morphometric analyses,”Integrative Organismal Biology, vol. 1, no. 1, pp. 1–34, 2019

  6. [6]

    How many specimens make a sufficient training set for automated three-dimensional feature extraction?,

    J. M. Mulqueeney, A. Searle-Barnes, A. Brombacher, M. Sweeney, A. Goswami, and T. H. G. Ezard, “How many specimens make a sufficient training set for automated three-dimensional feature extraction?,” Royal Society Open Science, vol. 11, no. 6, p. rsos.240113, 2024

  7. [7]

    Auto-Encoding Variational Bayes

    D. P. Kingma and M. Welling, “Auto-encoding variational Bayes,”arXiv preprint arXiv:1312.6114, 2013

  8. [8]

    Generative adversarial networks,

    I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,”Communications of the ACM, vol. 63, no. 11, pp. 139–144, 2020

  9. [9]

    Denoising diffusion probabilistic models,

    J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,”Advances in Neural Information Processing Systems, vol. 33, pp. 6840–6851, 2020

  10. [10]

    Flow matching for generative modeling,

    Y. Lipman, R. T. Chen, H. Ben-Hamu, M. Nickel, and M. Le, “Flow matching for generative modeling,” in11th International Conference on Learning Representations (ICLR 2023), 2023

  11. [11]

    On memorization in probabilistic deep generative models,

    G. van den Burg and C. Williams, “On memorization in probabilistic deep generative models,”Advances in Neural Information Processing Systems, vol. 34, pp. 27916–27928, 2021

  12. [12]

    Understanding deep learning (still) requires rethinking generalization,

    C. Zhang, S. Bengio, M. Hardt, B. Recht, and O. Vinyals, “Understanding deep learning (still) requires rethinking generalization,”Communications of the ACM, vol. 64, no. 3, pp. 107–115, 2021

  13. [13]

    Phylogenies and the comparative method,

    J. Felsenstein, “Phylogenies and the comparative method,”The American Naturalist, vol. 125, no. 1, pp. 1–15, 1985

  14. [14]

    Understanding posterior collapse in generative latent variable models,

    J. Lucas, G. Tucker, R. Grosse, and M. Norouzi, “Understanding posterior collapse in generative latent variable models,” inICLR 2019 Workshop DeepGenStruct, 2019

  15. [15]

    Effectively unbiased FID and inception score and where to find them,

    M. J. Chong and D. Forsyth, “Effectively unbiased FID and inception score and where to find them,” in2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6069–6078, 2020

  16. [16]

    DeepSDF: Learning continuous signed distance functions for shape representation,

    J. J. Park, P. Florence, J. Straub, R. Newcombe, and S. Lovegrove, “DeepSDF: Learning continuous signed distance functions for shape representation,” in2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 165–174, 2019

  17. [17]

    Computing the Fréchet distance between two polygonal curves,

    H. Alt and M. Godau, “Computing the Fréchet distance between two polygonal curves,”International Journal of Computational Geometry & Applications, vol. 5, no. 1–2, pp. 75–91, 1995

  18. [18]

    Occupancy networks: Learning 3D reconstruction in function space,

    L. Mescheder, M. Oechsle, M. Niemeyer, S. Nowozin, and A. Geiger, “Occupancy networks: Learning 3D reconstruction in function space,” in2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4455–4465, 2019

  19. [19]

    Learning implicit fields for generative shape modeling,

    Z. Chen and H. Zhang, “Learning implicit fields for generative shape modeling,” in2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (Los Alamitos, CA, USA), pp. 5932– 5941, IEEE Computer Society, June 2019

  20. [20]

    Neural unsigned distance fields for implicit function learning,

    J. Chibane, A. Mir, and G. Pons-Moll, “Neural unsigned distance fields for implicit function learning,” inProceedings of the 34th International Conference on Neural Information Processing Systems, (Red Hook, NY, USA), pp. 21638–21652, Curran Associates Inc., 2020

  21. [21]

    Differentiable volumetric rendering: Learning implicit 3D representations without 3D supervision,

    M. Niemeyer, L. Mescheder, M. Oechsle, and A. Geiger, “Differentiable volumetric rendering: Learning implicit 3D representations without 3D supervision,” in2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3501–3512, 2020

  22. [22]

    Learning representations and generative models for 3D point clouds,

    P. Achlioptas, O. Diamanti, I. Mitliagkas, and L. Guibas, “Learning representations and generative models for 3D point clouds,” inInternational Conference on Machine Learning, pp. 40–49, PMLR, 2018

  23. [23]

    PointFlow: 3D point cloud generation with continuous normalizing flows,

    G. Yang, X. Huang, Z. Hao, M.-Y. Liu, S. Belongie, and B. Hariharan, “PointFlow: 3D point cloud generation with continuous normalizing flows,” in2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4540–4549, 2019

  24. [24]

    VoxNet: A3Dconvolutionalneuralnetworkforreal-timeobjectrecognition,

    D.MaturanaandS.Scherer, “VoxNet: A3Dconvolutionalneuralnetworkforreal-timeobjectrecognition,” in2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 922–928, 2015

  25. [25]

    PolyGen: An autoregressive generative model of 3D meshes,

    C. Nash, Y. Ganin, S. M. A. Eslami, and P. Battaglia, “PolyGen: An autoregressive generative model of 3D meshes,” inProceedings of the 37th International Conference on Machine Learning(H. D. III and A. Singh, eds.), vol. 119 ofProceedings of Machine Learning Research, pp. 7220–7229, PMLR, 2020. 21

  26. [26]

    DualSDF: Semantic shape manipulation using a two-level representation,

    Z. Hao, H. Averbuch-Elor, N. Snavely, and S. Belongie, “DualSDF: Semantic shape manipulation using a two-level representation,” in2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), p. 7628–7638, IEEE, 2020

  27. [27]

    LION: Latent point diffusion models for 3D shape generation,

    A. Vahdat, F. Williams, Z. Gojcic, O. Litany, S. Fidler, K. Kreis,et al., “LION: Latent point diffusion models for 3D shape generation,”Advances in Neural Information Processing Systems, vol. 35, pp. 10021– 10039, 2022

  28. [28]

    Phylogenetic principal components analysis and geometric morphometrics,

    P. D. Polly, A. M. Lawing, A.-C. Fabre, and A. Goswami, “Phylogenetic principal components analysis and geometric morphometrics,”Hystrix, the Italian Journal of Mammalogy, vol. 24, pp. 33–41, Apr. 2013

  29. [29]

    A registration and deep learning approach to automated landmark detection for geometric morphometrics,

    J. Devine, J. D. Aponte, D. C. Katz, W. Liu, L. D. L. Vercio, N. D. Forkert, R. Marcucio, C. J. Percival, and B. Hallgrímsson, “A registration and deep learning approach to automated landmark detection for geometric morphometrics,”Evolutionary Biology, vol. 47, no. 3, pp. 246–259, 2020

  30. [30]

    The phylogenetic effective sample size and jumps,

    K. Bartoszek, “The phylogenetic effective sample size and jumps,”Mathematica Applicanda, vol. 46, July 2018

  31. [31]

    Phylogenies and community ecology,

    C. O. Webb, D. D. Ackerly, M. A. McPeek, and M. J. Donoghue, “Phylogenies and community ecology,” Annual Review of Ecology and Systematics, vol. 33, no. 1, pp. 475–505, 2002

  32. [32]

    Inferring the historical patterns of biological evolution,

    M. Pagel, “Inferring the historical patterns of biological evolution,”Nature, vol. 401, no. 6756, pp. 877–884, 1999

  33. [33]

    Implicit geometric regularization for learning shapes,

    A. Gropp, L. Yariv, N. Haim, M. Atzmon, and Y. Lipman, “Implicit geometric regularization for learning shapes,” inProceedings of the 37th International Conference on Machine Learning (ICML), pp. 3789–3799, 2020

  34. [34]

    Level set methods and dynamic implicit surfaces,

    S. Osher, R. Fedkiw, and K. Piechor, “Level set methods and dynamic implicit surfaces,”Applied Mechanics Reviews, vol. 57, no. 3, p. B15, 2004

  35. [35]

    Marching cubes: A high resolution 3D surface construction algorithm,

    W. E. Lorensen and H. E. Cline, “Marching cubes: A high resolution 3D surface construction algorithm,” inSeminal graphics: Pioneering efforts that shaped the field, pp. 347–353, 1998

  36. [36]

    Phylogenetics and diversification of tanagers (Passeriformes: Thraupidae), the largest radiation of Neotropical songbirds,

    K. J. Burns, A. J. Shultz, P. O. Title, N. A. Mason, F. K. Barker, J. Klicka, S. M. Lanyon, and I. J. Lovette, “Phylogenetics and diversification of tanagers (Passeriformes: Thraupidae), the largest radiation of Neotropical songbirds,”Molecular Phylogenetics and Evolution, vol. 75, pp. 41–77, 2014

  37. [37]

    Phylogenetic relationships and morphological diversity in Darwin’s finches and their relatives,

    K. J. Burns, S. J. Hackett, and N. K. Klein, “Phylogenetic relationships and morphological diversity in Darwin’s finches and their relatives,”Evolution, vol. 56, no. 6, pp. 1240–1252, 2002

  38. [38]

    Building normalizing flows with stochastic interpolants,

    M. S. Albergo and E. Vanden-Eijnden, “Building normalizing flows with stochastic interpolants,” in International Conference on Learning Representations (ICLR), 2023

  39. [39]

    Improvingandgeneralizingflow-basedgenerativemodelswithminibatchoptimaltransport,

    A. Tong, K. Fatras, N. Malkin, G. Huguet, Y. Zhang, J. Rector-Brooks, G. Wolf, and Y. Bengio, “Improvingandgeneralizingflow-basedgenerativemodelswithminibatchoptimaltransport,”Transactions on Machine Learning Research, pp. 1–34, 2024

  40. [40]

    The Fréchet distance between multivariate normal distributions,

    D. C. Dowson and B. V. Landau, “The Fréchet distance between multivariate normal distributions,” Journal of Multivariate Analysis, vol. 12, no. 3, pp. 450–455, 1982

  41. [41]

    Cranial shape evolution in adaptive radiations of birds: comparative morphometrics of Darwin’s finches and Hawaiian honeycreepers,

    M. Tokita, W. Yano, H. F. James, and A. Abzhanov, “Cranial shape evolution in adaptive radiations of birds: comparative morphometrics of Darwin’s finches and Hawaiian honeycreepers,”Philosophical Transactions of the Royal Society B: Biological Sciences, vol. 372, no. 1713, p. 20150481, 2017

  42. [42]

    Geometry and dynamics link form, function, and evolution of finch beaks,

    S. Al-Mosleh, G. P. T. Choi, A. Abzhanov, and L. Mahadevan, “Geometry and dynamics link form, function, and evolution of finch beaks,”Proceedings of the National Academy of Sciences, vol. 118, no. 46, p. e2105957118, 2021

  43. [43]

    Beak morphometry and morphogenesis across avian radiations,

    S. Mosleh, G. P. T. Choi, G. M. Musser, H. F. James, A. Abzhanov, and L. Mahadevan, “Beak morphometry and morphogenesis across avian radiations,”Proceedings of the Royal Society B: Biological Sciences, vol. 290, no. 2007, p. 20230420, 2023

  44. [44]

    Adam: A method for stochastic optimization,

    D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” inInternational Conference on Learning Representations (ICLR), 2015

  45. [45]

    Towards understanding convergence and generalization of AdamW,

    P. Zhou, X. Xie, Z. Lin, and S. Yan, “Towards understanding convergence and generalization of AdamW,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 46, no. 9, pp. 6486–6493, 2024

  46. [46]

    Rethinking network design and local geometry in point cloud: A simple residual MLP framework,

    X. Ma, C. Qin, H. You, H. Ran, and Y. Fu, “Rethinking network design and local geometry in point cloud: A simple residual MLP framework,” inInternational Conference on Learning Representations (ICLR), 2022. 22

  47. [47]

    Gaussian Error Linear Units (GELUs)

    D. Hendrycks and K. Gimpel, “Gaussian error linear units (GELUs),”arXiv preprint arXiv:1606.08415, 2016

  48. [48]

    GANs trained by a two time-scale update rule converge to a local Nash equilibrium,

    M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “GANs trained by a two time-scale update rule converge to a local Nash equilibrium,”Advances in Neural Information Processing Systems, vol. 30, 2017

  49. [49]

    Efficient Score Pre-computation for Diffusion Models via Cross-Matrix Krylov Projection

    K. Lau, A. S. Na, and J. W. L. Wan, “Efficient score pre-computation for diffusion models via cross-matrix Krylov projection,”arXiv preprint arXiv:2511.17634, 2025

  50. [50]

    Denoising diffusion implicit models,

    J. Song, C. Meng, and S. Ermon, “Denoising diffusion implicit models,” inInternational Conference on Learning Representations (ICLR), 2021

  51. [51]

    Sexual dimorphism,

    D. W. Frayer and M. H. Wolpoff, “Sexual dimorphism,”Annual Review of Anthropology, vol. 14, pp. 429–473, 1985

  52. [52]

    Exogenous dna loading into extracellular vesicles via electroporation is size-dependent and enables limited gene delivery,

    T. N. Lamichhane, R. S. Raiker, and S. M. Jay, “Exogenous dna loading into extracellular vesicles via electroporation is size-dependent and enables limited gene delivery,”Molecular Pharmaceutics, vol. 12, no. 10, p. 3650–3657, 2015. 23 Supplementary Information Layer Input dim Output dim Notes Linear 164 + 3 = 67512 Weight norm + ReLU Linear 2–4 512 512 W...