pith. machine review for the scientific record. sign in

arxiv: 2604.25439 · v1 · submitted 2026-04-28 · ❄️ cond-mat.dis-nn · cond-mat.soft

Recognition: unknown

Training cell stress patterns in 3D cellular packings

Authors on Pith no claims yet

Pith reviewed 2026-05-07 13:52 UTC · model grok-4.3

classification ❄️ cond-mat.dis-nn cond-mat.soft
keywords cellular packings3D vertex modelcontrastive learningcell stress patternstissue rigiditycell rearrangementsmechanical learning
0
0 comments X

The pith

A 3D vertex model shows cellular packings can learn prescribed stress patterns by updating hidden-cell shape indices with contrastive learning.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes that reconfigurable cellular tissues modeled in three dimensions can acquire specific internal stress distributions through an algorithm that adjusts cell shape parameters, rather than relying on fixed network weights. This process is collective and depends on the tissue's mechanical rigidity, which balances exploration of new cell arrangements against exploitation of current ones, with rearrangements creating jumps between different states in the learning landscape. Learnability slows and grows more variable as more cells must match targets or as constraint loads rise, and sequential training offers a steadier path than simultaneous updates. If the model holds, it indicates that living or engineered cell assemblies could serve as materials that adapt their mechanics internally without separate computation.

Core claim

Using a three-dimensional vertex model, cellular packings can be trained to realize prescribed cell stress patterns through a contrastive learning algorithm to update hidden-cell shape indices. Learning is intrinsically collective, requiring coordinated system-wide parameter adjustments, with learnability governed by an interplay between mechanical state, capacity, and training protocol. The rigidity of the tissue controls an effective exploration-exploitation tradeoff: fluid-like regimes enhance exploration through cellular rearrangements, while rigid regimes constrain dynamics and favor exploitation of existing configurations. These rearrangements introduce discontinuous learning dynamics,

What carries the argument

The contrastive learning algorithm that iteratively updates the shape indices of hidden cells in the 3D vertex model to reduce the mismatch between realized and target stress patterns.

If this is right

  • Learning slows and becomes more heterogeneous with increasing ratio of target cells or constraint load, relying on rare rearrangements to escape constrained states.
  • Sequential training of cells yields more robust outcomes than parallel protocols but requires more time for the loads examined.
  • Tissue rigidity sets an exploration-exploitation tradeoff that produces a phase diagram of learnability controlled by rigidity, load, and protocol.
  • Rearrangements create discontinuous jumps that let the system move between distinct local minima in the cost landscape.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Engineered cell-based materials could be designed to self-organize toward desired internal mechanical states by applying similar update rules.
  • Biological tissues may already exploit rigidity-dependent rearrangements to adapt stress patterns during growth or response to injury.
  • Varying the order of cell updates could be tested as a practical way to improve learning speed in physical analogs of the model.
  • The approach opens the possibility of using cell packings as physical reservoirs that store and process mechanical information.

Load-bearing premise

The 3D vertex model together with contrastive updates on cell shape indices accurately captures how real cellular tissues would adjust to achieve prescribed stress patterns.

What would settle it

An experiment on a real three-dimensional cell aggregate in which cells fail to produce the target stress pattern after shape-index adjustments predicted by the trained model, or succeed under conditions the model says should be impossible.

Figures

Figures reproduced from arXiv: 2604.25439 by J. M. Schwarz, Shabeeb Ameen, Tao Zhang.

Figure 1
Figure 1. Figure 1: FIG. 1 view at source ↗
Figure 2
Figure 2. Figure 2: FIG. 2 view at source ↗
Figure 3
Figure 3. Figure 3: FIG. 3 view at source ↗
Figure 4
Figure 4. Figure 4: FIG. 4 view at source ↗
Figure 2
Figure 2. Figure 2: We train NT = 1, 2, 4 randomly chosen target cells in systems with fixed total cell number, and track the evolution of error, overlap, and hidden-parameter distri￾butions. As NT increases, we observe a systematic shift toward larger numbers of iterations required for conver￾gence ( view at source ↗
Figure 5
Figure 5. Figure 5: FIG. 5 view at source ↗
Figure 6
Figure 6. Figure 6: FIG. 6 view at source ↗
Figure 6
Figure 6. Figure 6: For an intermediate s0(0), in the low-complexity regime, where the number of constraints is small relative to the hidden degrees of freedom, the system exhibits a highly degenerate solution space. In this regime, parallel training is most effective: all constraints can be addressed simultaneously, and the system converges efficiently with￾out requiring extensive exploration. Sequential training, by contras… view at source ↗
read the original abstract

The task of learning patterns is typically associated with systems that update parameters on fixed architectures, such as neural networks, where learning proceeds through continuous optimization. Here, we demonstrate that pattern learning can also emerge in reconfigurable cellular tissue, where both mechanical parameters and network topology evolve. Using a three-dimensional vertex model, we show that cellular packings can be trained to realize prescribed cell stress patterns through a contrastive learning algorithm to update hidden-cell shape indices. We find that learning is intrinsically collective, requiring coordinated, system-wide parameter adjustments, with learnability governed by an interplay between mechanical state, capacity, and training protocol. In particular, the rigidity of the tissue controls an effective exploration-exploitation tradeoff: fluid-like regimes enhance exploration through cellular rearrangements, while rigid regimes constrain dynamics and favor exploitation of existing configurations. These rearrangements introduce discontinuous learning dynamics, enabling the system to transition between distinct local minima in the cost function landscape. As the ratio of target cells to the total number cells in the packing or constraint load increases, learning becomes slower, more heterogeneous, and increasingly dependent on rare rearrangements that allow escape from geometrically constrained states. Finally, training cells in sequence, in contrast to parallel protocols, provides an alternative route that can be more robust but generally takes longer to train for the constraint loads studied. These results suggest a learning phase diagram governed by constraint load, cell packing rigidity, and training protocol. By enabling the training of localized internal states, this work positions tissues not only as adaptive materials, but as nonconventional AI platforms.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript uses a 3D vertex model of cellular packings to demonstrate that prescribed cell stress patterns can be realized by applying a contrastive learning algorithm that updates the shape indices of hidden cells. Learning is shown to be collective and governed by an interplay of mechanical rigidity (controlling exploration via rearrangements versus exploitation), constraint load (ratio of target to total cells), and training protocol (sequential versus parallel), with rearrangements producing discontinuous jumps between cost minima; the authors propose a learning phase diagram in the space of these parameters.

Significance. If the simulation results are quantitatively validated, the work establishes that reconfigurable mechanical networks with evolving topology can perform pattern-learning tasks, extending machine-learning concepts to active matter and suggesting tissues as adaptive or computational materials. The explicit mapping of rigidity to an exploration-exploitation tradeoff and the role of topological rearrangements in escaping local minima are potentially novel contributions.

major comments (2)
  1. [Results (rigidity and rearrangements subsection)] The central claim that rigidity gates an exploration-exploitation tradeoff and produces discontinuous learning dynamics rests on the observed correlation between rearrangement events and cost-function jumps. The manuscript must supply quantitative support (e.g., histograms of rearrangement times versus learning epochs, or the fraction of successful escapes from local minima) in the results section that compares fluid and rigid regimes at fixed constraint load; without these metrics the tradeoff remains qualitative.
  2. [Training protocols comparison] The statement that sequential training is 'more robust but generally takes longer' for the studied constraint loads requires explicit comparison data (learning curves, success rates, and variance across multiple initial conditions) between sequential and parallel protocols at the same target-cell fractions; the current description leaves open whether the robustness advantage is statistically significant or merely anecdotal.
minor comments (2)
  1. [Abstract and Introduction] The term 'hidden-cell shape indices' is introduced without an immediate definition or reference to the vertex-model energy functional; a brief equation or parenthetical clarification on first use would improve readability.
  2. [Figures] Figure captions should explicitly state the number of independent realizations, the precise definition of 'constraint load,' and the color scale or normalization used for stress-pattern visualizations.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their detailed and constructive report. We have addressed the major comments by planning to include additional quantitative data and analyses in the revised manuscript, as outlined in the point-by-point responses below.

read point-by-point responses
  1. Referee: [Results (rigidity and rearrangements subsection)] The central claim that rigidity gates an exploration-exploitation tradeoff and produces discontinuous learning dynamics rests on the observed correlation between rearrangement events and cost-function jumps. The manuscript must supply quantitative support (e.g., histograms of rearrangement times versus learning epochs, or the fraction of successful escapes from local minima) in the results section that compares fluid and rigid regimes at fixed constraint load; without these metrics the tradeoff remains qualitative.

    Authors: We agree with the referee that providing quantitative metrics will make the evidence for the rigidity-controlled exploration-exploitation tradeoff more rigorous. In the revised version, we will add a new figure or panel in the Results section showing histograms of rearrangement event timings relative to learning epochs for fluid and rigid regimes at the same constraint load. We will also compute and report the fraction of successful escapes from local minima facilitated by rearrangements in each regime, based on multiple simulation runs. This will quantify the discontinuous jumps and the difference in dynamics between the regimes. revision: yes

  2. Referee: [Training protocols comparison] The statement that sequential training is 'more robust but generally takes longer' for the studied constraint loads requires explicit comparison data (learning curves, success rates, and variance across multiple initial conditions) between sequential and parallel protocols at the same target-cell fractions; the current description leaves open whether the robustness advantage is statistically significant or merely anecdotal.

    Authors: We acknowledge that the current description of the training protocol comparison could be strengthened with statistical details. In the revision, we will include averaged learning curves with error bars representing variance across multiple independent initial conditions for both sequential and parallel protocols at identical target-cell fractions. We will also tabulate success rates (defined as achieving a cost below a certain threshold) and the average number of training steps required, allowing a direct statistical comparison of robustness and efficiency. revision: yes

Circularity Check

0 steps flagged

Simulation protocol is self-contained with no circular reductions

full rationale

The paper presents a computational demonstration in which a 3D vertex model is driven by an explicitly defined contrastive learning rule that updates hidden-cell shape indices to match target stress patterns. All reported outcomes (collective learning, rigidity-gated rearrangements, effects of sequential vs. parallel training, dependence on constraint load) follow directly from running the stated energy functional, update dynamics, and protocol on the model; none are obtained by fitting a parameter to a subset of the target data and then relabeling the fit as a prediction, nor by importing a uniqueness theorem or ansatz from self-citation that is itself unverified. The derivation chain is therefore a forward simulation whose results are independent of the target patterns once the model and algorithm are fixed.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The approach relies on the vertex model being a valid representation of cellular mechanics and the contrastive algorithm being applicable to parameter updates in evolving topologies. No explicit free parameters or invented entities are detailed in the abstract.

axioms (2)
  • domain assumption The 3D vertex model accurately captures cell mechanics and rearrangements in tissue packings.
    Central to all simulations and learning dynamics described.
  • domain assumption Contrastive learning can be implemented by updating hidden-cell shape indices to minimize a cost function for target stress patterns.
    Basis for the training process.

pith-pipeline@v0.9.0 · 5577 in / 1245 out tokens · 38991 ms · 2026-05-07T13:52:39.372544+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

47 extracted references · 7 canonical work pages · 1 internal anchor

  1. [1]

    Maintenance of this assignment of tar- get shape indices shall be referred to as thefree stateof the system

    Target Cells: For a training task we assignN T target cells with initial target shape index assign- ments{s T 0 }. Maintenance of this assignment of tar- get shape indices shall be referred to as thefree stateof the system. In contrast, aclamped stateis achieved by changing{s T 0 }. The training task seeks to achieve target maximum shear stresses{σ T }on ...

  2. [2]

    Training involves chang- ing these assignments using a learning rule

    Hidden Cells: These cells have initial target shape index assignments{s H 0 }. Training involves chang- ing these assignments using a learning rule. In other words,{s H 0 }constitute the learning, or mod- ifiable, parameters of the system

  3. [3]

    clamp- ing

    Passive Cells: These cells have initial target shape index assignments{s P 0 }, which will remain un- changed in all stages of the training algorithm. That is, while these cells appear in the total energy of the system, they are not part of the training al- gorithm. We note that the set of passive cells may be empty. As we will choose some fraction of cel...

  4. [4]

    Boussard, A

    A. Boussard, A. Fessel, C. Oettmeier, L. Briard, H.- G. D¨ obereiner, and A. Dussutour, Adaptive behaviour and learning in slime moulds: the role of oscillations, Philosophical Transactions of the Royal Society B376, 20190757 (2021)

  5. [5]

    Schick, E

    L. Schick, E. Eichenlaub, F. Drexel, A. Mayer, S. Chen, M. Roper, and K. Alim, Decision-making in light-trapped slime molds involves active mechanical processes, PRX Life , (2026)

  6. [6]

    Rajan, T

    D. Rajan, T. Makushok, A. Kalish, L. Acuna, A. Bonville, K. C. Almanza, B. Garibay, E. Tang, M. Voss, A. Lin,et al., Single-cell analysis of habituation in stentor coeruleus, Current Biology33, 241 (2023)

  7. [7]

    S. J. Gershman, P. E. Balbi, C. R. Gallistel, and J. Gu- nawardena, Reconsidering the evidence for learning in single cells, Elife10, e61907 (2021)

  8. [8]

    D. H. Rajan and W. F. Marshall, A receptor-inactivation model for single-celled habituation in stentor coeruleus, Current Biology35, 3327 (2025)

  9. [9]

    Zentner, E

    A. Zentner, E. V. Halingstad, C. Chalk, M. P. Bren- ner, A. Murugan, E. Winfree, and K. Shrinivas, Informa- tion processing driven by multicomponent surface con- densates, arXiv preprint arXiv:2509.08100 (2025)

  10. [10]

    C. G. Evans, J. O’Brien, E. Winfree, and A. Murugan, Pattern recognition in the nucleation kinetics of non- equilibrium self-assembly, Nature625, 500 (2024)

  11. [11]

    Bahri, J

    Y. Bahri, J. Kadmon, J. Pennington, S. S. Schoenholz, J. Sohl-Dickstein, and S. Ganguli, Statistical mechanics of deep learning, Annual Review of Condensed Matter Physics11, 501 (2020)

  12. [12]

    Gabrie, S

    M. Gabrie, S. Ganguli, C. Lucibello, and R. Zecchina, Neural networks: from the perceptron to deep networks, Journal of Statistical Mechanics: Theory and Experi- ment (2023), arXiv:2304.06636

  13. [13]

    L. G. Wright, T. Onodera, M. M. Stein, T. Wang, D. T. Schachter, Z. Hu, and P. L. McMahon, Deep physical neural networks trained with backpropagation, Nature 601, 549 (2022)

  14. [14]

    Stern and A

    M. Stern and A. Murugan, Learning without neurons in physical systems, Annual Review of Condensed Matter Physics14, 417 (2023)

  15. [15]

    V. R. Anisetti, A. Kandala, and J. Schwarz, Emergent learning in physical systems as feedback-based aging in a glassy landscape, arXiv preprint arXiv:2309.04382 (2023)

  16. [16]

    Momeni, B

    A. Momeni, B. Rahmani, B. Scellier, L. G. Wright, P. L. McMahon, C. C. Wanjura, Y. Li, A. Skalli, N. G. Berloff, T. Onodera,et al., Training of physical neural networks, Nature645, 53 (2025)

  17. [17]

    Scellier and Y

    B. Scellier and Y. Bengio, Equilibrium propagation: Bridging the gap between energy-based models and back- propagation, Frontiers in computational neuroscience11, 24 (2017). 16

  18. [18]

    V. R. Anisetti, B. Scellier, and J. M. Schwarz, Learning by non-interfering feedback chemical signaling in physical networks, Physical Review Research5, 023024 (2023)

  19. [19]

    V. R. Anisetti, A. Kandala, B. Scellier, and J. M. Schwarz, Frequency propagation: Multimechanism learn- ing in nonlinear physical networks, Neural Computation 36, 596 (2024)

  20. [20]

    Li and X

    S. Li and X. Mao, Training all-mechanical neural net- works for task learning through in situ backpropagation, Nature Communications15, 10528 (2024)

  21. [21]

    M. J. Falk, A. T. Strupp, B. Scellier, and A. Muru- gan, Temporal contrastive learning through implicit non- equilibrium memory, Nature Communications16, 2163 (2025)

  22. [22]

    Hexner, Training precise stress patterns, Soft Matter 19, 2120 (2023)

    D. Hexner, Training precise stress patterns, Soft Matter 19, 2120 (2023)

  23. [23]

    D. O. Hebb, The organization of behavior: A neuropsy- chological theory, Wiley (1949)

  24. [24]

    Gopinathan, K.-C

    A. Gopinathan, K.-C. Lee, J. M. Schwarz, and A. J. Liu, Branching, capping, and severing in dynamic actin struc- tures, Physical review letters99, 058103 (2007)

  25. [25]

    D. A. Fletcher and R. D. Mullins, Cell mechanics and the cytoskeleton, Nature463, 485 (2010)

  26. [26]

    W. Guo, V. R. Anisetti, K. Zhang, S. Ameen, A. Kan- dala, M. Stern, N. Pashine, J. D. Paulsen, J. M. Schwarz, and T. Zhang, Learning associations in recon- figurable particle packings via local cyclic driving (2026), arXiv:2603.14534 [cond-mat.dis-nn]

  27. [27]

    Berger-Tal, R

    O. Berger-Tal, R. Nathan, E. Meron, and D. Saltz, The exploration–exploitation dilemma: A multidisciplinary framework, PLoS ONE9, e95693 (2014)

  28. [28]

    Zhang, S

    T. Zhang, S. Gupta, M. Lancaster, and J. M. Schwarz, How human-derived brain organoids are built differently from brain organoids derived from genetically-close rela- tives: A multi-scale hypothesis, Soft Matter (2026)

  29. [29]

    Nasrollahi, C

    S. Nasrollahi, C. Walter, A. J. Loza, G. V. Schimizzi, G. D. Longmore, and A. Pathak, Past matrix stiffness primes epithelial cells and regulates their future collective migration through a mechanical memory, Biomaterials 146, 146 (2017)

  30. [30]

    Honda, M

    H. Honda, M. Tanemura, and T. Nagai, A three- dimensional vertex dynamics cell model of space-filling polyhedra simulating cell behavior in a cell aggregate, Journal of Theoretical Biology226, 439 (2004)

  31. [31]

    Okuda, Y

    S. Okuda, Y. Inoue, M. Eiraku, Y. Sasai, and T. Adachi, Reversible network reconnection model for simulating large deformation in dynamic tissue morphogenesis, Biomechanics and Modeling in Mechanobiology12, 627 (2013)

  32. [32]

    Okuda, Y

    S. Okuda, Y. Inoue, and T. Adachi, Three-dimensional vertex model for simulating multicellular morphogenesis, Biophysics and Physicobiology12, 13 (2015)

  33. [33]

    Zhang and J

    T. Zhang and J. M. Schwarz, Topologically-protected in- terior for three-dimensional confluent cellular collectives, Phys. Rev. Res.4, 043148 (2022)

  34. [34]

    Ameen, K

    S. Ameen, K. Kim, L. Theeyancheri, M. Thanh, M. Wu, A. E. Patteson, J. M. Schwarz, and T. Zhang, Cell strain- stiffening drives cell breakout from embedded spheroids (2026), arXiv:2602.08737 [physics.bio-ph]

  35. [35]

    Bitzek, P

    E. Bitzek, P. Koskinen, F. G¨ ahler, M. Moseler, and P. Gumbsch, Structural relaxation made simple, Phys. Rev. Lett.97, 170201 (2006)

  36. [36]

    Bitzek and P

    E. Bitzek and P. Gumbsch, An efficient algorithm for structural relaxation of atomic systems, Modelling and Simulation in Materials Science and Engineering15, S159 (2007)

  37. [37]

    Geman, E

    S. Geman, E. Bienenstock, and R. Doursat, Neural net- works and the bias/variance dilemma, Neural Computa- tion4, 1 (1992)

  38. [38]

    Hastie, R

    T. Hastie, R. Tibshirani, and J. Friedman,The Elements of Statistical Learning, 2nd ed. (Springer, 2009)

  39. [39]

    M. S. Advani, A. M. Saxe, and H. Sompolinsky, High- dimensional dynamics of generalization error in neural networks, Neural Networks132, 428 (2020)

  40. [40]

    Parker, M

    A. Parker, M. C. Marchetti, M. L. Manning, and J. Schwarz, How does the extracellular matrix affect the rigidity of an embedded spheroid?, Physical Review E 111, 044410 (2025)

  41. [41]

    Parker and J

    A. Parker and J. M. Schwarz, arXiv:2503.14319 (2025)

  42. [42]

    Arzash, I

    S. Arzash, I. Tah, A. J. Liu, and M. L. Manning, Rigidity of epithelial tissues as a double optimization problem, Physical Review Research7, 013157 (2025)

  43. [43]

    Arzash and S

    S. Arzash and S. Banerjee, Learning epithelial elasticity via local tension remodeling, bioRxiv , 2025 (2025)

  44. [44]

    Arzash, A

    S. Arzash, A. J. Liu, and M. L. Manning, Epithelial con- vergent extension as a tuning process, bioRxiv , 2025 (2025)

  45. [45]

    VertAX: a differentiable vertex model for learning epithelial tissue mechanics

    A. Pasqui, J. M. C. Ocana, A. Sinha, M. Perez, F. Del- bary, G. Gosti, M. Miotto, D. Caudo, M. Ernoult, and H. Turlier, Vertax: a differentiable vertex model for learning epithelial tissue mechanics, arXiv preprint arXiv:2604.06896 (2026)

  46. [46]

    D. S. Banerjee, M. J. Falk, M. L. Gardel, A. M. Wal- czak, T. Mora, and S. Vaikuntanathan, Learning via mechanosensitivity and activity in cytoskeletal networks, PRX Life4, 013001 (2026)

  47. [47]

    Shomar, O

    A. Shomar, O. Barak, and N. Brenner, Cancer progres- sion as a learning process, Iscience25(2022). 17 SUPPLEMENT AR Y FIGURES (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) (k) (l) FIG. S1:Reducing geometric frustration enhances learning.The same parameters are used as in Figure 1, with the exception ofk V = 1. The row and column layout logic, as well as color s...