pith. machine review for the scientific record. sign in

arxiv: 2605.07608 · v1 · submitted 2026-05-08 · 🧬 q-bio.QM · q-bio.BM

Recognition: 2 theorem links

· Lean Theorem

GoForth: Language Models for RNA Design under Structure, Sequence, and Coding Constraints

Michael Lindsey

Pith reviewed 2026-05-11 02:11 UTC · model grok-4.3

classification 🧬 q-bio.QM q-bio.BM
keywords RNA designinverse sequence designconditional language modelsgenerative modelsstructure constraintssequence constraintscoding constraintsdesignability
0
0 comments X

The pith

Encoder-decoder models trained on real RNA folds generate sequences satisfying mixed structure, sequence, and coding constraints.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper treats RNA inverse design as a conditional generation problem where sequences are sampled from a distribution given user-specified constraints on folds, fixed bases, and coding restrictions. It introduces GoForth, an encoder-decoder language model trained forward on observed RNA structures rather than on teacher-generated inverse designs. This separation of sequence prior, folding sampler, and oracle enables handling of arbitrary constraint combinations in unspecified regions. Sympathetic readers would care if this yields efficient generation without heavy post-processing for practical biological and engineering applications. The models also produce semantic embeddings and a measure of designability as byproducts.

Core claim

GoForth is a forward-trained RNA language model that conditions on structure, sequence, and coding targets. The basic object is a conditional law over RNA sequences given a user-specified condition, with full inverse folding as a special case. We train encoder-decoder models on witnessed folds rather than on outputs from an inverse-design teacher and validate our methodology on full inverse-folding benchmarks, as well as tasks involving constraints on structure, sequence, and coding. The resulting models achieve fast and high-quality candidate generation for mixed RNA design specifications. Moreover they furnish useful semantic embeddings of design tasks and a robust learned notion of design

What carries the argument

The GoForth conditional encoder-decoder model, trained on witnessed folds to learn distributions over sequences under combined structure, sequence, and coding constraints.

If this is right

  • Fast and high-quality candidate generation for mixed RNA design specifications.
  • Useful semantic embeddings of design tasks.
  • A robust learned notion of designability.
  • Effective performance on full inverse-folding benchmarks and constrained tasks.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The method could scale to larger RNA molecules or other biopolymers by expanding the training data.
  • Learned embeddings might support automated design space exploration or transfer learning across different constraint types.
  • Integration with experimental validation loops could use the designability score to guide iterations.

Load-bearing premise

That models trained on observed RNA folds generalize to generate valid sequences for any arbitrary combination of structure, sequence, and coding constraints without additional filtering or retraining.

What would settle it

Finding that on a benchmark of mixed constraints, the generated sequences violate the specified conditions at rates comparable to random sampling or require extensive filtering to achieve high quality.

Figures

Figures reproduced from arXiv: 2605.07608 by Michael Lindsey.

Figure 1
Figure 1. Figure 1: Best-of-100 ViennaRNA MFE layouts for the L = 450 two-hairpin demo. The specified structure is small relative to the sequence length. Blue and green denote constraints for left- and right-pairing, while yellow denotes unpaired constraint. 6 [PITH_FULL_IMAGE:figures/full_fig_p006_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Nearest semantic neighbors from disparate dataset chunks under the centered pooled em [PITH_FULL_IMAGE:figures/full_fig_p007_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Procedurally generated adversarial structure tuned toward many small loops. The left [PITH_FULL_IMAGE:figures/full_fig_p008_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Training and GRPO convergence curves for the principal GoForth-FS, GoForth-PSB, and [PITH_FULL_IMAGE:figures/full_fig_p012_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Illustrative length extrapolation diagnostics. Left: best-of- [PITH_FULL_IMAGE:figures/full_fig_p014_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Representative long-chain target and recovered MFE structure. Exact pair recovery is poor [PITH_FULL_IMAGE:figures/full_fig_p014_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Best-of-100 neural generation time as a function of target length on the full-structure benchmark. Points are mean per-target generation times within length bins. 14 [PITH_FULL_IMAGE:figures/full_fig_p014_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Predicted versus observed full-structure difficulty for GoForth-PSB, using centered and [PITH_FULL_IMAGE:figures/full_fig_p016_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Dataset-level similarity from centered, normalized GoForth condition embeddings, or [PITH_FULL_IMAGE:figures/full_fig_p016_9.png] view at source ↗
read the original abstract

RNA inverse sequence design has broad biological and engineering applications, but computational methods for practical design queries remain limited. Such queries may impose several constraints at once, including target folds or motifs, fixed bases, and coding restrictions, while leaving arbitrary sequence and structure in unspecified regions. Because these constraints may permit many acceptable sequences, we study RNA design as a conditional generative modeling problem. The basic object is a conditional law over RNA sequences given a user-specified condition, with full inverse folding as a special case. We introduce GoForth, a forward-trained RNA language model that conditions on structure, sequence, and coding targets. The formulation separates three ingredients that are often entangled in RNA design: a sequence prior, a forward folding sampler, and a reward or likelihood oracle. We train encoder-decoder models on witnessed folds rather than on outputs from an inverse-design teacher and validate our methodology on full inverse-folding benchmarks, as well as tasks involving constraints on structure, sequence, and coding. The resulting models achieve fast and high-quality candidate generation for mixed RNA design specifications. Moreover they furnish useful semantic embeddings of design tasks and a robust learned notion of designability.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces GoForth, a forward-trained encoder-decoder RNA language model for conditional sequence generation under user-specified combinations of structure, sequence, and coding constraints. It frames RNA inverse design as conditional generative modeling, separates the sequence prior, forward folding sampler, and reward/likelihood oracle, trains exclusively on witnessed natural folds rather than inverse-design teacher outputs, and reports validation on standard inverse-folding benchmarks plus mixed-constraint tasks, claiming fast high-quality candidate generation together with useful semantic embeddings of design tasks and a learned notion of designability.

Significance. If the central claims hold, the separation of prior/sampler/oracle components and training on witnessed folds would provide a clean, extensible framework for RNA design that avoids entanglement common in prior methods and supplies independent grounding. The resulting embeddings and designability measure could support downstream biological applications, while the fast generation under mixed constraints addresses a practical gap in handling arbitrary user-specified specifications.

major comments (2)
  1. [Abstract] Abstract and results sections: the central claim that encoder-decoder models trained only on witnessed folds directly produce valid sequences for arbitrary mixed constraints (without post-hoc filtering or retraining) is load-bearing yet unsupported by any reported quantitative metrics, ablation studies, or explicit verification that all constraints are satisfied simultaneously on out-of-distribution combinations; the abstract states validation occurs but provides no numbers, error analysis, or evidence isolating generalization from implicit rejection sampling.
  2. [Methods] Methods and evaluation: the separation into prior, sampler, and oracle is asserted as enabling the result, but no explicit description or experiment demonstrates that the learned conditional law covers user-specified constraint combinations outside the training distribution without requiring additional filtering steps or oracle calls at inference time.
minor comments (2)
  1. [Abstract] The abstract mentions 'full inverse-folding benchmarks' and 'tasks involving constraints on structure, sequence, and coding' but does not name the specific benchmarks or datasets used, which would aid reproducibility.
  2. [Introduction] Notation for the conditional law and the three separated ingredients (prior, sampler, oracle) is introduced but not formalized with equations or pseudocode, making the claimed separation harder to follow.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their constructive and detailed comments, which have identified important opportunities to strengthen the clarity and evidentiary support in our manuscript. We address each major comment point by point below, indicating the revisions we will undertake.

read point-by-point responses
  1. Referee: [Abstract] Abstract and results sections: the central claim that encoder-decoder models trained only on witnessed folds directly produce valid sequences for arbitrary mixed constraints (without post-hoc filtering or retraining) is load-bearing yet unsupported by any reported quantitative metrics, ablation studies, or explicit verification that all constraints are satisfied simultaneously on out-of-distribution combinations; the abstract states validation occurs but provides no numbers, error analysis, or evidence isolating generalization from implicit rejection sampling.

    Authors: We agree that the abstract would be improved by including specific quantitative metrics. In the revised manuscript we will update the abstract to report key figures from our experiments, including success rates for simultaneous satisfaction of structure, sequence, and coding constraints on mixed tasks. The results sections already contain quantitative evaluations on standard inverse-folding benchmarks (sequence recovery and structure accuracy) and mixed-constraint tasks (constraint satisfaction rates verified by external folding and coding checks). We will add an ablation comparing training on witnessed folds versus teacher-generated inverse-design data and expand the error analysis with a breakdown of per-constraint and joint satisfaction rates on held-out combinations that were not observed during training. These held-out sets serve as out-of-distribution test cases, and satisfaction is measured independently of the model likelihood, thereby isolating generalization from any implicit rejection sampling. revision: yes

  2. Referee: [Methods] Methods and evaluation: the separation into prior, sampler, and oracle is asserted as enabling the result, but no explicit description or experiment demonstrates that the learned conditional law covers user-specified constraint combinations outside the training distribution without requiring additional filtering steps or oracle calls at inference time.

    Authors: Section 3 of the manuscript describes the encoder-decoder architecture that receives the user-specified combination of structure, sequence, and coding constraints as input and is trained to model the conditional distribution over sequences given those constraints. Training uses only natural witnessed folds paired with their sequences; no inverse-design teacher outputs are involved. At inference the model directly samples from this conditional distribution by conditioning on the encoded constraints, with no filtering or oracle calls performed during generation. To address the referee's concern we will add a dedicated subsection in Methods that walks through the inference procedure step by step and will include a new experiment that generates sequences for novel constraint combinations absent from the training distribution, reporting the fraction of outputs that satisfy all constraints simultaneously as verified by independent external tools. This will explicitly demonstrate coverage of out-of-distribution combinations without post-hoc steps. revision: partial

Circularity Check

0 steps flagged

No significant circularity; training on external witnessed folds supplies independent grounding

full rationale

The paper trains encoder-decoder models directly on witnessed (natural) folds as the data source and evaluates generated candidates on standard inverse-folding benchmarks plus constraint tasks. No equations, parameters, or central claims are shown to reduce by construction to quantities fitted from the target result itself. The separation into prior, sampler, and oracle is presented as a modeling choice with external data grounding rather than a self-referential loop. No self-citations are invoked as load-bearing uniqueness theorems or ansatzes. This is the standard non-circular pattern for a conditional generative modeling paper.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

Abstract-only review supplies no explicit free parameters, axioms, or invented entities beyond the modeling framework itself.

axioms (1)
  • domain assumption RNA sequences admit a conditional generative distribution given partial structure, sequence, and coding constraints
    Foundational premise of the conditional modeling approach described in the abstract.

pith-pipeline@v0.9.0 · 5500 in / 1190 out tokens · 45124 ms · 2026-05-11T02:11:17.451784+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

71 extracted references · 71 canonical work pages · 3 internal anchors

  1. [1]

    Designing

    Gautam, Milan and Dai, Ning and Zhou, Tianshuo and Xie, Bowen and Mathews, David and Huang, Liang , year =. Designing. 2602.12470 , archivePrefix =

  2. [2]

    Advances in Neural Information Processing Systems , volume =

    Attention Is All You Need , author =. Advances in Neural Information Processing Systems , volume =. 2017 , url =

  3. [3]

    Shao, Zhihong and Wang, Peiyi and Zhu, Qihao and Xu, Runxin and Song, Junxiao and Bi, Xiao and Zhang, Haowei and Zhang, Mingchuan and Li, Y. K. and Wu, Y. and Guo, Daya , year =. 2402.03300 , archivePrefix =

  4. [4]

    2022 , doi =

    Adamczyk, Bartosz and Antczak, Maciej and Szachniuk, Marta , journal =. 2022 , doi =

  5. [5]

    2018 , doi =

    Danaee, Padideh and Rouches, Mason and Wiley, Michelle and Deng, Dezhong and Huang, Liang and Hendrix, David , journal =. 2018 , doi =

  6. [6]

    Kalvari, Ioanna and Nawrocki, Eric P. and Ontiveros-Palacios, Nancy and Argasinska, Joanna and Lamkiewicz, Kevin and Marz, Manja and Griffiths-Jones, Sam and Toffano-Nioche, Claire and Gautheret, Daniel and Weinberg, Zasha and Rivas, Elena and Eddy, Sean R. and Finn, Robert D. and Bateman, Alex and Petrov, Anton I. , journal =. 2021 , doi =

  7. [7]

    2407.19838 , archivePrefix =

    Gao, Letian and Lu, Zhi John , year =. 2407.19838 , archivePrefix =

  8. [8]

    2024 , url =

    Cole, Jack and Li, Fan and Wu, Liwen and Li, Ke , booktitle =. 2024 , url =

  9. [9]

    and Honer zu Siederdissen, Christian and Tafer, Hakim and Flamm, Christoph and Stadler, Peter F

    Lorenz, Ronny and Bernhart, Stephan H. and Honer zu Siederdissen, Christian and Tafer, Hakim and Flamm, Christoph and Stadler, Peter F. and Hofacker, Ivo L. , journal =. 2011 , doi =

  10. [10]

    and Steenberg, Conrad D

    Zadeh, Joseph N. and Steenberg, Conrad D. and Bois, Justin S. and Wolfe, Brian R. and Pierce, Marshall B. and Khan, Amardeep R. and Dirks, Robert M. and Pierce, Niles A. , journal =. 2011 , doi =

  11. [11]

    , journal =

    Eastman, Peter and Shi, Jade and Ramsundar, Bharath and Pande, Vijay S. , journal =. Solving the. 2018 , doi =

  12. [12]

    Learning to Design

    Runge, Frederic and Stoll, Danny and Falkner, Stefan and Hutter, Frank , booktitle =. Learning to Design. 2019 , url =

  13. [13]

    and Keep, Benjamin and Coppess, Katherine R

    Koodli, Rohan V. and Keep, Benjamin and Coppess, Katherine R. and Portela, Fernando and. PLOS Computational Biology , volume =. 2019 , doi =

  14. [14]

    and Mechery, Vinodh and Meyer, Michelle M

    Dotu, Ivan and Garcia-Martin, Juan Antonio and Slinger, Betty L. and Mechery, Vinodh and Meyer, Michelle M. and Clote, Peter , journal =. Complete. 2014 , doi =

  15. [15]

    2024 , doi =

    Huang, Han and Lin, Ziqian and He, Dongchen and Hong, Liang and Li, Yu , journal =. 2024 , doi =

  16. [16]

    2015 , doi =

    Kleinkauf, Robert and Mann, Martin and Backofen, Rolf , journal =. 2015 , doi =

  17. [17]

    and Huang, Liang , booktitle =

    Zhou, Tianshuo and Tang, Wei Yu and Mathews, David H. and Huang, Liang , booktitle =. Undesignable. 2024 , publisher =

  18. [18]

    and Huang, Liang , year =

    Zhou, Tianshuo and Mathews, David H. and Huang, Liang , year =. Fast and Versatile. 2603.02283 , archivePrefix =

  19. [19]

    Interpretable

    Chen, Jiayang and Hu, Zhihang and Sun, Siqi and Tan, Qingxiong and Wang, Yixuan and Yu, Qinze and Zong, Licheng and Hong, Liang and Xiao, Jin and Shen, Tao and King, Irwin and Li, Yu , year =. Interpretable. 2204.00300 , archivePrefix =

  20. [20]

    Nature Communications , volume =

    Peni. Nature Communications , volume =. 2025 , doi =

  21. [21]

    , year =

    Tan, Cheng and Zhang, Yijie and Gao, Zhangyang and Hu, Bozhen and Li, Siyuan and Liu, Zicheng and Li, Stan Z. , year =. 2301.10774 , archivePrefix =

  22. [22]

    and Jamasb, Arian R

    Joshi, Chaitanya K. and Jamasb, Arian R. and Vi. 2025 , eprint =

  23. [23]

    Journal of Computational Chemistry , volume =

    Nucleic Acid Sequence Design via Efficient Ensemble Defect Optimization , author =. Journal of Computational Chemistry , volume =. 2011 , doi =

  24. [24]

    Journal of the American Chemical Society , volume =

    Constrained Multistate Sequence Design for Nucleic Acid Reaction Pathway Engineering , author =. Journal of the American Chemical Society , volume =. 2017 , doi =

  25. [25]

    and Hutter, Frank and Hoos, Holger H

    Andronescu, Mirela and Fejes, Anthony P. and Hutter, Frank and Hoos, Holger H. and Condon, Anne , journal =. A New Algorithm for. 2004 , doi =

  26. [26]

    2006 , doi =

    Busch, Anke and Backofen, Rolf , journal =. 2006 , doi =

  27. [27]

    2013 , doi =

    Garcia-Martin, Juan Antonio and Clote, Peter and Dotu, Ivan , journal =. 2013 , doi =

  28. [28]

    2011 , doi =

    Taneda, Akito , journal =. 2011 , doi =

  29. [29]

    BMC Bioinformatics , volume =

    Lyngs. BMC Bioinformatics , volume =. 2012 , doi =

  30. [30]

    A Weighted Sampling Algorithm for the Design of

    Reinharz, Vladimir and Ponty, Yann and Waldisp. A Weighted Sampling Algorithm for the Design of. Bioinformatics , volume =. 2013 , doi =

  31. [31]

    2017 , doi =

    Yang, Xiufeng and Yoshizoe, Kazuki and Taneda, Akito and Tsuda, Koji , journal =. 2017 , doi =

  32. [32]

    An Unexpectedly Effective Monte Carlo Technique for the

    Portela, Fernando , year =. An Unexpectedly Effective Monte Carlo Technique for the

  33. [33]

    Solving the

    Rubio-Largo,. Solving the. Applied Soft Computing , volume =. 2023 , doi =

  34. [34]

    and Lach, Grzegorz and Badepally, Nagendar Goud and Moafinejad, S

    Wirecki, Tomasz K. and Lach, Grzegorz and Badepally, Nagendar Goud and Moafinejad, S. Naeim and Jaryani, Farhang and Klaudel, Gaja and Nec, Kalina and Baulin, Eugene F. and Bujnicki, Janusz M. , journal =. 2025 , doi =

  35. [35]

    and Huang, Liang , journal =

    Zhou, Tianshuo and Dai, Ning and Li, Sizhen and Ward, Max and Mathews, David H. and Huang, Liang , journal =. 2023 , doi =

  36. [36]

    and Huang, Liang , journal =

    Tang, Wei Yu and Dai, Ning and Zhou, Tianshuo and Mathews, David H. and Huang, Liang , journal =. 2026 , doi =

  37. [37]

    Principles for Predicting

    Anderson-Lee, Jeff and Fisker, Eli and Kosaraju, Vineet and Wu, Michelle and Kong, Justin and Lee, Jeehyung and Lee, Minjae and Zada, Mathew and Treuille, Adrien and Das, Rhiju , journal =. Principles for Predicting. 2016 , doi =

  38. [38]

    Design of

    Churkin, Alexander and Retwitzer, Matan Drory and Reinharz, Vladimir and Ponty, Yann and Waldisp. Design of. Briefings in Bioinformatics , volume =. 2018 , doi =

  39. [39]

    Fitness Functions for

    Ward, Max and Courtney, Eliot and Rivas, Elena , journal =. Fitness Functions for. 2023 , doi =

  40. [40]

    and Huang, Liang , year =

    Zhou, Tianshuo and Tang, Wei Yu and Malik, Apoorv and Mathews, David H. and Huang, Liang , year =. Scalable and Interpretable Identification of Minimal Undesignable. 2402.17206 , archivePrefix =

  41. [41]

    Probabilistic RNA Designability via Interpretable Ensemble Approximation and Dynamic Decomposition

    Zhou, Tianshuo and Mathews, David H. and Huang, Liang , year =. Probabilistic. 2602.13610 , archivePrefix =

  42. [42]

    and Rudolfs, Boris and Romano, Jonathan and Wayment-Steele, Hannah K

    Koodli, Rohan V. and Rudolfs, Boris and Romano, Jonathan and Wayment-Steele, Hannah K. and Dunlap, William A. and. Redesigning the. 2021 , doi =

  43. [43]

    2024 , publisher =

    Runge, Frederic and Farid, Karim and Franke, J. 2024 , publisher =

  44. [44]

    Runge, Frederic and Franke, J. Partial. Bioinformatics , volume =. 2024 , doi =

  45. [45]

    and Stadler, Peter F

    Hammer, Stefan and Tschiatschek, Birgit and Flamm, Christoph and Hofacker, Ivo L. and Stadler, Peter F. and Will, Sebastian , journal =. 2017 , doi =

  46. [46]

    Fixed-Parameter Tractable Sampling for

    Hammer, Stefan and Wang, Wei and Will, Sebastian and Ponty, Yann , journal =. Fixed-Parameter Tractable Sampling for. 2019 , doi =

  47. [47]

    , year =

    Shi, Jie and Das, Rhiju and Pande, Vijay S. , year =. 1803.03146 , archivePrefix =

  48. [48]

    Designing

    Obonyo, Stephen and Jouandeau, Nicolas and Owuor, Dickson , booktitle =. Designing. 2022 , publisher =

  49. [49]

    2025 , doi =

    Li, Chang and Pan, Liwei and Shen, Jianyang , journal =. 2025 , doi =

  50. [50]

    Informative

    Akiyama, Manato and Sakakibara, Yasubumi , journal =. Informative. 2022 , doi =

  51. [51]

    Multiple Sequence Alignment-Based

    Zhang, Yiqun and Lang, Mingyang and Jiang, Wenjun and others , journal =. Multiple Sequence Alignment-Based. 2024 , doi =

  52. [52]

    Deep Generative Design of

    Sumi, Shunsuke and Hamada, Michiaki and Saito, Hirohide , journal =. Deep Generative Design of. 2024 , doi =

  53. [53]

    2024 , doi =

    Zhao, Yuwen and Oono, Kentaro and Takizawa, Hiroaki and Kotera, Masaaki , journal =. 2024 , doi =

  54. [54]

    Ercument , year =

    Ozden, Furkan and Barazandeh, Sina and Akboga, Dogus and Tabrizi, Sobhan Shokoueian and Seker, Urartu Ozgur Safak and Cicek, A. Ercument , year =

  55. [55]

    Ercument , journal =

    Shukueian Tabrizi, Sobhan and Barazandeh, Sina and Hashemi Aghdam, Hamed and Cicek, A. Ercument , journal =. 2025 , doi =

  56. [56]

    2025 , publisher =

    Klypa, Roman and Bietti, Alberto and Grudinin, Sergei , booktitle =. 2025 , publisher =

  57. [57]

    2024 , doi =

    Wang, Zhen and Liu, Ziqi and Zhang, Wei and Li, Yanjun and Feng, Yizhen and Lv, Shaokang and Diao, Han and Luo, Zhaofeng and Yan, Pengju and He, Min and Li, Xiaolin , journal =. 2024 , doi =

  58. [58]

    and Wang, Jiuming and Hu, Zhihang and Omori, Satotaka and Li, Alicia and Rao, Jiahua and Yu, Qinze and Jin, Wengong and Zhang, Tianqing and Ilia, Katherine and Chen, Jack X

    Wong, Felix and He, Dongchen and Krishnan, Aarti and Hong, Liang and Wang, Alexander Z. and Wang, Jiuming and Hu, Zhihang and Omori, Satotaka and Li, Alicia and Rao, Jiahua and Yu, Qinze and Jin, Wengong and Zhang, Tianqing and Ilia, Katherine and Chen, Jack X. and Zheng, Shuangjia and King, Irwin and Li, Yu and Collins, James J. , journal =. Deep Generat...

  59. [59]

    and Zhang, Yujian and Huang, Liang , journal =

    Zhang, He and Zhang, Liang and Lin, Ang and Xu, Congcong and Li, Ziyu and Liu, Kaibo and Liu, Boxiang and Ma, Xiaopin and Zhao, Fanfan and Jiang, Huiling and Chen, Chunxiu and Shen, Haifa and Li, Hangwen and Mathews, David H. and Zhang, Yujian and Huang, Liang , journal =. Algorithm for Optimized. 2023 , doi =

  60. [60]

    2023 , url =

    Li, Sizhen and Moayedpour, Saeed and Li, Ruijiang and Bailey, Michael and Riahi, Saleh and Miladi, Milad and Miner, Jacob and Zheng, Dinghai and Wang, Jun and Balsubramani, Akshay and others , booktitle =. 2023 , url =

  61. [61]

    and Huang, Liang , year =

    Dai, Ning and Tang, Wei Yu and Zhou, Tianshuo and Mathews, David H. and Huang, Liang , year =. Messenger. 2401.00037 , archivePrefix =

  62. [62]

    and Huang, Liang , journal =

    Dai, Ning and Zhou, Tianshuo and Tang, Wei Yu and Mathews, David H. and Huang, Liang , journal =. 2025 , doi =

  63. [63]

    Fornace, Mark and Wang, Christina Wuyan and Lindsey, Michael , year =. Direct. 2604.19718 , archivePrefix =

  64. [64]

    Joint Design of 5' Untranslated Region and Coding Sequence of

    Liu, Yang and Gao, Jie and Zhang, Xiaonan and Fang, Xiaomin , year =. Joint Design of 5' Untranslated Region and Coding Sequence of. 2410.20781 , archivePrefix =

  65. [65]

    and Cao, Jicong , journal =

    Zhang, He and Liu, Hailong and Xu, Yushan and Huang, Haoran and Liu, Yiming and Wang, Jia and Qin, Yan and Wang, Haiyan and Ma, Lili and Xun, Zhiyuan and Hou, Xuzhuang and Lu, Timothy K. and Cao, Jicong , journal =. Deep Generative Models Design. 2025 , doi =

  66. [66]

    and Huang, Liang , year =

    Yue, Feipeng and Dai, Ning and Tang, Wei Yu and Zhou, Tianshuo and Mathews, David H. and Huang, Liang , year =. Sampling-Based Continuous Optimization for Messenger. 2603.06559 , archivePrefix =

  67. [67]

    2405.18768 , archivePrefix =

    Nori, Divya and Jin, Wengong , year =. 2405.18768 , archivePrefix =

  68. [68]

    2026 , url =

    Hu, Tianmeng and Cui, Yongzheng and Luo, Biao and Li, Ke , booktitle =. 2026 , url =

  69. [69]

    2510.21161 , archivePrefix =

    Sun, Minghao and Cao, Hanqun and Zhang, Zhou and Wei, Chen and Wang, Liang and Jia, Tianrui and Liu, Zhiyuan and Fu, Tianfan and Tang, Xiangru and Choi, Yejin and Heng, Pheng-Ann and Wu, Fang and Zhang, Yang , year =. 2510.21161 , archivePrefix =

  70. [70]

    Patil, Vaibhav and others , year =

  71. [71]

    Secondary-Structure-Informed

    Manzourolajdad, Amirhossein and Mohebbi, Mohammad , journal =. Secondary-Structure-Informed. 2025 , doi =