Recognition: no theorem link
Towards A Generative Protein Evolution Machine with DPLM-Evo
Pith reviewed 2026-05-14 21:00 UTC · model grok-4.3
The pith
DPLM-Evo models protein evolution by predicting explicit substitutions, insertions, and deletions in a discrete diffusion process.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
DPLM-Evo is an evolutionary discrete diffusion framework that explicitly predicts substitution, insertion, and deletion operations during denoising. It decouples an upsampled-length latent alignment space from the variable-length observed sequence space to make indel-aware generation tractable and enable adaptive scaffold growth. A contextualized evolutionary noising kernel produces biologically informed, context-dependent mutation patterns. This results in state-of-the-art mutation effect prediction on ProteinGym in the single-sequence setting and enables variable-length simulated evolution and post-editing of proteins via explicit edit trajectories.
What carries the argument
The decoupled upsampled latent alignment space combined with a contextualized evolutionary noising kernel that predicts explicit edit operations instead of masks.
If this is right
- Improves sequence understanding across protein tasks
- Achieves state-of-the-art mutation effect prediction performance on ProteinGym using only single sequences
- Enables variable-length simulated evolution of proteins
- Allows post-editing and optimization of existing proteins through explicit edit trajectories with negligible overhead
Where Pith is reading between the lines
- Such explicit edit modeling could integrate with lab-based directed evolution to guide experimental protein optimization
- The framework might generalize to other sequence types like nucleic acids for evolutionary simulations
- By producing edit trajectories, the model offers a way to interpret and control the steps in generative protein design
Load-bearing premise
The contextualized evolutionary noising kernel must produce biologically realistic, context-dependent mutation patterns, and decoupling the latent alignment space from the observed sequence must not introduce artifacts in indel generation.
What would settle it
An experiment that measures whether the mutation patterns and indel frequencies generated by DPLM-Evo match those observed in natural protein family alignments or deep mutational scanning experiments.
read the original abstract
Proteins are shaped by gradual evolution under biophysical and functional constraints. Protein language models learn rich evolutionary constraints from large-scale sequences, and discrete diffusion-based protein language models~(\eg, DPLMs) are promising for both understanding and generation. However, existing DPLMs typically rely on masking-based absorbing diffusion that contradicts a simple biological intuition: proteins evolve through accumulated edits, not by emerging from masks. Consequently, these frameworks lack explicit pretraining objectives for substitution and insertion/deletion (indel) operations, limiting both optimization-style post-editing and flexible guided generation. To address these limitations, we present DPLM-Evo, an evolutionary discrete diffusion framework that explicitly predicts substitution, insertion, and deletion operations during denoising. DPLM-Evo decouples an upsampled-length latent alignment space from the variable-length observed sequence space, which makes indel-aware generation tractable and enables adaptive scaffold growth throughout the process with negligible computational overhead. To better align substitutions with real evolution, we further introduce a contextualized evolutionary noising kernel that produces biologically informed, context-dependent mutation patterns. Across tasks, DPLM-Evo improves sequence understanding and achieves state-of-the-art mutation effect prediction performance on ProteinGym in the single-sequence setting. It also enables variable-length simulated evolution, and post-editing/optimization of existing proteins via explicit edit trajectories.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces DPLM-Evo, a discrete diffusion framework for protein generation that replaces masking-based absorbing diffusion with explicit modeling of substitution, insertion, and deletion operations. It uses a contextualized evolutionary noising kernel to produce context-dependent mutations and decouples an upsampled-length latent alignment space from the observed variable-length sequence space to enable tractable indel-aware generation, adaptive scaffold growth, simulated evolution, and post-editing via explicit edit trajectories. The work claims state-of-the-art mutation effect prediction on ProteinGym in the single-sequence setting along with improved sequence understanding.
Significance. If the central claims hold, DPLM-Evo would advance generative protein models by aligning the diffusion process more closely with biological evolution, potentially enabling more realistic variable-length sequence generation and optimization trajectories. The explicit edit modeling and contextual noising could strengthen applications in mutation effect prediction and protein engineering, provided the noising kernel matches real evolutionary statistics and the latent decoupling introduces no systematic artifacts.
major comments (2)
- [Abstract] Abstract: the claim of state-of-the-art mutation effect prediction performance on ProteinGym in the single-sequence setting is presented without any numerical metrics, baselines, error bars, ablation details, or validation procedures, preventing assessment of whether the improvement is load-bearing or driven by post-hoc choices.
- [Abstract] Abstract: the assertion that the contextualized evolutionary noising kernel produces biologically informed, context-dependent mutation patterns and that the upsampled-length latent alignment space introduces no indel artifacts is central to the variable-length evolution and post-editing claims, yet the abstract supplies no direct empirical match to observed substitution matrices or ablation isolating decoupling effects on indel distributions.
minor comments (1)
- [Abstract] Abstract: consider adding one or two key quantitative results (e.g., ProteinGym Spearman correlation or AUROC) to ground the SOTA claim.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback on our manuscript. We address each major comment point by point below, agreeing that the abstract can be made more informative while preserving its brevity. Revisions will be incorporated in the next version.
read point-by-point responses
-
Referee: [Abstract] Abstract: the claim of state-of-the-art mutation effect prediction performance on ProteinGym in the single-sequence setting is presented without any numerical metrics, baselines, error bars, ablation details, or validation procedures, preventing assessment of whether the improvement is load-bearing or driven by post-hoc choices.
Authors: We agree that the abstract would benefit from greater specificity to facilitate immediate assessment. The full manuscript reports these details extensively, including Spearman correlations on ProteinGym, comparisons against baselines such as ESM-1v and Tranception, error bars from multiple independent runs, ablation studies isolating model components, and the exact single-sequence evaluation protocol (see Section 4.1 and Table 2). To address the referee's concern directly, we will revise the abstract to include concise key metrics and a brief reference to the evaluation setup, ensuring the SOTA claim is presented with supporting context while respecting length constraints. revision: yes
-
Referee: [Abstract] Abstract: the assertion that the contextualized evolutionary noising kernel produces biologically informed, context-dependent mutation patterns and that the upsampled-length latent alignment space introduces no indel artifacts is central to the variable-length evolution and post-editing claims, yet the abstract supplies no direct empirical match to observed substitution matrices or ablation isolating decoupling effects on indel distributions.
Authors: We acknowledge that the abstract summarizes these design choices without inline empirical references. The manuscript provides the requested evidence in full: Section 3.2 quantifies the noising kernel's alignment with observed substitution matrices (e.g., BLOSUM and evolutionary statistics), and Section 4.4 presents targeted ablations demonstrating that the latent alignment decoupling produces indel distributions statistically indistinguishable from ground-truth data with no systematic artifacts. We will revise the abstract to include a brief clause noting this empirical grounding (e.g., 'empirically matched to evolutionary statistics with ablations confirming no indel artifacts'), thereby strengthening the claims without expanding beyond typical abstract limits. revision: yes
Circularity Check
No significant circularity detected in DPLM-Evo framework
full rationale
The paper proposes new components—an evolutionary discrete diffusion process with explicit substitution/insertion/deletion prediction, a contextualized evolutionary noising kernel, and decoupling of upsampled latent alignment space from observed sequences—presented as independent architectural innovations rather than reductions of prior fitted quantities or self-citations. No equations or claims in the abstract reduce the central results (ProteinGym SOTA in single-sequence setting, variable-length evolution) to inputs by construction. The derivation chain remains self-contained, relying on new pretraining objectives and empirical validation without load-bearing self-citation chains or ansatz smuggling.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption Proteins evolve through accumulated edits (substitutions and indels) rather than emerging from masks
invented entities (1)
-
upsampled-length latent alignment space
no independent evidence
Reference graph
Works this paper leans on
-
[1]
Protein generation with evolutionary diffusion: sequence is all you need.bioRxiv, pages 2023–09, 2023
Sarah Alamdari, Nitya Thakkar, Rianne van den Berg, Alex Xijie Lu, Nicolo Fusi, Ava Pardis Amini, and Kevin K Yang. Protein generation with evolutionary diffusion: sequence is all you need.bioRxiv, pages 2023–09, 2023
2023
-
[2]
Structured denoising diffusion models in discrete state-spaces
Jacob Austin, Daniel D Johnson, Jonathan Ho, Daniel Tarlow, and Rianne van den Berg. Structured denoising diffusion models in discrete state-spaces. InAdvances in Neural Information Processing Systems, volume 34, pages 17981–17993, 2021
2021
-
[3]
Ethan Baron, Alan N Amin, Ruben Weitzman, Debora Marks, and Andrew Gordon Wilson. A diffusion model to shrink proteins while maintaining their function.arXiv preprint arXiv:2511.07390, 2025
-
[4]
Protein sequence profile prediction using protalbert transformer.Computational Biology and Chemistry, 99:107717, 2022
Armin Behjati, Fatemeh Zare-Mirakabad, Seyed Shahriar Arab, and Abbas Nowzari-Dalini. Protein sequence profile prediction using protalbert transformer.Computational Biology and Chemistry, 99:107717, 2022
2022
-
[5]
Proteinbert: a universal deep- learning model of protein sequence and function.Bioinformatics, 38(8):2102–2110, 2022
Nadav Brandes, Dan Ofer, Yam Peleg, Nadav Rappoport, and Michal Linial. Proteinbert: a universal deep- learning model of protein sequence and function.Bioinformatics, 38(8):2102–2110, 2022
2022
-
[6]
Famsa: Fast and accurate multiple se- quence alignment of huge protein families.Scientific reports, 6(1):33964, 2016
Sebastian Deorowicz, Agnieszka Debudaj-Grabysz, and Adam Gudyś. Famsa: Fast and accurate multiple se- quence alignment of huge protein families.Scientific reports, 6(1):33964, 2016
2016
-
[7]
Prottrans: Toward understanding the language of life through self-supervised learning.IEEE transactions on pattern analysis and machine intelligence, 44(10): 7112–7127, 2021
Ahmed Elnaggar, Michael Heinzinger, Christian Dallago, Ghalia Rehawi, Yu Wang, Llion Jones, Tom Gibbs, Tamas Feher, Christoph Angerer, Martin Steinegger, et al. Prottrans: Toward understanding the language of life through self-supervised learning.IEEE transactions on pattern analysis and machine intelligence, 44(10): 7112–7127, 2021
2021
-
[8]
Esm cambrian: Revealing the mysteries of proteins with unsupervised learning, 2024
ESM Team. Esm cambrian: Revealing the mysteries of proteins with unsupervised learning, 2024. URLhttps: //evolutionaryscale.ai/blog/esm-cambrian
2024
-
[9]
Disease variant prediction with deep generative models of evolutionary data.Nature, 599(7883):91–95, 2021
JonathanFrazer, PascalNotin, MafaldaDias, AidanGomez, JosephKMin, KellyBuss, DanielHZuber, JosephN Glover, and Debora S Marks. Disease variant prediction with deep generative models of evolutionary data.Nature, 599(7883):91–95, 2021
2021
-
[10]
Scaling diffusion language models via adaptation from autoregressive models
Shansan Gong, Shivam Agarwal, Yizhe Zhang, Jiacheng Ye, Lin Zheng, Mukai Li, Chenxin An, Peilin Zhao, Wei Bi, Jiawei Han, Hao Peng, and Lingpeng Kong. Scaling diffusion language models via adaptation from autoregressive models. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=j1tSLYKwg8
2025
-
[11]
Connectionist temporal classifica- tion: labellingunsegmentedsequencedatawithrecurrentneuralnetworks
Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. Connectionist temporal classifica- tion: labellingunsegmentedsequencedatawithrecurrentneuralnetworks. InProceedings of the 23rd international conference on Machine learning, pages 369–376, 2006
2006
-
[12]
Protein design with guided discrete diffusion
Nate Gruver, Samuel Stanton, Nathan C Frey, Tim GJ Rudner, Isidro Hotber, Julien Lafrance-Vanasse, Arvind Rajpal, Kyunghyun Cho, and Andrew Gordon Wilson. Protein design with guided discrete diffusion. InAdvances in Neural Information Processing Systems, 2023
2023
-
[13]
Fully non-autoregressive neural machine translation: Tricks of the trade
Jiatao Gu and Xiang Kong. Fully non-autoregressive neural machine translation: Tricks of the trade. InFindings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 120–133, Online, August 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.findings-acl.11. URLhttps://aclanthology. org/2021.findings-acl.11
-
[14]
Levenshtein transformer
Jiatao Gu, Changhan Wang, and Junbo Zhao. Levenshtein transformer. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett, editors,Advances in Neu- ral Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC,...
2019
-
[15]
Nat: Neural architecture transformer for accurate and compact architectures.Advances in Neural Information Processing Systems, 32, 2019
Yong Guo, Yin Zheng, Mingkui Tan, Qi Chen, Jian Chen, Peilin Zhao, and Junzhou Huang. Nat: Neural architecture transformer for accurate and compact architectures.Advances in Neural Information Processing Systems, 32, 2019
2019
-
[16]
Edit flows: Flow matching with edit operations.arXiv preprint arXiv:2506.09018,
Marton Havasi, Brian Karrer, Itai Gat, and Ricky TQ Chen. Edit flows: Flow matching with edit operations. arXiv preprint arXiv:2506.09018, 2025
-
[17]
Simulating 500 million years of evolution with a language model
TomasHayes, RoshanRao, HalilAkin, NicholasJSofroniew, DenizOktay, ZemingLin, RobertVerkuil, VincentQ Tran, Jonathan Deaton, Marius Wiggert, et al. Simulating 500 million years of evolution with a language model. bioRxiv, pages 2024–07, 2024
2024
-
[18]
Diffusionbert: Improving gen- erative masked language models with diffusion models
Zhengfu He, Tianxiang Sun, Kuanning Wang, Xuanjing Huang, and Xipeng Qiu. Diffusionbert: Improving gen- erative masked language models with diffusion models. InAnnual Meeting of the Association for Computational Linguistics, 2023
2023
-
[19]
Denoising diffusion probabilistic models.Advances in Neural In- formation Processing Systems, 33:6840–6851, 2020
Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models.Advances in Neural In- formation Processing Systems, 33:6840–6851, 2020. URLhttps://proceedings.neurips.cc/paper/2020/file/ 4c5bcfec8584af0d967f1ab10179ca4b-Paper.pdf
2020
-
[20]
Argmax flows and multinomial diffusion: Learning categorical distributions.Advances in Neural Information Processing Systems, 34:12454– 12465, 2021
Emiel Hoogeboom, Didrik Nielsen, Priyank Jaini, Patrick Forré, and Max Welling. Argmax flows and multinomial diffusion: Learning categorical distributions.Advances in Neural Information Processing Systems, 34:12454– 12465, 2021
2021
-
[21]
Elucidatingthedesignspaceofmultimodalproteinlanguagemodels
Cheng-Yen Hsieh, Xinyou Wang, Daiheng Zhang, Dongyu Xue, Fei Ye, Shujian Huang, Zaixiang Zheng, and QuanquanGu. Elucidatingthedesignspaceofmultimodalproteinlanguagemodels. InForty-second International Conference on Machine Learning, 2025
2025
-
[22]
Learning inverse folding from millions of predicted structures
Chloe Hsu, Robert Verkuil, Jason Liu, Zeming Lin, Brian Hie, Tom Sercu, Adam Lerer, and Alexander Rives. Learning inverse folding from millions of predicted structures. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors,Proceedings of the 39th International Conference on Machine Learning, volume 162 ofP...
-
[23]
URLhttps://proceedings.mlr.press/v162/hsu22a.html
-
[24]
Gemme: a simple and fast global epistatic model predicting mutational effects.Molecular biology and evolution, 36(11):2604–2619, 2019
Elodie Laine, Yasaman Karami, and Alessandra Carbone. Gemme: a simple and fast global epistatic model predicting mutational effects.Molecular biology and evolution, 36(11):2604–2619, 2019
2019
-
[25]
Language models of protein sequences at the scale of evolution enable accurate structure prediction.BioRxiv, 2022
Zeming Lin, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Allan dos Santos Costa, Maryam Fazel-Zarandi, Tom Sercu, Sal Candido, et al. Language models of protein sequences at the scale of evolution enable accurate structure prediction.BioRxiv, 2022
2022
-
[26]
Evolutionary-scale prediction of atomic-level protein structure with a language model.Science, 379(6637):1123–1130, 2023
Zeming Lin, Halil Akin, Roshan Rao, Brian Hie, Zhongkai Zhu, Wenting Lu, Nikita Smetanin, Robert Verkuil, Ori Kabeli, Yaniv Shmueli, et al. Evolutionary-scale prediction of atomic-level protein structure with a language model.Science, 379(6637):1123–1130, 2023
2023
-
[27]
Sequential diffusion language models.arXiv preprint arXiv:2509.24007, 2025
Yangzhou Liu, Yue Cao, Hao Li, Gen Luo, Zhe Chen, Weiyun Wang, Xiaobo Liang, Biqing Qi, Lijun Wu, Changyao Tian, et al. Sequential diffusion language models.arXiv preprint arXiv:2509.24007, 2025
-
[28]
Expert-guided pro- tein language models enable accurate and blazingly fast fitness prediction.Bioinformatics, 40(11):btae621, 11
Céline Marquet, Julius Schlensok, Marina Abakarova, Burkhard Rost, and Elodie Laine. Expert-guided pro- tein language models enable accurate and blazingly fast fitness prediction.Bioinformatics, 40(11):btae621, 11
-
[29]
doi: 10.1093/bioinformatics/btae621
ISSN 1367-4811. doi: 10.1093/bioinformatics/btae621. URLhttps://doi.org/10.1093/bioinformatics/ btae621
-
[30]
Language models enable zero-shot prediction of the effects of mutations on protein function
Joshua Meier, Roshan Rao, Robert Verkuil, Jason Liu, Tom Sercu, and Alex Rives. Language models enable zero-shot prediction of the effects of mutations on protein function. InAdvances in Neural Information Processing Systems, pages 29287–29303, 2021
2021
-
[31]
Alexey Meshchaninov, Daniil Zinchenko, Andrey Golovin, Sergey Evfratov, Alexey Chertkov, and Nikita Nikitin. Dima: Diffusion mamba – a diffusion model with state space backbone for protein design.arXiv preprint arXiv:2410.13514, 2024
-
[32]
Transforming the language of life: transformer neural networks for protein prediction tasks
Ananthan Nambiar, Maeve Heflin, Simon Liu, Sergei Maslov, Mark Hopkins, and Anna Ritz. Transforming the language of life: transformer neural networks for protein prediction tasks. InProceedings of the 11th ACM international conference on bioinformatics, computational biology and health informatics, pages 1–8, 2020. 14
2020
-
[33]
Scaling up masked diffusion models on text
Shen Nie, Fengqi Zhu, Chao Du, Tianyu Pang, Qian Liu, Guangtao Zeng, Min Lin, and Chongxuan Li. Scaling up masked diffusion models on text. InThe Thirteenth International Conference on Learning Representations, 2024
2024
-
[34]
Large Language Diffusion Models
Shen Nie, Fengqi Zhu, Zebin You, Xiaolu Zhang, Jingyang Ou, Jun Hu, Jun Zhou, Yankai Lin, Ji-Rong Wen, and Chongxuan Li. Large language diffusion models.arXiv preprint arXiv:2502.09992, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[35]
Progen2: exploring the boundaries of protein language models.arXiv preprint arXiv:2206.13517, 2022
Erik Nijkamp, Jeffrey Ruffolo, Eli N Weinstein, Nikhil Naik, and Ali Madani. Progen2: exploring the boundaries of protein language models.arXiv preprint arXiv:2206.13517, 2022
-
[36]
Proteingym: Large-scale benchmarks for protein fitness prediction and design
Pascal Notin, Aaron W Kollasch, Daniel Ritter, Lood Van Niekerk, Steffan Paul, Han Spinner, Nathan J Rollins, Ada Shaw, Ruben Weitzman, Jonathan Frazer, et al. Proteingym: Large-scale benchmarks for protein fitness prediction and design. InThirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2023
2023
-
[37]
Your absorbing discrete diffusion secretly models the conditional distributions of clean data
Jingyang Ou, Shen Nie, Kaiwen Xue, Fengqi Zhu, Jiacheng Sun, Zhenguo Li, and Chongxuan Li. Your absorbing discrete diffusion secretly models the conditional distributions of clean data. InThe Thirteenth International Conference on Learning Representations, 2024
2024
-
[38]
Evaluating protein transfer learning with tape.Advances in neural information processing systems, 32, 2019
Roshan Rao, Nicholas Bhattacharya, Neil Thomas, Yan Duan, Peter Chen, John Canny, Pieter Abbeel, and Yun Song. Evaluating protein transfer learning with tape.Advances in neural information processing systems, 32, 2019
2019
-
[39]
Diffuser: Diffusion via edit-based reconstruction
Machel Reid, Vincent Josua Hellendoorn, and Graham Neubig. Diffuser: Diffusion via edit-based reconstruction. InInternational Conference on Learning Representations, 2022
2022
-
[40]
Lawrence Zitnick, Jerry Ma, and Rob Fergus
Alexander Rives, Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Jason Liu, Demi Guo, Myle Ott, C. Lawrence Zitnick, Jerry Ma, and Rob Fergus. Biological structure and function emerge from scal- ing unsupervised learning to 250 million protein sequences.PNAS, 2019. doi: 10.1101/622803. URL https://www.biorxiv.org/content/10.1101/622803v4
-
[41]
Simpleandeffectivemaskeddiffusionlanguagemodels.Advances in Neural Information Processing Systems, 37:130136–130184, 2024
Subham Sahoo, Marianne Arriola, Yair Schiff, Aaron Gokaslan, Edgar Marroquin, Justin Chiu, Alexander Rush, andVolodymyrKuleshov. Simpleandeffectivemaskeddiffusionlanguagemodels.Advances in Neural Information Processing Systems, 37:130136–130184, 2024
2024
-
[42]
The diffusion duality
Subham Sekhar Sahoo, Justin Deschenaux, Aaron Gokaslan, Guanghan Wang, Justin T Chiu, and Volodymyr Kuleshov. The diffusion duality. InForty-second International Conference on Machine Learning, 2025
2025
-
[43]
Simple guidance mechanisms for discrete diffusion models
Yair Schiff, Subham Sekhar Sahoo, Hao Phung, Guanghan Wang, Sam Boshar, Hugo Dalla-torre, Bernardo P de Almeida, Alexander M Rush, Thomas PIERROT, and Volodymyr Kuleshov. Simple guidance mechanisms for discrete diffusion models. InThe Thirteenth International Conference on Learning Representations, 2025
2025
-
[44]
Simplified and generalized masked diffusion for discrete data.Advances in neural information processing systems, 37:103131–103167, 2024
Jiaxin Shi, Kehang Han, Zhe Wang, Arnaud Doucet, and Michalis Titsias. Simplified and generalized masked diffusion for discrete data.Advances in neural information processing systems, 37:103131–103167, 2024
2024
-
[45]
Deep unsupervised learning using nonequilibrium thermodynamics
Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In Francis Bach and David Blei, editors,International Conference on Machine Learning, volume 37 ofProceedings of Machine Learning Research, pages 2256–2265, Lille, France, 07–09 Jul
-
[46]
URLhttps://proceedings.mlr.press/v37/sohl-dickstein15.html
PMLR, PMLR. URLhttps://proceedings.mlr.press/v37/sohl-dickstein15.html
-
[47]
Denoising diffusion implicit models
Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. InInternational Confer- ence on Learning Representations, 2020
2020
-
[48]
Score- based generative modeling through stochastic differential equations
Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score- based generative modeling through stochastic differential equations. InInternational Conference on Learning Representations, 2020
2020
-
[49]
Saprot: Protein language modeling with structure-aware vocabulary.bioRxiv, pages 2023–10, 2023
Jin Su, Chenchen Han, Yuyang Zhou, Junjie Shan, Xibin Zhou, and Fajie Yuan. Saprot: Protein language modeling with structure-aware vocabulary.bioRxiv, pages 2023–10, 2023
2023
-
[50]
PoET: A generative model of protein families as sequences-of-sequences
Timothy F Truong Jr and Tristan Bepler. PoET: A generative model of protein families as sequences-of-sequences. InAdvances in Neural Information Processing Systems, 2024
2024
-
[51]
Language models generalize beyond natural proteins.bioRxiv, pages 2022–12, 2022
Robert Verkuil, Ori Kabeli, Yilun Du, Basile IM Wicky, Lukas F Milles, Justas Dauparas, David Baker, Sergey Ovchinnikov, Tom Sercu, and Alexander Rives. Language models generalize beyond natural proteins.bioRxiv, pages 2022–12, 2022. 15
2022
-
[52]
Gen- eralized interpolating discrete diffusion
Dimitri von Rütte, Janis Fluri, Yuhui Ding, Antonio Orvieto, Bernhard Schölkopf, and Thomas Hofmann. Gen- eralized interpolating discrete diffusion. InForty-second International Conference on Machine Learning, 2025
2025
-
[53]
Dplm-2: A multimodal diffusion protein language model.arXiv preprint arXiv:2410.13782, 2024
Xinyou Wang, Zaixiang Zheng, Fei Ye, Dongyu Xue, Shujian Huang, and Quanquan Gu. Dplm-2: A multimodal diffusion protein language model.arXiv preprint arXiv:2410.13782, 2024
-
[54]
Diffusion language models are versatile protein learners.arXiv preprint arXiv:2402.18567, 2024
Xinyou Wang, Zaixiang Zheng, Fei Ye, Dongyu Xue, Shujian Huang, and Quanquan Gu. Diffusion language models are versatile protein learners.arXiv preprint arXiv:2402.18567, 2024
-
[55]
Dreamon: Diffusion language models for code infilling beyond fixed-size canvas, 2025
Zirui Wu, Lin Zheng, Zhihui Xie, Jiacheng Ye, Jiahui Gao, Yansong Feng, Zhenguo Li, Victoria W., Guorui Zhou, and Lingpeng Kong. Dreamon: Diffusion language models for code infilling beyond fixed-size canvas, 2025. URL https://hkunlp.github.io/blog/2025/dreamon
2025
-
[56]
Modeling protein using large-scale pretrain language model.arXiv preprint arXiv:2108.07435, 2021
Yijia Xiao, Jiezhong Qiu, Ziang Li, Chang-Yu Hsieh, and Jie Tang. Modeling protein using large-scale pretrain language model.arXiv preprint arXiv:2108.07435, 2021
-
[57]
Convolutions are competitive with transformers for protein sequence pretraining.bioRxiv, pages 2022–05, 2022
Kevin K Yang, Alex X Lu, and Nicolo Fusi. Convolutions are competitive with transformers for protein sequence pretraining.bioRxiv, pages 2022–05, 2022
2022
-
[58]
Dream 7B: Diffusion Large Language Models
Jiacheng Ye, Zhihui Xie, Lin Zheng, Jiahui Gao, Zirui Wu, Xin Jiang, Zhenguo Li, and Lingpeng Kong. Dream 7b: Diffusion large language models.arXiv preprint arXiv:2508.15487, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[59]
Jiasheng Ye, Zaixiang Zheng, Yu Bao, Lihua Qian, and Quanquan Gu. Diffusion language models can perform many tasks with scaling and instruction-finetuning.arXiv preprint arXiv:2308.12219, 2023
-
[60]
Jiasheng Ye, Zaixiang Zheng, Yu Bao, Lihua Qian, and Mingxuan Wang. Dinoiser: Diffused conditional sequence learning by manipulating noises.arXiv preprint arXiv:2302.10025, 2023
-
[61]
A reparameterized discrete diffusion model for text generation.arXiv preprint arXiv:2302.05737, 2023
Lin Zheng, Jianbo Yuan, Lei Yu, and Lingpeng Kong. A reparameterized discrete diffusion model for text generation.arXiv preprint arXiv:2302.05737, 2023
-
[62]
Structure-informed language models are protein designers
Zaixiang Zheng, Yifan Deng, Dongyu Xue, Yi Zhou, Fei YE, and Quanquan Gu. Structure-informed language models are protein designers. InInternational Conference on Machine Learning, 2023. 16 Appendix A Training Details. A.1 Substitution Learning with Contextualized Evolutionary Noise The quality of the DPLM-Evo heavily depends on how the substitution proces...
2023
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.