pith. machine review for the scientific record. sign in

arxiv: 2605.13101 · v1 · submitted 2026-05-13 · 💻 cs.LG · cs.AI

Recognition: unknown

Margin-calibrated Classifier Guidance for Property-driven Synthesis Planning

Authors on Pith no claims yet

Pith reviewed 2026-05-14 19:38 UTC · model grok-4.3

classification 💻 cs.LG cs.AI
keywords retrosynthesissynthesis planningclassifier guidancecontrastive learningmargin lossbeam searchchemical reactionsproperty-driven generation
0
0 comments X

The pith

Margin-calibrated classifiers let single-step retrosynthesis models reach valid multi-step routes for targets that unguided search misses.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper demonstrates that auxiliary classifiers trained with ordinary cross-entropy loss cannot reliably steer a pretrained autoregressive retrosynthesis model toward desired properties during repeated decoding steps. It introduces Sequence Completion Ranking, a training procedure that uses contrastive pairs and a margin loss to calibrate the classifier so it ranks which reaction continuation better satisfies a chemist-specified target. With this calibrated guidance inside beam search, multi-step solve rates on USPTO-190 rise from 16.8 percent unguided to 78.4 percent under reaction-type guidance and 95.3 percent under Tanimoto guidance, recovering valid routes for 33 targets that no baseline solved.

Core claim

Sequence Completion Ranking trains margin-calibrated classifiers via contrastive argumentation on single-disconnection data so that the classifiers can discriminate among candidate continuations during guided beam search of a fixed single-step retrosynthesis model. The calibration formally enlarges the set of property-satisfying sequences reachable under decoding. On USPTO-190 this yields solve rates of 78.4 percent with reaction-type guidance and 95.3 percent with Tanimoto guidance, unlocking routes for 17.4 percent of targets that remained unsolvable without the method and narrowing the diversity gap with template-based planners.

What carries the argument

Sequence Completion Ranking (SCR), a margin-based contrastive loss that trains an auxiliary classifier to rank which of two sequence completions better satisfies the guidance target.

If this is right

  • Guided beam search recovers valid multi-step routes for 17.4 percent of targets unreachable by unguided or standard-guided baselines.
  • The same pretrained generator can be steered toward chemist-specified reaction types or molecular similarity without any retraining.
  • Template-free retrosynthesis closes the long-standing diversity gap with template-based methods under calibrated guidance.
  • Property-driven planning becomes practical at inference time for any target property that can be expressed as a ranking signal.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same calibration technique could be applied to other autoregressive generators that must respect sparse training data while satisfying downstream constraints.
  • Chemists could supply new guidance signals at test time without collecting additional multi-step reaction data.
  • If the margin ranking remains reliable across datasets, the need for large labeled multi-step corpora in synthesis planning is reduced.

Load-bearing premise

Margin-calibrated classifiers trained on sparse single-disconnection data can discriminate among continuations in multi-step decoding without overfitting to the guidance targets or introducing systematic bias.

What would settle it

Measure solve rates on a new set of targets whose reaction types or similarity profiles differ from those used to train the classifier; if rates fall back near the unguided baseline, the calibration does not transfer.

Figures

Figures reproduced from arXiv: 2605.13101 by Najwa Laabid, Vikas Garg.

Figure 1
Figure 1. Figure 1: Autoregressive generation under three configurations: unguided generator, generator + [PITH_FULL_IMAGE:figures/full_fig_p003_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: CE training fails to deliver token-level discriminability on retrosynthesis data; SCR installs the gap by construction. (a) Excess CE signal ∆ˆ CE k − Gk vs. per-prefix samples N (at ϵ=0.05); curves are the 10th percentile of ∆ˆ CE k across simulated training draws, the value reached in 90% of runs, matching δ=0.1 in the bound. The η=0 (USPTO) curve stays at Gk for any N; at η=0.05 the signal clears γ only… view at source ↗
Figure 3
Figure 3. Figure 3: Ablations of classifier training methods on reaction type prediction. Accuracy is evaluated [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Steering breadth for all baselines. SCRrxn achieves the highest steering breadth on class accuracy and matches template-based methods when additionally filtering for round-trip validity (+RT) or RXN-Insight reaction naming (+RXN-I). (SGrxn, Section E), the Retro* value function is additionally biased toward precursors whose predicted property matches the target, with a multiplicative factor balancing synth… view at source ↗
Figure 5
Figure 5. Figure 5: Case studies on USPTO-190. (a) Wieland-Gumlich Aldehyde admits two retrosynthetic disconnections: reduction to an ester precursor (recovered by the unguided generator) and deprotection to a TMS-ether precursor (recovered only by SCR with rxntype guidance). (b) A route for Target 49 found exclusively by rxntype guidance; all eight baselines fail to produce any valid route for this target. (c) A 3-step route… view at source ↗
Figure 6
Figure 6. Figure 6: Effect of guidance on the samples of the generator. The tables show metrics computed over [PITH_FULL_IMAGE:figures/full_fig_p021_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: The difference between guided and baseline results for steering towards all non-ground-truth [PITH_FULL_IMAGE:figures/full_fig_p023_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Steering breadth on USPTO-50k for all baselines and SCR [PITH_FULL_IMAGE:figures/full_fig_p023_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Difference between SCRrxn and baseline results for steering towards all non-ground-truth reaction types on USPTO-50k. The heatmaps show the difference in the number of products with precursors from non-ground-truth target classes, stratified per class, against the unguided generator ablation, RetroKNN (template-based), and Megan (template-free). As on USPTO-190-Steps, our guidance scheme improves steering … view at source ↗
Figure 10
Figure 10. Figure 10: Effect of the two main hyperparameters on class accuracy (steering, blue) and round-trip [PITH_FULL_IMAGE:figures/full_fig_p028_10.png] view at source ↗
read the original abstract

Synthesis planning seeks an efficient sequence of chemical reactions that produce a target molecule. Typically, a pretrained single-step (autoregressive) retrosynthesis model is repeatedly invoked to generate such a sequence. Classifier guidance can, in principle, help steer the output of single-step model toward reactions that satisfy specific constraints or accommodate chemist's preferences during inference without having to retrain the autoregressive generator. We expose the insufficiency of auxiliary classifiers trained with cross-entropy loss to override the unconditional token-level distributions learned from typical sparse single-disconnection reaction datasets. We overcome this issue with a novel method called Sequence Completion Ranking (SCR), which employs contrastive argumentation and a margin-based loss to calibrate the classifier so that it can meaningfully discriminate between continuations during decoding. We formally establish that margin-calibrated classifiers can expand the set of property-satisfying sequences reachable under guided beam search. Empirically, on USPTO-190, given chemist-specified guidance targets, SCR substantially improves multi-step solve rates from $16.8\%$ (unguided generator) to $78.4\%$ with reaction-type guidance and $95.3\%$ with Tanimoto guidance, unlocking valid routes for 33 targets ($17.4\%$) previously unsolvable with baselines. Our method also effectively closes the long-standing diversity gap between template-free and template-based methods.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper introduces Sequence Completion Ranking (SCR), which trains auxiliary classifiers via contrastive argumentation and a margin-based loss to guide pretrained single-step retrosynthesis models during beam search. It claims that cross-entropy classifiers are insufficient on sparse single-disconnection data, that margin calibration formally expands the reachable set of property-satisfying sequences, and that SCR raises multi-step solve rates on USPTO-190 from 16.8% (unguided) to 78.4% (reaction-type guidance) and 95.3% (Tanimoto guidance), solving 33 previously unsolvable targets while closing the diversity gap with template-based methods.

Significance. If the formal reachability argument and empirical gains are confirmed without overfitting to guidance targets, the work would meaningfully advance property-driven synthesis planning by enabling effective inference-time guidance without retraining the generator. The reported ability to unlock valid routes for 17.4% of targets and improve diversity would be a concrete step toward more flexible, chemist-preference-aware retrosynthesis.

major comments (3)
  1. [Formal Result] The formal claim that margin-calibrated classifiers expand reachable property-satisfying sequences under guided beam search is load-bearing; the manuscript should supply the explicit derivation steps showing how the margin loss prevents the classifier scores from being dominated by the unconditional token distribution on sparse data (see the reachability argument section).
  2. [Empirical Evaluation] The empirical solve-rate jumps (16.8% to 78.4%/95.3%) and the 33 newly solved targets rest on the assumption that classifiers generalize to new guidance targets; the experiments section does not report whether the single-disconnection training data was strictly separated from the guidance targets used at test time, leaving open the possibility that gains reflect memorization rather than the claimed expansion of reachable sequences.
  3. [Method] The contrastive argumentation procedure and margin loss are presented as overcoming cross-entropy insufficiency, yet no ablation isolates the contribution of the margin term versus standard contrastive losses, which is needed to substantiate that the calibration is what enables meaningful discrimination during decoding.
minor comments (2)
  1. [Abstract / Experiments] USPTO-190 is referenced in the abstract and results without a concise definition or citation to its construction; add a short description in §2 or the experimental setup.
  2. [Notation] Notation for sequences, property functions, and classifier scores should be unified across the formal argument and the beam-search implementation to avoid ambiguity.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We are grateful to the referee for the detailed and insightful comments, which have helped us improve the clarity and rigor of our work. We address each major comment below and have incorporated revisions accordingly to strengthen the formal argument, clarify the experimental setup, and provide additional ablations.

read point-by-point responses
  1. Referee: The formal claim that margin-calibrated classifiers expand reachable property-satisfying sequences under guided beam search is load-bearing; the manuscript should supply the explicit derivation steps showing how the margin loss prevents the classifier scores from being dominated by the unconditional token distribution on sparse data (see the reachability argument section).

    Authors: We agree that explicit derivation steps are necessary for the load-bearing formal claim. In the revised manuscript, we have added a step-by-step derivation in Section 3.3 and Appendix A. The key insight is that the margin-based loss L_margin = max(0, m - (f(x_pos) - f(x_neg))) ensures that the classifier output f(x) for positive (property-satisfying) sequences exceeds that for negatives by at least m, which, when added to the beam search score, overcomes the dominance of the base model's token probabilities on sparse single-disconnection data. This formally expands the reachable set as shown in the updated proof. revision: yes

  2. Referee: The empirical solve-rate jumps (16.8% to 78.4%/95.3%) and the 33 newly solved targets rest on the assumption that classifiers generalize to new guidance targets; the experiments section does not report whether the single-disconnection training data was strictly separated from the guidance targets used at test time, leaving open the possibility that gains reflect memorization rather than the claimed expansion of reachable sequences.

    Authors: Thank you for raising this critical point about potential data leakage. We confirm that the classifier training data consists of single-disconnection reactions from the USPTO training set (1,000,000+ reactions), while the 190 guidance targets in USPTO-190 are from a completely disjoint test set with no overlap in molecules or reactions. We have revised the manuscript to explicitly report this separation in Section 4.1 and added a statement ruling out memorization. Furthermore, we include results on a secondary benchmark to demonstrate generalization. revision: yes

  3. Referee: The contrastive argumentation procedure and margin loss are presented as overcoming cross-entropy insufficiency, yet no ablation isolates the contribution of the margin term versus standard contrastive losses, which is needed to substantiate that the calibration is what enables meaningful discrimination during decoding.

    Authors: We acknowledge the need for an ablation to isolate the margin term's contribution. In the revised paper, we have included a new ablation study (Table 4) comparing the full SCR (contrastive + margin) against a baseline using only contrastive loss without margin. The margin term provides an additional 14.2% absolute improvement in multi-step solve rate, validating that the calibration is essential for effective discrimination on sparse data. We discuss this in the updated Section 4.2. revision: yes

Circularity Check

0 steps flagged

No circularity: derivation relies on novel loss and formal reachability argument

full rationale

The paper defines Sequence Completion Ranking (SCR) via contrastive argumentation plus margin loss to address cross-entropy insufficiency, then proves that such margin-calibrated classifiers expand the reachable set of property-satisfying sequences under guided beam search. These steps are introduced as new constructions rather than reductions of outputs to fitted inputs or prior self-citations. Empirical gains on USPTO-190 are reported as downstream validation, not as predictions forced by construction from the training data. No load-bearing step equates a claimed result to its own definition or to a self-cited uniqueness theorem.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

Review performed on abstract only; therefore the ledger is necessarily incomplete. The paper appears to rely on standard domain assumptions about pretrained autoregressive retrosynthesis models and sparse reaction datasets.

axioms (1)
  • domain assumption Pretrained single-step retrosynthesis models produce token-level distributions that auxiliary cross-entropy classifiers cannot reliably override on sparse datasets.
    Stated as the core insufficiency the new method addresses.

pith-pipeline@v0.9.0 · 5534 in / 1165 out tokens · 52227 ms · 2026-05-14T19:38:30.855131+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

123 extracted references · 32 canonical work pages · 2 internal anchors

  1. [2]

    Advances in Neural Information Processing Systems , editor=

    Controllable Text Generation with Neurally-Decomposed Oracle , author=. Advances in Neural Information Processing Systems , editor=. 2022 , url=

  2. [3]

    and Maier, Josef M

    Wołos, Agnieszka and Koszelewski, Dominik and Roszak, Rafał and Szymkuć, Sara and Moskal, Martyna and Ostaszewski, Ryszard and Herrera, Brenden T. and Maier, Josef M. and Brezicki, Gordon and Samuel, Jonathon and Lummiss, Justin A. M. and McQuade, D. Tyler and Rogers, Luke and Grzybowski, Bartosz A. , year =. Computer-designed repurposing of chemical wast...

  3. [4]

    Why is Tanimoto index an appropriate choice for fingerprint-based similarity calculations? , volume =

    Bajusz, Dávid and Rácz, Anita and Héberger, Károly , year =. Why is Tanimoto index an appropriate choice for fingerprint-based similarity calculations? , volume =. Journal of Cheminformatics , publisher =. doi:10.1186/s13321-015-0069-3 , number =

  4. [5]

    Organic & Biomolecular Chemistry , volume=

    Analysis of the reactions used for the preparation of drug candidate molecules , author=. Organic & Biomolecular Chemistry , volume=. 2006 , publisher=

  5. [6]

    CoRR , volume =

    Prafulla Dhariwal and Alex Nichol , title =. CoRR , volume =

  6. [7]

    AiZynthFinder : a fast, robust and flexible open-source software for retrosynthetic planning

    Genheden, Samuel and Thakkar, Amol and Chadimov \'a , Veronika and Reymond, Jean-Louis and Engkvist, Ola and Bjerrum, Esben. AiZynthFinder : a fast, robust and flexible open-source software for retrosynthetic planning. J. Cheminform

  7. [8]

    2020 , eprint=

    Plug and Play Language Models: A Simple Approach to Controlled Text Generation , author=. 2020 , eprint=

  8. [9]

    2022 , archivePrefix=

    Classifier-Free Diffusion Guidance , author=. 2022 , archivePrefix=

  9. [10]

    Journal of Chemical Theory and Computation , volume =

    Guo, Jiasheng and Yu, Chenning and Li, Kenan and Zhang, Yijian and Wang, Guoqiang and Li, Shuhua and Dong, Hao , title =. Journal of Chemical Theory and Computation , volume =. 2024 , doi =

  10. [11]

    Organic Process Research & Development , volume=

    Opportunities for Machine Learning and Artificial Intelligence to Advance Synthetic Drug Substance Process Development , author=. Organic Process Research & Development , volume=. 2023 , publisher=. doi:10.1021/acs.oprd.3c00229 , url=

  11. [12]

    Journal of Cheminformatics , volume=

    Improving the performance of models for one-step retrosynthesis through re-ranking , author=. Journal of Cheminformatics , volume=. 2022 , publisher=

  12. [13]

    1996 , publisher=

    Classics in Total Synthesis: Targets, Strategies, Methods , author=. 1996 , publisher=

  13. [14]

    and Bostr

    Brown, Dean G. and Bostr. Analysis of Past and Present Synthetic Methodologies on Medicinal Chemistry: Where Have All the New Reactions Gone? , journal =. 2016 , doi =

  14. [15]

    and Maggiora, Gerald M

    Lajiness, Michael S. and Maggiora, Gerald M. and Shanmugasundaram, Veerabahu , title =. Journal of Medicinal Chemistry , volume =. 2004 , doi =

  15. [16]

    ACS central science , volume=

    Molecular transformer: A model for uncertainty-calibrated chemical reaction prediction , author=. ACS central science , volume=. 2019 , publisher=

  16. [17]

    Rxn-INSIGHT : fast chemical reaction analysis using bond-electron matrices

    Dobbelaere, Maarten R and Lengyel, Istv \'a n and Stevens, Christian V and Van Geem, Kevin M. Rxn-INSIGHT : fast chemical reaction analysis using bond-electron matrices. J. Cheminform

  17. [18]

    and Torres, Elizabeth E

    Morgan, John P. and Torres, Elizabeth E. and Averill, Richard and Brody, Alex M. , title =. Journal of Chemical Education , volume =. 2023 , doi =

  18. [19]

    Demir and Haluk Hamamci and Peruze Ayhan and A

    Ayhan S. Demir and Haluk Hamamci and Peruze Ayhan and A. Nese Duygu and A. Cigdem İgdir and Doga Capanoglu , abstract =. Fungi mediated conversion of benzil to benzoin and hydrobenzoin , journal =. 2004 , issn =. doi:https://doi.org/10.1016/j.tetasy.2004.07.010 , url =

  19. [20]

    PubChem Compound Summary for CID 8400, Benzoin , year =

  20. [21]

    A platform for automated nanomole-scale reaction screening and micromole-scale synthesis in flow

    Perera, Damith and Tucker, Joseph W and Brahmbhatt, Shalini and Helal, Christopher J and Chong, Ashley and Farrell, William and Richardson, Paul and Sach, Neal W. A platform for automated nanomole-scale reaction screening and micromole-scale synthesis in flow. Science

  21. [22]

    Ahneman and Jesús G

    Derek T. Ahneman and Jesús G. Estrada and Shishi Lin and Spencer D. Dreher and Abigail G. Doyle , title =. Science , volume =. 2018 , doi =

  22. [23]

    Practical high-throughput experimentation for chemists

    Shevlin, Michael. Practical high-throughput experimentation for chemists. ACS Med. Chem. Lett

  23. [24]

    SuperCYP : a comprehensive database on Cytochrome P450 enzymes including a tool for analysis of CYP-drug interactions

    Preissner, Saskia and Kroll, Katharina and Dunkel, Mathias and Senger, Christian and Goldsobel, Gady and Kuzman, Daniel and Guenther, Stefan and Winnenburg, Rainer and Schroeder, Michael and Preissner, Robert. SuperCYP : a comprehensive database on Cytochrome P450 enzymes including a tool for analysis of CYP-drug interactions. Nucleic Acids Res

  24. [25]

    and Ye, Lin and Sedykh, Alexander and Young, Douglas M

    Zhu, Hao and Martin, Todd M. and Ye, Lin and Sedykh, Alexander and Young, Douglas M. and Tropsha, Alexander , title =. Chemical Research in Toxicology , volume =. 2009 , doi =

  25. [26]

    2025 , eprint=

    PyTDC: A multimodal machine learning training, evaluation, and inference platform for biomedical foundation models , author=. 2025 , eprint=

  26. [27]

    2025 , primaryClass=

    Chemical reasoning in LLMs unlocks steerable synthesis planning and reaction mechanism elucidation , author=. 2025 , primaryClass=

  27. [28]

    2021 , eprint=

    Diffusion Models Beat GANs on Image Synthesis , author=. 2021 , eprint=

  28. [29]

    Journal of Chemical Information and Computer Sciences , volume =

    Weininger, David , title =. Journal of Chemical Information and Computer Sciences , volume =. 1988 , doi =

  29. [30]

    2020 , archivePrefix=

    GeDi: Generative Discriminator Guided Sequence Generation , author=. 2020 , archivePrefix=

  30. [31]

    FUDGE: Controlled Text Generation With Future Discriminators , DOI=

    Yang, Kevin and Klein, Dan , year=. FUDGE: Controlled Text Generation With Future Discriminators , DOI=. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies , publisher=

  31. [32]

    2018 , archivePrefix=

    Pragmatically Informative Image Captioning with Character-Level Inference , author=. 2018 , archivePrefix=

  32. [33]

    2018 , eprint=

    Learning to Write with Cooperative Discriminators , author=. 2018 , eprint=

  33. [34]

    Proceedings of

    Ghazvininejad, Marjan and Shi, Xing and Priyadarshi, Jay and Knight, Kevin , editor =. Proceedings of. 2017 , pages =

  34. [35]

    2020 , doi =

    A Structure-Based Platform for Predicting Chemical Reactivity , journal =. 2020 , doi =

  35. [36]

    Prediction of chemical reaction yields using deep learning , author =. Mach. Learn. Sci. Technol. , volume =

  36. [37]

    Journal of Medicinal Chemistry , volume =

    Xiong, Zhaoping and Wang, Dingyan and Liu, Xiaohong and Zhong, Feisheng and Wan, Xiaozhe and Li, Xutong and Li, Zhaojun and Luo, Xiaomin and Chen, Kaixian and Jiang, Hualiang and Zheng, Mingyue , title =. Journal of Medicinal Chemistry , volume =. 2020 , doi =

  37. [38]

    Journal of Chemical Information and Modeling , volume =

    Yang, Kevin and Swanson, Kyle and Jin, Wengong and Coley, Connor and Eiden, Philipp and Gao, Hua and Guzman-Perez, Angel and Hopper, Timothy and Kelley, Brian and Mathea, Miriam and Palmer, Andrew and Settels, Volker and Jaakkola, Tommi and Jensen, Klavs and Barzilay, Regina , title =. Journal of Chemical Information and Modeling , volume =. 2019 , doi =

  38. [39]

    2018 , eprint =

    MoleculeNet: A Benchmark for Molecular Machine Learning , author =. 2018 , eprint =

  39. [40]

    Journal of Chemical Information and Modeling , volume =

    Li, Xiao and Chen, Lei and Cheng, Feixiong and Wu, Zengrui and Bian, Hanping and Xu, Congying and Li, Weihua and Liu, Guixia and Shen, Xu and Tang, Yun , title =. Journal of Chemical Information and Modeling , volume =. 2014 , doi =

  40. [41]

    Quantitative structure-activity relationship modeling of rat acute toxicity by oral exposure , author =. Chem. Res. Toxicol. , publisher =

  41. [42]

    Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1) , year =

    Therapeutics Data Commons: Machine Learning Datasets and Tasks for Drug Discovery and Development , author =. Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1) , year =

  42. [43]

    Acute oral toxicity , author =. Environ. Health Perspect. , publisher =

  43. [44]

    DeepTox : Toxicity Prediction using Deep Learning

    Mayr, Andreas and Klambauer, G. DeepTox : Toxicity Prediction using Deep Learning. Front. Environ. Sci. , volume =

  44. [45]

    Nature , volume =

    Toxicology for the twenty-first century , author =. Nature , volume =

  45. [46]

    Expert Opin

    Predicting the carcinogenicity of chemicals with alternative approaches: recent advances , author =. Expert Opin. Drug Metab. Toxicol. , volume =

  46. [47]

    The application of discovery toxicology and pathology towards the design of safer pharmaceutical lead candidates , author =. Nat. Rev. Drug Discov. , volume =

  47. [48]

    2023 , url =

    RDKit: Open-source cheminformatics , author =. 2023 , url =. doi:10.5281/zenodo.591637 , note =

  48. [49]

    1978 , doi =

    Hose — a novel substructure code , journal =. 1978 , doi =

  49. [50]

    Journal of Chemical Information and Modeling , volume =

    Ertl, Peter and Roggo, Silvio and Schuffenhauer, Ansgar , title =. Journal of Chemical Information and Modeling , volume =. 2008 , doi =

  50. [51]

    Journal of Cheminformatics , year=

    Ertl, Peter and Schuffenhauer, Ansgar , title=. Journal of Cheminformatics , year=

  51. [52]

    Improving Few- and Zero-Shot Reaction Template Prediction Using Modern Hopfield Networks , journal =

    Seidl, Philipp and Renz, Philipp and Dyubankova, Natalia and Neves, Paulo and Verhoeven, Jonas and Segler, Marwin and Wegner, J. Improving Few- and Zero-Shot Reaction Template Prediction Using Modern Hopfield Networks , journal =. 2022 , doi =

  52. [53]

    Journal of Chemical Information and Modeling , volume =

    Sacha, Mikołaj and Błaż, Mikołaj and Byrski, Piotr and Dąbrowski-Tumański, Paweł and Chromiński, Mikołaj and Loska, Rafał and Włodarczyk-Pruszyński, Paweł and Jastrzębski, Stanisław , title =. Journal of Chemical Information and Modeling , volume =. 2021 , doi =

  53. [54]

    JACS Au , volume =

    Chen, Shuan and Jung, Yousung , title =. JACS Au , volume =. 2021 , doi =

  54. [56]

    Attention is All you Need , url =

    Vaswani, Ashish and Shazeer, Noam and Parmar, Niki and Uszkoreit, Jakob and Jones, Llion and Gomez, Aidan N and Kaiser, ukasz and Polosukhin, Illia , booktitle =. Attention is All you Need , url =

  55. [58]

    Multistep retrosynthesis combining a disconnection aware triple transformer loop with a route penalty score guided tree search , volume =

    Kreutter, David and Reymond, Jean-Louis , year =. Multistep retrosynthesis combining a disconnection aware triple transformer loop with a route penalty score guided tree search , volume =. Chemical Science , publisher =. doi:10.1039/d3sc01604h , number =

  56. [59]

    2024 , eprint=

    RetroGFN: Diverse and Feasible Retrosynthesis using GFlowNets , author=. 2024 , eprint=

  57. [60]

    2022 , eprint=

    Mind the Retrosynthesis Gap: Bridging the divide between Single-step and Multi-step Retrosynthesis Prediction , author=. 2022 , eprint=

  58. [61]

    2024 , eprint=

    Alignment is Key for Applying Diffusion Models to Retrosynthesis , author=. 2024 , eprint=

  59. [63]

    2024 , eprint=

    DirectMultiStep: Direct Route Generation for Multi-Step Retrosynthesis , author=. 2024 , eprint=

  60. [64]

    PaRoutes: towards a framework for benchmarking retrosynthesis route predictions , volume =

    Genheden, Samuel and Bjerrum, Esben , year =. PaRoutes: towards a framework for benchmarking retrosynthesis route predictions , volume =. Digital Discovery , publisher =. doi:10.1039/d2dd00015f , number =

  61. [65]

    NeurIPS 2023 AI for Science Workshop , year=

    Re-evaluating Retrosynthesis Algorithms with Syntheseus , author=. NeurIPS 2023 AI for Science Workshop , year=

  62. [66]

    Torren-Peraire, Paula and Hassen, Alan Kai and Genheden, Samuel and Verhoeven, Jonas and Clevert, Djork-Arné and Preuss, Mike and Tetko, Igor V. , year=. Models Matter: the impact of single-step retrosynthesis on synthesis planning , volume=. Digital Discovery , publisher=. doi:10.1039/d3dd00252g , number=

  63. [67]

    2023 , volume =

    Liu, Songtao and Tu, Zhengkai and Xu, Minkai and Zhang, Zuobai and Lin, Lu and Ying, Rex and Tang, Jian and Zhao, Peilin and Wu, Dinghao , booktitle =. 2023 , volume =

  64. [68]

    Communications Chemistry , volume=

    Retrosynthetic planning with experience-guided Monte Carlo tree search , author=. Communications Chemistry , volume=. 2023 , publisher=

  65. [69]

    International conference on machine learning , pages=

    Retro*: learning retrosynthetic planning with neural guided A* search , author=. International conference on machine learning , pages=. 2020 , organization=

  66. [70]

    International Conference on Machine Learning , pages=

    Self-improved retrosynthetic planning , author=. International Conference on Machine Learning , pages=. 2021 , organization=

  67. [71]

    34th Conference on Neural Information Processing Systems (NeurIPS) , year=

    Barking up the right tree: an approach to search over molecule synthesis DAGs , author=. 34th Conference on Neural Information Processing Systems (NeurIPS) , year=

  68. [72]

    Digital Discovery , volume=

    Models matter: The impact of single-step retrosynthesis on synthesis planning , author=. Digital Discovery , volume=. 2024 , publisher=

  69. [73]

    Lowe, Daniel Mark , title =

  70. [74]

    Journal of Chemical Information and Modeling , volume=

    What’s what: The (nearly) definitive guide to reaction role assignment , author=. Journal of Chemical Information and Modeling , volume=. 2016 , publisher=

  71. [75]

    Predicting Organic Reaction Outcomes with

    Jin, Wengong and Coley, Connor and Barzilay, Regina and Jaakkola, Tommi , booktitle =. Predicting Organic Reaction Outcomes with

  72. [76]

    Schwaller, Philippe and Gaudin, Theophile and Lanyi, David and Bekas, Costas and Laino, Teodoro , journal=. “. 2018 , publisher=

  73. [77]

    2024 , note =

    eMolecules , title =. 2024 , note =

  74. [78]

    , keywords =

    Yu, Kevin and Roh, Jihye and Li, Ziang and Gao, Wenhao and Wang, Runzhong and Coley, Connor W. , keywords =. Double-Ended Synthesis Planning with Goal-Constrained Bidirectional Search , publisher =. 2024 , copyright =. doi:10.48550/ARXIV.2407.06334 , url =

  75. [79]

    2024 , note =

    NextMoveSoftware , title =. 2024 , note =

  76. [80]

    Proceedings of the 40 th International Conference on Machine Learning , year=

    Retrosynthetic Planning with Dual Value Networks , author=. Proceedings of the 40 th International Conference on Machine Learning , year=

  77. [81]

    36th Conference on Neural Information Processing Systems , year=

    GRASP: Navigating Retrosynthetic Planning with Goal-driven Policy , author=. 36th Conference on Neural Information Processing Systems , year=

  78. [82]

    Tango*: Constrained synthesis planning using chemically informed value functions, 2024

    Armstrong, Daniel and Joncev, Zlatko and Guo, Jeff and Schwaller, Philippe , keywords =. Tango*: Constrained synthesis planning using chemically informed value functions , publisher =. 2024 , copyright =. doi:10.48550/ARXIV.2412.03424 , url =

  79. [83]

    ICLR 2024 , year =

    Tripp, Austin and Maziarz, Krzysztof and Lewis, Sarah and Segler, Marwin and Hernández-Lobato, José Miguel , title =. ICLR 2024 , year =

  80. [84]

    NeurIPS 2022 AI for Science Workshop , year =

    Tripp, Austin and Maziarz, Krzysztof and Lewis, Sarah and Liu, Guoqing and Segler, Marwin , title =. NeurIPS 2022 AI for Science Workshop , year =

Showing first 80 references.