Recognition: unknown
ECHO: Efficient Chest X-ray Report Generation with One-step Block Diffusion
Pith reviewed 2026-05-10 17:51 UTC · model grok-4.3
The pith
A one-step block diffusion model generates clinically accurate chest X-ray reports eight times faster than autoregressive methods by distilling joint token dependencies.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
ECHO is a diffusion-based vision-language model for chest X-ray report generation that performs stable one-step-per-block inference via a Direct Conditional Distillation framework. The framework constructs unfactorized supervision from on-policy diffusion trajectories to encode joint token dependencies and thereby mitigates the mean-field bias of token-factorized denoisers. A Response-Asymmetric Diffusion training strategy further improves efficiency. Experiments show ECHO surpasses state-of-the-art autoregressive methods, raising RaTE by 64.33 percent and SemScore by 60.58 percent while delivering an eightfold inference speedup with no loss of clinical accuracy.
What carries the argument
Direct Conditional Distillation (DCD) framework that supplies unfactorized supervision drawn from on-policy diffusion trajectories to capture joint token dependencies during single-step block generation.
If this is right
- One-step-per-block inference becomes practical for high-quality report generation without coherence loss.
- Radiology workflows can handle substantially higher imaging volumes within existing time budgets.
- The Response-Asymmetric Diffusion strategy reduces training compute while preserving final model quality.
- Clinical accuracy remains comparable to autoregressive baselines on standard semantic and clinical metrics.
Where Pith is reading between the lines
- The same trajectory-distillation idea could shorten generation in other medical report tasks that involve longer sequences or additional modalities.
- If the method transfers to non-medical domains, diffusion-based language models might replace multi-step sampling across a wider range of sequence tasks.
- Real-time report drafting during image acquisition becomes feasible once latency drops to the reported level.
Load-bearing premise
The Direct Conditional Distillation framework successfully encodes joint token dependencies from on-policy trajectories to overcome mean-field bias in one-step generation without introducing new coherence failures not captured by the reported metrics.
What would settle it
A side-by-side clinical review in which radiologists assign lower factual consistency or coherence scores to ECHO reports than to autoregressive reports on the same set of chest X-ray cases with complex pathologies.
read the original abstract
Chest X-ray report generation (CXR-RG) has the potential to substantially alleviate radiologists' workload. However, conventional autoregressive vision--language models (VLMs) suffer from high inference latency due to sequential token decoding. Diffusion-based models offer a promising alternative through parallel generation, but they still require multiple denoising iterations. Compressing multi-step denoising to a single step could further reduce latency, but often degrades textual coherence due to the mean-field bias introduced by token-factorized denoisers. To address this challenge, we propose \textbf{ECHO}, an efficient diffusion-based VLM (dVLM) for chest X-ray report generation. ECHO enables stable one-step-per-block inference via a novel Direct Conditional Distillation (DCD) framework, which mitigates the mean-field limitation by constructing unfactorized supervision from on-policy diffusion trajectories to encode joint token dependencies. In addition, we introduce a Response-Asymmetric Diffusion (RAD) training strategy that further improves training efficiency while maintaining model effectiveness. Extensive experiments demonstrate that ECHO surpasses state-of-the-art autoregressive methods, improving RaTE and SemScore by \textbf{64.33\%} and \textbf{60.58\%} respectively, while achieving an \textbf{$8\times$} inference speedup without compromising clinical accuracy.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces ECHO, a diffusion-based vision-language model (dVLM) for chest X-ray report generation. It proposes a Direct Conditional Distillation (DCD) framework that constructs unfactorized supervision from on-policy diffusion trajectories to enable stable one-step-per-block inference while mitigating mean-field bias in token-factorized denoisers, along with a Response-Asymmetric Diffusion (RAD) training strategy for efficiency. The central claims are that ECHO surpasses state-of-the-art autoregressive VLMs, with reported gains of 64.33% on RaTE and 60.58% on SemScore, an 8× inference speedup, and no compromise to clinical accuracy.
Significance. If the empirical claims hold under rigorous verification, the work could meaningfully advance practical deployment of VLMs in radiology by enabling low-latency parallel report generation. The core technical idea—distilling joint token dependencies from on-policy trajectories rather than relying on factorized denoising—is a targeted response to a known limitation of one-step diffusion and merits further exploration if properly validated.
major comments (3)
- [Abstract, §4] Abstract and §4 (Experiments): The reported metric gains (64.33% RaTE, 60.58% SemScore) and 8× speedup are presented without any description of the experimental protocol, baseline implementations, number of runs, statistical significance tests, or the procedure used to verify clinical accuracy (e.g., radiologist review or specific clinical metrics). This absence prevents assessment of whether the improvements are robust or reproducible.
- [§3.2] §3.2 (Direct Conditional Distillation): The manuscript asserts that DCD encodes joint token dependencies from on-policy trajectories to overcome mean-field bias, yet provides no ablation or targeted diagnostic (e.g., analysis of contradictory clinical findings or broken logical chains across blocks) demonstrating that the unfactorized supervision actually prevents coherence failures that aggregate metrics like RaTE and SemScore may miss.
- [§4.3] §4.3 (Ablations): There are no ablation studies isolating the contribution of DCD from RAD, architecture modifications, or the on-policy sampling procedure itself. Without these, it is impossible to determine whether the performance gains are attributable to the proposed distillation framework or to other uncontrolled factors.
minor comments (2)
- [§3] The notation and algorithmic description of the DCD objective and the on-policy trajectory sampling could be made more precise with explicit equations or pseudocode.
- [§2] The paper would benefit from additional citations to recent one-step diffusion and distillation literature for text generation to better situate the novelty of DCD.
Simulated Author's Rebuttal
We sincerely thank the referee for their detailed and constructive feedback on our manuscript. We appreciate the acknowledgment of the potential significance of our work for practical VLM deployment in radiology. Below, we provide point-by-point responses to the major comments and describe the revisions we plan to implement.
read point-by-point responses
-
Referee: [Abstract, §4] Abstract and §4 (Experiments): The reported metric gains (64.33% RaTE, 60.58% SemScore) and 8× speedup are presented without any description of the experimental protocol, baseline implementations, number of runs, statistical significance tests, or the procedure used to verify clinical accuracy (e.g., radiologist review or specific clinical metrics). This absence prevents assessment of whether the improvements are robust or reproducible.
Authors: We agree that the experimental details were not adequately described in the abstract and Section 4. In the revised version, we will provide a full account of the experimental protocol, including how baselines were implemented, the number of independent runs performed, results of statistical significance tests, and the specific procedure used to verify clinical accuracy, which included review by board-certified radiologists confirming preservation of clinical fidelity. revision: yes
-
Referee: [§3.2] §3.2 (Direct Conditional Distillation): The manuscript asserts that DCD encodes joint token dependencies from on-policy trajectories to overcome mean-field bias, yet provides no ablation or targeted diagnostic (e.g., analysis of contradictory clinical findings or broken logical chains across blocks) demonstrating that the unfactorized supervision actually prevents coherence failures that aggregate metrics like RaTE and SemScore may miss.
Authors: We acknowledge the value of targeted diagnostics to validate the mechanism of DCD. Although aggregate metrics indicate improved coherence, we will incorporate additional analyses in the revised manuscript, including ablations and case studies examining contradictory clinical findings and logical consistency across generated blocks, to directly demonstrate the benefits of the unfactorized supervision from on-policy trajectories. revision: yes
-
Referee: [§4.3] §4.3 (Ablations): There are no ablation studies isolating the contribution of DCD from RAD, architecture modifications, or the on-policy sampling procedure itself. Without these, it is impossible to determine whether the performance gains are attributable to the proposed distillation framework or to other uncontrolled factors.
Authors: We recognize that the existing ablations do not sufficiently isolate the individual contributions. We will perform and report additional ablation studies in the revised §4.3 that systematically vary DCD, RAD, and the on-policy sampling procedure while controlling for other factors, thereby clarifying the source of the observed performance improvements. revision: yes
Circularity Check
No significant circularity in derivation chain
full rationale
The paper presents ECHO as a new diffusion-based VLM using the proposed Direct Conditional Distillation (DCD) framework and Response-Asymmetric Diffusion (RAD) strategy. These are introduced as novel training procedures that construct unfactorized supervision from on-policy trajectories and improve efficiency, respectively. The central claims of improved RaTE/SemScore and 8x speedup are supported by empirical experiments rather than reducing by construction to fitted inputs, self-definitions, or self-citation chains. No equations or sections in the provided text exhibit the target result being equivalent to its own inputs; the derivation remains self-contained against external benchmarks.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption Mean-field bias in token-factorized denoisers degrades textual coherence in one-step generation
- ad hoc to paper On-policy diffusion trajectories provide sufficient joint token dependency information for distillation
Reference graph
Works this paper leans on
-
[1]
Block diffusion: Interpolating between autoregressive and diffusion language models
Marianne Arriola, Aaron Gokaslan, Justin T Chiu, Zhihan Yang, Zhixuan Qi, Jiaqi Han, Subham Sekhar Sahoo, and Volodymyr Kuleshov. Block diffusion: Interpolating between autoregressive and diffusion language models. InThe Thirteenth International Conference on Learning Representations, 2025
2025
-
[2]
Structured denoising diffusion models in discrete state-spaces.Advances in neural information processing systems, 2021
Jacob Austin, Daniel D Johnson, Jonathan Ho, Daniel Tarlow, and Rianne Van Den Berg. Structured denoising diffusion models in discrete state-spaces.Advances in neural information processing systems, 2021
2021
-
[3]
Shuai Bai, Yuxuan Cai, Ruizhe Chen, Keqin Chen, Xionghui Chen, Zesen Cheng, Lianghao Deng, Wei Ding, Chang Gao, Chunjiang Ge, et al. Qwen3-vl technical report.arXiv preprint arXiv:2511.21631, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[4]
Tiwei Bie, Maosong Cao, Kun Chen, Lun Du, Mingliang Gong, Zhuochen Gong, Yanmei Gu, Jiaqi Hu, Zenan Huang, Zhenzhong Lan, et al. Llada2. 0: Scaling up diffusion language models to 100b.arXiv preprint arXiv:2512.15745, 2025
work page internal anchor Pith review arXiv 2025
-
[5]
Thomas, Jerome Ku, Federico Berto, Jae Myung Kim, Garyk Brixi, Eric Nguyen, Stefano Massaroli, and Michael Poli
Keshigeyan Chandrasegaran, Armin W. Thomas, Jerome Ku, Federico Berto, Jae Myung Kim, Garyk Brixi, Eric Nguyen, Stefano Massaroli, and Michael Poli. Rnd1: Simple, scalable ar-to-diffusion conversion. 2025
2025
-
[6]
Towards injecting medical visual knowledge into multimodal llms at scale
Junying Chen, Chi Gui, Ruyi Ouyang, Anningzhe Gao, Shunian Chen, Guiming Hardy Chen, Xidong Wang, Zhenyang Cai, Ke Ji, Xiang Wan, et al. Towards injecting medical visual knowledge into multimodal llms at scale. InProceedings of the 2024 conference on empirical methods in natural language processing, 2024
2024
-
[7]
Zhihong Chen, Maya Varma, Justin Xu, Magdalini Paschali, Dave Van Veen, Andrew Johnston, Alaa Youssef, Louis Blankemeier, Christian Bluethgen, Stephan Altmayer, et al. A vision-language foundation model to enhance efficiency of chest x-ray interpretation.arXiv preprint arXiv:2401.12208, 2024
-
[8]
dparallel: Learnable parallel decoding for dllms.arXiv preprint arXiv:2509.26488,
Zigeng Chen, Gongfan Fang, Xinyin Ma, Ruonan Yu, and Xinchao Wang. dparallel: Learnable parallel decoding for dllms.arXiv preprint arXiv:2509.26488, 2025
-
[9]
Shuang Cheng, Yuhua Jiang, Zineng Zhou, Dawei Liu, Wang Tao, Linfeng Zhang, Biqing Qi, and Bowen Zhou. Sdar-vl: Stable and efficient block-wise diffusion for vision-language understanding.arXiv preprint arXiv:2512.14068, 2025
-
[10]
Speculative diffusion decoding: Accelerating language generation through diffusion
Jacob K Christopher, Brian R Bartoldson, Tal Ben-Nun, Michael Cardei, Bhavya Kailkhura, and Ferdinando Fioretto. Speculative diffusion decoding: Accelerating language generation through diffusion. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volum...
2025
-
[11]
Preparing a collection of radiology examinations for distribution and retrieval.Journal of the American Medical Informatics Association, 2016
Dina Demner-Fushman, Marc D Kohli, Marc B Rosenman, Sonya E Shooshan, Laritza Rodriguez, Sameer Antani, George R Thoma, and Clement J McDonald. Preparing a collection of radiology examinations for distribution and retrieval.Journal of the American Medical Informatics Association, 2016
2016
-
[12]
Beyond autoregression: Fast llms via self-distillation through time, 2025
Justin Deschenaux and Caglar Gulcehre. Beyond autoregression: Fast llms via self-distillation through time, 2025
2025
-
[13]
Xuanzhao Dong, Wenhui Zhu, Xiwen Chen, Zhipeng Wang, Peijie Qiu, Shao Tang, Xin Li, and Yalin Wang. Llada-medv: Exploring large language diffusion models for biomedical image understanding.arXiv preprint arXiv:2508.01617, 2025
-
[14]
Unifying autoregressive and diffusion-based sequence generation
Nima Fathi, Torsten Scholak, and Pierre-Andre Noel. Unifying autoregressive and diffusion-based sequence generation. InICLR 2025 Workshop on Deep Generative Model in Machine Learning: Theory, Principle and Efficacy, 2025
2025
-
[15]
Scaling diffusion language models via adaptation from autoregressive models
Shansan Gong, Shivam Agarwal, Yizhe Zhang, Jiacheng Ye, Lin Zheng, Mukai Li, Chenxin An, Peilin Zhao, Wei Bi, Jiawei Han, et al. Scaling diffusion language models via adaptation from autoregressive models. InThe Thirteenth International Conference on Learning Representations, 2025
2025
-
[16]
Gemini 3 pro model card, 2025
Google DeepMind. Gemini 3 pro model card, 2025
2025
-
[17]
Ssd-lm: Semi-autoregressive simplex-based diffusion lan- guagemodelfortextgenerationandmodularcontrol
Xiaochuang Han, Sachin Kumar, and Yulia Tsvetkov. Ssd-lm: Semi-autoregressive simplex-based diffusion lan- guagemodelfortextgenerationandmodularcontrol. InProceedings of the 61st Annual Meeting of the Association for Computational Linguistics, 2023
2023
-
[18]
Accelerating diffusion language model inference via efficient kv caching and guided diffusion.arXiv e-prints, 2025
Zhanqiu Hu, Jian Meng, Yash Akhauri, Mohamed S Abdelfattah, Jae-sun Seo, Zhiru Zhang, and Udit Gupta. Accelerating diffusion language model inference via efficient kv caching and guided diffusion.arXiv e-prints, 2025. 13
2025
-
[19]
Aaron Hurst, Adam Lerer, Adam P Goucher, Adam Perelman, Aditya Ramesh, Aidan Clark, AJ Ostrow, Akila Welihinda, Alan Hayes, Alec Radford, et al. Gpt-4o system card.arXiv preprint arXiv:2410.21276, 2024
work page internal anchor Pith review Pith/arXiv arXiv 2024
-
[20]
Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison
Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Henrik Marklund, Be- hzad Haghgoo, Robyn Ball, Katie Shpanskaya, et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. InProceedings of the AAAI conference on artificial intelligence, 2019
2019
-
[21]
arXiv preprint arXiv:2510.08668 (2025)
Songtao Jiang, Yuan Wang, Sibo Song, Tianxiang Hu, Chenyi Zhou, Bin Pu, Yan Zhang, Zhibo Yang, Yang Feng, Joey Tianyi Zhou, et al. Hulu-med: A transparent generalist model towards holistic medical vision-language understanding.arXiv preprint arXiv:2510.08668, 2025
-
[22]
Mimic-cxr, a de-identified publicly available database of chest radiographs with free-text reports.Scientific data, 2019
Alistair EW Johnson, Tom J Pollard, Seth J Berkowitz, Nathaniel R Greenbaum, Matthew P Lungren, Chih- ying Deng, Roger G Mark, and Steven Horng. Mimic-cxr, a de-identified publicly available database of chest radiographs with free-text reports.Scientific data, 2019
2019
-
[23]
Cxr-llava: a multimodal large language model for interpreting chest x-ray images.European Radiology, 2025
Seowoo Lee, Jiwon Youn, Hyungjin Kim, Mansu Kim, and Soon Ho Yoon. Cxr-llava: a multimodal large language model for interpreting chest x-ray images.European Radiology, 2025
2025
-
[24]
Llava-onevision: Easy visual task transfer.Transactions on Machine Learning Research, 2024
Bo Li, Yuanhan Zhang, Dong Guo, Renrui Zhang, Feng Li, Hao Zhang, Kaichen Zhang, Peiyuan Zhang, Yanwei Li, Ziwei Liu, et al. Llava-onevision: Easy visual task transfer.Transactions on Machine Learning Research, 2024
2024
-
[25]
Llava-med: Training a large language-and-vision assistant for biomedicine in one day
Chunyuan Li, Cliff Wong, Sheng Zhang, Naoto Usuyama, Haotian Liu, Jianwei Yang, Tristan Naumann, Hoifung Poon, and Jianfeng Gao. Llava-med: Training a large language-and-vision assistant for biomedicine in one day. Advances in Neural Information Processing Systems, 2023
2023
-
[26]
LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models
Feng Li, Renrui Zhang, Hao Zhang, Yuanhan Zhang, Bo Li, Wei Li, Zejun Ma, and Chunyuan Li. Llava-next- interleave: Tackling multi-image, video, and 3d in large multimodal models.arXiv preprint arXiv:2407.07895, 2024
work page internal anchor Pith review arXiv 2024
-
[27]
Guanghao Li, Zhihui Fu, Min Fang, Qibin Zhao, Ming Tang, Chun Yuan, and Jun Wang. Diffuspec: Unlocking diffusion language models for speculative decoding.arXiv preprint arXiv:2510.02358, 2025
-
[28]
Lavida: A large diffusion language model for multimodal under- standing
Shufan Li, Konstantinos Kallidromitis, Hritik Bansal, Akash Gokul, Yusuke Kato, Kazuki Kozuka, Jason Kuen, Zhe Lin, Kai-Wei Chang, and Aditya Grover. Lavida: A large diffusion language model for multimodal under- standing. InThe Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025
2025
-
[29]
Yihao Liang, Ze Wang, Hao Chen, Ximeng Sun, Jialian Wu, Xiaodong Yu, Jiang Liu, Emad Barsoum, Zicheng Liu, and Niraj K Jha. Cd4lm: Consistency distillation and adaptive decoding for diffusion language models. arXiv preprint arXiv:2601.02236, 2026
-
[30]
Rouge: A package for automatic evaluation of summaries
Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. InText summarization branches out, 2004
2004
-
[31]
Visual instruction tuning.Advances in neural information processing systems, 2023
Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction tuning.Advances in neural information processing systems, 2023
2023
-
[32]
Zhiyuan Liu, Yicun Yang, Yaojie Zhang, Junjie Chen, Chang Zou, Qingyuan Wei, Shaobo Wang, and Lin- feng Zhang. dllm-cache: Accelerating diffusion large language models with adaptive caching.arXiv preprint arXiv:2506.06295, 2025
-
[33]
Large language diffusion models
Shen Nie, Fengqi Zhu, Zebin You, Xiaolu Zhang, Jingyang Ou, Jun Hu, JUN ZHOU, Yankai Lin, Ji-Rong Wen, and Chongxuan Li. Large language diffusion models. InThe Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025
2025
-
[34]
Medvlm-r1: Incentivizing medical reasoning capability of vision-language models (vlms) via reinforcement learning
Jiazhen Pan, Che Liu, Junde Wu, Fenglin Liu, Jiayuan Zhu, Hongwei Bran Li, Chen Chen, Cheng Ouyang, and Daniel Rueckert. Medvlm-r1: Incentivizing medical reasoning capability of vision-language models (vlms) via reinforcement learning. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, 2025
2025
-
[35]
Chantal Pellegrini, Ege Özsoy, Benjamin Busam, Nassir Navab, and Matthias Keicher. Radialog: A large vision- language model for radiology report generation and conversational assistance.arXiv preprint arXiv:2311.18681, 2023
-
[36]
Yu-Yang Qian, Junda Su, Lanxiang Hu, Peiyuan Zhang, Zhijie Deng, Peng Zhao, and Hao Zhang. d3llm: Ultra- fast diffusion llm using pseudo-trajectory distillation.arXiv preprint arXiv:2601.07568, 2026. 14
-
[37]
arXiv preprint arXiv:2404.18416 (2024)
Khaled Saab, Tao Tu, Wei-Hung Weng, Ryutaro Tanno, David Stutz, Ellery Wulczyn, Fan Zhang, Tim Strother, Chunjong Park, Elahe Vedadi, et al. Capabilities of gemini models in medicine.arXiv preprint arXiv:2404.18416, 2024
-
[38]
Simpleandeffectivemaskeddiffusionlanguagemodels.Advances in Neural Information Processing Systems, 2024
Subham Sahoo, Marianne Arriola, Yair Schiff, Aaron Gokaslan, Edgar Marroquin, Justin Chiu, Alexander Rush, andVolodymyrKuleshov. Simpleandeffectivemaskeddiffusionlanguagemodels.Advances in Neural Information Processing Systems, 2024
2024
-
[39]
Andrew Sellergren, Sahar Kazemzadeh, Tiam Jaroensri, Atilla Kiraly, Madeleine Traverse, Timo Kohlberger, Shawn Xu, Fayaz Jamil, Cían Hughes, Charles Lau, et al. Medgemma technical report.arXiv preprint arXiv:2507.05201, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[40]
Simplified and generalized masked diffusion for discrete data.Advances in neural information processing systems, 2024
Jiaxin Shi, Kehang Han, Zhe Wang, Arnaud Doucet, and Michalis Titsias. Simplified and generalized masked diffusion for discrete data.Advances in neural information processing systems, 2024
2024
-
[41]
Combining automatic labelers and expert annotations for accurate radiology report labeling using bert
Akshay Smit, Saahil Jain, Pranav Rajpurkar, Anuj Pareek, Andrew Y Ng, and Matthew Lungren. Combining automatic labelers and expert annotations for accurate radiology report labeling using bert. InProceedings of the 2020 conference on empirical methods in natural language processing (EMNLP), pages 1500–1519, 2020
2020
-
[42]
Gemini: A Family of Highly Capable Multimodal Models
Gemini Team, Rohan Anil, Sebastian Borgeaud, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalk- wyk, Andrew M Dai, Anja Hauth, Katie Millican, et al. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805, 2023
work page internal anchor Pith review Pith/arXiv arXiv 2023
-
[43]
Xraygpt: Chest radiographs summarization using large medical vision-language models
Omkar Chakradhar Thawakar, Abdelrahman M Shaker, Sahal Shaji Mullappilly, Hisham Cholakkal, Rao Muham- mad Anwer, Salman Khan, Jorma Laaksonen, and Fahad Khan. Xraygpt: Chest radiographs summarization using large medical vision-language models. InProceedings of the 23rd workshop on biomedical natural language processing, 2024
2024
-
[44]
Towards generalist biomedical ai.Nejm Ai, 2024
Tao Tu, Shekoofeh Azizi, Danny Driess, Mike Schaekermann, Mohamed Amin, Pi-Chuan Chang, Andrew Carroll, Charles Lau, Ryutaro Tanno, Ira Ktena, et al. Towards generalist biomedical ai.Nejm Ai, 2024
2024
-
[45]
Cider: Consensus-based image description evaluation
Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. Cider: Consensus-based image description evaluation. InProceedings of the IEEE conference on computer vision and pattern recognition, 2015
2015
-
[46]
Qwen2-VL: Enhancing Vision-Language Model's Perception of the World at Any Resolution
Peng Wang, Shuai Bai, Sinan Tan, Shijie Wang, Zhihao Fan, Jinze Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, et al. Qwen2-vl: Enhancing vision-language model’s perception of the world at any resolution.arXiv preprint arXiv:2409.12191, 2024
work page internal anchor Pith review Pith/arXiv arXiv 2024
-
[47]
Diffusion LLMs Can Do Faster- Than-AR Inference via Discrete Diffusion Forcing, August 2025c
Xu Wang, Chenkai Xu, Yijie Jin, Jiachun Jin, Hao Zhang, and Zhijie Deng. Diffusion llms can do faster-than-ar inference via discrete diffusion forcing.arXiv preprint arXiv:2508.09192, 2025
-
[48]
Towards generalist foundation model for radiology by leveraging web-scale 2d&3d medical data.Nature Communications, 2025
Chaoyi Wu, Xiaoman Zhang, Ya Zhang, Hui Hui, Yanfeng Wang, and Weidi Xie. Towards generalist foundation model for radiology by leveraging web-scale 2d&3d medical data.Nature Communications, 2025
2025
-
[49]
Fast-dllm v2: Efficient block-diffusion llm.arXiv preprint arXiv:2509.26328, 2025a
Chengyue Wu, Hao Zhang, Shuchen Xue, Shizhe Diao, Yonggan Fu, Zhijian Liu, Pavlo Molchanov, Ping Luo, Song Han, and Enze Xie. Fast-dllm v2: Efficient block-diffusion llm.arXiv preprint arXiv:2509.26328, 2025
-
[50]
Chengyue Wu, Hao Zhang, Shuchen Xue, Zhijian Liu, Shizhe Diao, Ligeng Zhu, Ping Luo, Song Han, and Enze Xie. Fast-dllm: Training-free acceleration of diffusion llm by enabling kv cache and parallel decoding.arXiv preprint arXiv:2505.22618, 2025
-
[51]
Energy-based diffusion language models for text generation
Minkai Xu, Tomas Geffner, Karsten Kreis, Weili Nie, Yilun Xu, Jure Leskovec, Stefano Ermon, and Arash Vahdat. Energy-based diffusion language models for text generation. InThe Thirteenth International Conference on Learning Representations, 2025
2025
-
[52]
Weiwen Xu, Hou Pong Chan, Long Li, Mahani Aljunied, Ruifeng Yuan, Jianyu Wang, Chenghao Xiao, Guizhen Chen, Chaoqun Liu, Zhaodonghui Li, et al. Lingshu: A generalist foundation model for unified multimodal medical understanding and reasoning.arXiv preprint arXiv:2506.07044, 2025
-
[53]
An Yang, Anfeng Li, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Gao, Chengen Huang, Chenxu Lv, et al. Qwen3 technical report.arXiv preprint arXiv:2505.09388, 2025
work page internal anchor Pith review Pith/arXiv arXiv 2025
-
[54]
Medxchat: A unified multimodal large language model framework towards cxrs understanding and generation
Ling Yang, Zhanyu Wang, Zhenghao Chen, Xinyu Liang, and Luping Zhou. Medxchat: A unified multimodal large language model framework towards cxrs understanding and generation. In2025 IEEE 22nd International Symposium on Biomedical Imaging (ISBI), 2025. 15
2025
-
[55]
Dream 7B: Diffusion Large Language Models
Jiacheng Ye, Zhihui Xie, Lin Zheng, Jiahui Gao, Zirui Wu, Xin Jiang, Zhenguo Li, and Lingpeng Kong. Dream 7b: Diffusion large language models.arXiv preprint arXiv:2508.15487, 2025
work page internal anchor Pith review arXiv 2025
-
[56]
Redi: Rectified discrete flow
Jaehoon Yoo, Wonjung Kim, and Seunghoon Hong. Redi: Rectified discrete flow. InThe Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025
2025
-
[57]
Zebin You, Shen Nie, Xiaolu Zhang, Jun Hu, Jun Zhou, Zhiwu Lu, Ji-Rong Wen, and Chongxuan Li. Llada-v: Large language diffusion models with visual instruction tuning.arXiv preprint arXiv:2505.16933, 2025
-
[58]
Dimple: Discrete diffusion multimodal large language model with parallel decoding
Runpeng Yu, Xinyin Ma, and Xinchao Wang. Dimple: Discrete diffusion multimodal large language model with parallel decoding.arXiv preprint arXiv:2505.16990, 2025
-
[59]
A generalist vision–language foundation model for diverse biomedical tasks.Nature medicine, 2024
Kai Zhang, Rong Zhou, Eashan Adhikarla, Zhiling Yan, Yixin Liu, Jun Yu, Zhengliang Liu, Xun Chen, Brian D Davison, Hui Ren, et al. A generalist vision–language foundation model for diverse biomedical tasks.Nature medicine, 2024
2024
-
[60]
Tunyu Zhang, Xinxi Zhang, Ligong Han, Haizhou Shi, Xiaoxiao He, Zhuowei Li, Hao Wang, Kai Xu, Akash Srivastava, Vladimir Pavlovic, et al. T3d: Few-step diffusion language models via trajectory self-distillation with direct discriminative optimization.arXiv preprint arXiv:2602.12262, 2026
-
[61]
Acosta, Josh Miller, Ouwen Huang, and Pranav Rajpurkar
Xiaoman Zhang, Julián N Acosta, Josh Miller, Ouwen Huang, and Pranav Rajpurkar. Rexgradient-160k: A large-scale publicly available dataset of chest radiographs with free-text reports.arXiv preprint arXiv:2505.00228, 2025
-
[62]
Variational masked diffusion models.arXiv preprint arXiv:2510.23606, 2025
Yichi Zhang, Alex Schwing, and Zhizhen Zhao. Variational masked diffusion models.arXiv preprint arXiv:2510.23606, 2025
-
[63]
Review this chest X-ray and write a report. Use this format: Findings: {}, Impression: {}
Weike Zhao, Chaoyi Wu, Xiaoman Zhang, Ya Zhang, Yanfeng Wang, and Weidi Xie. Ratescore: A metric for radiology report generation. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024. 16 Appendix This supplementary material provides additional details and results to complement the main paper, organized as follows....
2024
-
[64]
Findings
Incomplete "Findings" sections—often omitting descriptions of normal (negative) findings
-
[65]
Findings
Inclusion of interpretive or inferential statements within the "Findings" section that should belong in the "Impression" section. Your input will be an original report containing both "Findings" and "Impression" sections. Your output must be a standardized report in JSON format, without any additional explanations or comments. Standardization requirements:
-
[66]
- Retain all organ/structure descriptions present in the original report
Findings: - Structure the findings in the following anatomical order: thorax, mediastinum and trachea, lung fields, cardiac silhouette, hila, diaphragm and costophrenic angles, and bony structures. - Retain all organ/structure descriptions present in the original report. - Retain any additional relevant details (e.g., presence of tubes or lines). - Retain...
-
[67]
findings
Impression: - Retain the original impression. - You may add diagnostic conclusions or clinical recommendations based on the standardized findings and original impression. Below are examples of standardized negative reports for reference: Standardized Negative Report (1): Findings: The thorax is symmetric bilaterally. The mediastinum and trachea are midlin...
-
[68]
- FINDINGS must contain only descriptive radiological observations
Extract and summarize the objective imaging observations into the FINDINGS section. - FINDINGS must contain only descriptive radiological observations. - Do NOT include any diagnostic conclusions in FINDINGS
-
[69]
- IMPRESSION should reflect the clinician’s overall diagnostic assessment
Extract and summarize the diagnostic interpretation into the IMPRESSION section. - IMPRESSION should reflect the clinician’s overall diagnostic assessment. - Do NOT introduce any new diagnoses that are not explicitly stated in the original report
-
[70]
increased bronchovascular markings
Translate the content accurately into English using standard radiology terminology. - Avoid literal word-by-word translation. - Use clinically accepted expressions (e.g., “increased bronchovascular markings” instead of “lung texture thickened”)
-
[71]
suggestive of
Preserve all expressions of uncertainty (e.g., “suggestive of”, “cannot exclude”, “likely”, “consider”). - Do NOT convert uncertain statements into definitive conclusions
-
[72]
- Leave the missing section empty if necessary
If the original Chinese report contains only FINDINGS or only IMPRESSION, do NOT fabricate the missing section. - Leave the missing section empty if necessary
-
[73]
Standardized Output Format (strict): FINDINGS: <content> IMPRESSION: <content>
-
[74]
Wrap your final output strictly within: ```output <your standardized report> ```
-
[75]
- Do NOT include any explanation, notes, or additional commentary
Output ONLY the standardized English report. - Do NOT include any explanation, notes, or additional commentary. Here is the Chinese medical report to be processed: ```input {content} ``` - If the original content is ambiguous, incomplete, or poorly structured, you must translate it faithfully without attempting to correct or improve it. here is the output...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.