Recognition: unknown
HypEHR: Hyperbolic Modeling of Electronic Health Records for Efficient Question Answering
Pith reviewed 2026-05-09 23:49 UTC · model grok-4.3
The pith
A Lorentzian hyperbolic model embeds electronic health records to answer clinical questions nearly as accurately as large language models but with far fewer parameters.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
HypEHR is a compact model that embeds medical codes, visits, and questions into hyperbolic space using Lorentzian geometry, pretrains via next-visit diagnosis prediction with hierarchy-aware regularization to align with the ICD ontology, and answers queries through geometry-consistent cross-attention and type-specific pointer heads, approaching the accuracy of LLM methods on MIMIC-IV-based EHR-QA benchmarks while using far fewer parameters.
What carries the argument
Lorentzian hyperbolic embeddings with hierarchy-aware regularization and geometry-consistent cross-attention for question answering.
If this is right
- Clinical question-answering systems become feasible to run on modest hardware without relying on large language model APIs.
- Hierarchical relationships in medical ontologies are explicitly respected rather than learned implicitly through scale.
- Pretraining on next-visit prediction transfers useful structure for downstream query tasks.
- Pointer heads allow direct grounding of answers in specific patient records or codes.
Where Pith is reading between the lines
- Similar hyperbolic techniques could apply to other tree-structured domains such as biological taxonomies or legal case hierarchies.
- Geometric distances in the embedding space might provide natural measures of clinical similarity or uncertainty for explanations.
- The method suggests that domain-specific geometry can reduce the need for massive parameter counts in specialized applications.
Load-bearing premise
Medical ontologies and patient trajectories exhibit hyperbolic geometry that a Lorentzian embedding plus hierarchy-aware regularization can capture sufficiently for question answering.
What would settle it
Running the same benchmarks with a Euclidean version of the model and showing that accuracy drops significantly below the hyperbolic version at the same parameter count.
Figures
read the original abstract
Electronic health record (EHR) question answering is often handled by LLM-based pipelines that are costly to deploy and do not explicitly leverage the hierarchical structure of clinical data. Motivated by evidence that medical ontologies and patient trajectories exhibit hyperbolic geometry, we propose HypEHR, a compact Lorentzian model that embeds codes, visits, and questions in hyperbolic space and answers queries via geometry-consistent cross-attention with type-specific pointer heads. HypEHR is pretrained with next-visit diagnosis prediction and hierarchy-aware regularization to align representations with the ICD ontology. On two MIMIC-IV-based EHR-QA benchmarks, HypEHR approaches LLM-based methods while using far fewer parameters. Our code is publicly available at https://github.com/yuyuliu11037/HypEHR.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes HypEHR, a compact Lorentzian model for electronic health record question answering. It embeds codes, visits, and questions in hyperbolic space, performs geometry-consistent cross-attention with type-specific pointer heads, and pretrains via next-visit diagnosis prediction plus hierarchy-aware regularization aligned to the ICD ontology. The central claim is that on two MIMIC-IV-based EHR-QA benchmarks HypEHR approaches LLM-based methods while using far fewer parameters; code is released at https://github.com/yuyuliu11037/HypEHR.
Significance. If the performance claims hold after verification, the work would demonstrate a parameter-efficient alternative to LLMs for clinical QA by explicitly exploiting hyperbolic geometry for hierarchical medical data. The public code release supports reproducibility and is a clear strength.
major comments (2)
- [§4 Experiments] §4 Experiments: benchmark results are summarized but the text provides no concrete accuracy/F1 numbers, error bars, statistical tests, or direct comparisons to the specific LLM baselines referenced, so the headline claim that HypEHR 'approaches' LLM performance cannot be evaluated.
- [§4.3 Ablation Studies] §4.3 Ablation Studies: no controlled Euclidean counterpart (identical architecture, same hierarchy-aware regularization, same parameter budget) is reported. This is load-bearing for the claim that Lorentzian geometry plus hierarchy regularization drives the efficiency, because the observed scores could arise from model compactness or the pointer-head design alone.
minor comments (2)
- [Abstract] Abstract: the phrase 'approaches LLM-based methods' should name the concrete LLMs and report the exact performance delta.
- [§3.2] §3.2: the weighting coefficient between the next-visit loss and the hierarchy-aware regularization term is not stated; add the value or schedule used.
Simulated Author's Rebuttal
We thank the referee for the constructive and detailed feedback. We address each major comment point by point below. We have revised the manuscript to strengthen the experimental reporting and will incorporate additional ablations as described.
read point-by-point responses
-
Referee: [§4 Experiments] §4 Experiments: benchmark results are summarized but the text provides no concrete accuracy/F1 numbers, error bars, statistical tests, or direct comparisons to the specific LLM baselines referenced, so the headline claim that HypEHR 'approaches' LLM performance cannot be evaluated.
Authors: We agree that the narrative in §4 would benefit from explicit quantitative details to allow direct evaluation of the claims. Tables 1 and 2 already report accuracy, F1 scores, and comparisons against the referenced LLM baselines (including GPT-4 and Med-PaLM variants) on the two MIMIC-IV EHR-QA benchmarks. In the revised manuscript we will expand the text of §4 to quote the key numbers, report standard deviations across multiple random seeds as error bars, and include statistical significance tests (paired t-tests with p-values) against the LLM baselines. This change will make the 'approaches LLM performance' claim directly verifiable from the text while leaving the underlying results unchanged. revision: yes
-
Referee: [§4.3 Ablation Studies] §4.3 Ablation Studies: no controlled Euclidean counterpart (identical architecture, same hierarchy-aware regularization, same parameter budget) is reported. This is load-bearing for the claim that Lorentzian geometry plus hierarchy regularization drives the efficiency, because the observed scores could arise from model compactness or the pointer-head design alone.
Authors: This observation is correct and highlights a gap in isolating the contribution of the Lorentzian geometry. Our current ablations remove the hierarchy-aware regularization or the hyperbolic components individually, but do not include a fully matched Euclidean model with identical architecture, regularization, and parameter count. To address this, we will add a controlled Euclidean ablation (same pointer heads, same hierarchy regularization applied in Euclidean space, same total parameters) to the revised §4.3. The new results will be presented alongside the existing ablations to better substantiate the role of Lorentzian geometry for the hierarchical EHR data. revision: yes
Circularity Check
No circularity: model architecture and empirical results are independent of inputs
full rationale
The paper defines a new Lorentzian embedding architecture with cross-attention and pointer heads, pretrained via next-visit diagnosis prediction plus hierarchy-aware regularization on ICD ontology. These are standard training choices applied to the proposed model rather than re-derivations. Reported QA performance on MIMIC-IV benchmarks is measured empirically after training; no equation or claim reduces a 'prediction' to a fitted parameter by construction, nor does any load-bearing step rely on self-citation chains or imported uniqueness theorems. The derivation from model definition to benchmark scores is self-contained and externally falsifiable.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption Medical ontologies and patient trajectories exhibit hyperbolic geometry
Reference graph
Works this paper leans on
-
[1]
Agarwal, Dhruv and Arivazhagan, Manoj Ghuhan and Das, Rajarshi and Swamy, Sandesh and Khosla, Sopan and Gangadharaiah, Rashmi , year =
-
[2]
Liu, Yuyu and Yang, Jiannan and Yu, Ziyang and Pan, Weishen and Wang, Fei and Ma, Tengfei , month = jan, year =. Efficient. doi:10.48550/arXiv.2601.14653 , abstract =
-
[3]
Liu, Songtao and Peng, Hongwu and Zhang, Zhiwei and Chen, Zhengyu and Guo, Yue , month = mar, year =. Multi-. doi:10.48550/arXiv.2603.02188 , abstract =
-
[4]
arXiv.org , author =
The. arXiv.org , author =
-
[5]
Adjusting batch effects in microarray expression data using empirical. Biostatistics , author =. 2007 , keywords =. doi:10.1093/biostatistics/kxj037 , abstract =
-
[6]
Fast, sensitive and accurate integration of single-cell data with
Korsunsky, Ilya and Millard, Nghia and Fan, Jean and Slowikowski, Kamil and Zhang, Fan and Wei, Kevin and Baglaenko, Yuriy and Brenner, Michael and Loh, Po-ru and Raychaudhuri, Soumya , month = dec, year =. Fast, sensitive and accurate integration of single-cell data with. Nature Methods , publisher =. doi:10.1038/s41592-019-0619-0 , abstract =
-
[7]
SCANPY: large-scale single-cell gene expression data analysis
Genome Biology , author =. 2018 , keywords =. doi:10.1186/s13059-017-1382-0 , abstract =
-
[8]
Diffusion of
Ye, Jiacheng and Gong, Shansan and Chen, Liheng and Zheng, Lin and Gao, Jiahui and Shi, Han and Wu, Chuan and Jiang, Xin and Li, Zhenguo and Bi, Wei and Kong, Lingpeng , year =. Diffusion of
-
[9]
Efficient
Li, Yulin and Tu, Tengyao and Ding, Li and Wang, Junjie and Zhen, Hui-Ling and Chen, Yixin and Li, Yong and Tian, Zhuotao , month = oct, year =. Efficient
-
[10]
arXiv preprint arXiv:2505.23648 , year=
Gozeten, Halil Alperen and Ildiz, M. Emrullah and Zhang, Xuechen and Harutyunyan, Hrayr and Rawat, Ankit Singh and Oymak, Samet , month = mar, year =. Continuous. doi:10.48550/arXiv.2505.23648 , abstract =
-
[11]
Yang, Wen and Liao, Minpeng and Fan, Kai , month = mar, year =. Markov. doi:10.48550/arXiv.2410.17635 , abstract =
-
[12]
Ctrls: Chain-of-thought rea- soning via latent state-transition.arXiv:2507.08182,
Wu, Junda and Xiong, Yuxin and Li, Xintong and Yu, Sheldon and Hu, Zhengmian and Yu, Tong and Wang, Rui and Chen, Xiang and Shang, Jingbo and McAuley, Julian , month = jan, year =. doi:10.48550/arXiv.2507.08182 , abstract =
-
[13]
Wang, Chaojie and Deng, Yanchen and Lv, Zhiyi and Liang, Zeng and He, Jujie and Yan, Shuicheng and Bo, An , month = jun, year =. Q*:. doi:10.48550/arXiv.2406.14283 , abstract =
-
[14]
Reasoning with language model is planning with world model
Hao, Shibo and Gu, Yi and Ma, Haodi and Hong, Joshua and Wang, Zhen and Wang, Daisy and Hu, Zhiting , editor =. Reasoning with. Proceedings of the 2023. 2023 , pages =. doi:10.18653/v1/2023.emnlp-main.507 , abstract =
-
[15]
Moon, Seungyong and Park, Bumsoo and Song, Hyun Oh , month = oct, year =. Learning to. doi:10.48550/arXiv.2410.02992 , abstract =
-
[16]
Stream of search (sos): Learning to search in language, 2024,
Gandhi, Kanishk and Lee, Denise and Grand, Gabriel and Liu, Muxin and Cheng, Winson and Sharma, Archit and Goodman, Noah D. , month = apr, year =. Stream of. doi:10.48550/arXiv.2404.03683 , abstract =
-
[17]
DeepSeek-R1 incentivizes reasoning in LLMs through reinforcement learning , volume=
Nature , author =. 2025 , note =. doi:10.1038/s41586-025-09422-z , abstract =
-
[18]
Hassid, Michael and Synnaeve, Gabriel and Adi, Yossi and Schwartz, Roy , month = feb, year =. Don't. doi:10.48550/arXiv.2505.17813 , abstract =
-
[19]
Xu, Zhongxing and Wang, Zhonghua and Qian, Zhe and Shi, Dachuan and Tang, Feilong and Hu, Ming and Su, Shiyan and Zou, Xiaocheng and Feng, Wei and Mahapatra, Dwarikanath and Peng, Yifan and Lin, Mingquan and Ge, Zongyuan , month = mar, year =. Thinking in. doi:10.48550/arXiv.2603.13366 , abstract =
-
[20]
Ye, Wengao and Liang, Yan and Shan, Lianlei , month = jan, year =. Thinking on the. doi:10.48550/arXiv.2510.04182 , abstract =
-
[21]
A survey on large language models for code generation,
A. ACM Transactions on Software Engineering and Methodology , author =. 2026 , pages =. doi:10.1145/3747588 , abstract =
-
[22]
He, Yinhan and Zheng, Wendy and Zhu, Yaochen and Zheng, Zaiyi and Su, Lin and Vasudevan, Sriram and Guo, Qi and Hong, Liangjie and Li, Jundong , month = oct, year =
-
[23]
Softcot: Soft chain-of-thought for efficient reasoning with llms
Xu, Yige and Guo, Xu and Zeng, Zhiwei and Miao, Chunyan , month = may, year =. doi:10.48550/arXiv.2502.12134 , abstract =
-
[24]
Sui, Yang and Chuang, Yu-Neng and Wang, Guanchu and Zhang, Jiamu and Zhang, Tianyi and Yuan, Jiayi and Liu, Hongyi and Wen, Andrew and Zhong, Shaochen and Zou, Na and Chen, Hanjie and Hu, Xia , month = aug, year =. Stop. doi:10.48550/arXiv.2503.16419 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2503.16419
-
[25]
Nature Communications , publisher =
Cao, Yichuan and Zhao, Xiamiao and Tang, Songming and Jiang, Qun and Li, Sijie and Li, Siyu and Chen, Shengquan , month = apr, year =. Nature Communications , publisher =. doi:10.1038/s41467-024-47418-x , abstract =
-
[26]
Eraslan, Gökcen and Simon, Lukas M. and Mircea, Maria and Mueller, Nikola S. and Theis, Fabian J. , month = jan, year =. Single-cell. Nature Communications , publisher =. doi:10.1038/s41467-018-07931-2 , abstract =
-
[27]
Briefings in Bioinformatics , author =
Single-cell. Briefings in Bioinformatics , author =. 2024 , keywords =. doi:10.1093/bib/bbae209 , abstract =
-
[28]
Genome Biology , author =. 2019 , keywords =. doi:10.1186/s13059-019-1837-6 , abstract =
-
[29]
Scientific Reports , publisher =
Talwar, Divyanshu and Mongia, Aanchal and Sengupta, Debarka and Majumdar, Angshul , month = nov, year =. Scientific Reports , publisher =. doi:10.1038/s41598-018-34688-x , abstract =
-
[30]
and Cuturi, Marco , month = oct, year =
Klein, Dominik and Uscidda, Théo and Theis, Fabian J. and Cuturi, Marco , month = oct, year =. Generative
-
[31]
Proceedings of the AAAI Conference on Artificial Intelligence , author =
Jointly. Proceedings of the AAAI Conference on Artificial Intelligence , author =. 2023 , keywords =. doi:10.1609/aaai.v37i4.25599 , abstract =
-
[32]
Muzellec, Boris and Josse, Julie and Boyer, Claire and Cuturi, Marco , month = nov, year =. Missing. Proceedings of the 37th
-
[33]
Elicitability of Instance and Object Ranking , url =
Demetci, Pinar and Santorella, Rebecca and Sandstede, Björn and Noble, William Stafford and Singh, Ritambhara , month = jan, year =. Journal of Computational Biology , publisher =. doi:10.1089/cmb.2021.0446 , abstract =
-
[34]
and Raj, Arjun and Li, Mingyao and Zhang, Nancy R
Huang, Mo and Wang, Jingshu and Torre, Eduardo and Dueck, Hannah and Shaffer, Sydney and Bonasio, Roberto and Murray, John I. and Raj, Arjun and Li, Mingyao and Zhang, Nancy R. , month = jul, year =. Nature Methods , publisher =. doi:10.1038/s41592-018-0033-z , abstract =
-
[35]
An accurate and robust imputation method
Li, Wei Vivian and Li, Jingyi Jessica , month = mar, year =. An accurate and robust imputation method. Nature Communications , publisher =. doi:10.1038/s41467-018-03405-7 , abstract =
-
[36]
and Streets, Aaron and Yosef, Nir , month = mar, year =
Gayoso, Adam and Steier, Zoë and Lopez, Romain and Regier, Jeffrey and Nazor, Kristopher L. and Streets, Aaron and Yosef, Nir , month = mar, year =. Joint probabilistic modeling of single-cell multi-omic data with. Nature Methods , publisher =. doi:10.1038/s41592-020-01050-x , abstract =
-
[37]
and Burdziak, Cassandra and Moon, Kevin R
Dijk, David van and Sharma, Roshan and Nainys, Juozas and Yim, Kristina and Kathail, Pooja and Carr, Ambrose J. and Burdziak, Cassandra and Moon, Kevin R. and Chaffer, Christine L. and Pattabiraman, Diwakar and Bierie, Brian and Mazutis, Linas and Wolf, Guy and Krishnaswamy, Smita and Pe’er, Dana , month = jul, year =. Recovering. Cell , publisher =. doi:...
-
[38]
Genome Biology , author =. 2023 , keywords =. doi:10.1186/s13059-023-02989-8 , abstract =
-
[39]
Journal of Computational Biology , publisher =
Demetci, Pinar and Santorella, Rebecca and Chakravarthy, Manav and Sandstede, Bjorn and Singh, Ritambhara , month = nov, year =. Journal of Computational Biology , publisher =. doi:10.1089/cmb.2022.0270 , abstract =
-
[40]
Spectral clustering of single-cell multi-omics data on multilayer graphs , volume =. Bioinformatics , author =. 2022 , pages =. doi:10.1093/bioinformatics/btac378 , abstract =
-
[41]
Ashuach, Tal and Gabitto, Mariano I. and Koodli, Rohan V. and Saldi, Giuseppe-Antonio and Jordan, Michael I. and Yosef, Nir , month = aug, year =. Nature Methods , publisher =. doi:10.1038/s41592-023-01909-9 , abstract =
-
[42]
doi:10.48550/arXiv.2505.02322 , abstract =
Gui, Runquan and Wang, Zhihai and Wang, Jie and Ma, Chi and Zhen, Huiling and Yuan, Mingxuan and Hao, Jianye and Lian, Defu and Chen, Enhong and Wu, Feng , month = may, year =. doi:10.48550/arXiv.2505.02322 , abstract =
-
[43]
Hierarchical reasoning model, 2025
Wang, Guan and Li, Jin and Sun, Yuhao and Chen, Xing and Liu, Changling and Wu, Yue and Lu, Meng and Song, Sen and Yadkori, Yasin Abbasi , month = aug, year =. Hierarchical. doi:10.48550/arXiv.2506.21734 , abstract =
-
[44]
Enhancing
Xiong, Siheng and Payani, Ali and Fekri, Faramarz , month = oct, year =. Enhancing
-
[45]
Chen, Sijia and Niu, Di , month = dec, year =. doi:10.48550/arXiv.2512.24014 , abstract =
-
[46]
Chen, Qiguang and Qin, Libo and Liu, Jinhao and Peng, Dengyun and Guan, Jiannan and Wang, Peng and Hu, Mengkang and Zhou, Yuhang and Gao, Te and Che, Wanxiang , month = jul, year =. Towards. doi:10.48550/arXiv.2503.09567 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2503.09567
-
[47]
Jiang, Cong and Zhang, Xiaofeng and Zhu, Fangzhi and Chen, Xiaowei and Zhu, Junxiong and Zhang, Zheng , year =
-
[48]
Hyperbolic. 2018 , note =. doi:10.48550/arXiv.1805.09112 , abstract =
-
[49]
Hyperbolic-based. ACM Trans. Knowl. Discov. Data , author =. 2026 , pages =. doi:10.1145/3786587 , abstract =
-
[50]
Wang, Xinyi and Caccia, Lucas and Ostapenko, Oleksiy and Yuan, Xingdi and Wang, William Yang and Sordoni, Alessandro , month = aug, year =. Guiding. doi:10.48550/arXiv.2310.05707 , abstract =
-
[51]
Scientific Reports , publisher =
Mao, Runze and Zhang, Zhengyuan and Yang, Mengyao and Xie, Hui and Wei, Shengjun and Hu, Changzhen , month = jan, year =. Scientific Reports , publisher =. doi:10.1038/s41598-026-37328-x , abstract =
-
[52]
Sun, Hanshi and Haider, Momin and Zhang, Ruiqi and Yang, Huitao and Qiu, Jiahao and Yin, Ming and Wang, Mengdi and Bartlett, Peter and Zanette, Andrea , month = oct, year =. Fast. doi:10.48550/arXiv.2410.20290 , abstract =
-
[53]
Efficient reasoning models: A survey
Feng, Sicheng and Fang, Gongfan and Ma, Xinyin and Wang, Xinchao , month = sep, year =. Efficient. doi:10.48550/arXiv.2504.10903 , abstract =
-
[54]
Wu, Taiqiang and Xu, Zenan and Zhou, Bo and Wong, Ngai , month = feb, year =. The. doi:10.48550/arXiv.2602.20945 , abstract =
-
[55]
Li, You and Chen, Chi and Li, Yanghao and Zeng, Fanhu and Huang, Kaiyu and Xu, Jinan and Sun, Maosong , month = feb, year =. Imagination. doi:10.48550/arXiv.2602.22766 , abstract =
-
[56]
Huang, Zixuan and Xia, Xin and Ren, Yuxi and Zheng, Jianbin and Wang, Xuanda and Zhang, Zhixia and Xie, Hongyan and Liang, Songshi and Chen, Zehao and Xiao, Xuefeng and Zhuang, Fuzhen and Li, Jianxin and Ban, Yikun and Wang, Deqing , month = feb, year =. Does. doi:10.48550/arXiv.2602.08354 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2602.08354
-
[57]
Tan, Wenhui and Li, Jiaze and Ju, Jianzhong and Luo, Zhenbo and Song, Ruihua and Luan, Jian , month = feb, year =. Think. doi:10.48550/arXiv.2505.16552 , abstract =
-
[58]
Li, Bangzheng and Sun, Ximeng and Liu, Jiang and Wang, Ze and Wu, Jialian and Yu, Xiaodong and Chen, Hao and Barsoum, Emad and Chen, Muhao and Liu, Zicheng , month = oct, year =. Latent. doi:10.48550/arXiv.2509.24251 , abstract =
-
[59]
Render-of-Thought: Rendering Textual Chain-of-Thought as Images for Visual Latent Reasoning
Wang, Yifan and Li, Shiyu and Li, Peiming and Yang, Xiaochen and Tang, Yang and Wei, Zheng , month = jan, year =. Render-of-. doi:10.48550/arXiv.2601.14750 , abstract =
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2601.14750
-
[60]
Goyal, Sachin and Ji, Ziwei and Rawat, Ankit Singh and Menon, Aditya Krishna and Kumar, Sanjiv and Nagarajan, Vaishnavh , month = apr, year =. Think before you speak:. doi:10.48550/arXiv.2310.02226 , abstract =
-
[61]
Wang, Jiecong and Peng, Hao and Liu, Chunyang , month = feb, year =. Latent. doi:10.48550/arXiv.2601.21358 , abstract =
-
[62]
Jeddi, Ahmadreza and Ciccone, Marco and Taati, Babak , month = feb, year =. doi:10.48550/arXiv.2602.11451 , abstract =
-
[63]
and Reddy, Chandra and Nguyen, Lam M
Wen, Yunshi and Gifford, Wesley M. and Reddy, Chandra and Nguyen, Lam M. and Kalagnanam, Jayant and Julius, Anak Agung , month = feb, year =. Revisiting the. doi:10.48550/arXiv.2602.06909 , abstract =
-
[64]
DeepSeek-OCR: Contexts Optical Compression
Wei, Haoran and Sun, Yaofeng and Li, Yukun , month = oct, year =. doi:10.48550/arXiv.2510.18234 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2510.18234
-
[65]
arXiv preprint arXiv:2601.23184
Wang, Fanmeng and Liu, Haotian and Zhao, Guojiang and Xu, Hongteng and Gao, Zhifeng , month = jan, year =. doi:10.48550/arXiv.2601.23184 , abstract =
-
[66]
Rea- soning models can be effective without thinking
Ma, Wenjie and He, Jingxuan and Snell, Charlie and Griggs, Tyler and Min, Sewon and Zaharia, Matei , month = apr, year =. Reasoning. doi:10.48550/arXiv.2504.09858 , abstract =
-
[67]
Jiang, Tianyi and An, Arctanx and Feng, Hengyi and Zhai, Naixin and Li, Haodong and Yu, Xiaomin and Liu, Jiahui and Du, Hanwen and Zhang, Shuo and Yang, Zhi and Huang, Jie and Li, Yuhua and Ni, Yongxin and Wang, Huacan and Chen, Ronghao , month = feb, year =. Chain of. doi:10.48550/arXiv.2602.10063 , abstract =
-
[68]
Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models,
Wang, Lei and Xu, Wanyu and Lan, Yihuai and Hu, Zhiqiang and Lan, Yunshi and Lee, Roy Ka-Wei and Lim, Ee-Peng , month = may, year =. Plan-and-. doi:10.48550/arXiv.2305.04091 , abstract =
-
[69]
Fundamental reasoning paradigms induce out-of-domain generalization in language models, 2026
Cao, Mingzi and Tan, Xingwei and Akhter, Mahmud Elahi and Valentino, Marco and Liakata, Maria and Wang, Xi and Aletras, Nikolaos , month = feb, year =. Fundamental. doi:10.48550/arXiv.2602.08658 , abstract =
-
[70]
team, L. C. M. and Barrault, Loïc and Duquenne, Paul-Ambroise and Elbayad, Maha and Kozhevnikov, Artyom and Alastruey, Belen and Andrews, Pierre and Coria, Mariano and Couairon, Guillaume and Costa-jussà, Marta R. and Dale, David and Elsahar, Hady and Heffernan, Kevin and Janeiro, João Maria and Tran, Tuan and Ropers, Christophe and Sánchez, Eduardo and R...
-
[71]
arXiv.org , author =
Emergence of. arXiv.org , author =
-
[72]
arXiv.org , author =
Reasoning by. arXiv.org , author =
-
[73]
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Dao, Tri and Fu, Daniel Y. and Ermon, Stefano and Rudra, Atri and Ré, Christopher , month = jun, year =. doi:10.48550/arXiv.2205.14135 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2205.14135
-
[74]
Zhuang, Ren and Wang, Ben and Sun, Shuifa , month = jan, year =. The. doi:10.48550/arXiv.2601.18832 , abstract =
work page internal anchor Pith review doi:10.48550/arxiv.2601.18832
-
[75]
Zhang, Junyu and Sun, Yifan and Leng, Tianang and Shen, Jingyan and Ziyin, Liu and Liang, Paul Pu and Zhang, Huan , month = dec, year =. When. doi:10.48550/arXiv.2512.17901 , abstract =
-
[76]
Sketch-of-thought: Efficient llm reasoning with adaptive cognitive-inspired sketching
Aytes, Simon A. and Baek, Jinheon and Hwang, Sung Ju , month = oct, year =. Sketch-of-. doi:10.48550/arXiv.2503.05179 , abstract =
-
[77]
Li, Kelvin and Shang, Chuyi and Karlinsky, Leonid and Feris, Rogerio and Darrell, Trevor and Herzig, Roei , month = dec, year =. Latent. doi:10.48550/arXiv.2512.21218 , abstract =
-
[78]
Wei, Tianxin and Li, Ting-Wei and Liu, Zhining and Ning, Xuying and Yang, Ze and Zou, Jiaru and Zeng, Zhichen and Qiu, Ruizhong and Lin, Xiao and Fu, Dongqi and Li, Zihao and Ai, Mengting and Zhou, Duo and Bao, Wenxuan and Li, Yunzhe and Li, Gaotang and Qian, Cheng and Wang, Yu and Tang, Xiangru and Xiao, Yin and Fang, Liri and Liu, Hui and Tang, Xianfeng...
-
[79]
Xu, Liyan and Yu, Mo and Meng, Fandong and Zhou, Jie , month = feb, year =. No. doi:10.48550/arXiv.2602.02103 , abstract =
-
[80]
Hong, Bin and Liu, Jiayu and Huang, Zhenya and Zhang, Kai and Zhang, Mengdi , month = aug, year =. Pruning. doi:10.48550/arXiv.2508.10164 , abstract =
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2508.10164
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.