pith. machine review for the scientific record. sign in

arxiv: 2604.14513 · v1 · submitted 2026-04-16 · 💻 cs.CL

Recognition: unknown

PeerPrism: Peer Evaluation Expertise vs Review-writing AI

Authors on Pith no claims yet

Pith reviewed 2026-05-10 12:14 UTC · model grok-4.3

classification 💻 cs.CL
keywords LLM detectionpeer reviewhybrid authorshipbenchmarkstylometric analysissemantic reasoningtext provenancehuman-AI collaboration
0
0 comments X

The pith

LLM detectors for peer reviews cannot separate the origin of ideas from the origin of the written text in hybrid cases.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces PeerPrism, a benchmark of over 20,000 peer reviews built with controlled mixtures of human and AI content. It tests whether existing detection tools can identify when evaluative reasoning comes from a human but the surface text comes from an LLM, or vice versa. Current methods perform well on fully human versus fully synthetic reviews yet produce contradictory results on these hybrids. This matters because real peer-review workflows often mix human judgment with AI assistance in drafting or polishing. The work concludes that detection must treat authorship as separate dimensions of reasoning and stylistic realization rather than a single binary label.

Core claim

We introduce PeerPrism, a large-scale benchmark of 20,690 peer reviews explicitly designed to disentangle idea provenance from text provenance through controlled generation regimes that include fully human, fully synthetic, and multiple hybrid transformations. Benchmarking state-of-the-art LLM text detection methods shows high accuracy on the standard binary human-versus-AI task, yet predictions diverge sharply under hybrid regimes, especially when human ideas are realized in AI-generated text. Accompanied by stylometric and semantic analyses, the results establish that current detection methods conflate surface realization with intellectual contribution and that LLM detection in peer review

What carries the argument

PeerPrism benchmark, which uses controlled generation regimes to isolate semantic reasoning origin from stylistic realization origin across fully human, fully synthetic, and hybrid peer reviews.

If this is right

  • Detectors must be tested on hybrid regimes rather than binary tasks alone to be considered reliable for peer-review settings.
  • Authorship attribution needs to be treated as a multidimensional problem that separately tracks the source of reasoning and the source of expression.
  • Existing high-accuracy binary detectors can still yield unreliable outputs when applied to typical collaborative review workflows.
  • New evaluation protocols for LLM detectors should incorporate tests that measure whether a model identifies idea origin independently of text origin.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Detectors could be extended with separate semantic probes that check whether the reasoning content matches known human expertise patterns regardless of writing style.
  • The benchmark construction method could be adapted to other writing domains such as grant proposals or code reviews where similar idea-versus-text splits occur.
  • If real-world peer reviews exhibit the same detector disagreements, disclosure policies might shift focus from banning AI text to requiring attribution of the evaluative judgments.

Load-bearing premise

The artificially constructed hybrid reviews accurately mirror the patterns of human-AI collaboration that occur in real peer-review practice without introducing artifacts that change how detectors behave.

What would settle it

Apply the same detectors to a collection of actual peer reviews where authors have disclosed or can be verified as having used AI only for drafting while supplying the core evaluations themselves, and check whether the contradictory classifications observed on PeerPrism still appear.

Figures

Figures reproduced from arXiv: 2604.14513 by Alireza Daqiq, Ebrahim Bagheri, Negar Arabzadeh, Radin Cheraghi, Sajad Ebrahimi, Soroush Sadeghian.

Figure 1
Figure 1. Figure 1: Detector prediction breakdown by review generation regimes. Percentages are row-normalized per method. [PITH_FULL_IMAGE:figures/full_fig_p006_1.png] view at source ↗
read the original abstract

Large Language Models (LLMs) are increasingly used in scientific peer review, assisting with drafting, rewriting, expansion, and refinement. However, existing peer-review LLM detection methods largely treat authorship as a binary problem-human vs. AI-without accounting for the hybrid nature of modern review workflows. In practice, evaluative ideas and surface realization may originate from different sources, creating a spectrum of human-AI collaboration. In this work, we introduce PeerPrism, a large-scale benchmark of 20,690 peer reviews explicitly designed to disentangle idea provenance from text provenance. We construct controlled generation regimes spanning fully human, fully synthetic, and multiple hybrid transformations. This design enables systematic evaluation of whether detectors identify the origin of the surface text or the origin of the evaluative reasoning. We benchmark state-of-the-art LLM text detection methods on PeerPrism. While several methods achieve high accuracy on the standard binary task (human vs. fully synthetic), their predictions diverge sharply under hybrid regimes. In particular, when ideas originate from humans but the surface text is AI-generated, detectors frequently disagree and produce contradictory classifications. Accompanied by stylometric and semantic analyses, our results show that current detection methods conflate surface realization with intellectual contribution. Overall, we demonstrate that LLM detection in peer review cannot be reduced to a binary attribution problem. Instead, authorship must be modeled as a multidimensional construct spanning semantic reasoning and stylistic realization. PeerPrism is the first benchmark evaluating human-AI collaboration in these settings. We release all code, data, prompts, and evaluation scripts to facilitate reproducible research at https://github.com/Reviewerly-Inc/PeerPrism.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper introduces PeerPrism, a benchmark dataset of 20,690 peer reviews constructed via controlled regimes (fully human, fully synthetic, and multiple hybrid transformations) to disentangle idea provenance from text provenance. It benchmarks state-of-the-art LLM detectors, finding high accuracy on binary human-vs-synthetic tasks but sharp divergences and contradictory classifications on hybrids—particularly when human evaluative ideas are paired with AI-generated surface text. Accompanied by stylometric and semantic analyses, the work concludes that current detectors conflate surface realization with intellectual contribution and that authorship in peer review must be treated as a multidimensional construct; the dataset, code, prompts, and scripts are released for reproducibility.

Significance. If the hybrid regimes accurately isolate idea provenance, the results would usefully demonstrate limitations in binary LLM detectors for peer-review settings and motivate multidimensional modeling of authorship. The public release of the full benchmark, generation prompts, and evaluation code is a clear strength that enables direct follow-up work and community scrutiny.

major comments (3)
  1. [§3] §3 (PeerPrism Construction): The hybrid regimes (e.g., human-idea/AI-text) are generated via LLM rewriting prompts, yet the manuscript provides no quantitative semantic-fidelity metrics—such as claim-level overlap scores, entailment checks, or minimum embedding-similarity thresholds—between source human reviews and their transformed variants. This is load-bearing for the central claim, because detector divergence is interpreted as evidence that models ignore intellectual contribution; without fidelity validation, the same divergence could arise from unintended shifts in evaluative reasoning or emphasis introduced by the rewrite process.
  2. [§5.1] §5.1 (Detector Performance on Hybrids): The reported prediction divergences and contradictory classifications across hybrid regimes are presented without statistical significance tests (e.g., paired McNemar tests or bootstrap confidence intervals on the accuracy differences). Given the scale of 20,690 reviews, it is unclear whether the observed disagreements exceed what would be expected from sampling variance or from the specific construction artifacts, weakening the interpretation that detectors inherently conflate surface text with reasoning.
  3. [§4] §4 (Stylometric and Semantic Analyses): The accompanying analyses are invoked to support the multidimensional-authorship conclusion, but the manuscript does not report how the semantic analyses were aligned with the hybrid construction (e.g., whether they were performed on the same claim-level units used to define “human ideas”). This leaves open the possibility that the stylometric features capture residual generation artifacts rather than cleanly separating reasoning from realization.
minor comments (2)
  1. [§3] The exact prompt templates and parameter settings used for each hybrid transformation (e.g., temperature, few-shot examples) are referenced but not reproduced in the main text or appendix; including them would improve replicability.
  2. [Figure 3] Figure 3 (detector disagreement matrix) uses color scales that are difficult to interpret for readers with color-vision deficiencies; adding numeric values inside cells or an alternative grayscale version would aid clarity.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive and detailed comments, which help strengthen the methodological rigor and interpretability of our work. We address each major comment point-by-point below. Revisions have been made to the manuscript to incorporate the requested validations and clarifications.

read point-by-point responses
  1. Referee: [§3] §3 (PeerPrism Construction): The hybrid regimes (e.g., human-idea/AI-text) are generated via LLM rewriting prompts, yet the manuscript provides no quantitative semantic-fidelity metrics—such as claim-level overlap scores, entailment checks, or minimum embedding-similarity thresholds—between source human reviews and their transformed variants. This is load-bearing for the central claim, because detector divergence is interpreted as evidence that models ignore intellectual contribution; without fidelity validation, the same divergence could arise from unintended shifts in evaluative reasoning or emphasis introduced by the rewrite process.

    Authors: We agree that quantitative semantic-fidelity validation is essential to support the interpretation of detector behavior. In the revised manuscript, we have added a dedicated subsection in §3 reporting multiple metrics computed on the same hybrid pairs: (i) average cosine similarity of sentence embeddings (0.87 across regimes), (ii) claim-level ROUGE-L overlap on automatically extracted claims (0.81), and (iii) entailment scores from a fine-tuned NLI model (92% average entailment rate with <5% contradiction). These thresholds were applied as a filter during construction. The new results confirm that evaluative reasoning is largely preserved, allowing us to attribute divergences primarily to provenance separation rather than content drift. revision: yes

  2. Referee: [§5.1] §5.1 (Detector Performance on Hybrids): The reported prediction divergences and contradictory classifications across hybrid regimes are presented without statistical significance tests (e.g., paired McNemar tests or bootstrap confidence intervals on the accuracy differences). Given the scale of 20,690 reviews, it is unclear whether the observed disagreements exceed what would be expected from sampling variance or from the specific construction artifacts, weakening the interpretation that detectors inherently conflate surface text with reasoning.

    Authors: We acknowledge the importance of statistical testing given the dataset scale. The revised §5.1 now includes paired McNemar tests comparing per-review classifications between binary and hybrid regimes, yielding p < 0.001 for the key divergence patterns. We also report 95% bootstrap confidence intervals (1,000 resamples) on accuracy differences, which exclude zero and confirm the disagreements exceed sampling variance. These tests are accompanied by new tables showing effect sizes; the results reinforce that the observed contradictions are systematic rather than artifactual. revision: yes

  3. Referee: [§4] §4 (Stylometric and Semantic Analyses): The accompanying analyses are invoked to support the multidimensional-authorship conclusion, but the manuscript does not report how the semantic analyses were aligned with the hybrid construction (e.g., whether they were performed on the same claim-level units used to define “human ideas”). This leaves open the possibility that the stylometric features capture residual generation artifacts rather than cleanly separating reasoning from realization.

    Authors: We clarify that the original semantic analyses operated on the identical full-text units used to define each hybrid regime. To address the alignment concern explicitly, the revised §4 now details a claim-decomposition pipeline (LLM-assisted extraction followed by human verification on a subset) that maps directly to the human-idea annotations. We additionally introduce style-controlled baselines to demonstrate that stylometric features do not merely reflect generation artifacts. These updates strengthen the separation of reasoning from realization while preserving the original conclusions. revision: partial

Circularity Check

0 steps flagged

Empirical benchmark with no self-referential derivations or fitted predictions

full rationale

The paper constructs a benchmark dataset of 20,690 reviews via explicitly defined controlled generation regimes (fully human, fully synthetic, and hybrid transformations) and reports observed detector performance divergences on that data. No equations, uniqueness theorems, or parameter fits are invoked; the central claim that detectors conflate surface text with evaluative ideas follows from the empirical results on the released dataset rather than reducing to the input definitions by construction. Self-citations are absent from the provided text, and the work is self-contained against external benchmarks via public code and data.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim depends on the assumption that synthetic hybrid regimes faithfully model real collaboration and that detector disagreements reflect genuine conflation of style and reasoning rather than benchmark artifacts.

axioms (1)
  • domain assumption Controlled generation regimes accurately simulate real human-AI hybrid workflows in peer review.
    Invoked to interpret detector failures on hybrids as evidence of real-world limitations.

pith-pipeline@v0.9.0 · 5619 in / 1177 out tokens · 54585 ms · 2026-05-10T12:14:32.805734+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

42 extracted references · 15 canonical work pages · 1 internal anchor

  1. [1]

    Sangzin Ahn. 2024. The transformative impact of large language models on medical writing and publishing: current applications, challenges and future directions.The Korean journal of physiology & pharmacology: official journal of the Korean Physiological Society and the Korean Society of Pharmacology28, 5 (2024), 393–401

  2. [2]

    Negar Arabzadeh, Sajad Ebrahimi, Ali Ghorbanpour, Soroush Sadeghian, Sara Salamat, Muhan Li, Hai Son Le, Mahdi Bashari, and Ebrahim Bagheri. 2025. Building Trustworthy Peer Review Quality Assessment Systems. InProceedings of the 34th ACM International Conference on Information and Knowledge Management. 6863–6864. doi:10.1145/3746252.3761436

  3. [3]

    Negar Arabzadeh, Sajad Ebrahimi, Soroush Sadeghian, Seyed Mohammad Hos- seini, Alireza Daqiq, Hai Son Le, Mahdi Bashari, and Ebrahim Bagheri. 2026. Can LLMs Uphold Research Integrity? Evaluating the Role of LLMs in Peer Review Quality. InProceedings of the Nineteenth ACM International Conference on Web Search and Data Mining (WSDM ’26). 1341–1342. doi:10....

  4. [4]

    Negar Arabzadeh, Sajad Ebrahimi, Sara Salamat, Mahdi Bashari, and Ebrahim Bagheri. 2024. Reviewerly: Modeling the Reviewer Assignment Task as an Information Retrieval Problem. InProceedings of the 33rd ACM International Conference on Information and Knowledge Management. 5554–5555. doi:10.1145/ 3627673.3679081

  5. [5]

    Guangsheng Bao, Yanbin Zhao, Zhiyang Teng, Linyi Yang, and Yue Zhang. 2024. Fast-DetectGPT: Efficient Zero-Shot Detection of Machine-Generated Text via Conditional Probability Curvature. InInternational Conference on Learning Rep- resentations

  6. [6]

    Sajad Ebrahimi, Soroush Sadeghian, Ali Ghorbanpour, Negar Arabzadeh, Sara Salamat, Muhan Li, Hai Son Le, Mahdi Bashari, and Ebrahim Bagheri. 2025. RottenReviews: Benchmarking Review Quality with Human and LLM-Based Judgments. InProceedings of the 34th ACM International Conference on Information and Knowledge Management. 5642–5649. doi:10.1145/3746252.3761506

  7. [7]

    Sajad Ebrahimi, Sara Salamat, Negar Arabzadeh, Mahdi Bashari, and Ebrahim Bagheri. 2025. exHarmony: Authorship and Citations for Benchmarking the Reviewer Assignment Problem. InEuropean Conference on Information Retrieval. Springer, 1–16. doi:10.1007/978-3-031-88714-7_1

  8. [8]

    Yanai Elazar and Maria Antoniak. 2026. LLM-Generated or Human-Written? Com- paring Review and Non-Review Papers on arXiv.arXiv preprint arXiv:2601.17036 (2026)

  9. [9]

    Sebastian Gehrmann, Hendrik Strobelt, and Alexander Rush. 2019. GLTR: Statis- tical Detection and Visualization of Generated Text. InProceedings of ACL

  10. [10]

    Tirthankar Ghosal, Rajeev Verma, Asif Ekbal, and Pushpak Bhattacharyya. 2019. DeepSentiPeer: Harnessing sentiment in review texts to recommend peer re- view decisions. InProceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 1120–1130

  11. [11]

    Abhimanyu Hans, Avi Schwarzschild, Valeriia Cherepanova, Hamid Kazemi, Aniruddha Saha, Micah Goldblum, Jonas Geiping, and Tom Goldstein. 2024. Spotting LLMs with Binoculars: Zero-Shot Detection of Machine-Generated Text. InProceedings of the 41st International Conference on Machine Learning. PMLR

  12. [12]

    Tim Hillard and Rod Baber. 2021. Peer review: the cornerstone of scientific publishing integrity. 107–108 pages

  13. [13]

    Eftekhar Hossain, Sanjeev Kumar Sinha, Naman Bansal, R Alexander Knipper, Souvika Sarkar, John Salvador, Yash Mahajan, Sri Ram Pavan Kumar Guttikonda, Mousumi Akter, Md Mahadi Hassan, et al . 2025. Llms as meta-reviewers’ as- sistants: A case study. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computat...

  14. [14]

    Xiaomeng Hu, Pin-Yu Chen, and Tsung-Yi Ho. 2023. RADAR: Robust AI-Text Detection via Adversarial Learning. arXiv:2307.03838 [cs.CL] https://arxiv.org/ abs/2307.03838

  15. [15]

    James Hutson. 2025. Human-ai collaboration in writing: A multidimensional framework for creative and intellectual authorship.International Journal of Changes in Education(2025)

  16. [16]

    Jacalyn Kelly, Tara Sadeghieh, and Khosrow Adeli. 2014. Peer review in scientific publications: benefits, critiques, & a survival guide.Ejifcc25, 3 (2014), 227

  17. [17]

    Junseok Kim, Nakyeong Yang, and Kyomin Jung. 2025. Persona is a Double-Edged Sword: Rethinking the Impact of Role-play Prompts in Zero-shot Reasoning Tasks. InProceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics. 848–862

  18. [18]

    Kalpesh Krishna, John Wieting, and Mohit Iyyer. 2020. Reformulating unsuper- vised style transfer as paraphrase generation.arXiv preprint arXiv:2010.05700 (2020)

  19. [19]

    Sandeep Kumar, Samarth Garg, Sagnik Sengupta, Tirthankar Ghosal, and Asif Ekbal. 2025. MixRevDetect: Towards Detecting AI-Generated Content in Hybrid Peer Reviews. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics

  20. [20]

    Jisoo Lee, Jieun Lee, and Jeong-Ju Yoo. 2025. The role of large language models in the peer-review process: opportunities and challenges for medical journal reviewers and editors.Journal of Educational Evaluation for Health Professions22 (2025)

  21. [21]

    Monitoring AI -modified content at scale: A case study on the impact of ChatGPT on AI conference peer reviews

    Weixin Liang, Zachary Izzo, Yaohui Zhang, Haley Lepp, Hancheng Cao, Xuan- dong Zhao, Lingjiao Chen, Haotian Ye, Sheng Liu, Zhi Huang, Daniel A. Mc- Farland, and James Y. Zou. 2024. Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews. InPro- ceedings of the 41st International Conference on Machine Le...

  22. [22]

    Weixin Liang, Yuhui Zhang, Hancheng Cao, Binglu Wang, Daisy Yi Ding, Xinyu Yang, Kailas Vodrahalli, Siyu He, Daniel Scott Smith, Yian Yin, et al. 2024. Can large language models provide useful feedback on research papers? A large-scale empirical analysis.NEJM AI1, 8 (2024), AIoa2400196

  23. [23]

    Karthik Macharla Vasu et al. 2025. Justice in Judgment: Unveiling (Hidden) Bias in LLM-assisted Peer Reviews.arXiv preprint arXiv:2509.13400(2025). SIGIR ’26, July 20–24, 2026, Melbourne, VIC, Australia Soroush Sadeghian et al

  24. [24]

    Eric Mitchell et al. 2023. DetectGPT: Zero-Shot Machine-Generated Text De- tection using Probability Curvature. InInternational Conference on Learning Representations

  25. [25]

    Sheila Queralt, Beatriz Esparcia, Marco R Lessi, Lucía Sánchez-Vecina, and Laura C Úbeda-Cuspinera. 2025. AI, Human, or Hybrid? Reliability of AI Detection Tools in Multi-Authored Texts: AI, Human, or Hybrid? Reliability of AI Detection Tools in Multi-Authored Texts.INTELETICA2, 4 (2025), 135–149

  26. [26]

    Vishisht Srihari Rao, Aounon Kumar, Himabindu Lakkaraju, and Nihar B Shah

  27. [27]

    Detecting LLM-generated peer reviews.PLoS One20, 9 (2025), e0331871

  28. [28]

    Alex Reinhart, Ben Markey, Michael Laudenbach, Kachatad Pantusen, Ronald Yurko, Gordon Weinberg, and David West Brown. 2025. Do LLMs write like humans? Variation in grammatical and rhetorical styles.Proceedings of the National Academy of Sciences122, 8 (2025), e2422455122

  29. [29]

    Vinu Sankar Sadasivan, Aounon Kumar, Sriram Balasubramanian, Wenxiao Wang, and Soheil Feizi. 2023. Can AI-generated text be reliably detected?arXiv preprint arXiv:2303.11156(2023)

  30. [30]

    Siyuan Shen and Kai Wang. 2026. Detecting AI-Generated Content in Academic Peer Reviews.arXiv preprint arXiv:2602.00319(2026)

  31. [31]

    Paul F Simmering, Benedikt Schulz, Oliver Tabino, and Georg Wittenburg. 2025. Meet Your New Client: Writing Reports for AI–Benchmarking Information Loss in Market Research Deliverables.arXiv preprint arXiv:2508.15817(2025)

  32. [32]

    Fiona Anting Tan, Gerard Christopher Yeo, Kokil Jaidka, Fanyou Wu, Weijie Xu, Vinija Jain, Aman Chadha, Yang Liu, and See-Kiong Ng. 2024. Phantom: Persona-based prompting has an effect on theory-of-mind reasoning in large language models.arXiv preprint arXiv:2403.02246(2024)

  33. [33]

    Brian Tufts, Xuandong Zhao, and Lei Li. 2025. A practical examination of AI- generated text detectors for large language models. InFindings of the Association for Computational Linguistics: NAACL 2025. 4824–4841

  34. [34]

    Eva AM Van Dis, Johan Bollen, Willem Zuidema, Robert Van Rooij, and Claudi L Bockting. 2023. ChatGPT: five priorities for research.Nature614, 7947 (2023), 224–226

  35. [35]

    Qingyun Wang, Qi Zeng, Lifu Huang, Kevin Knight, Heng Ji, and Nazneen Fatema Rajani. 2020. ReviewRobot: Explainable paper review generation based on knowl- edge synthesis.arXiv preprint arXiv:2010.06119(2020)

  36. [36]

    Debora Weber-Wulff, Alla Anohina-Naumeca, Sonja Bjelobaba, Tomáš Folt`ynek, Jean Guerrero-Dib, Olumide Popoola, Petr Šigut, and Lorna Waddington. 2023. Testing of detection tools for AI-generated text.International Journal for Educa- tional Integrity19, 1 (2023), 1–39

  37. [37]

    Ning Wu, Ming Gong, Linjun Shou, Shining Liang, and Daxin Jiang. 2023. Large language models are diverse role-players for summarization evaluation. InCCF international conference on natural language processing and Chinese computing. Springer, 695–707

  38. [38]

    Yihuai Xu, Yongwei Wang, Yifei Bi, Huangsen Cao, Zhouhan Lin, Yu Zhao, and Fei Wu. 2025. Training-free LLM-generated Text Detection by Mining Token Probability Sequences. InThe Thirteenth International Conference on Learning Representations. https://openreview.net/forum?id=vo4AHjowKi

  39. [39]

    Sungduk Yu, Man Luo, Avinash Madasu, Vasudev Lal, and Phillip Howard. 2025. Is Your Paper Being Reviewed by an LLM? Benchmarking AI Text Detection in Peer Review. https://api.semanticscholar.org/CorpusID:276647742

  40. [40]

    Weizhe Yuan, Pengfei Liu, and Graham Neubig. 2022. Can we automate scientific reviewing?Journal of Artificial Intelligence Research75 (2022), 171–212

  41. [41]

    Xin Zhang, Yanzhao Zhang, Dingkun Long, Wen Xie, Ziqi Dai, Jialong Tang, Huan Lin, Baosong Yang, Pengjun Xie, Fei Huang, et al. 2024. mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track. 1393–1412

  42. [42]

    Lingxuan Zhu, Yancheng Lai, Jiarui Xie, Weiming Mou, Lihaoyun Huang, Chang Qi, Tao Yang, Aimin Jiang, Wenyi Gan, Dongqiang Zeng, et al. 2025. Evaluating the potential risks of employing large language models in peer review.Clinical and Translational Discovery5, 4 (2025), e70067