pith. machine review for the scientific record. sign in

arxiv: 2604.07344 · v1 · submitted 2026-04-08 · 💻 cs.HC

Recognition: unknown

NIRVANA: A Comprehensive Dataset for Reproducing How Students Use Generative AI for Essay Writing

Alice Jang, Andrew Jelson, Daniel Dunlap, Daniel Manesh, Sang Won Lee, Sangwook Lee, Tamara Maddox, Young-Ho Kim

Pith reviewed 2026-05-10 16:58 UTC · model grok-4.3

classification 💻 cs.HC
keywords student writinggenerative AIChatGPTwriting processessay datasetbehavioral profileskeystroke loggingAI assistance
0
0 comments X

The pith

A dataset of 77 students writing essays with ChatGPT logs every keystroke, prompt, and copy action so the full writing process can be replayed and analyzed.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces a dataset gathered from university students who wrote an analytical essay while using ChatGPT. It records keystroke-level behavior, complete conversation histories with the AI, and every piece of text copied from ChatGPT. This level of detail makes it possible to reconstruct the exact sequence of actions students took during writing. Analysis of the data reveals patterns such as how often students query the AI and how that frequency relates to the length and readability of the final essay. The authors group students into four profiles based on how much original writing and revision each person contributes.

Core claim

By capturing keystroke-level writing behavior, full ChatGPT conversation histories, and all text copied from ChatGPT, the dataset enables a complete reconstruction of the writing process and reveals how AI assistance shapes student work, including variation in query frequency and its link to essay length and readability, along with four distinct writing profiles defined by contribution and revision patterns: Lead Authors, Collaborators, Drafters, and Vibe Writers.

What carries the argument

The NIRVANA dataset, which combines keystroke logging, full AI conversation records, and copied text segments to support exact replay of each student's writing process.

If this is right

  • Educators can examine specific moments when students request AI help to design better guidance on when and how to use such tools.
  • The four profiles indicate that different students rely on AI in distinct ways, suggesting tailored feedback rather than uniform rules.
  • Correlations between query frequency and essay traits such as length and readability provide measurable indicators of AI influence on output quality.
  • The replay interface allows systematic review of individual interactions, turning raw logs into observable sequences of student-AI collaboration.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar detailed logging could be extended to other writing tasks or AI tools to test whether the same profiles appear across subjects.
  • If the profiles prove stable over time, assignment design could deliberately steer students toward profiles that emphasize original contribution.
  • The dataset opens the possibility of comparing monitored versus unmonitored conditions to quantify any effect of awareness on behavior.
  • Policy discussions on AI in education could use the observed patterns to set limits on acceptable AI use based on measurable student actions rather than self-reports.

Load-bearing premise

Students used ChatGPT in their usual way even though they knew every action was being recorded and that the patterns seen in this group of 77 participants apply to students more broadly.

What would settle it

A follow-up study that logs the same essay task without any monitoring and finds markedly different query rates, revision amounts, or profile distributions would show that the recorded behaviors do not match natural use.

Figures

Figures reproduced from arXiv: 2604.07344 by Alice Jang, Andrew Jelson, Daniel Dunlap, Daniel Manesh, Sang Won Lee, Sangwook Lee, Tamara Maddox, Young-Ho Kim.

Figure 1
Figure 1. Figure 1: A screenshot of the Editor Interface Mean = 381.9 Median = 344 0 5 10 15 20 0 250 500 750 1000 1250 1500 Total Words (20 bins) Participants (a) Word count distribution across participants Mean = 4.5 Median = 2 0 5 10 15 20 0 5 10 15 20 25 30 GPT Inquiries (binwidth of 1 query) Participants (b) Inquiry distribution across participants [PITH_FULL_IMAGE:figures/full_fig_p004_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Distribution of word count and inquiry counts across participants [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Spearman Correlation for word count and Dale-Chall Readability [PITH_FULL_IMAGE:figures/full_fig_p006_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Distributions of HCR and HER interpreted as an approximation of observable editing effort rather than a comprehensive measure of cognitive contribution. 5.3 Clustering Writer’s Profile Based on Contribution and Edit We used HCR and HER to identify distinct patterns of ChatGPT usage by positioning each participant within a two-dimensional space defined by the two metrics (see [PITH_FULL_IMAGE:figures/full_… view at source ↗
Figure 6
Figure 6. Figure 6: Inquiry usage across clusters asked ChatGPT to rewrite it, subsequently copying and pasting the revised version into the editor. 5.4 Behavioral and Perceptual Differences Across Writer Clusters Using the previously identified clusters, we conducted a Kruskal–Wallis test to examine whether the writer groups differed significantly across measures of the writing process. Post hoc analyses were performed using… view at source ↗
Figure 7
Figure 7. Figure 7: Time taken across clusters *** *** *** 0 2 4 6 8 10 12 14 Lead Authors Collaborators Drafters Vibe Writers Cluster Dale−Chall Score [PITH_FULL_IMAGE:figures/full_fig_p008_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Dale–Chall scores across clusters * *** 0 1 2 3 4 5 6 7 Lead Authors Collaborators Drafters Vibe Writers Cluster Perceived Ownership [PITH_FULL_IMAGE:figures/full_fig_p008_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Perceived Ownership across clusters 6 Replay Interface Development for In-depth Analysis Although this clustering provides a useful framework for character￾izing usage patterns reflected in editing activity, it does not capture other important forms of interaction that do not involve direct edits. For example, P47, who belongs to the Lead Authors group, drew ideas from ChatGPT’s response to a copy-pasted w… view at source ↗
Figure 10
Figure 10. Figure 10: Writing page of our application P26’s case [PITH_FULL_IMAGE:figures/full_fig_p009_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: SSE for our clusters [PITH_FULL_IMAGE:figures/full_fig_p015_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: A screenshot of in-house ChatGPT provided to participants [PITH_FULL_IMAGE:figures/full_fig_p015_12.png] view at source ↗
read the original abstract

With the rapid adoption of AI writing assistants in education, educators and researchers need empirical evidence to understand the impact on student writing and inform effective pedagogical design. Despite widespread use, we lack systematic understanding of how students engage with these tools during authentic writing tasks: when they seek assistance, what they ask, and how they incorporate AI-generated content into their essays. This gap limits evidence-based policy development and rigorous evaluation of generative AI's learning effects. To address this gap, we introduce NIRVANA, a dataset capturing how university students use generative AI while writing an analytical essay. The dataset includes 77 students who completed an essay task with access to ChatGPT, recording keystroke-level writing behavior, full ChatGPT conversation histories, and all text copied from ChatGPT, enabling a complete reconstruction of the writing process and revealing how AI assistance shapes student work. Our analysis identifies key behavioral patterns, including variation in ChatGPT query frequency and its relationship to essay characteristics such as length and readability. We identify four writing profiles based on students' contribution and revision patterns: Lead Authors, Collaborators, Drafters, and Vibe Writers. To support deeper investigation, we developed a replay interface that reconstructs the writing process; qualitative analysis of sampled replays demonstrates how this tool enables systematic examination of student-AI interactions.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript introduces the NIRVANA dataset collected from 77 university students who completed an analytical essay task with access to ChatGPT. It records keystroke-level writing behavior, full ChatGPT conversation histories, and all text copied from ChatGPT, enabling complete reconstruction of the writing process. The authors report behavioral patterns such as variation in query frequency and its relation to essay length and readability, and classify students into four profiles (Lead Authors, Collaborators, Drafters, and Vibe Writers) based on contribution and revision patterns. A replay interface is provided to support qualitative examination of student-AI interactions.

Significance. If the data collection procedures are sound and the observed patterns are not substantially distorted by the monitored setting, NIRVANA would constitute a valuable resource for HCI and education research. The multimodal logs (keystrokes, full histories, copied text) allow fine-grained process reconstruction that is rare in this domain and could support reproducible studies of how generative AI shapes writing strategies and outcomes.

major comments (3)
  1. [Data collection and experimental setup] The manuscript provides no discussion of potential observer or Hawthorne effects arising from the explicit recording of every keystroke and conversation. This is load-bearing for the central claim that the dataset captures how students 'use generative AI' in a manner that reveals authentic shaping of student work, as the controlled, monitored task may have systematically altered query frequency, revision behavior, or strategy selection.
  2. [Analysis and profile identification] The identification of the four writing profiles is described only at a high level in the abstract and analysis summary. No quantitative definitions of 'contribution and revision patterns,' clustering or classification method, threshold criteria, or stability checks across subsamples are provided, undermining the claim that these profiles are stable and reproducible.
  3. [Methods] Essential methodological details are absent: participant recruitment procedures, demographics, informed consent process, IRB/ethics approval, data cleaning steps, and any controls for task-specific effects. These omissions prevent evaluation of selection bias, generalizability beyond the single essay task, and reproducibility of the reported patterns.
minor comments (2)
  1. The replay interface is introduced as enabling systematic examination, but no details on its implementation, data format, or public release are given; adding these would strengthen the dataset's utility without altering the core claims.
  2. [Abstract] The abstract references relationships between query frequency and essay characteristics (length, readability) but does not preview any quantitative results, effect sizes, or figures; including a brief summary would improve clarity.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive and detailed feedback on our manuscript. We address each major comment point by point below, indicating the revisions we will make to strengthen the paper's clarity, completeness, and transparency.

read point-by-point responses
  1. Referee: [Data collection and experimental setup] The manuscript provides no discussion of potential observer or Hawthorne effects arising from the explicit recording of every keystroke and conversation. This is load-bearing for the central claim that the dataset captures how students 'use generative AI' in a manner that reveals authentic shaping of student work, as the controlled, monitored task may have systematically altered query frequency, revision behavior, or strategy selection.

    Authors: We agree that the absence of an explicit discussion of observer or Hawthorne effects is a limitation in the current manuscript. The data collection occurred in a monitored setting with keystroke logging and full conversation capture, which may have influenced participants' query strategies or revision behaviors compared to unmonitored use. In the revised version, we will add a dedicated paragraph in the Limitations section acknowledging this effect, describing the steps taken to make the task feel as natural as possible (e.g., framing it as a standard course assignment), and cautioning readers about generalizability to fully private writing contexts. This addition will provide a more balanced interpretation of the dataset without overstating its ecological validity. revision: yes

  2. Referee: [Analysis and profile identification] The identification of the four writing profiles is described only at a high level in the abstract and analysis summary. No quantitative definitions of 'contribution and revision patterns,' clustering or classification method, threshold criteria, or stability checks across subsamples are provided, undermining the claim that these profiles are stable and reproducible.

    Authors: The referee correctly identifies that the profile descriptions lack sufficient quantitative detail and methodological transparency. The four profiles (Lead Authors, Collaborators, Drafters, and Vibe Writers) were derived from two primary dimensions: (1) contribution, measured as the percentage of final essay text originating from the student versus copied AI output, and (2) revision patterns, quantified via edit distance and number of post-copy modifications per paragraph. These features were normalized and used for classification. However, we acknowledge the manuscript does not specify the exact clustering procedure, any thresholds applied, or validation metrics. In the revision, we will expand the Analysis section with a new subsection providing precise definitions of all metrics, the classification algorithm and parameters, example threshold values, and stability assessments (e.g., consistency across bootstrap subsamples). This will make the profiles fully reproducible. revision: yes

  3. Referee: [Methods] Essential methodological details are absent: participant recruitment procedures, demographics, informed consent process, IRB/ethics approval, data cleaning steps, and any controls for task-specific effects. These omissions prevent evaluation of selection bias, generalizability beyond the single essay task, and reproducibility of the reported patterns.

    Authors: We accept that the Methods section in the submitted manuscript is insufficiently detailed on these points. The revised manuscript will expand this section to include: recruitment via university-wide email lists and course announcements targeting students in writing-related classes; summary demographics from a pre-task survey (age, gender, major, and prior generative AI experience); the electronic informed consent procedure with specifics on data anonymization and usage; reference to the IRB approval (including protocol number); data cleaning protocols (e.g., exclusion criteria for incomplete logs or technical errors); and rationale for the analytical essay task with discussion of potential task-specific effects. These additions will allow readers to better evaluate selection bias, generalizability, and reproducibility. revision: yes

Circularity Check

0 steps flagged

No circularity: purely empirical dataset and observational profile identification

full rationale

The paper is an observational study that collects keystroke logs, ChatGPT histories, and copied text from 77 students completing a single essay task, then reports behavioral patterns and four profiles (Lead Authors, Collaborators, Drafters, Vibe Writers) derived directly from those observations. No equations, fitted parameters, predictions, or derivation chains exist that could reduce to self-definition or self-citation. Profile labels are post-hoc categorizations of measured contribution/revision metrics; the central claim that the dataset enables reconstruction rests on the recorded data itself rather than any circular reduction. External-validity concerns (observer effects) are separate from circularity and do not appear in any load-bearing step.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The contribution rests on the dataset itself rather than on new theory; the analysis assumes standard HCI observation methods suffice to extract meaningful profiles.

axioms (2)
  • domain assumption The recorded interactions reflect students' typical behavior with generative AI outside the study setting.
    The paper claims the data reveals how AI assistance shapes student work without discussing observer effects or task artificiality.
  • domain assumption The sample of 77 students is adequate to identify stable behavioral profiles applicable beyond this cohort.
    Profiles are presented as general findings without reported validation on held-out data or discussion of demographic limits.

pith-pipeline@v0.9.0 · 5557 in / 1530 out tokens · 73567 ms · 2026-05-10T16:58:43.219121+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

55 extracted references · 42 canonical work pages · 1 internal anchor

  1. [1]

    [n. d.]. https://www.act.org/content/act/en/products-and-services/the-act/test- preparation/writing-sample-essays.html?page=0&chapter=0

  2. [2]

    Jisuanke/CodeMirror-Record

    2025. Jisuanke/CodeMirror-Record. https://github.com/Jisuanke/CodeMirror- Record original-date: 2019-05-08T13:32:58Z

  3. [3]

    Catherine Adams, Patti Pente, Gillian Lemermeyer, Joni Turville, and Geoffrey Rockwell. 2022. Artificial Intelligence and Teachers’ New Ethical Obligations.The International Review of Information Ethics31, 1 (Nov. 2022). doi:10.29173/irie483 Number: 1

  4. [4]

    Tazin Afrin, Omid Kashefi, Christopher Olshefski, Diane Litman, Rebecca Hwa, and Amanda Godley. 2021. Effective Interfaces for Student-Driven Revision Sessions for Argumentative Writing. InProceedings of the 2021 CHI Conference on Human Factors in Computing Systems (CHI ’21). Association for Computing Machinery, New York, NY, USA, 1–13. doi:10.1145/341176...

  5. [5]

    Avey, Bruce J

    James B. Avey, Bruce J. Avolio, Craig D. Crossley, and Fred Luthans. 2009. Psy- chological ownership: theoretical extensions, measurement and relation to work outcomes.Journal of Organizational Behavior30, 2 (2009), 173–191. doi:10.1002/ job.583 _eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/job.583

  6. [6]

    Alex Barrett and Austin Pack. 2023. Not quite eye to A.I.: student and teacher perspectives on the use of generative artificial intelligence in the writing process. International Journal of Educational Technology in Higher Education20, 1 (Nov. 2023), 59. doi:10.1186/s41239-023-00427-0

  7. [7]

    Matt Bower, Jodie Torrington, Jennifer W. M. Lai, Peter Petocz, and Mark Alfano

  8. [8]

    2024), 15403–15439

    How should we change teaching and assessment in response to increasingly powerful generative Artificial Intelligence? Outcomes of the ChatGPT teacher survey.Education and Information Technologies29, 12 (Aug. 2024), 15403–15439. doi:10.1007/s10639-023-12405-0

  9. [9]

    Kauffman, Courtney McKim, and Sharon Zumbrunn

    Roger Bruning, Michael Dempsey, Douglas F. Kauffman, Courtney McKim, and Sharon Zumbrunn. 2013. Examining dimensions of self-efficacy for writing. Journal of Educational Psychology105, 1 (2013), 25–38. doi:10.1037/a0029692 Place: US Publisher: American Psychological Association

  10. [10]

    Tuhin Chakrabarty, Philippe Laban, Divyansh Agarwal, Smaranda Muresan, and Chien-Sheng Wu. 2024. Art or Artifice? Large Language Models and the False Promise of Creativity. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24). Association for Computing Machinery, New York, NY, USA, 1–34. doi:10.1145/3613904.3642731

  11. [11]

    Tuhin Chakrabarty, Vishakh Padmakumar, Faeze Brahman, and Smaranda Mure- san. 2024. Creativity Support in the Age of Large Language Models: An Empirical Study Involving Professional Writers. InProceedings of the 16th Conference on Creativity & Cognition (C&C ’24). Association for Computing Machinery, New York, NY, USA, 132–155. doi:10.1145/3635636.3656201

  12. [12]

    Olckers Chantal. 2012. Psychological ownership: Development of an instrument. SA Journal of Industrial Psychology39 (Dec. 2012), 1–13. doi:10.4102/sajip.v39i2. 1105

  13. [13]

    Fumian Chen, Sotheara Veng, Joshua Wilson, Xiaoming Li, and Hui Fang. 2025. CoachGPT: A Scaffolding-based Academic Writing Assistant. InProceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’25). Association for Computing Machinery, New York, NY, USA, 4051–4055. doi:10.1145/3726302.3730143

  14. [14]

    Erin Cherry and Celine Latulipe. 2014. Quantifying the Creativity Support of Digital Tools through the Creativity Support Index.ACM Trans. Comput.-Hum. Interact.21, 4 (June 2014), 21:1–21:25. doi:10.1145/2617588

  15. [15]

    William Condon and Diane Kelly-Riley. 2004. Assessing and teaching what we value: The relationship between college-level writing and critical thinking abilities.Assessing Writing9, 1 (2004), 56–75

  16. [16]

    Debby R. E. Cotton, Peter A. Cotton, and J. Reuben Shipway. 2023. Chatting and cheating: Ensuring academic integrity in the era of ChatGPT.Innovations in Education and Teaching International0, 0 (2023), 1–12. doi:10.1080/14703297. 2023.2190148 _eprint: https://doi.org/10.1080/14703297.2023.2190148

  17. [17]

    Helen Crompton and Diane Burke. 2023. Artificial intelligence in higher edu- cation: the state of the field.International Journal of Educational Technology in Higher Education20, 1 (April 2023), 22. doi:10.1186/s41239-023-00392-8

  18. [18]

    Edgar Dale and Jeanne S. Chall. 1948. A Formula for Predicting Readability. Educational Research Bulletin27, 1 (1948), 11–28. https://www.jstor.org/stable/ 1473169

  19. [19]

    Fred D. Davis. 1989. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology.MIS Quarterly13, 3 (1989), 319–340. doi:10.2307/249008 Publisher: Management Information Systems Research Center, University of Minnesota

  20. [20]

    Damian Okaibedi Eke. 2023. ChatGPT and the rise of generative AI: Threat to academic integrity?Journal of Responsible Technology13 (2023), 100060. doi:10.1016/j.jrt.2023.100060

  21. [21]

    Linda Flower and John R. Hayes. 1981. A Cognitive Process Theory of Writing. College Composition and Communication32, 4 (1981), 365–387. doi:10.2307/356600 Publisher: National Council of Teachers of English

  22. [22]

    Katy Ilonka Gero, Tao Long, and Lydia B Chilton. 2023. Social Dynamics of AI Support in Creative Writing. InProceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI ’23). Association for Computing Machinery, New York, NY, USA, 1–15. doi:10.1145/3544548.3580782

  23. [23]

    grammarly. 2023. grammarly. https://www.grammarly.com/. Accessed: 2024-10- 29

  24. [24]

    Alicia Guo, Shreya Sathyanarayanan, Leijie Wang, Jeffrey Heer, and Amy Zhang

  25. [25]

    doi:10.48550/arXiv.2411.03137 arXiv:2411.03137 [cs]

    From Pen to Prompt: How Creative Writers Integrate AI into their Writing Practice. doi:10.48550/arXiv.2411.03137 arXiv:2411.03137 [cs]

  26. [26]

    Andreas Göldi, Thiemo Wambsganss, Seyed Parsa Neshaei, and Roman Rietsche

  27. [27]

    InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24)

    Intelligent Support Engages Writers Through Relevant Cognitive Processes. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24). Association for Computing Machinery, New York, NY, USA, 1–12. doi:10.1145/3613904.3642549

  28. [28]

    Jieun Han, Haneul Yoo, Yoonsu Kim, Junho Myung, Minsun Kim, Hyunseung Lim, Juho Kim, Tak Yeon Lee, Hwajung Hong, So-Yeon Ahn, and Alice Oh. 2023. RECIPE: How to Integrate ChatGPT into EFL Writing Education. InProceedings of the Tenth ACM Conference on Learning @ Scale. ACM, Copenhagen Denmark, 416–420. doi:10.1145/3573051.3596200

  29. [29]

    Hui-Wen Huang, Zehui Li, and Linda Taylor. 2020. The Effectiveness of Using Grammarly to Improve Students’ Writing Skills. InProceedings of the 5th Interna- tional Conference on Distance Education and Learning (ICDEL ’20). Association for Computing Machinery, New York, NY, USA, 122–127. doi:10.1145/3402569. 3402594

  30. [30]

    Andrew Jelson, Daniel Manesh, Alice Jang, Daniel Dunlap, Young-Ho Kim, and Sang Won Lee. 2025. An Empirical Study to Understand How Students Use ChatGPT for Writing Essays. doi:10.48550/arXiv.2501.10551 arXiv:2501.10551 [cs]

  31. [31]

    Nikhita Joshi and Daniel Vogel. 2025. Writing with AI Lowers Psychologi- cal Ownership, but Longer Prompts Can Help. doi:10.48550/arXiv.2404.03108 Conference’17, July 2017, Washington, DC, USA Jelson, Manesh, Sangwook Lee, Sang won Lee arXiv:2404.03108 [cs]

  32. [32]

    George R. Klare. 2000. The measurement of readability: useful information for communicators.ACM J. Comput. Doc.24, 3 (Aug. 2000), 107–121. doi:10.1145/ 344599.344630

  33. [33]

    Simon Knight, Antonette Shibani, Sophie Abel, Andrew Gibson, Philippa Ryan, Nicole Sutton, Raechel Wight, Cherie Lucas, Ágnes Sándor, Kirsty Kitto, Ming Liu, Radhika Vijay Mogarkar, and Simon Buckingham Shum. 2020. AcaWriter: A learning analytics tool for formative feedback on academic writing.Journal of Writing Research12, 1 (June 2020), 141–186. doi:10....

  34. [34]

    Svetlana Koltovskaia. 2020. Student engagement with automated written correc- tive feedback (AWCF) provided byGrammarly: A multiple case study.Assessing Writing44 (April 2020), 100450. doi:10.1016/j.asw.2020.100450

  35. [35]

    Nataliya Kosmyna, Eugene Hauptmann, Ye Tong Yuan, Jessica Situ, Xian-Hao Liao, Ashly Vivian Beresnitzky, Iris Braunstein, and Pattie Maes. 2025. Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task. arXiv:2506.08872 [cs.AI] https://arxiv.org/abs/2506.08872

  36. [36]

    Mina Lee, Percy Liang, and Qian Yang. 2022. CoAuthor: Designing a Human-AI Collaborative Writing Dataset for Exploring Language Model Capabilities. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (CHI ’22). Association for Computing Machinery, New York, NY, USA, 1–19. doi:10.1145/3491102.3502030

  37. [37]

    Rongxin Liu, Carter Zenke, Charlie Liu, Andrew Holmes, Patrick Thornton, and David J. Malan. 2024. Teaching CS50 with AI: Leveraging Generative Artificial Intelligence in Computer Science Education. InProceedings of the 55th ACM Technical Symposium on Computer Science Education V. 1 (SIGCSE 2024). Association for Computing Machinery, New York, NY, USA, 75...

  38. [38]

    Victoria Livingstone. 2024. I Quit Teaching Because of ChatGPT.TIME(30 September 2024). https://time.com/7026050/chatgpt-quit-teaching-ai-essay/

  39. [39]

    Fiammetta Marulli, Lelio Campanile, Maria Stella de Biase, Stefano Marrone, Laura Verde, and Marianna Bifulco. 2024. Understanding Readability of Large Language Models Output: An Empirical Analysis.Procedia Computer Science246 (Jan. 2024), 5273–5282. doi:10.1016/j.procs.2024.09.636

  40. [40]

    Peterson, Sohaib Z

    Melissa Romoff, Madison Brunette, Melanie K. Peterson, Sohaib Z. Hashmi, and Michael S. Kim. 2025. The role of large language models in improving the read- ability of orthopaedic spine patient educational material.Journal of Orthopaedic Surgery and Research20 (May 2025), 531. doi:10.1186/s13018-025-05955-1

  41. [41]

    Advait Sarkar and Ian Drosos. 2025. Vibe coding: programming through conver- sation with artificial intelligence. doi:10.48550/arXiv.2506.23253 arXiv:2506.23253 [cs]

  42. [42]

    Turing Test

    Peter Scarfe, Kelly Watcham, Alasdair Clarke, and Etienne Roesch. 2024. A real-world test of artificial intelligence infiltration of a university examinations system: A “Turing Test” case study.PloS one19, 6 (2024), e0305354

  43. [43]

    Mei Tan, Christopher Mah, and Dorottya Demszky. 2024. Reframing Authority: A Computational Measure of Power-Affirming Feedback on Student Writing. In Proceedings of the Eleventh ACM Conference on Learning @ Scale. ACM, Atlanta GA USA, 417–421. doi:10.1145/3657604.3664680

  44. [44]

    Thorndike

    Robert L. Thorndike. 1953. Who Belongs in the Family?Psychometrika18, 4 (Dec. 1953), 267–276. doi:10.1007/BF02289263

  45. [45]

    Don Vandewalle, Linn Van Dyne, and Tatiana Kostova. 1995. Psychological Own- ership: An Empirical Examination of its Consequences.Group & Organization Management20, 2 (June 1995), 210–226. doi:10.1177/1059601195202008 Publisher: SAGE Publications Inc

  46. [46]

    Carole Wade. 1995. Using writing to develop and assess critical thinking.Teaching of psychology22, 1 (1995), 24–28

  47. [47]

    Jin Wang and Wenxiang Fan. 2025. The effect of ChatGPT on students’ learning performance, learning perception, and higher-order thinking: insights from a meta-analysis.Humanities and Social Sciences Communications12, 1 (May 2025),

  48. [48]

    doi:10.1057/s41599-025-04787-y

  49. [49]

    Azmine Toushik Wasi, Mst Rafia Islam, and Raima Islam. 2024. LLMs as Writing Assistants: Exploring Perspectives on Sense of Ownership and Reasoning. In Proceedings of the Third Workshop on Intelligent and Interactive Writing Assistants. 38–42. doi:10.1145/3690712.3690723 arXiv:2404.00027 [cs]

  50. [50]

    Florian Weber, Thiemo Wambsganss, Seyed Parsa Neshaei, and Matthias Soellner

  51. [51]

    InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems(Honolulu, HI, USA) (CHI ’24)

    LegalWriter: An Intelligent Writing Support System for Structured and Persuasive Legal Case Writing for Novice Law Students. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems (CHI ’24). Association for Computing Machinery, New York, NY, USA, 1–23. doi:10.1145/3613904.3642743

  52. [52]

    Simon Wilbers, Johanna Gröpler, Bastian Prell, and Jörg Reiff-Stephan. 2024. Overall Writing Effectiveness: Exploring Students’ Use of LLMs, Pushing the Limits of Automated Text Generation. InSmart Technologies for a Sustainable Future, Michael E. Auer, Reinhard Langmann, Dominik May, and Kim Roos (Eds.). Springer Nature Switzerland, Cham, 11–22. doi:10.1...

  53. [53]

    Ye Xiong and Yi-Fang Brook Wu. 2019. An Automated Feedback System to Support Student Learning in Writing-to-Learn Activities. InProceedings of the Sixth (2019) ACM Conference on Learning @ Scale (L@S ’19). Association for Computing Machinery, New York, NY, USA, 1–4. doi:10.1145/3330430.3333657

  54. [54]

    Rzes- zotarski

    Chao Zhang, Kexin Ju, Peter Bidoshi, Yu-Chun Grace Yen, and Jeffrey M. Rzes- zotarski. 2025. Friction: Deciphering Writing Feedback into Writing Revisions through LLM-Assisted Reflection. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems (CHI ’25). Association for Computing Ma- chinery, New York, NY, USA, 1–27. doi:10.1145/370...

  55. [55]

    1989.Writing to learn

    William K Zinsser. 1989.Writing to learn. Harper & Row, New York. A Appendix A.1 Full List of Dataset Columns Here, we provide a full description of all the dataset columns. •Participant Sheet – id- Participant id to correspond to the essay_num in the Essay Data sheet – Gender- Participant gender – Age- Participant age in age bands – Race- Participant rac...