pith. machine review for the scientific record. sign in

arxiv: 2605.08943 · v1 · submitted 2026-05-09 · 💻 cs.CY · cs.HC

Recognition: 2 theorem links

· Lean Theorem

Understanding Student Effort Using Response-Time Propensities During Problem Solving

Authors on Pith no claims yet

Pith reviewed 2026-05-12 01:55 UTC · model grok-4.3

classification 💻 cs.CY cs.HC
keywords student effortresponse timelearning efficiencyadaptive learning systemsalgebra tutoringhierarchical modelingproficiency differencespersistence
0
0 comments X

The pith

Response-time propensities from algebra tutoring logs form stable effort measures whose link to learning efficiency depends on student proficiency.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Adaptive learning systems need better ways to gauge effort because raw time on task mixes careful work with hard problems. This study fits hierarchical models to step-level response times from 794 students across eight algebra tutoring deployments to isolate student-specific propensities while controlling for skill difficulty. The resulting propensities show moderate to strong stability within students across problems. Slower propensities predict higher learning efficiency for higher-proficiency students but weaker or negative relations for lower-proficiency students. The associations appear strongest early in practice sequences and weaken later.

Core claim

Response-time propensities estimated via hierarchical models on step-level logs from algebra tutoring systems capture trait-like individual differences in effort beyond correctness. Slower propensities associate with greater performance improvement per completed step for higher-proficiency students, consistent with constructive processing, whereas for lower-proficiency students the relation is weak or negative, consistent with unproductive struggle or idling. These associations are strongest early in practice sequences and attenuate later in the class period.

What carries the argument

Student- and knowledge-component-level response-time propensities: adjustments to typical step timing extracted from a hierarchical model that accounts for skill difficulty and serves as a scalable proxy for trait-like effort during multi-step problem solving.

If this is right

  • Response-time propensities offer a practical signal for incorporating temporal process data into learner models beyond correctness or raw time.
  • Adaptive supports can be timed and differentiated by early-session propensities and proficiency level to address emerging disengagement.
  • Learning efficiency gains from slower propensities are most detectable at the start of practice sequences.
  • The measure supports targeting interventions when effort is most diagnostic of persistence.
  • Stability across problems allows use as an individual-differences variable in educational technology.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar hierarchical modeling of response times could be tested in other tutoring domains if difficulty adjustments are made domain-specifically.
  • Real-time propensity estimates might enable dynamic problem selection or hint timing before disengagement sets in.
  • Tracking changes in propensities over longer periods could reveal whether effort patterns predict broader course outcomes.

Load-bearing premise

That step-to-step response time, after adjusting for skill difficulty in the hierarchical model, primarily captures trait-like effort differences rather than momentary factors, unmodeled problem features, or idling.

What would settle it

If within-student stability of the propensities disappeared across separate tutoring sessions or if their conditional links to learning efficiency vanished after adding controls for hint usage and exact problem features, the interpretation as stable effort traits would not hold.

Figures

Figures reproduced from arXiv: 2605.08943 by Benjamin W. Domingue, Conrad Borchers, Kexin Yang, Lijin Zhang, Tomohiro Nagashima.

Figure 1
Figure 1. Figure 1: The Lynnette intelligent tutoring system interface [PITH_FULL_IMAGE:figures/full_fig_p005_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Relationship between student learning parameters [PITH_FULL_IMAGE:figures/full_fig_p007_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Relationship between student-level response-time [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
read the original abstract

Adaptive learning systems can produce substantial learning gains, yet many students engage for too brief or too superficial a period to benefit. A central obstacle is measuring effort. Effort during multi-step problem solving is rarely directly observed, and common log-based proxies, such as time on task, cannot distinguish between a student working carefully and a student encountering a harder problem. We examine step-to-step response time as a scalable effort signal by modeling trait-like differences in students' typical response timing during tutoring (while adjusting for skill difficulty). Using step-level logs from eight classroom deployments of algebra tutoring systems (2020 to 2023) across six U.S. schools (794 students), we estimate student- and knowledge-component-level propensities using hierarchical models and relate them to learning efficiency, defined as performance improvement per completed solution step. Response-time propensities show moderate to strong stability within students, supporting their use as an individual differences measure beyond correctness. At the same time, their relationship to learning is not uniform but conditional on the learner and context. Slower propensities predict greater learning efficiency for higher-proficiency students, consistent with constructive processing, whereas for lower-proficiency students, slower propensities are weakly related or even negative, consistent with unproductive struggle or idling. These associations are strongest early in practice sequences and attenuate later in the class period, highlighting an actionable window for detecting emerging disengagement and low persistence. Overall, response-time propensities provide a practical way to incorporate temporal process data into learner models and to target adaptive supports when effort is most diagnostic.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper analyzes step-level response time logs from eight classroom deployments of algebra tutoring systems (794 students across six U.S. schools, 2020-2023). It estimates student- and KC-level response-time propensities via hierarchical models that adjust for skill difficulty, reports moderate-to-strong within-student stability of these propensities, and relates them to learning efficiency (performance gains per completed step). The key findings are conditional: slower propensities predict higher learning efficiency among higher-proficiency students but are weakly or negatively related among lower-proficiency students, with associations strongest early in practice sequences and attenuating later.

Significance. If the propensities can be shown to index stable effort rather than residual confounds, the work supplies a scalable, log-based process measure that augments correctness data in learner models for adaptive systems. The conditional proficiency and temporal effects are potentially actionable for real-time intervention design. The multi-deployment classroom dataset and focus on within-student stability are concrete strengths that support ecological validity.

major comments (3)
  1. [Methods (hierarchical modeling)] The hierarchical model (described in the methods) adjusts solely for KC-level skill difficulty. Without reported checks for within-KC problem-feature variation, ability-speed correlations, or transient state effects, the student propensities remain vulnerable to confounding; this directly undermines the interpretation that they primarily capture trait-like effort and therefore the differential links to learning efficiency.
  2. [Results (learning efficiency regressions)] The results relating propensities to learning efficiency (performance improvement per step) report conditional associations but provide no effect sizes, confidence intervals, model-fit diagnostics, or sensitivity analyses to alternative specifications. These omissions are load-bearing because the central claim of non-uniform, proficiency-moderated effects rests on the statistical robustness of those associations.
  3. [Results (stability analysis)] Stability of propensities within students is asserted as supporting their use as an individual-differences measure, yet no comparison to alternative effort proxies, split-half reliability, or explicit test of whether stability survives additional controls for proficiency is shown. This weakens the claim that the measure adds information beyond correctness.
minor comments (2)
  1. [Abstract] The abstract would be clearer if it briefly stated the random-effects structure and estimation method of the hierarchical model.
  2. [Figures] Figures displaying propensity distributions or regression coefficients should include error bars or credible intervals to aid interpretation of the reported associations.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive and detailed comments, which identify several opportunities to strengthen the clarity and robustness of our analyses. We address each major comment below and outline the revisions we will make.

read point-by-point responses
  1. Referee: [Methods (hierarchical modeling)] The hierarchical model (described in the methods) adjusts solely for KC-level skill difficulty. Without reported checks for within-KC problem-feature variation, ability-speed correlations, or transient state effects, the student propensities remain vulnerable to confounding; this directly undermines the interpretation that they primarily capture trait-like effort and therefore the differential links to learning efficiency.

    Authors: We agree that the primary model adjusts for KC-level difficulty and that explicit checks for additional confounds would improve interpretability. In the revision we will add (1) within-KC analyses incorporating available problem-feature metadata, (2) reported correlations between estimated propensities and proficiency to quantify ability-speed relations, and (3) a sensitivity analysis with session-level controls to address transient state effects. These will appear in an expanded Methods section and supplementary materials. We note that the observed stability of propensities across independent deployments already provides some protection against purely transient or context-specific confounds, but the additional checks will make this explicit. revision: partial

  2. Referee: [Results (learning efficiency regressions)] The results relating propensities to learning efficiency (performance improvement per step) report conditional associations but provide no effect sizes, confidence intervals, model-fit diagnostics, or sensitivity analyses to alternative specifications. These omissions are load-bearing because the central claim of non-uniform, proficiency-moderated effects rests on the statistical robustness of those associations.

    Authors: We will revise the Results section to report standardized effect sizes for the key propensity-by-proficiency interactions, 95% confidence intervals for all coefficients, model-fit statistics (R-squared, AIC), and residual diagnostics. We will also add sensitivity analyses using alternative proficiency thresholds and model specifications (with and without random slopes). These elements will be included in the main text and a new supplementary table. revision: yes

  3. Referee: [Results (stability analysis)] Stability of propensities within students is asserted as supporting their use as an individual-differences measure, yet no comparison to alternative effort proxies, split-half reliability, or explicit test of whether stability survives additional controls for proficiency is shown. This weakens the claim that the measure adds information beyond correctness.

    Authors: We will augment the stability analysis with split-half reliability estimates, direct comparisons of propensity stability to that of correctness-based measures and other available log proxies, and a supplementary test of stability after residualizing propensities on proficiency. These additions will demonstrate incremental validity beyond correctness while preserving the multi-deployment ecological validity of the original results. revision: yes

Circularity Check

0 steps flagged

No significant circularity: propensities and learning efficiency are distinct empirical measures

full rationale

The paper estimates student- and KC-level response-time propensities via hierarchical modeling of step-level logs (adjusting only for skill difficulty), then reports empirical associations between those propensities and an independently defined learning-efficiency metric (performance improvement per completed step). Stability within students is likewise an observed property of the fitted propensities, not a definitional identity. No equation reduces the reported conditional relationships (slower propensities predicting efficiency for high-proficiency students) to the propensity definition itself, no self-citation chain supplies a load-bearing uniqueness theorem, and no ansatz is smuggled in. The use of the same logs for time-based and correctness-based quantities does not create circularity when the quantities remain distinct.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 1 invented entities

Abstract-only review prevents exhaustive enumeration; key derived quantity is the propensity itself, treated as a new effort signal without external validation.

free parameters (1)
  • student response-time propensity
    Trait-like parameter per student estimated via hierarchical model from response times, adjusted for difficulty.
axioms (1)
  • domain assumption Response time after difficulty adjustment primarily reflects effort or persistence
    Invoked to interpret propensities as effort signals rather than other cognitive or contextual factors.
invented entities (1)
  • response-time propensity no independent evidence
    purpose: Individual-differences measure of effort during multi-step problem solving
    Constructed from logs as a stable trait; no independent evidence such as external validation or falsifiable prediction provided in abstract.

pith-pipeline@v0.9.0 · 5590 in / 1291 out tokens · 64562 ms · 2026-05-12T01:55:03.094617+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

64 extracted references · 64 canonical work pages

  1. [1]

    Kamil Akhuseyinoglu and Peter Brusilovsky. 2021. Data Driven Modeling of Learners Individual Differences for Predicting Engagement and Success in Online Learning. InProceedings of the 29th ACM Conference on User Modeling Adaptation and Personalization. Association for Computing Machinery, New York, 201 212. doi:10.1145/3450613.3456834

  2. [2]

    Vincent Aleven, Conrad Borchers, Yun Huang, Tomohiro Nagashima, Bruce McLaren, Paulo Carvalho, Octav Popescu, Jonathan Sewall, and Kenneth Koedinger. 2025. An integrated platform for studying learning with intelligent tutoring systems: CTAT+ TutorShop.arXiv preprint arXiv:2502.10395(2025). Understanding Student Effort Using Response-Time Propensities L@S ...

  3. [3]

    Husni Almoubayyed, Stephen E Fancsali, and Steve Ritter. 2023. Instruction- embedded assessment for reading ability in adaptive mathematics software. In LAK23: 13th International Learning Analytics and Knowledge Conference. 366–377

  4. [4]

    Şeyhmus Aydoğdu. 2020. Predicting Student Final Performance Using Artificial Neural Networks in Online Learning Environments.Education and Information Technologies25 (2020), 1913–1927. doi:10.1007/s10639-019-10053-x

  5. [5]

    gaming the system

    Ryan Baker, Jason Walonoski, Neil Heffernan, Ido Roll, Albert Corbett, and Kenneth Koedinger. 2008. Why students engage in “gaming the system” behavior in interactive learning environments.Journal of Interactive Learning Research19, 2 (2008), 185–224

  6. [6]

    Ryan SJd Baker. 2007. Is gaming the system state-or-trait? Educational data mining through the multi-contextual application of a validated behavioral model. InComplete On-Line Proceedings of the Workshop on Data Mining for User Mod- eling at the 11th International Conference on User Modeling 2007, Vol. 2007. User Modeling Inc Boston, MA, 76–80

  7. [7]

    Ryan S Baker et al. 2019. Challenges for the future of educational data mining: The Baker learning analytics prizes.Journal of educational data mining11, 1 (2019), 1–17

  8. [8]

    gaming the system

    Ryan SJD Baker, Adriana MJB De Carvalho, Jay Raspat, Vincent Aleven, Al- bert T Corbett, and Kenneth R Koedinger. 2009. Educational software features that encourage and discourage “gaming the system”. InArtificial Intelligence in Education. Ios Press, 475–482

  9. [9]

    Douglas Bates, Martin Mächler, Ben Bolker, and Steve Walker. 2015. Fitting linear mixed-effects models using lme4.Journal of statistical software67 (2015), 1–48

  10. [10]

    Conrad Borchers, Paulo F Carvalho, Meng Xia, Pinyang Liu, Kenneth R Koedinger, and Vincent Aleven. 2023. What makes problem-solving practice effective? Comparing paper and AI tutoring. InEuropean conference on technology enhanced learning. Springer, 44–59

  11. [11]

    Conrad Borchers, Hannah Deininger, and Zachary A Pardos. 2026. Toward Trait-Aware Learning Analytics.arXiv preprint arXiv:2602.00018(2026)

  12. [12]

    Conrad Borchers, Tomohiro Nagashima, Pinyang Liu, Martha W Alibali, and Vincent Aleven. 2025. Is More Gamification Better? Evaluating Playful Interac- tions and Narratives for Algebra Learning. InEuropean Conference on Technology Enhanced Learning. Springer, 76–90

  13. [13]

    Conrad Borchers, Kexin Yang, Jionghao Lin, Nikol Rummel, Kenneth R Koedinger, and Vincent Aleven. 2024. Combining dialog acts and skill modeling: What chat interactions enhance learning rates during ai-supported peer tutoring?. InProceedings of the 17th International Conference on Educational Data Mining. 117–130

  14. [14]

    Bråten, N

    I. Bråten, N. Latini, and Y. E. Haverkamp. 2022. Predictors and outcomes of behavioral engagement in the context of text comprehension: when quantity means quality.Reading and Writing35 (2022), 687–711. doi:10.1007/s11145-021- 10174-4

  15. [15]

    Jenny Yun-Chen Chan, Erin R Ottmar, and Ji-Eun Lee. 2022. Slow down to speed up: Longer pause time before solving problems relates to higher strategy efficiency.Learning and Individual Differences93 (2022), 102109

  16. [16]

    Albert Corbett, Megan McLaughlin, and K Christine Scarpinatto. 2000. Modeling student knowledge: Cognitive tutors in high school and college.User modeling and user-adapted interaction10, 2 (2000), 81–108

  17. [17]

    Albert T Corbett and John R Anderson. 1994. Knowledge tracing: Modeling the acquisition of procedural knowledge.User modeling and user-adapted interaction 4, 4 (1994), 253–278

  18. [18]

    Steve C. Dang. 2022.Exploring Behavioral Measurement Models of Learner Moti- vation. Ph. D. Dissertation. Carnegie Mellon University

  19. [19]

    Steven C Dang and Kenneth R Koedinger. 2020. The Ebb and Flow of Student Engagement: Measuring Motivation through Temporal Pattern of Self-Regulation. International Educational Data Mining Society(2020)

  20. [20]

    Benjamin W Domingue, Klint Kanopka, Ben Stenhaug, Michael J Sulik, Tanesia Beverly, Matthieu Brinkhuis, Ruhan Circi, Jessica Faul, Dandan Liao, Bruce Mc- Candliss, et al. 2022. Speed–accuracy trade-off? Not so fast: Marginal changes in speed have inconsistent relationships with accuracy in real-world settings. Journal of Educational and Behavioral Statist...

  21. [21]

    Taryn Eames, Emma Brunskill, Bogdan Yamkovenko, Kodi Weatherholtz, and Philip Oreopoulos. 2026. Computer-assisted learning in the real world: How Khan Academy influences student math learning.Proceedings of the National Academy of Sciences123, 1 (2026), e2507708123

  22. [22]

    Tino Endres, Oliver Lovell, David Morkunas, Werner Rieß, and Alexander Renkl

  23. [23]

    doi:10.1111/bjep.12563

    Can prior knowledge increase task complexity? Cases in which higher prior knowledge leads to higher intrinsic cognitive load.British Journal of Educational Psychology93 (2022), 305–317. doi:10.1111/bjep.12563

  24. [24]

    Ashish Gurung, Anthony F Botelho, and Neil T Heffernan. 2021. Examining student effort on help through response time decomposition. InLAK21: 11th international learning analytics and knowledge conference. 292–301

  25. [25]

    Ashish Gurung, Jionghao Lin, Zhongtian Huang, Conrad Borchers, Ryan S Baker, Vincent Aleven, and Kenneth R Koedinger. 2025. Starting Seatwork Earlier as a Valid Measure of Student Engagement.arXiv preprint arXiv:2505.13341(2025)

  26. [26]

    LAURENCE Holt. 2024. The 5 percent problem.Education Next(2024)

  27. [27]

    Yun Huang, Steven Dang, J Elizabeth Richey, Pallavi Chhabra, Danielle R Thomas, Michael W Asher, Nikki G Lobczowski, Elizabeth A McLaughlin, Judith M Harack- iewicz, Vincent Aleven, et al. 2023. Using latent variable models to make gaming- the-system detection robust to context variations.User Modeling and User-Adapted Interaction33, 5 (2023), 1211

  28. [28]

    Militsa Ivanova, Michalis Michaelides, and Hanna Eklöf. 2020. How does the number of actions on constructed-response items relate to test-taking effort and performance?Educational Research and Evaluation26, 5–6 (2020), 252–274

  29. [29]

    Ivanova and Michalis P

    Militsa G. Ivanova and Michalis P. Michaelides. 2023. Measuring Test-Taking Effort on Constructed-Response Items with Item Response Time and Number of Actions.Practical Assessment, Research & Evaluation28, 15 (2023)

  30. [30]

    Slava Kalyuga. 2009. The expertise reversal effect. InManaging cognitive load in adaptive multimedia learning. IGI Global Scientific Publishing, 58–80

  31. [31]

    Kenneth R Koedinger and Vincent Aleven. 2007. Exploring the assistance dilemma in experiments with cognitive tutors.Educational psychology review19, 3 (2007), 239–264

  32. [32]

    Kenneth R Koedinger, Ryan SJd Baker, Kyle Cunningham, Alida Skogsholm, Brett Leber, and John Stamper. 2010. A data repository for the EDM community: The PSLC DataShop.Handbook of educational data mining43 (2010), 43–56

  33. [33]

    Kenneth R Koedinger, Paulo F Carvalho, Ran Liu, and Elizabeth A McLaughlin

  34. [34]

    An astonishing regularity in student learning rate.Proceedings of the National Academy of Sciences120, 13 (2023), e2221311120

  35. [35]

    Kenneth R Koedinger, Albert T Corbett, and Charles Perfetti. 2012. The Knowledge-Learning-Instruction framework: Bridging the science-practice chasm to enhance robust student learning.Cognitive science36, 5 (2012), 757–798

  36. [36]

    Kenneth R Koedinger, Jihee Kim, Julianna Zhuxin Jia, Elizabeth A McLaughlin, and Norman L Bier. 2015. Learning is not a spectator sport: Doing is better than watching for learning from a MOOC. InProceedings of the second (2015) ACM conference on learning@ scale. 111–120

  37. [37]

    Xiaojing J Kong, Steven L Wise, and Dennison S Bhola. 2007. Setting the response time threshold parameter to differentiate solution behavior from rapid-guessing behavior.Educational and Psychological Measurement67, 4 (2007), 606–619

  38. [38]

    Vitomir Kovanovic, Dragan Gašević, Shane Dawson, Srećko Joksimovic, and Ryan Baker. 2015. Does time-on-task estimation matter? Implications on validity of learning analytics findings.Journal of Learning Analytics2, 3 (2015), 81–110

  39. [39]

    Vitomir Kovanović, Dragan Gašević, Shane Dawson, Srećko Joksimović, Ryan S Baker, and Marek Hatala. 2015. Penetrating the black box of time-on-task esti- mation. InProceedings of the fifth international conference on learning analytics and knowledge. 184–193

  40. [40]

    Chia-An Lee, Jian-Wei Tzeng, Nen-Fu Huang, and Yu-Sheng Su. 2021. Prediction of Student Performance in Massive Open Online Courses Using Deep Learning System Based on Learning Behaviors.Educational Technology & Society24, 3 (2021), 130–146

  41. [41]

    Ji-Eun Lee, Carly Siegel Thorp, Arba Kamberi, and Erin Ottmar. 2025. Unpacking strategy efficiency: Examining the relations between pre-solving pause time and productivity in a digital mathematics game.Metacognition and Learning20, 1 (2025), 31

  42. [42]

    Qiujie Li, Di Xu, Rachel Baker, Amanda Holton, and Mark Warschauer. 2022. Can student-facing analytics improve online students’ effort and success by affecting how they explain the cause of past performance?Computers & Education185 (2022), 104517

  43. [43]

    Ran Liu and Kenneth RK Koedinger. 2017. Towards Reliable and Valid Mea- surement of Individualized Student Parameters.International Educational Data Mining Society(2017)

  44. [44]

    Wu, Xingyu Zhang, Melanie Stefan, Johanna Gutlerner, and Chanmin Kim

    Yan Liu, Audrey Béliveau, Henrike Besche, Amery D. Wu, Xingyu Zhang, Melanie Stefan, Johanna Gutlerner, and Chanmin Kim. 2021. Bayesian Mixed Effects Model and Data Visualization for Understanding Item Response Time and Response Order in Open Online Assessment.Frontiers in Education5 (2021), 607260. doi:10. 3389/feduc.2020.607260

  45. [45]

    Edwin A Locke and Gary P Latham. 2019. The development of goal setting theory: A half century retrospective.Motivation science5, 2 (2019), 93

  46. [46]

    Yanjin Long, Kenneth Holstein, and Vincent Aleven. 2018. What exactly do students learn when they practice equation solving? refining knowledge com- ponents with the additive factors model. InProceedings of the 8th International Conference on Learning Analytics and Knowledge. 399–408

  47. [47]

    Raymond Lynch, Adrian Hurley, Olivia Cumiskey, Brian Nolan, and Bridgeen McGlynn. 2019. Exploring the relationship between homework task difficulty, student engagement and performance.Irish Educational Studies38, 1 (2019), 89–103

  48. [48]

    Mehrdad Mirzaei, Shaghayegh Sahebi, and Peter Brusilovsky. 2020. Detecting Trait versus Performance Student Behavioral Patterns Using Discriminative Non- Negative Matrix Factorization. InProceedings of the 33rd International Florida Artificial Intelligence Research Society Conference. Association for the Advance- ment of Artificial Intelligence, 439–444

  49. [49]

    Jaclyn Ocumpaugh et al. 2015. Baker Rodrigo Ocumpaugh monitoring protocol (BROMP) 2.0 technical and training manual.New York, NY and Manila, Philippines: Teachers College, Columbia University and Ateneo Laboratory for the Learning Sciences60 (2015). L@S ’26, June 29-July 03, 2026, Seoul, Republic of Korea Conrad Borchers, Lijin Zhang, Kexin Yang, Tomohiro...

  50. [50]

    Ernesto Panadero. 2017. A review of self-regulated learning: Six models and four directions for research.Frontiers in Psychology8 (2017), 422. doi:10.3389/fpsyg. 2017.00422

  51. [51]

    Jan Papoušek, Vít Stanislav, and Radek Pelánek. 2016. Impact of question difficulty on engagement and learning. InInternational Conference on Intelligent Tutoring Systems. Springer, 267–272

  52. [52]

    Simonsmeier

    Michael Schneider and Bianca A. Simonsmeier. 2025. How does prior knowledge affect learning? A review of 16 mechanisms and a framework for future research. Learning and Individual Differences122 (2025), 102744. doi:10.1016/j.lindif.2025. 102744

  53. [53]

    Sharmistha Self. 2013. Utilizing online tools to measure effort: Does it really improve student outcome?International Review of Economics Education14 (2013), 36–45

  54. [54]

    Yee Lee Shing and Garvin Brod. 2016. Effects of Prior Knowledge on Memory: Implications for Education.Mind, Brain, and Education10, 3 (2016), 153–161. doi:10.1111/mbe.12110

  55. [55]

    Dirk Tempelaar, Quan Nguyen, and Bart Rienties. 2020. Learning analytics and the measurement of learning engagement. InAdoption of data analytics in higher education learning and teaching. Springer, 159–176

  56. [56]

    David Thissen. 1983. Timed testing: An approach using item response theory. In New horizons in testing. Elsevier, 179–203

  57. [57]

    Kirk Vanacore, Amanda Barany, Shruti Mehta, Jaclyn Ocumpaugh, Ryan S Baker, Nabil Al Nahin, Vishal Kuvar, Caitlin Mills, and Owen Henkel. 2025. Down- shifting: The Nuanced Antecedents of Seeking Easier Content. InProceedings of the 19th International Conference of the Learning Sciences-ICLS 2025, pp. 637-645. International Society of the Learning Sciences

  58. [58]

    Vanacore, Ashish Gurung, Andrew A

    Kirk P. Vanacore, Ashish Gurung, Andrew A. McReynolds, Allison Liu, Stacy T. Shaw, and Neil T. Heffernan. 2023. Impact of Non-Cognitive Interventions on Student Learning Behaviors and Outcomes: An Analysis of Seven Large-Scale Experimental Interventions. InProceedings of the 13th International Learning Analytics and Knowledge Conference. ACM, 165–174. doi...

  59. [59]

    Kurt VanLehn. 2011. The relative effectiveness of human tutoring, intelligent tutoring systems, and other tutoring systems.Educational psychologist46, 4 (2011), 197–221

  60. [60]

    Shiyu Wang and Yinghan Chen. 2020. Using response times and response accu- racy to measure fluency within cognitive diagnosis models.psychometrika85, 3 (2020), 600–629

  61. [61]

    Wise and Xiaojing Kong

    Steven L. Wise and Xiaojing Kong. 2005. Response Time Effort: A New Measure of Examinee Motivation in Computer-Based Tests.Applied Measurement in Education18, 2 (2005), 163–183

  62. [62]

    A. E. Witherby and S. K. Carpenter. 2022. The rich-get-richer effect: Prior knowl- edge predicts new learning of domain-relevant information.Journal of Ex- perimental Psychology: Learning, Memory, and Cognition48, 4 (2022), 483–498. doi:10.1037/xlm0001026

  63. [63]

    Zambrano, Femke Kirschner, John Sweller, and Paul A

    Jimmy R. Zambrano, Femke Kirschner, John Sweller, and Paul A. Kirschner. 2019. Effects of prior knowledge on collaborative and individual learning.Learning and Instruction63 (2019), 101214. doi:10.1016/j.learninstruc.2019.05.011

  64. [64]

    Zimmerman

    Barry J. Zimmerman. 2000. Attaining self-regulation: A social cognitive perspec- tive.Handbook of Self-Regulation(2000), 13–39. doi:10.1016/B978-012109890- 2/50031-7