pith. machine review for the scientific record. sign in

arxiv: 2604.05163 · v1 · submitted 2026-04-06 · 💻 cs.CL · cs.AI

Recognition: 2 theorem links

· Lean Theorem

What Makes a Good Response? An Empirical Analysis of Quality in Qualitative Interviews

Anjalie Field, Jonathan Ivey, Ziang Xiao

Authors on Pith no claims yet

Pith reviewed 2026-05-10 19:02 UTC · model grok-4.3

classification 💻 cs.CL cs.AI
keywords qualitative interviewsresponse qualityinterview corpusNLP evaluationrelevanceclarityinformativenessautomated interviews
0
0 comments X

The pith

Direct relevance to a key research question is the strongest predictor of response quality in qualitative interviews.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Qualitative interviews are valuable only when participant responses contribute meaningfully to the research goals, yet existing ways to measure response quality have not been checked against that standard. This paper builds a new dataset of over 16,000 responses from real interview projects and tests ten different quality measures to see which ones actually track contribution to study findings. The analysis shows that how directly a response addresses a key research question stands out as the strongest signal, while two measures often used in NLP—clarity and surprisal-based informativeness—show no predictive power. These results give concrete guidance for designing better interview protocols and for building automated systems that can conduct or evaluate interviews.

Core claim

By constructing the Qualitative Interview Corpus and having responses annotated for their contribution to study findings, the authors demonstrate that among ten candidate quality measures, only direct relevance to a key research question reliably predicts high contribution scores. Measures of clarity and surprisal-based informativeness, which are commonly applied to NLP-generated interviews, show no significant correlation with actual contribution.

What carries the argument

The Qualitative Interview Corpus of 343 transcripts containing 16,940 responses, together with human annotations of each response's contribution to the study's goals, used to correlate against ten proposed quality metrics.

Load-bearing premise

That trained annotators can consistently and objectively determine how much an individual response contributes to the overall study findings.

What would settle it

Re-annotating a subset of the responses using a different set of criteria for contribution or different annotators and finding that relevance no longer ranks as the top predictor would undermine the central result.

Figures

Figures reproduced from arXiv: 2604.05163 by Anjalie Field, Jonathan Ivey, Ziang Xiao.

Figure 1
Figure 1. Figure 1: We compare the distributions of quality in [PITH_FULL_IMAGE:figures/full_fig_p006_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: We compare the distribution of quality re [PITH_FULL_IMAGE:figures/full_fig_p007_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: LLM Judge prompt used to evaluate the attributed meaning of participant statements. [PITH_FULL_IMAGE:figures/full_fig_p012_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: LLM Judge prompt used to evaluate the clarity of participant statements. [PITH_FULL_IMAGE:figures/full_fig_p012_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: LLM Judge prompt used to evaluate the relevance of participant statements to the interviewer’s question. [PITH_FULL_IMAGE:figures/full_fig_p013_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: LLM Judge prompt used to evaluate the self-reportedness of participant statements. [PITH_FULL_IMAGE:figures/full_fig_p013_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: LLM Judge prompt used to evaluate the specificity of participant statements. [PITH_FULL_IMAGE:figures/full_fig_p014_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: LLM Judge prompt used to evaluate the spontaneity of participant statements. [PITH_FULL_IMAGE:figures/full_fig_p014_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: LLM Judge prompt used to evaluate the relevance of participant statements to a key research question. [PITH_FULL_IMAGE:figures/full_fig_p015_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: LLM Judge prompt used to evaluate the likelihood that participant statements contributed to the results [PITH_FULL_IMAGE:figures/full_fig_p016_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: LLM Judge prompt used to identify the techniques used in interviewer statements. [PITH_FULL_IMAGE:figures/full_fig_p017_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: An example of the interface used to collect human annotations. The participant response is redacted to [PITH_FULL_IMAGE:figures/full_fig_p019_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: P-values from Dunn’s post-hoc test for difference in median response quality between pairs of interview [PITH_FULL_IMAGE:figures/full_fig_p023_13.png] view at source ↗
Figure 14
Figure 14. Figure 14: Correlations observed in the Qualitative Interview Corpus between each pair of characteristics in our [PITH_FULL_IMAGE:figures/full_fig_p024_14.png] view at source ↗
read the original abstract

Qualitative interviews provide essential insights into human experiences when they elicit high-quality responses. While qualitative and NLP researchers have proposed various measures of interview quality, these measures lack validation that high-scoring responses actually contribute to the study's goals. In this work, we identify, implement, and evaluate 10 proposed measures of interview response quality to determine which are actually predictive of a response's contribution to the study findings. To conduct our analysis, we introduce the Qualitative Interview Corpus, a newly constructed dataset of 343 interview transcripts with 16,940 participant responses from 14 real research projects. We find that direct relevance to a key research question is the strongest predictor of response quality. We additionally find that two measures commonly used to evaluate NLP interview systems, clarity and surprisal-based informativeness, are not predictive of response quality. Our work provides analytic insights and grounded, scalable metrics to inform the design of qualitative studies and the evaluation of automated interview systems.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper introduces the Qualitative Interview Corpus (343 transcripts, 16,940 responses from 14 real projects) and evaluates 10 proposed measures of response quality to identify which predict a response's contribution to study findings. It reports that direct relevance to a key research question is the strongest predictor, while clarity and surprisal-based informativeness are not predictive of quality.

Significance. If the central results hold after addressing the annotation and analysis details, the work supplies the first large-scale empirical validation of quality measures for qualitative interviews. This could shift evaluation practices in both qualitative research and NLP interview systems away from unvalidated proxies toward relevance-focused metrics, with direct implications for study design and automated system assessment.

major comments (3)
  1. [Corpus construction and annotation (likely §3)] The annotation protocol for the dependent variable ('contribution to the study findings') is not described: no coding rubric, number of annotators per response, disagreement resolution procedure, or inter-rater reliability statistics (Cohen’s κ or Krippendorff’s α) are reported. Because every correlation and ranking among the 10 measures rests on these human judgments as ground truth, the absence of this information renders the headline claims (relevance as strongest predictor; clarity and surprisal as non-predictive) impossible to evaluate for reliability or bias.
  2. [Results and statistical analysis (likely §5)] No information is supplied on the statistical tests, effect sizes, sample sizes per measure, or controls for confounds (question type, interviewer experience, project domain). Without these, it is unclear whether the reported superiority of relevance over the other nine measures survives multiple-comparison correction or is an artifact of unmodeled dependencies in the 16,940 responses.
  3. [Measure definitions and implementation (likely §4)] The operational definitions and implementation details for the ten measures themselves are insufficiently specified to permit replication or to diagnose why clarity and surprisal-based informativeness fail to correlate with contribution scores.
minor comments (2)
  1. [Abstract] The abstract would benefit from a single sentence stating the number of annotators and the reliability metric used for the contribution scores.
  2. [Tables and figures] Table or figure captions should explicitly note the exact statistical test and any covariates included when ranking the ten measures.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed feedback. The comments highlight important gaps in reporting that affect the reproducibility and interpretability of our results. We address each point below and will revise the manuscript accordingly to strengthen the paper.

read point-by-point responses
  1. Referee: [Corpus construction and annotation (likely §3)] The annotation protocol for the dependent variable ('contribution to the study findings') is not described: no coding rubric, number of annotators per response, disagreement resolution procedure, or inter-rater reliability statistics (Cohen’s κ or Krippendorff’s α) are reported. Because every correlation and ranking among the 10 measures rests on these human judgments as ground truth, the absence of this information renders the headline claims (relevance as strongest predictor; clarity and surprisal as non-predictive) impossible to evaluate for reliability or bias.

    Authors: We agree that these annotation details are critical for assessing the validity of the ground-truth labels and were omitted from the submitted manuscript. In the revision, we will add a dedicated subsection describing the full annotation protocol: the coding rubric provided to annotators, the number of annotators per response, the disagreement resolution procedure, and the inter-rater reliability statistics. This will enable readers to evaluate potential biases or limitations in the dependent variable. revision: yes

  2. Referee: [Results and statistical analysis (likely §5)] No information is supplied on the statistical tests, effect sizes, sample sizes per measure, or controls for confounds (question type, interviewer experience, project domain). Without these, it is unclear whether the reported superiority of relevance over the other nine measures survives multiple-comparison correction or is an artifact of unmodeled dependencies in the 16,940 responses.

    Authors: We acknowledge that the statistical reporting was insufficient. The revised results section will specify the exact statistical tests performed, effect sizes, sample sizes used for each measure, and any controls or covariates included for confounds such as question type, interviewer experience, and project domain. We will also report whether the key findings hold after appropriate multiple-comparison corrections and discuss potential dependencies in the data. revision: yes

  3. Referee: [Measure definitions and implementation (likely §4)] The operational definitions and implementation details for the ten measures themselves are insufficiently specified to permit replication or to diagnose why clarity and surprisal-based informativeness fail to correlate with contribution scores.

    Authors: We agree that greater specificity is needed for replication. The revised manuscript will expand the measure definitions section with precise operational details, including formulas or pseudocode where applicable, preprocessing steps, and the exact tools or libraries used for implementation. This will also help clarify why certain measures (such as clarity and surprisal-based informativeness) did not predict contribution scores. revision: yes

Circularity Check

0 steps flagged

No significant circularity in empirical evaluation

full rationale

The paper performs an empirical analysis by constructing a new corpus of 16,940 responses from 14 projects and correlating 10 implemented measures against human-annotated contribution scores. No equations, parameter-fitting steps, or derivations are present that would reduce any reported predictor (such as direct relevance) to a quantity computed from the same contribution labels by construction. The measures are drawn from prior literature and applied independently to the text, while the annotations serve as an external ground truth; the comparison therefore remains self-contained and does not collapse into its inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The study is empirical and relies on standard practices in corpus construction and statistical prediction without introducing new mathematical axioms or fitted parameters beyond ordinary model fitting.

axioms (1)
  • domain assumption Human annotators can consistently judge whether an individual response contributes to a study's overall findings
    This judgment is required to label the target variable against which the ten measures are evaluated.

pith-pipeline@v0.9.0 · 5461 in / 1253 out tokens · 90666 ms · 2026-05-10T19:02:31.980988+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

36 extracted references · 1 canonical work pages · 1 internal anchor

  1. [1]

    The Pile: An 800GB Dataset of Diverse Text for Language Modeling

    SparkMe: Adaptive Semi-Structured Inter- viewing for Qualitative Insight Discovery. Alemitu Mequanint Bezabih and C. Estelle Smith. 2025. Expanding Models of Delivery for Online Spiritual Care. Kathy Charmaz. 2014.Constructing Grounded Theory. SAGE Publications Ltd, London ; Thousand Oaks, Calif. Deepro Choudhury, Sinead Williamson, Adam Goli´nski, Ning M...

  2. [2]

    State Contexts during the COVID-19 Pandemic

    Data for: What Extraordinary Times Tell Us about Ordinary Ones: A Multiple Case Study of Pre- cariously Employed Food Retail and Service Work- ers in Two U.S. State Contexts during the COVID-19 Pandemic. Ziang Xiao, Michelle X. Zhou, Wenxi Chen, Huahai Yang, and Changyan Chi. 2020a. If I Hear You Cor- rectly: Building and Evaluating Interview Chatbots 10 ...

  3. [3]

    The participant statement does not provide information indicating how significant an action or experience is to the participant

  4. [4]

    The participant statement indicates some level of significance of an action or experience to a participant but does not provide information about what that significance is

  5. [5]

    The participant statement directly shows the significance of an action or experience to the participant and explains what that significance is. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write ...

  6. [6]

    The participant statement is incoherent, or the meaning is completely unclear

  7. [7]

    The participant statement is ambiguous or requires guessing to understand

  8. [8]

    The participant statement is clear to read and understand. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text. Figure 4: LLM Judge prompt used to evaluate the clarity of parti...

  9. [9]

    The participant statement is completely unrelated to the question asked, avoids the question entirely, or addresses a totally different topic

  10. [10]

    The participant statement is related to the general topic of the question but drifts or answers a different question than the one posed

  11. [11]

    The participant statement directly answers the specific question posed by the interviewer. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text. Figure 5: LLM Judge prompt used ...

  12. [12]

    The participant statement is not interpretable without additional context

  13. [13]

    The core idea of the participant statement is understandable but may require additional context for full understanding

  14. [14]

    The participant statement is fully self-contained, not requiring any additional context to be interpretable. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text. Figure 6: LLM ...

  15. [15]

    The participant statement is generic or abstract, providing only high-level summaries or vague descriptions

  16. [16]

    The participant statement describes a particular event, action, or opinion without concrete details or examples

  17. [17]

    The participant statement describes a particular event, action, or opinion with concrete details or examples. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text. Figure 7: LLM...

  18. [18]

    The participant statement only confirms or reiterates information provided in the interviewer’s statement

  19. [19]

    The participant statement adds additional information beyond what was provided in the interviewer’s statement but remains within the topic posed

  20. [20]

    The participant statement introduces a new topic that may be related but was not introduced in the interviewer’s statement. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text....

  21. [21]

    The participant statement is unrelated to the research question or discusses a completely different topic

  22. [22]

    The participant statement is tangentially related to the topic of the research question

  23. [23]

    The participant statement directly addresses the research question. RESEARCH QUESTION: {research_question} CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, or 3). Do not write any additional text. Figure 9: LLM Ju...

  24. [24]

    The statement is unrelated to the results section or contradicts it

  25. [25]

    Tangential relation; discusses the topic but offers no specific substance

  26. [26]

    Aligns with the results section but is general or vague

  27. [27]

    Provides an example or sentiment that matches the results section’s conclusion

  28. [28]

    Appears in the results section and likely served as a primary source for it. RESULTS SECTION: {result} CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {excerpt} CRITICAL: Output only a single digit (1, 2, 3, 4, or 5). Do not write any additional text. Figure 10: LLM...

  29. [29]

    Introductory/Contextualization Questions: Open-ended questions that may be unrelated to the overall research questions but are designed to give an understanding of the participant or context

  30. [30]

    Support and Rapport Building: A statement designed to make a connection with the participant, provide support, or let the participant know that the purpose of the interview is being fulfilled

  31. [31]

    uh-huh” or “mmm,

    Follow-up / Elaboration Probe: A statement designed to encourage a participant to continue talking. It may be a simple “uh-huh” or “mmm,” or it could also be a direct call such as, “Could you say some more about that?”

  32. [32]

    Specifying / Detail-Oriented Probe: Questions that follow up to ask who, where, what, when, or how to obtain a complete and detailed picture of an activity or experience

  33. [33]

    Direct Questioning: A question that directly introduces topics or dimensions and asks the respondent about them

  34. [34]

    Indirect / Projective Questioning: Indirect questions that may ask about the attitudes of others or encourage an indirect statement of the participant’s own motivations, attitudes, or emotions

  35. [35]

    Structuring: A statement that controls the structure of the interview by transitioning topics, redirecting respondents, or breaking off participant answers that may be irrelevant to the purpose of the interview

  36. [36]

    Interpreting: A statement that rephrases or interprets answers provided by the participant to get clarification or reach common ground with the participant. CONTEXT BLURB (context only): {context_blurb} PREVIOUS INTERVIEW EXCERPT (context_only): {previous} CURRENT INTERVIEW EXCERPT (rate this): {current_excerpt} CRITICAL: Output only digits (1, 2, 3, 4, 5...