pith. machine review for the scientific record. sign in

arxiv: 2604.13899 · v2 · submitted 2026-04-15 · 💻 cs.CL · cs.AI

Recognition: unknown

Do We Still Need Humans in the Loop? Comparing Human and LLM Annotation in Active Learning for Hostility Detection

Authors on Pith no claims yet

Pith reviewed 2026-05-10 13:33 UTC · model grok-4.3

classification 💻 cs.CL cs.AI
keywords LLM annotationactive learninghostility detectionerror analysiscost comparisonsocial media classificationTikTok commentsanti-immigrant hostility
0
0 comments X

The pith

LLM-generated labels train hostility detectors to the same F1-Macro level as human labels but at far lower cost.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper examines whether instruction-tuned LLMs can substitute for human annotators inside active learning loops and whether active learning itself remains useful once an entire corpus can be labeled cheaply. On a collection of German political TikTok comments, seven annotation strategies are tested across four encoders for the task of detecting anti-immigrant hostility. A model trained on the full set of 25,974 GPT-5.2 labels at a cost of $43 reaches F1-Macro scores comparable to one trained on 3,800 human labels costing $316. Active learning yields little gain over random sampling in the pre-enriched pool and produces lower performance than labeling the whole set with the LLM at equal cost. Aggregate scores nevertheless conceal a consistent difference: classifiers trained on LLM labels over-predict the positive class relative to the human gold standard, especially on topically ambiguous comments that blend hostility with policy critique.

Core claim

The central claim is that LLM labels can replace human labels inside the active learning loop for hostility detection without loss of aggregate performance and at substantially reduced cost. In the reported experiments, full LLM annotation outperforms both human-only and hybrid active learning strategies on the same budget. The work further shows that this performance parity does not extend to error structure: LLM-trained models systematically over-predict hostility in cases where the boundary between anti-immigrant hostility and legitimate policy discussion is subtle.

What carries the argument

The side-by-side comparison of seven annotation strategies (human, LLM, and active learning variants) on a fixed pre-enriched pool of TikTok comments, evaluated by both aggregate F1-Macro and per-class error profiles.

If this is right

  • Full LLM labeling of large corpora becomes feasible at costs low enough to replace selective human annotation for many classification tasks.
  • Annotation strategy should be chosen according to the error profile required by the downstream application rather than aggregate F1 alone.
  • Active learning offers diminishing returns once the initial pool has already been enriched for the target class.
  • Hybrid human-LLM pipelines may still be needed when the distinction between classes is subtle and context-dependent.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The observed cost advantage could make continuous monitoring of hostility feasible across many more languages and platforms than human-only annotation allows.
  • Similar divergences in error structure are likely to appear in other subjective labeling tasks such as misinformation or sentiment detection.
  • Developers should test LLM-trained models on domain-specific ambiguous examples before deployment rather than relying on overall accuracy figures.

Load-bearing premise

The 5,000 human-annotated comments function as an unbiased gold standard and the pre-enriched comment pool does not artificially restrict the possible gains from active learning.

What would settle it

A follow-up experiment on a fresh, non-enriched corpus of comments or a different platform in which classifiers trained on LLM labels fall significantly below the F1-Macro of human-trained classifiers, or in which the over-prediction of the positive class disappears.

Figures

Figures reproduced from arXiv: 2604.13899 by Ahmad Dawar Hakimi, Hinrich Sch\"utze, Isabelle Augenstein, Lea Hirlimann.

Figure 1
Figure 1. Figure 1: Experimental framework. Human and LLM (Llama-3.3-70B prefiltering + GPT-5.2 classification) annotations feed into the same active learning loop, eval￾uated across seven conditions, four German encoder models, and 10 seeds on a party-stratified test set. longer AL versus human annotation but AL versus LLM annotation at scale. We test whether AL’s cost premium survives that comparison. We investigate this qu… view at source ↗
Figure 2
Figure 2. Figure 2: Learning curves for german_bert (top) and [PITH_FULL_IMAGE:figures/full_fig_p005_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Annotation cost–performance frontier for ger [PITH_FULL_IMAGE:figures/full_fig_p006_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Left: FP/FN counts per condition for german_bert (mean ± 1 SD across 10 seeds; dots = individual seeds). Right: Errors ranked by prediction confidence for FULL-HUMAN and FULL-LLM-26K. Confidence on errors [PITH_FULL_IMAGE:figures/full_fig_p007_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: LLM prompts. A.1: Prefiltering prompt ap￾plied to all 277,902 comments (temperature = 0, JSON output). A.2 Stage 2: Anti-Immigrant Annotation Prompt Prompt A.2: GPT-5.2 Anti-Immigrant Annotation You are an expert in sentiment analysis and political discourse, specializing in German-language text. Your task is to classify TikTok comments posted under videos from German political parties. Classify the commen… view at source ↗
Figure 6
Figure 6. Figure 6: A.2: Classification prompt applied to 25,974 prefiltered comments via GPT-5.2 Batch API (tempera￾ture = 0, structured JSON output). B LLM Infrastructure Details Llama-3.3-70B-Instruct (Stage 1). The model was run locally on four NVIDIA A100 80GB GPUs using vLLM (Kwon, 2025) with tensor par￾allelism across all four devices, bfloat16 preci￾sion, and a maximum sequence length of 512 tokens. Filtering all 277,… view at source ↗
Figure 7
Figure 7. Figure 7: t-SNE projection of all 25,974 immigration-relevant comments, clustered into 22 named discourse [PITH_FULL_IMAGE:figures/full_fig_p015_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Learning curves for gbert-base (top) and xlm-r-base (bottom) on F1-Macro (left) and F1-Anti (right). [PITH_FULL_IMAGE:figures/full_fig_p018_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: shows the predicted P(ANTI￾IMMIGRANT) distributions split by error type for german_bert. FP distributions for LLM-trained conditions peak substantially higher than for human-trained conditions, confirming that LLM-trained classifiers assign high confidence to their false positives. The pattern is consistent across all four encoder models. Correct FP FN 0.0 0.2 0.4 0.6 0.8 1.0 P(ANTI-IMMIGRANT) Full-Human C… view at source ↗
Figure 10
Figure 10. Figure 10: FP–FN imbalance by topic and condition for [PITH_FULL_IMAGE:figures/full_fig_p020_10.png] view at source ↗
read the original abstract

Instruction-tuned LLMs can annotate thousands of instances from a short prompt at negligible cost. This raises two questions for active learning (AL): can LLM labels replace human labels within the AL loop, and does AL remain necessary when entire corpora can be labelled at once? We investigate both questions on a new dataset of 277,902 German political TikTok comments (25,974 LLM-labelled, 5,000 human-annotated), comparing seven annotation strategies across four encoders to detect anti-immigrant hostility. A classifier trained on 25,974 GPT-5.2 labels (\$43) achieves comparable F1-Macro to one trained on 3,800 human annotations (\$316). Active learning offers little advantage over random sampling in our pre-enriched pool and delivers lower F1 than full LLM annotation at the same cost. However, comparable aggregate F1 masks a systematic difference in error structure: LLM-trained classifiers over-predict the positive class relative to the human gold standard. This divergence concentrates in topically ambiguous discussions where the distinction between anti-immigrant hostility and policy critique is most subtle, suggesting that annotation strategy should be guided not by aggregate F1 alone but by the error profile acceptable for the target application.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The paper claims that LLM-based annotation (specifically 25,974 GPT-5.2 labels at $43) can produce classifiers with F1-Macro comparable to those trained on 3,800 human annotations ($316) for anti-immigrant hostility detection in a corpus of 277,902 German TikTok comments. It further finds that active learning provides little benefit over random sampling within the pre-enriched pool and that full LLM annotation outperforms AL at equivalent cost, while noting systematic differences in error profiles where LLM models over-predict the positive class, particularly in topically ambiguous cases.

Significance. If the central empirical comparisons hold, the work has clear practical significance for scaling annotation in applied NLP tasks by quantifying cost savings and emphasizing error-profile analysis over aggregate metrics alone. It contributes concrete evidence to the human-in-the-loop debate through multi-encoder comparisons and qualitative discussion of ambiguity, with strengths in the scale of the dataset and the direct cost-F1 trade-off analysis.

major comments (2)
  1. [Dataset] Dataset section: The 5,000 human-annotated instances are used as the gold standard for all F1-Macro evaluations of LLM-trained classifiers. The manuscript does not specify the sampling procedure (random, uncertainty-based, or stratified) used to select these 5,000 from the pre-enriched pool of 277,902 comments. This detail is load-bearing for the comparability claim, because any over-representation of borderline or positive cases would directly affect measured F1 and could explain the reported LLM over-prediction of the positive class in ambiguous discussions.
  2. [Methods] Methods and Experimental Protocol: The abstract and results lack the exact prompt template used for GPT-5.2, inter-annotator agreement statistics for the human labels, and the full specification of the seven annotation strategies (including how active learning uncertainty was operationalized). These omissions undermine assessment of whether the $43 vs. $316 cost comparison is reproducible and whether the 'comparable F1' result is robust to prompt variation or label noise.
minor comments (1)
  1. [Abstract] Abstract: The claim of 'comparable F1-Macro' would be more informative if the actual F1 values, standard deviations, and any statistical test results were reported directly in the abstract rather than left to the full text.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments, which help strengthen the clarity and reproducibility of the work. We address each major point below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Dataset] Dataset section: The 5,000 human-annotated instances are used as the gold standard for all F1-Macro evaluations of LLM-trained classifiers. The manuscript does not specify the sampling procedure (random, uncertainty-based, or stratified) used to select these 5,000 from the pre-enriched pool of 277,902 comments. This detail is load-bearing for the comparability claim, because any over-representation of borderline or positive cases would directly affect measured F1 and could explain the reported LLM over-prediction of the positive class in ambiguous discussions.

    Authors: We agree that the sampling procedure for the 5,000-instance gold standard must be stated explicitly, as it bears on the validity of the F1 comparisons and error analysis. This information was omitted from the submitted manuscript. We will revise the Dataset section to describe the procedure used (a stratified random sample drawn from the pre-enriched pool to balance topical coverage while preserving the overall class distribution). This addition directly addresses the concern about potential bias in the evaluation set and the observed differences in positive-class predictions. revision: yes

  2. Referee: [Methods] Methods and Experimental Protocol: The abstract and results lack the exact prompt template used for GPT-5.2, inter-annotator agreement statistics for the human labels, and the full specification of the seven annotation strategies (including how active learning uncertainty was operationalized). These omissions undermine assessment of whether the $43 vs. $316 cost comparison is reproducible and whether the 'comparable F1' result is robust to prompt variation or label noise.

    Authors: We acknowledge that full reproducibility requires the prompt template, IAA statistics, and precise operationalization of each annotation strategy. These elements were not included in the submitted version. In the revision we will (1) place the complete GPT-5.2 prompt template in an appendix, (2) report inter-annotator agreement for the human annotations in the Methods section, and (3) expand the description of all seven strategies, including the exact uncertainty measure (entropy over the positive-class probability) used for active learning. These changes will allow readers to evaluate both the cost-F1 trade-offs and robustness to prompt or label variation. revision: yes

Circularity Check

0 steps flagged

No circularity: purely empirical comparisons without derivations or self-referential logic

full rationale

The paper conducts an empirical study comparing human and LLM annotation strategies for training hostility classifiers on a fixed dataset of TikTok comments. All claims rest on direct experimental results (F1-Macro scores from classifiers trained on varying label sets), with no equations, fitted parameters, uniqueness theorems, or ansatzes. The central result—that 25,974 GPT-5.2 labels yield comparable aggregate F1 to 3,800 human labels—is a straightforward performance comparison against the human-annotated subset, not a reduction to its own inputs by construction. No self-citations are load-bearing for the methodology or conclusions.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The paper relies on standard assumptions from machine learning about label quality and sampling representativeness rather than introducing new theoretical elements.

axioms (2)
  • domain assumption Human annotations represent the ground truth for model evaluation
    F1 scores are computed against the human-annotated test set assumed to be correct.
  • domain assumption The pre-enriched pool is representative for testing AL effectiveness
    Conclusions about AL vs random sampling are drawn from this specific pool.

pith-pipeline@v0.9.0 · 5529 in / 1379 out tokens · 93710 ms · 2026-05-10T13:33:05.188713+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

10 extracted references · 5 canonical work pages

  1. [1]

    InProceedings of The 18th Linguistic Annotation Workshop (LAW-XVIII), pages 77–86

    Class balancing for efficient active learning in imbalanced datasets. InProceedings of The 18th Linguistic Annotation Workshop (LAW-XVIII), pages 77–86. Jan Fillies, Esther Theisen, Michael Hoffmann, Robert Jung, Elena Jung, Nele Fischer, and Adrian Paschke

  2. [2]

    Paula Fortuna and Sérgio Nunes

    A novel german tiktok hate speech dataset: far-right comments against politicians, women, and others.Discover Data, 3(1):4. Paula Fortuna and Sérgio Nunes. 2018. A survey on automatic detection of hate speech in text.ACM Comput. Surv., 51(4):85:1–85:30. Yarin Gal and Zoubin Ghahramani. 2016. Dropout as a bayesian approximation: Representing model uncertai...

  3. [3]

    Chatgpt outperforms crowd-workers for text-annotation tasks

    Chatgpt outperforms crowd-workers for text- annotation tasks.CoRR, abs/2303.15056. Maarten Grootendorst. 2022. Bertopic: Neural topic modeling with a class-based TF-IDF procedure. CoRR, abs/2203.05794. Xingwei He, Zhenghao Lin, Yeyun Gong, A-Long Jin, Hang Zhang, Chen Lin, Jian Jiao, Siu Ming Yiu, Nan Duan, and Weizhu Chen. 2024. Annollm: Making large lan...

  4. [4]

    Bayesian active learning for classification and preferenc e learning,

    Large language models as a substitute for human experts in annotating political text.Research & Politics, 11(1):20531680241236239. Neil Houlsby, Ferenc Huszar, Zoubin Ghahramani, and Máté Lengyel. 2011. Bayesian active learning for classification and preference learning.CoRR, abs/1112.5745. Amir Hossein Kargaran, Ayyoob Imani, François Yvon, and Hinrich S...

  5. [5]

    InMachine Learning and Knowledge Discovery in Databases

    Llms in the loop: Leveraging large language model annotations for active learning in low-resource languages. InMachine Learning and Knowledge Discovery in Databases. Applied Data Science Track - European Conference, ECML PKDD 2024, Vilnius, Lithuania, September 9-13, 2024, Proceedings, Part X, Lecture Notes in Computer Science, pages 397–

  6. [6]

    Aida Kostikova, Dominik Beese, Benjamin Paassen, Ole Pütz, Gregor Wiedemann, and Steffen Eger

    Springer. Aida Kostikova, Dominik Beese, Benjamin Paassen, Ole Pütz, Gregor Wiedemann, and Steffen Eger. 2024. Fine-grained detection of solidarity for women and migrants in 155 years of german parliamentary de- bates. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024, Miami, FL, USA, November 12-16, 2024...

  7. [7]

    Jingwei Ni, Yu Fan, Vilém Zouhar, Donya Rooein, Alexander Miserlis Hoyle, Mrinmaya Sachan, Markus Leippold, Dirk Hovy, and Elliott Ash

    Association for Computational Linguistics. Jingwei Ni, Yu Fan, Vilém Zouhar, Donya Rooein, Alexander Miserlis Hoyle, Mrinmaya Sachan, Markus Leippold, Dirk Hovy, and Elliott Ash. 2025. Can large language models capture human annotator disagreements?CoRR, abs/2506.19467. Nicholas Pangakis, Samuel Wolken, and Neil Fasching

  8. [8]

    Moore, Daniel M

    Automated annotation with generative AI re- quires validation.CoRR, abs/2306.00176. Barbara Plank. 2022. The "problem" of human label variation: On ground truth in data, modeling and eval- uation. InProceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, De- cember 7-11, 2022, pa...

  9. [9]

    label":

    Measuring the reliability of hate speech an- notations: The case of the european refugee crisis. CoRR, abs/1701.08118. Paul Röttger, Bertie Vidgen, Dirk Hovy, and Janet B. Pierrehumbert. 2022. Two contrasting data annota- tion paradigms for subjective NLP tasks. InProceed- ings of the 2022 Conference of the North American Chapter of the Association for Co...

  10. [10]

    wann kommen die Grenzkontrollen?

    wir vergessen nicht” (‘Merkel 2015. . . we don’t forget’). The remainder (6/18) demand bor- der controls without referencing any group: “wann kommen die Grenzkontrollen?” (‘When are the bor- der controls coming?’), “Diese Grenze war vorher auch geschützt” (‘This border was protected before too’). In all 18 cases, the negativity targets political actors or...