pith. machine review for the scientific record. sign in

arxiv: 2605.13663 · v1 · submitted 2026-05-13 · 💻 cs.CL · cs.CY

Recognition: unknown

Fine-tuning with Hierarchical Prompting for Robust Propaganda Classification Across Annotation Schemas

Authors on Pith no claims yet

Pith reviewed 2026-05-14 19:38 UTC · model grok-4.3

classification 💻 cs.CL cs.CY
keywords propaganda detectionhierarchical promptingfine-tuninglanguage modelsannotation schemasintent taxonomysocial media classificationzero-shot vs fine-tuned
0
0 comments X

The pith

Fine-tuning with hierarchical prompting turns weak zero-shot propaganda classifiers into competitive systems that handle ambiguous intent-based taxonomies better.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows that fine-tuning language models is essential for propaganda detection on noisy social media texts, where it converts poor zero-shot baselines into strong performers and exposes real differences between models and methods. It introduces a new intent-focused taxonomy of propaganda techniques and compares it to a higher-agreement established schema using four models. Hierarchical prompting, which first identifies fine-grained techniques before aggregating them, proves especially helpful after fine-tuning and on the more ambiguous new taxonomy. This setup yields stronger overall results with Qwen models leading and provides a richer, more challenging benchmark through the new HQP dataset.

Core claim

Fine-tuning transforms weak zero-shot baselines into competitive systems and reveals methodological differences hidden in base models. The hierarchical prompting method (HiPP) predicts fine-grained techniques before aggregating them and is especially beneficial after fine-tuning and on the more ambiguous, low-agreement taxonomy. Across schemas the Qwen models achieve the strongest performance while Phi-4 14B consistently outperforms GPT-4.1-nano, and the new intent-based labels in the HQP dataset offer a richer lens on propaganda's strategic goals.

What carries the argument

Hierarchical prompting (HiPP), which first predicts fine-grained propaganda techniques before aggregating them into higher-level labels.

If this is right

  • Fine-tuning exposes model and method differences that zero-shot evaluations hide.
  • HiPP improves results most on low-agreement, ambiguous taxonomies after fine-tuning.
  • Qwen models deliver the strongest performance across both taxonomies.
  • Phi-4 14B outperforms GPT-4.1-nano consistently in this setting.
  • The HQP dataset serves as a harder benchmark for future robust detection work.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar hierarchical breakdown could help other noisy text classification tasks with low inter-annotator agreement.
  • Intent-focused labels might support downstream applications such as targeted content moderation or campaign analysis.
  • Testing the same fine-tuning plus HiPP pipeline on multilingual datasets would check if the gains hold beyond English.
  • The performance gap between fine-tuned and base models suggests zero-shot evaluations may systematically underestimate practical utility.

Load-bearing premise

The new intent-focused taxonomy captures propaganda's strategic goals in a way that actually improves real-world robustness beyond the specific datasets tested.

What would settle it

An experiment showing that models trained on the new intent taxonomy perform worse than those using the established schema when evaluated on held-out real-world social media posts with human-verified ground truth.

Figures

Figures reproduced from arXiv: 2605.13663 by Ariana Sahitaj, Charlott Jakob, David Kaplan, Lukas St\"ahelin, Max Upravitelev, Premtim Sahitaj, Sebastian M\"oller, Vera Schmitt, Veronika Solopova.

Figure 1
Figure 1. Figure 1: Weighted-F1 improvements from Direct-High to Main-High. For each backbone, lines connect zero-shot (Baseline) and fine-tuned scores. On both schemata, Main→High lags behind Direct-High in zero-shot but gains more from fine-tuning, especially on our lower-IAA taxonomy. (GPT-4.1 here refers to the GPT-4.1-nano variant.) LMs. Weighted-F1 scores are systematically higher than macro-F1, reflecting the skewed la… view at source ↗
Figure 2
Figure 2. Figure 2: PCA of confusion-cell TF–IDF vectors for Qwen3-14B in its best setting. Each point represents a (true, [PITH_FULL_IMAGE:figures/full_fig_p008_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Qwen3 confusion matrices on the Sahitaj benchmark. [PITH_FULL_IMAGE:figures/full_fig_p015_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Qwen3 confusion matrices on our dataset. [PITH_FULL_IMAGE:figures/full_fig_p016_4.png] view at source ↗
read the original abstract

Propaganda detection in social media is challenging due to noisy, short texts and low annotation agreements. We introduce a new intent-focused taxonomy of propaganda techniques and compare it against an established, higher-agreement schema. Along three dimensions (model portfolio, schema effects, and prompting strategy) we evaluate the taxonomies as a classification task with the help of four language models (GPT-4.1-nano, Phi-4 14B, Qwen2.5-14B, Qwen3-14B). Our results show that fine-tuning is essential, since it transforms weak zero-shot baselines into competitive systems and reveals methodological differences that are hidden using base models. Across schemas, the Qwen models achieve the strongest overall performance, and Phi-4 14B consistently outperforms GPT-4.1-nano. Our hierarchical prompting method (HiPP), which predicts fine-grained techniques before aggregating them, is especially beneficial after fine-tuning and on the more ambiguous, low-agreement taxonomy, while remaining competitive on the simpler schema. The HQP dataset, annotated with the new intent-based labels, provides a richer lens on propaganda's strategic goals and a challenging benchmark for future work on robust, real-world detection.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

0 major / 0 minor

Summary. The manuscript introduces a new intent-focused taxonomy for propaganda techniques in social media and compares it to an established higher-agreement schema. Using four LLMs (GPT-4.1-nano, Phi-4 14B, Qwen2.5-14B, Qwen3-14B), it evaluates performance across zero-shot and fine-tuned regimes with standard prompting versus the proposed hierarchical prompting method (HiPP), which first predicts fine-grained techniques before aggregation. The central claims are that fine-tuning is essential to convert weak zero-shot baselines into competitive systems and to surface model and schema differences, that HiPP is especially beneficial after fine-tuning on the low-agreement taxonomy, that Qwen models perform strongest overall, and that the released HQP dataset offers a richer benchmark for real-world robustness.

Significance. If the empirical comparisons hold, the work provides concrete evidence that fine-tuning combined with hierarchical prompting improves robustness in propaganda classification across annotation schemas with differing agreement levels. The new intent-focused taxonomy and the HQP dataset constitute a useful contribution for future benchmarking, while the finding that base-model differences are masked until fine-tuning occurs has practical implications for model selection in noisy social-media settings.

Simulated Author's Rebuttal

0 responses · 0 unresolved

We thank the referee for their careful summary of the manuscript and for recommending minor revision. We are pleased that the empirical findings on fine-tuning, hierarchical prompting, and the new HQP dataset are viewed as providing concrete evidence for improved robustness across schemas.

Circularity Check

0 steps flagged

No significant circularity in empirical evaluation

full rationale

The paper is an empirical study comparing four language models, two annotation schemas, and prompting strategies (including the introduced HiPP method) on propaganda detection tasks. All central claims—such as the necessity of fine-tuning to reveal differences and the benefits of hierarchical prompting on low-agreement taxonomies—are presented as direct outcomes of experimental results across datasets. There are no equations, derivations, fitted parameters renamed as predictions, or load-bearing self-citations that reduce any result to its inputs by construction. The work remains self-contained as a benchmark comparison without circular reductions.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The paper is an empirical machine learning study. It relies on standard assumptions about data splits and model training but introduces no new mathematical derivations or free parameters beyond typical hyperparameter choices in fine-tuning.

axioms (1)
  • domain assumption Standard i.i.d. assumptions for train/test splits in supervised classification
    Implicit in all reported fine-tuning and evaluation results.

pith-pipeline@v0.9.0 · 5546 in / 1218 out tokens · 42204 ms · 2026-05-14T19:38:05.464624+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

24 extracted references · 2 canonical work pages

  1. [1]

    Fine-grained analysis of propaganda in news articles. InProceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natu- ral Language Processing (EMNLP-IJCNLP), pages 5636–5646, Hong Kong, China. Association for Com- putational Linguistics. A. P. Dawid and A. M. Skene. 1979. Maximum...

  2. [2]

    Jews for Nazis

    Semeval-2020 task 11: Detection of propa- ganda techniques in news articles.arXiv preprint arXiv:2009.02696. OpenAI. 2023. Gpt-4 technical report.arXiv preprint arXiv:2303.08774. E Perez. 2022. Strategic disinformation: Russia, ukraine and crisis communication in digital era. Jakub Piskorski, Nicolas Stefanovitch, Giovanni Da San Martino, and Preslav Nako...

  3. [3]

    Subtypes: 2, 11, 15, 17 (False Origin Attribution, Circular Reasoning, Appeal to Ignorance)

    Distort Reality and Rewrite the Past Goal: Undermine truth and legitimize the present through distortion. Subtypes: 2, 11, 15, 17 (False Origin Attribution, Circular Reasoning, Appeal to Ignorance). Rationale: Denial, fabrication, and invented evidence challenge fact-based narratives

  4. [4]

    Subtypes: 4, 5, 13, 17 (False Causality, False Balance)

    Shift Blame and Justify Aggression Goal: Reframe aggressor as victim or rational actor. Subtypes: 4, 5, 13, 17 (False Causality, False Balance). Rationale: Inverts moral frameworks to rationalize wrongdoing or redirect guilt

  5. [5]

    Subtypes: 1, 3, 8, 9, 12, 6

    Delegitimize the Opponent Goal: Undermine credibility and morality of adversaries. Subtypes: 1, 3, 8, 9, 12, 6. Rationale: Frames enemies as liars, extremists, or inhuman entities

  6. [6]

    Subtypes: 12, 14, 16, 10, 7

    Manufacture Consent and Identity Goal: Rally support and polarize identities. Subtypes: 12, 14, 16, 10, 7. Rationale: Builds loyalty through fear, pride, and tribal solidarity

  7. [7]

    Subtypes: 7, 15, 6, 17 (Red Herrings, Appeal to Probability, False Di- chotomy)

    Confuse and Distract Goal: Overwhelm critical thinking through noise and uncertainty. Subtypes: 7, 15, 6, 17 (Red Herrings, Appeal to Probability, False Di- chotomy). Rationale: Undermines clarity, trust, and consensus. Table 8: Overview of the five high-level propaganda categories and their purposes. Low-Level Fine-Grained Labels

  8. [8]

    Includes WWII analogies, Hitler/SS references, burning people like Nazis

    Guilt-by-association fallacy Most frequently as nazi analogies. Includes WWII analogies, Hitler/SS references, burning people like Nazis. Could also include other reductio ad Hitlerum instances

  9. [9]

    Historical Distortion / Revisionism Cherry-picking historical events, WWII nostalgia, Soviet glorification, fake quotes

  10. [10]

    Dehumanization / Demonization Calling opponents subhuman, monsters, or animals; atrocity narratives

  11. [11]

    Deflect and Justify I: Victimhood / Gaslighting Aggressor framed as victim, blame-shifting, moral inversion, denial of wrongdoing

  12. [12]

    but the West invaded Iraq

    Deflect and Justify II: Whataboutism False moral equivalence, hypocrisy framing (“but the West invaded Iraq”)

  13. [13]

    the other side lies,

    Accusation of Propaganda / Media Distrust Asserting “the other side lies,” mocking mainstream narratives, distrust in institutions

  14. [14]

    Conspiracy Narratives CIA/Nuland plots, puppet governments, Western coups, assertion-as-proof

  15. [15]

    Guilt by Association Azov = entire military is Nazi; guilt through affiliation or past ties

  16. [16]

    Sarcasm / Ridicule / Strawman Mocking tone, exaggeration, caricatures, eye-roll emojis

  17. [17]

    Emotional Manipulation / Shock Appeal Rape/torture narratives, fearmongering, nuclear threats, mass graves

  18. [18]

    False Authority / Fabrication Fake quotes, fabricated stats, unverified claims framed as facts

  19. [19]

    with us or against us

    Us vs. Them Framing / Identity Dichotomy Nationalist binaries, tribal solidarity, “with us or against us” rhetoric

  20. [20]

    Realpolitik Framing / Moral Detachment Cool, analytical justification of aggression; geopolitical realism

  21. [21]

    Russia is winning,

    Triumphalism / Victory Framing “Russia is winning,” inevitability rhetoric, “truth will prevail.”

  22. [22]

    Unconfirmed reports,

    Information Laundering / Rumor Seeding “Unconfirmed reports,” plausible deniability, sowing doubt

  23. [23]

    Anti-Establishment / Anti-Globalist Framing Criticism of NATO, EU, UN, WEF; anti-West framing

  24. [24]

    Table 9: Overview of the 17 fine-grained propaganda techniques

    Logical Fallacies False dichotomy, causal oversimplification, circular reasoning, false balance, appeal to ignorance. Table 9: Overview of the 17 fine-grained propaganda techniques. E Token-level evidence for confusion patterns (example file) True label Pred label #tokens Most frequent tokens (count) Non-Propaganda (0) Non-Propaganda (0) 905 ukraine (27),...