pith. machine review for the scientific record. sign in

arxiv: 2604.18112 · v2 · submitted 2026-04-20 · 💻 cs.CL · cs.MM

Recognition: unknown

Retrieval-Augmented Multimodal Model for Fake News Detection

Hanyi Yu, Weihai Lu, Yiheng Li, Yue Wang

Authors on Pith no claims yet

Pith reviewed 2026-05-10 05:14 UTC · model grok-4.3

classification 💻 cs.CL cs.MM
keywords fake news detectionmultimodal modelsretrieval-augmentednarrative alignmentanalogical reasoningmultimodal large language modelscross-domain generalizationcross-instance consistency
0
0 comments X

The pith

RAMM improves fake news detection by retrieving similar instances to align their abstract narratives and enable analogical reasoning on a multimodal LLM backbone.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Current multimodal detectors examine each news item alone and depend only on knowledge baked into their parameters, so they miss repeated story patterns across related posts and falter on new topics or scarce data. The paper presents RAMM, which adds an Abstract Narrative Alignment Module to pull high-level narrative structures from retrieved examples across domains and a Semantic Representation Alignment Module to change the model's process from direct feature judgment to reasoning by analogy with those examples. Both modules sit on top of a multimodal large language model. Experiments on three public datasets confirm better results than earlier approaches. A reader would care because social media often spreads clusters of related false stories and models must handle emerging events without fresh training data.

Core claim

RAMM employs a Multimodal Large Language Model as its backbone to capture cross-modal semantics, incorporates an Abstract Narrative Alignment Module that adaptively extracts and aggregates abstract narrative consistency from diverse instances across domains, and introduces a Semantic Representation Alignment Module that shifts reasoning from direct inference on multimodal features to instance-based analogical reasoning, thereby addressing the failure to capture cross-instance narrative consistency and the lack of domain-specific knowledge in multimodal multidomain fake news detection.

What carries the argument

The Abstract Narrative Alignment Module, which extracts and aggregates high-level narrative consistency from retrieved instances, together with the Semantic Representation Alignment Module, which realigns the decision process toward human-like analogy to those instances.

Load-bearing premise

That the Abstract Narrative Alignment Module can reliably extract and aggregate high-level narrative consistency across diverse instances and domains while the Semantic Representation Alignment Module actually improves generalization by moving the model to instance-based analogical reasoning.

What would settle it

Running RAMM on a new test set of clustered fake news stories in an unseen domain and finding no statistically significant gain over a plain multimodal LLM baseline that lacks the two alignment modules.

Figures

Figures reproduced from arXiv: 2604.18112 by Hanyi Yu, Weihai Lu, Yiheng Li, Yue Wang.

Figure 1
Figure 1. Figure 1: The issue of cluster-based propagation of fake news: [PITH_FULL_IMAGE:figures/full_fig_p002_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: The architecture of RAMM. Semantic Representation Alignment Module selects examples based on image-text [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Parameter sensitivity analysis of four key hyper [PITH_FULL_IMAGE:figures/full_fig_p008_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Performance comparison of domain generalization [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: T-SNE showcase the classification outcomes in [PITH_FULL_IMAGE:figures/full_fig_p009_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Performance comparison of LLM backbones. [PITH_FULL_IMAGE:figures/full_fig_p009_6.png] view at source ↗
read the original abstract

In recent years, multimodal multidomain fake news detection has garnered increasing attention. Nevertheless, this direction presents two significant challenges: (1) Failure to Capture Cross-Instance Narrative Consistency: existing models usually evaluate each news in isolation, fail to capture cross-instance narrative consistency, and thus struggle to address the spread of cluster based fake news driven by social media; (2) Lack of Domain Specific Knowledge for Reasoning: conventional models, which rely solely on knowledge encoded in their parameters during training, struggle to generalize to new or data-scarce domains (e.g., emerging events or niche topics). To tackle these challenges, we introduce Retrieval-Augmented Multimodal Model for Fake News Detection (RAMM). First, RAMM employs a Multimodal Large Language Model (MLLM) as its backbone to capture cross-modal semantic information from news samples. Second, RAMM incorporates an Abstract Narrative Alignment Module. This component adaptively extracts abstract narrative consistency from diverse instances across distinct domains, aggregates relevant knowledge, and thereby enables the modeling of high-level narrative information. Finally, RAMM introduces a Semantic Representation Alignment Module, which aligns the model's decision-making paradigm with that of humans - specifically, it shifts the model's reasoning process from direct inference on multimodal features to an instance-based analogical reasoning process. Extensive experimental results on three public datasets validate the efficacy of our proposed approach. Our code is available at the following link: https://github.com/li-yiheng/RAMM

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 0 minor

Summary. The paper proposes RAMM, a retrieval-augmented multimodal model for fake news detection. It employs an MLLM backbone to capture cross-modal semantics, an Abstract Narrative Alignment Module that adaptively extracts and aggregates high-level narrative consistency across diverse instances and domains, and a Semantic Representation Alignment Module that shifts reasoning from direct multimodal feature inference to instance-based analogical reasoning. The approach is claimed to address failures in capturing cross-instance consistency and lack of domain-specific knowledge, with efficacy validated by extensive experiments on three public datasets.

Significance. If the modules can be shown to deliver measurable gains in cross-instance narrative modeling and domain generalization beyond a standard MLLM backbone, the work could advance multimodal fake news detection by providing a practical way to handle cluster-based misinformation on social media and improve robustness in data-scarce domains. The open release of code is a positive step toward reproducibility.

major comments (3)
  1. [Abstract Narrative Alignment Module] Abstract Narrative Alignment Module description: The module is presented as adaptively extracting 'abstract narrative consistency from diverse instances across distinct domains' and aggregating 'relevant knowledge,' yet no retrieval corpus, aggregation procedure, adaptive mechanism, or equations are supplied. This detail is load-bearing for the central claim of capturing cross-instance narrative consistency.
  2. [Semantic Representation Alignment Module] Semantic Representation Alignment Module description: The module is said to align decision-making with human-like 'instance-based analogical reasoning' rather than direct inference on multimodal features, but no alignment loss, procedure for incorporating retrieved instances, or differentiation from standard feature-based inference is provided. This is essential for substantiating the second main contribution.
  3. [Experiments] Experimental validation: The abstract states that 'extensive experimental results on three public datasets validate the efficacy,' but the manuscript contains no information on the specific datasets, baselines, metrics, ablation studies, or any quantitative results. Without these, it is impossible to assess whether observed gains support the claimed mechanisms or stem from other factors such as the MLLM backbone.

Simulated Author's Rebuttal

3 responses · 0 unresolved

Thank you for the detailed review and constructive feedback on our manuscript. We appreciate the opportunity to clarify and strengthen our work. Below, we provide point-by-point responses to the major comments. We will revise the manuscript to address the concerns regarding insufficient technical details and experimental information.

read point-by-point responses
  1. Referee: [Abstract Narrative Alignment Module] Abstract Narrative Alignment Module description: The module is presented as adaptively extracting 'abstract narrative consistency from diverse instances across distinct domains' and aggregating 'relevant knowledge,' yet no retrieval corpus, aggregation procedure, adaptive mechanism, or equations are supplied. This detail is load-bearing for the central claim of capturing cross-instance narrative consistency.

    Authors: We acknowledge that the current description of the Abstract Narrative Alignment Module is high-level and lacks the specific technical details necessary to fully substantiate its operation. In the revised version, we will provide a detailed description of the retrieval corpus used, the aggregation procedure, the adaptive mechanism, and include the mathematical formulations and equations for the module. This will better illustrate how it extracts and aggregates abstract narrative consistency across instances and domains. revision: yes

  2. Referee: [Semantic Representation Alignment Module] Semantic Representation Alignment Module description: The module is said to align decision-making with human-like 'instance-based analogical reasoning' rather than direct inference on multimodal features, but no alignment loss, procedure for incorporating retrieved instances, or differentiation from standard feature-based inference is provided. This is essential for substantiating the second main contribution.

    Authors: We agree that more details are needed for the Semantic Representation Alignment Module to clearly differentiate it from standard approaches and to explain the instance-based analogical reasoning. In the revision, we will elaborate on the alignment loss function, the procedure for incorporating retrieved instances into the reasoning process, and how this shifts the paradigm from direct multimodal feature inference. We will also include relevant equations and diagrams if appropriate. revision: yes

  3. Referee: [Experiments] Experimental validation: The abstract states that 'extensive experimental results on three public datasets validate the efficacy,' but the manuscript contains no information on the specific datasets, baselines, metrics, ablation studies, or any quantitative results. Without these, it is impossible to assess whether observed gains support the claimed mechanisms or stem from other factors such as the MLLM backbone.

    Authors: We apologize for this oversight in the manuscript preparation. Although the abstract mentions the experimental validation, the detailed experimental setup, including the three public datasets used, the baseline methods, evaluation metrics, ablation studies, and quantitative results, were inadvertently omitted from the main text. In the revised manuscript, we will include a comprehensive Experiments section with all this information, along with tables and figures presenting the results to demonstrate the efficacy of the proposed modules. revision: yes

Circularity Check

0 steps flagged

No significant circularity; model architecture proposed and validated empirically without self-referential derivations or reductions to inputs

full rationale

The paper proposes RAMM to address two stated challenges in multimodal fake news detection by combining an MLLM backbone with an Abstract Narrative Alignment Module (described as adaptively extracting and aggregating narrative consistency) and a Semantic Representation Alignment Module (described as shifting to instance-based analogical reasoning). No equations, loss functions, retrieval procedures, or mathematical derivations are provided in the abstract or described text that could create self-definition or allow a claimed result to reduce to its own inputs by construction. Efficacy is asserted solely through standard empirical results on three public datasets, which does not constitute a 'prediction' that is statistically forced from fitted parameters. No self-citations are used to justify uniqueness theorems, ansatzes, or load-bearing premises, and no known results are renamed as novel organization. The derivation chain is therefore self-contained as an empirical model proposal rather than a circular reduction.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract provides no technical details on any free parameters, axioms, or invented entities; model components are described only at a conceptual level.

pith-pipeline@v0.9.0 · 5561 in / 1179 out tokens · 46249 ms · 2026-05-10T05:14:16.831107+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

72 extracted references · 6 canonical work pages · 1 internal anchor

  1. [1]

    Yixuan Chen, Dongsheng Li, Peng Zhang, Jie Sui, Qin Lv, Lu Tun, and Li Shang

  2. [2]

    In Proceedings of the ACM web conference 2022

    Cross-modal ambiguity learning for multimodal fake news detection. In Proceedings of the ACM web conference 2022. 2897–2905

  3. [3]

    Zhenghan Chen, Changzeng Fu, and Xunzhu Tang. 2023. Multi-domain fake news detection with fuzzy labels. InInternational Conference on Database Systems for Advanced Applications. Springer, 331–343

  4. [4]

    Limeng Cui, Kai Shu, Suhang Wang, Dongwon Lee, and Huan Liu. 2019. defend: A system for explainable fake news detection. InProceedings of the 28th ACM international conference on information and knowledge management. 2961–2964

  5. [5]

    Xiaoxi Cui, Weihai Lu, Yu Tong, Yiheng Li, and Zhejun Zhao. 2025. Diffusion- based multi-modal synergy interest network for click-through rate prediction. InProceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval. 581–591

  6. [6]

    Jannatul Ferdush, Joarder Kamruzzaman, Gour Karmakar, Iqbal Gondal, and Rajkumar Das. 2025. Cross-Domain Fake News Detection Through Fusion of Evidence from Multiple Social Media Platforms.Future Internet17, 2 (2025), 61

  7. [7]

    Tianyu Gao, Xingcheng Yao, and Danqi Chen. 2021. Simcse: Simple contrastive learning of sentence embeddings. InProceedings of the 2021 conference on empirical methods in natural language processing. 6894–6910

  8. [8]

    Shuzhi Gong, Richard Sinnott, Jianzhong Qi, and Cecile Paris. 2025. Unseen fake news detection through casual debiasing. InCompanion Proceedings of the ACM on Web Conference 2025. 981–985

  9. [9]

    Daniel A González-Padilla and Leonardo Tortolero-Blanco. 2020. Social media influence in the COVID-19 Pandemic.International braz j urol46 (2020), 120–124

  10. [10]

    Fiza Gulzar Hussain, Muhammad Wasim, Seemab Hameed, Abdur Rehman, Muhammad Nabeel Asim, and Andreas Dengel. 2025. Fake news detection landscape: Datasets, data modalities, AI approaches, their challenges, and future perspectives.IEEE Access(2025)

  11. [11]

    Korir Nancy Jeptoo and Chengjie Sun. 2024. Enhancing fake news detection with large language models through multi-agent debates. InCCF International Conference on Natural Language Processing and Chinese Computing. Springer, 474–486

  12. [12]

    Ye Jiang, Taihang Wang, Xiaoman Xu, Yimin Wang, Xingyi Song, and Diana Maynard. 2025. Cross-modal augmentation for few-shot multimodal fake news detection.Engineering Applications of Artificial Intelligence142 (2025), 109931

  13. [13]

    Betul Keles, Niall McCrae, and Annmarie Grealish. 2020. A systematic review: the influence of social media on depression, anxiety and psychological distress in adolescents.International journal of adolescence and youth25, 1 (2020), 79–93

  14. [14]

    Dhruv Khattar, Jaipal Singh Goud, Manish Gupta, and Vasudeva Varma. 2019. Mvae: Multimodal variational autoencoder for fake news detection. InThe world wide web conference. 2915–2921

  15. [15]

    Sejeong Kwon, Meeyoung Cha, Kyomin Jung, Wei Chen, and Yajun Wang. 2013. Prominent features of rumor propagation in online social media. In2013 IEEE 13th international conference on data mining. IEEE, 1103–1108

  16. [16]

    Batool Lakzaei, Mostafa Haghir Chehreghani, and Alireza Bagheri. 2025. Neighborhood-Order Learning Graph Attention Network for Fake News De- tection.arXiv preprint arXiv:2502.06927(2025)

  17. [17]

    Chenhui Li and Weihai Lu. 2026. Decoding the Market’s Pulse: Context-Enriched Agentic Retrieval Augmented Generation for Predicting Post-Earnings Price Shocks. InProceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). 3055–3073

  18. [18]

    Junnan Li, Dongxu Li, Silvio Savarese, and Steven Hoi. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. InInternational conference on machine learning. PMLR, 19730–19742

  19. [19]

    Jinjin LI, Guoming SANG, and Yijia ZHANG. 2024. APK-CNN and Transformer- enhanced multi-domain fake news detection model.Journal of Computer Appli- cations(2024), 1930–1939

  20. [20]

    Xinyi Li, Yongfeng Zhang, and Edward C Malthouse. 2024. Large language model agent for fake news detection.arXiv preprint arXiv:2405.01593(2024)

  21. [21]

    Yuyuan Li, Chaochao Chen, Yizhao Zhang, Weiming Liu, Lingjuan Lyu, Xiaolin Zheng, Dan Meng, and Jun Wang. 2023. Ultrare: Enhancing receraser for recom- mendation unlearning via error decomposition.Advances in Neural Information Processing Systems36 (2023), 12611–12625

  22. [22]

    Yuyuan Li, Yizhao Zhang, Weiming Liu, Xiaohua Feng, Zhongxuan Han, Chaochao Chen, and Chenggang Yan. 2025. Multi-Objective Unlearning in Rec- ommender Systems via Preference Guided Pareto Exploration.IEEE Transactions on Services Computing(2025)

  23. [23]

    Shijie Liu, Ruixin Ding, Weihai Lu, Jun Wang, Mo Yu, Xiaoming Shi, and Wei Zhang. 2025. Coherency improved explainable recommendation via large lan- guage model. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 12201–12209

  24. [24]

    Wayne Lu and Xiaoxi Cui. 2026. DEALT: LLM-driven Diversity-Enhanced Data Augmentation for Long-Tail Text Classification. InProceedings of the AAAI Con- ference on Artificial Intelligence, Vol. 40. 32338–32346

  25. [25]

    Wayne Lu and Yiheng Li. 2026. From Blind Transfer to Wise Selection: Prototype- Driven Neighbor-Domain Adaptation for Fake News Detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 40. 818–826

  26. [26]

    Weihai Lu, Yu Tong, and Zhiqiu Ye. 2025. Dammfnd: Domain-aware multimodal multi-view fake news detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 559–567

  27. [27]

    Weihai Lu and Li Yin. 2025. Dmmd4sr: Diffusion model-based multi-level multi- modal denoising for sequential recommendation. InProceedings of the 33rd ACM International Conference on Multimedia. 6363–6372. Conference acronym ’XX, June 03–05, 2018, Woodstock, NY Trovato et al

  28. [28]

    Xiaoxiao Ma, Yuchen Zhang, Kaize Ding, Jian Yang, Jia Wu, and Hao Fan. 2024. On fake news detection with LLM enhanced semantics mining. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. 508–521

  29. [29]

    Despoina Mouratidis, Andreas Kanavos, and Katia Kermanidis. 2025. From misinformation to insight: machine learning strategies for fake news detection. Information16, 3 (2025), 189

  30. [30]

    Qiong Nan, Juan Cao, Yongchun Zhu, Yanyan Wang, and Jintao Li. 2021. MD- FEND: Multi-domain fake news detection. InProceedings of the 30th ACM inter- national conference on information & knowledge management. 3343–3347

  31. [31]

    Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding.arXiv preprint arXiv:1807.03748(2018)

  32. [32]

    Eleftheria Papageorgiou, Christos Chronis, Iraklis Varlamis, and Yassine Himeur

  33. [33]

    A survey on the use of large language models (llms) in fake news.Future Internet16, 8 (2024), 298

  34. [34]

    Eleftheria Papageorgiou, Iraklis Varlamis, and Christos Chronis. 2025. Harnessing Large Language Models and deep neural networks for fake news detection. Information16, 4 (2025), 297

  35. [35]

    Zitao Peng, Nankai Lin, Yongmei Zhou, Dong Zhou, and Aimin Yang. 2023. Prompt learning for low-resource multi-domain fake news detection. In2023 international conference on asian Language Processing (IALP). IEEE, 314–319

  36. [36]

    Gordon Pennycook and David G Rand. 2021. The psychology of fake news.Trends in cognitive sciences25, 5 (2021), 388–402

  37. [37]

    Ben Poole, Sherjil Ozair, Aaron Van Den Oord, Alex Alemi, and George Tucker

  38. [38]

    InInternational conference on machine learning

    On variational bounds of mutual information. InInternational conference on machine learning. PMLR, 5171–5180

  39. [39]

    Peng Qi, Zehong Yan, Wynne Hsu, and Mong Li Lee. 2024. Sniffer: Multimodal large language model for explainable out-of-context misinformation detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 13052–13062

  40. [40]

    Shaina Raza, Drai Paulen-Patterson, and Chen Ding. 2025. Fake news detection: comparative evaluation of BERT-like models and large language models with generative AI-annotated data.Knowledge and Information Systems67, 4 (2025), 3267–3292

  41. [41]

    Amila Silva, Ling Luo, Shanika Karunasekera, and Christopher Leckie. 2021. Em- bracing domain differences in fake news: Cross-domain fake news detection using multi-modal data. InProceedings of the AAAI conference on artificial intelligence, Vol. 35. 557–565

  42. [42]

    Chenguang Song, Nianwen Ning, Yunlei Zhang, and Bin Wu. 2021. Knowl- edge augmented transformer for adversarial multidomain multiclassification multimodal fake news detection.Neurocomputing462 (2021), 88–100

  43. [43]

    Malliga Subramanian, B Premjith, Kogilavani Shanmugavadivel, Santhiya Pandiyan, Balasubramanian Palani, and Bharathi Raja Chakravarthi. 2025. Overview of the shared task on fake news detection in dravidian languages- DravidianLangTech@ NAACL 2025. InProceedings of the fifth workshop on speech, vision, and language technologies for dravidian languages. 759–767

  44. [44]

    Naftali Tishby, Fernando C Pereira, and William Bialek. 2000. The information bottleneck method.arXiv preprint physics/0004057(2000)

  45. [45]

    Yu Tong, Weihai Lu, Xiaoxi Cui, Yifan Mao, and Zhejun Zhao. 2025. Dapt: Domain- aware prompt-tuning for multimodal fake news detection. InProceedings of the 33rd ACM International Conference on Multimedia. 7902–7911

  46. [46]

    Yu Tong, Weihai Lu, Zhe Zhao, Song Lai, and Tong Shi. 2024. Mmdfnd: Multi- modal multi-domain fake news detection. InProceedings of the 32nd ACM Inter- national Conference on Multimedia. 1178–1186

  47. [47]

    Zhao Tong, Yimeng Gu, Huidong Liu, Qiang Liu, Shu Wu, Haichao Shi, and Xiao-Yu Zhang. 2025. Generate first, then sample: Enhancing fake news detection with LLM-augmented reinforced sampling. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 24276–24290

  48. [48]

    Bo Wang, Jing Ma, Hongzhan Lin, Zhiwei Yang, Ruichao Yang, Yuan Tian, and Yi Chang. 2024. Explainable fake news detection with large language model via defense among competing wisdom. InProceedings of the ACM web conference

  49. [49]

    Yaqing Wang, Fenglong Ma, Zhiwei Jin, Ye Yuan, Guangxu Xun, Kishlay Jha, Lu Su, and Jing Gao. 2018. Eann: Event adversarial neural networks for multi-modal fake news detection. InProceedings of the 24th acm sigkdd international conference on knowledge discovery & data mining. 849–857

  50. [50]

    Xiaolong Wei, Yuehu Dong, Xingliang Wang, Xingyu Zhang, Zhejun Zhao, Dong- dong Shen, Long Xia, and Dawei Yin. 2026. Beyond react: A planner-centric framework for complex tool-augmented llm reasoning. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 40. 33845–33853

  51. [51]

    Xiaolong Wei, Bo Lu, Xingyu Zhang, Zhejun Zhao, Dongdong Shen, Long Xia, and Dawei Yin. 2025. Igniting creative writing in small language models: Llm-as- a-judge versus multi-agent refined rewards. InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing. 17171–17197

  52. [52]

    Shitao Xiao, Zheng Liu, Peitian Zhang, Niklas Muennighoff, Defu Lian, and Jian-Yun Nie. 2024. C-pack: Packed resources for general chinese embeddings. InProceedings of the 47th international ACM SIGIR conference on research and development in information retrieval. 641–649

  53. [53]

    Aolin Xu and Maxim Raginsky. 2017. Information-theoretic analysis of gen- eralization capability of learning algorithms.Advances in neural information processing systems30 (2017)

  54. [54]

    Junxiao Xue, Yabo Wang, Yichen Tian, Yafei Li, Lei Shi, and Lin Wei. 2021. De- tecting fake news by exploring the consistency of multimodal data.Information Processing & Management58, 5 (2021), 102610

  55. [55]

    Kaiying Yan, Moyang Liu, Yukun Liu, Ruibo Fu, Zhengqi Wen, Jianhua Tao, Xuefei Liu, and Guanjun Li. 2025. MTPareto: A MultiModal Targeted Pareto Framework for Fake News Detection. InICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1–5

  56. [56]

    Xuankai Yang, Yan Wang, Xiuzhen Zhang, Shoujin Wang, Huaxiong Wang, and Kwok Yan Lam. 2025. A macro-and micro-hierarchical transfer learning framework for cross-domain fake news detection. InProceedings of the ACM on Web Conference 2025. 5297–5307

  57. [57]

    Jingyuan Yi, Zeqiu Xu, Tianyi Huang, and Peiyang Yu. 2025. Challenges and innovations in llm-powered fake news detection: A synthesis of approaches and future directions. InProceedings of the 2025 2nd international conference on generative artificial intelligence and information security. 87–93

  58. [58]

    Xinquan Yu, Ziqi Sheng, Wei Lu, Xiangyang Luo, and Jiantao Zhou. 2025. Racmc: Residual-aware compensation network with multi-granularity constraints for fake news detection. InProceedings of the AAAI Conference on Artificial Intelli- gence, Vol. 39. 986–994

  59. [59]

    Zhi Zeng, Minnan Luo, Xiangzheng Kong, Huan Liu, Hao Guo, Hao Yang, Zihan Ma, and Xiang Zhao. 2024. Mitigating World Biases: A Multimodal Multi-View Debiasing Framework for Fake News Video Detection. InProceedings of the 32nd ACM International Conference on Multimedia. 6492–6500

  60. [60]

    Zhi Zeng, Jiaying Wu, Minnan Luo, Xiangzheng Kong, Zihan Ma, Guang Dai, and Qinghua Zheng. 2025. Understand, Refine and Summarize: Multi-View Knowledge Progressive Enhancement Learning for Fake News Video Detection. InProceedings of the 33rd ACM International Conference on Multimedia. 9216– 9225

  61. [61]

    Zhi Zeng, Jiaying Wu, Minnan Luo, Herun Wan, Xiangzheng Kong, Zihan Ma, Guang Dai, and Qinghua Zheng. 2025. Imol: Incomplete-modality-tolerant learn- ing for multi-domain fake news video detection. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 30921–30933

  62. [62]

    Zhi Zeng, Yifei Yang, Jiaying Wu, Xulang Zhang, Xiangzheng Kong, Herun Wan, Zihan Ma, and Minnan Luo. 2026. From Manipulation to Mistrust: Explaining Diverse Micro-Video Misinformation for Robust Debunking in the Wild.arXiv preprint arXiv:2603.25423(2026)

  63. [63]

    Chaowei Zhang, Zongling Feng, Zewei Zhang, Jipeng Qiang, Guandong Xu, and Yun Li. 2025. Is llms hallucination usable? llm-based negative reasoning for fake news detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 1031–1039

  64. [64]

    Xing Zhou, Juan Cao, Zhiwei Jin, Fei Xie, Yu Su, Dafeng Chu, Xuehui Cao, and Junqiang Zhang. 2015. Real-time news cer tification system on sina weibo. In Proceedings of the 24th international conference on world wide web. 983–988

  65. [65]

    Ziyi Zhou, Xiaoming Zhang, Litian Zhang, Jiacheng Liu, Senzhang Wang, Zheng Liu, Xi Zhang, Chaozhuo Li, and Philip S Yu. 2024. Finefake: A knowledge- enriched dataset for fine-grained multi-domain fake news detection.arXiv preprint arXiv:2404.01336(2024)

  66. [66]

    Yongchun Zhu, Qiang Sheng, Juan Cao, Shuokai Li, Danding Wang, and Fuzhen Zhuang. 2022. Generalizing to the future: Mitigating entity bias in fake news detection. InProceedings of the 45th international ACM SIGIR conference on research and development in information retrieval. 2120–2125

  67. [67]

    Explosive footage! A poll worker in XX County was caught on a hidden camera shredding pro-government ballots to rig the election. The proof is undeniable!

    Yongchun Zhu, Qiang Sheng, Juan Cao, Qiong Nan, Kai Shu, Minghui Wu, Jindong Wang, and Fuzhen Zhuang. 2022. Memory-guided multi-view multi-domain fake news detection.IEEE Transactions on Knowledge and Data Engineering35, 7 (2022), 7178–7191. Retrieval-Augmented Multimodal Model for Fake News Detection Conference acronym ’XX, June 03–05, 2018, Woodstock, N...

  68. [68]

    Let the information content of the repre- sentations be decomposed ash 𝑢 ≜𝐶∪𝑈 andh + 𝑢 ≜𝐶∪𝑈 +, where 𝐶 is the common narrative and 𝑈 , 𝑈+ are unique components

    Limitations of Standard Contrastive Learning (CL).The stan- dard contrastive objective, such as InfoNCE, maximizes a variational lower bound on the mutual information (MI) between the encoded views: LCL ∝ −𝐼(𝑓 𝜃 (h𝑢 );𝑓 𝜃 (h+ 𝑢 ))(22) where 𝑓𝜃 is the encoder. Let the information content of the repre- sentations be decomposed ash 𝑢 ≜𝐶∪𝑈 andh + 𝑢 ≜𝐶∪𝑈 +, wh...

  69. [69]

    Limitations of the Canonical Information Bottleneck (IB).The canonical IB principle [41] is formulated as: max z 𝐼(z;y) −𝛽𝐼(z;x)(24) Adapting this to our self-supervised setting by settingx ≡ h𝑢 and y≡h + 𝑢 , the objective becomes: max z𝑢 𝐼(z 𝑢 ;h + 𝑢 ) −𝛽𝐼(z 𝑢 ;h 𝑢 )(25) This formulation is inherently flawed. The compression term, −𝛽𝐼( z𝑢 ;h 𝑢 ), penaliz...

  70. [70]

    The CIBL as a Variational Information-Theoretic Objective.CIBL resolves the aforementioned limitations by jointly optimizing three information-theoretic desiderata. The full objective is a variational upper bound on the following functional: F (𝜙, 𝜓)=−𝜆 1𝐼(z 𝑢 ;h + 𝑢 ) | {z } Alignment + (−𝜆2E𝑞𝜙 [log𝑝 𝜓 (h𝑢 |z𝑢 )]) | {z } Reconstruction +𝜆 3KL(𝑞𝜙 (z𝑢 |h𝑢,...

  71. [71]

    Minimizing Lrecon prevents the loss of unique, essential information fromh 𝑢, directly addressing the limitation of standard CL

    This term maximizes a lower bound on the MI required to reconstruct the original input: 𝐼(z 𝑢 ;h 𝑢 ) ≥ H (h 𝑢 ) −E z𝑢 ∼𝑞𝜙 [−log𝑝 𝜓 (h𝑢 |z𝑢 )]=H (h 𝑢 ) − L recon (28) where H (·) is the differential entropy. Minimizing Lrecon prevents the loss of unique, essential information fromh 𝑢, directly addressing the limitation of standard CL. • Compression: Lcompr...

  72. [72]

    Generalization Bound Perspective.The superiority of CIBL is further substantiated by learning theory. From a PAC-Bayesian standpoint, the generalization error Egen is bounded by the empiri- cal error Eemp plus a complexity term related to the KL divergence between the posterior and prior distributions over model param- eters [50]. This complexity term can...