Recognition: unknown
Retrieval-Augmented Multimodal Model for Fake News Detection
Pith reviewed 2026-05-10 05:14 UTC · model grok-4.3
The pith
RAMM improves fake news detection by retrieving similar instances to align their abstract narratives and enable analogical reasoning on a multimodal LLM backbone.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
RAMM employs a Multimodal Large Language Model as its backbone to capture cross-modal semantics, incorporates an Abstract Narrative Alignment Module that adaptively extracts and aggregates abstract narrative consistency from diverse instances across domains, and introduces a Semantic Representation Alignment Module that shifts reasoning from direct inference on multimodal features to instance-based analogical reasoning, thereby addressing the failure to capture cross-instance narrative consistency and the lack of domain-specific knowledge in multimodal multidomain fake news detection.
What carries the argument
The Abstract Narrative Alignment Module, which extracts and aggregates high-level narrative consistency from retrieved instances, together with the Semantic Representation Alignment Module, which realigns the decision process toward human-like analogy to those instances.
Load-bearing premise
That the Abstract Narrative Alignment Module can reliably extract and aggregate high-level narrative consistency across diverse instances and domains while the Semantic Representation Alignment Module actually improves generalization by moving the model to instance-based analogical reasoning.
What would settle it
Running RAMM on a new test set of clustered fake news stories in an unseen domain and finding no statistically significant gain over a plain multimodal LLM baseline that lacks the two alignment modules.
Figures
read the original abstract
In recent years, multimodal multidomain fake news detection has garnered increasing attention. Nevertheless, this direction presents two significant challenges: (1) Failure to Capture Cross-Instance Narrative Consistency: existing models usually evaluate each news in isolation, fail to capture cross-instance narrative consistency, and thus struggle to address the spread of cluster based fake news driven by social media; (2) Lack of Domain Specific Knowledge for Reasoning: conventional models, which rely solely on knowledge encoded in their parameters during training, struggle to generalize to new or data-scarce domains (e.g., emerging events or niche topics). To tackle these challenges, we introduce Retrieval-Augmented Multimodal Model for Fake News Detection (RAMM). First, RAMM employs a Multimodal Large Language Model (MLLM) as its backbone to capture cross-modal semantic information from news samples. Second, RAMM incorporates an Abstract Narrative Alignment Module. This component adaptively extracts abstract narrative consistency from diverse instances across distinct domains, aggregates relevant knowledge, and thereby enables the modeling of high-level narrative information. Finally, RAMM introduces a Semantic Representation Alignment Module, which aligns the model's decision-making paradigm with that of humans - specifically, it shifts the model's reasoning process from direct inference on multimodal features to an instance-based analogical reasoning process. Extensive experimental results on three public datasets validate the efficacy of our proposed approach. Our code is available at the following link: https://github.com/li-yiheng/RAMM
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper proposes RAMM, a retrieval-augmented multimodal model for fake news detection. It employs an MLLM backbone to capture cross-modal semantics, an Abstract Narrative Alignment Module that adaptively extracts and aggregates high-level narrative consistency across diverse instances and domains, and a Semantic Representation Alignment Module that shifts reasoning from direct multimodal feature inference to instance-based analogical reasoning. The approach is claimed to address failures in capturing cross-instance consistency and lack of domain-specific knowledge, with efficacy validated by extensive experiments on three public datasets.
Significance. If the modules can be shown to deliver measurable gains in cross-instance narrative modeling and domain generalization beyond a standard MLLM backbone, the work could advance multimodal fake news detection by providing a practical way to handle cluster-based misinformation on social media and improve robustness in data-scarce domains. The open release of code is a positive step toward reproducibility.
major comments (3)
- [Abstract Narrative Alignment Module] Abstract Narrative Alignment Module description: The module is presented as adaptively extracting 'abstract narrative consistency from diverse instances across distinct domains' and aggregating 'relevant knowledge,' yet no retrieval corpus, aggregation procedure, adaptive mechanism, or equations are supplied. This detail is load-bearing for the central claim of capturing cross-instance narrative consistency.
- [Semantic Representation Alignment Module] Semantic Representation Alignment Module description: The module is said to align decision-making with human-like 'instance-based analogical reasoning' rather than direct inference on multimodal features, but no alignment loss, procedure for incorporating retrieved instances, or differentiation from standard feature-based inference is provided. This is essential for substantiating the second main contribution.
- [Experiments] Experimental validation: The abstract states that 'extensive experimental results on three public datasets validate the efficacy,' but the manuscript contains no information on the specific datasets, baselines, metrics, ablation studies, or any quantitative results. Without these, it is impossible to assess whether observed gains support the claimed mechanisms or stem from other factors such as the MLLM backbone.
Simulated Author's Rebuttal
Thank you for the detailed review and constructive feedback on our manuscript. We appreciate the opportunity to clarify and strengthen our work. Below, we provide point-by-point responses to the major comments. We will revise the manuscript to address the concerns regarding insufficient technical details and experimental information.
read point-by-point responses
-
Referee: [Abstract Narrative Alignment Module] Abstract Narrative Alignment Module description: The module is presented as adaptively extracting 'abstract narrative consistency from diverse instances across distinct domains' and aggregating 'relevant knowledge,' yet no retrieval corpus, aggregation procedure, adaptive mechanism, or equations are supplied. This detail is load-bearing for the central claim of capturing cross-instance narrative consistency.
Authors: We acknowledge that the current description of the Abstract Narrative Alignment Module is high-level and lacks the specific technical details necessary to fully substantiate its operation. In the revised version, we will provide a detailed description of the retrieval corpus used, the aggregation procedure, the adaptive mechanism, and include the mathematical formulations and equations for the module. This will better illustrate how it extracts and aggregates abstract narrative consistency across instances and domains. revision: yes
-
Referee: [Semantic Representation Alignment Module] Semantic Representation Alignment Module description: The module is said to align decision-making with human-like 'instance-based analogical reasoning' rather than direct inference on multimodal features, but no alignment loss, procedure for incorporating retrieved instances, or differentiation from standard feature-based inference is provided. This is essential for substantiating the second main contribution.
Authors: We agree that more details are needed for the Semantic Representation Alignment Module to clearly differentiate it from standard approaches and to explain the instance-based analogical reasoning. In the revision, we will elaborate on the alignment loss function, the procedure for incorporating retrieved instances into the reasoning process, and how this shifts the paradigm from direct multimodal feature inference. We will also include relevant equations and diagrams if appropriate. revision: yes
-
Referee: [Experiments] Experimental validation: The abstract states that 'extensive experimental results on three public datasets validate the efficacy,' but the manuscript contains no information on the specific datasets, baselines, metrics, ablation studies, or any quantitative results. Without these, it is impossible to assess whether observed gains support the claimed mechanisms or stem from other factors such as the MLLM backbone.
Authors: We apologize for this oversight in the manuscript preparation. Although the abstract mentions the experimental validation, the detailed experimental setup, including the three public datasets used, the baseline methods, evaluation metrics, ablation studies, and quantitative results, were inadvertently omitted from the main text. In the revised manuscript, we will include a comprehensive Experiments section with all this information, along with tables and figures presenting the results to demonstrate the efficacy of the proposed modules. revision: yes
Circularity Check
No significant circularity; model architecture proposed and validated empirically without self-referential derivations or reductions to inputs
full rationale
The paper proposes RAMM to address two stated challenges in multimodal fake news detection by combining an MLLM backbone with an Abstract Narrative Alignment Module (described as adaptively extracting and aggregating narrative consistency) and a Semantic Representation Alignment Module (described as shifting to instance-based analogical reasoning). No equations, loss functions, retrieval procedures, or mathematical derivations are provided in the abstract or described text that could create self-definition or allow a claimed result to reduce to its own inputs by construction. Efficacy is asserted solely through standard empirical results on three public datasets, which does not constitute a 'prediction' that is statistically forced from fitted parameters. No self-citations are used to justify uniqueness theorems, ansatzes, or load-bearing premises, and no known results are renamed as novel organization. The derivation chain is therefore self-contained as an empirical model proposal rather than a circular reduction.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
Yixuan Chen, Dongsheng Li, Peng Zhang, Jie Sui, Qin Lv, Lu Tun, and Li Shang
-
[2]
In Proceedings of the ACM web conference 2022
Cross-modal ambiguity learning for multimodal fake news detection. In Proceedings of the ACM web conference 2022. 2897–2905
2022
-
[3]
Zhenghan Chen, Changzeng Fu, and Xunzhu Tang. 2023. Multi-domain fake news detection with fuzzy labels. InInternational Conference on Database Systems for Advanced Applications. Springer, 331–343
2023
-
[4]
Limeng Cui, Kai Shu, Suhang Wang, Dongwon Lee, and Huan Liu. 2019. defend: A system for explainable fake news detection. InProceedings of the 28th ACM international conference on information and knowledge management. 2961–2964
2019
-
[5]
Xiaoxi Cui, Weihai Lu, Yu Tong, Yiheng Li, and Zhejun Zhao. 2025. Diffusion- based multi-modal synergy interest network for click-through rate prediction. InProceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval. 581–591
2025
-
[6]
Jannatul Ferdush, Joarder Kamruzzaman, Gour Karmakar, Iqbal Gondal, and Rajkumar Das. 2025. Cross-Domain Fake News Detection Through Fusion of Evidence from Multiple Social Media Platforms.Future Internet17, 2 (2025), 61
2025
-
[7]
Tianyu Gao, Xingcheng Yao, and Danqi Chen. 2021. Simcse: Simple contrastive learning of sentence embeddings. InProceedings of the 2021 conference on empirical methods in natural language processing. 6894–6910
2021
-
[8]
Shuzhi Gong, Richard Sinnott, Jianzhong Qi, and Cecile Paris. 2025. Unseen fake news detection through casual debiasing. InCompanion Proceedings of the ACM on Web Conference 2025. 981–985
2025
-
[9]
Daniel A González-Padilla and Leonardo Tortolero-Blanco. 2020. Social media influence in the COVID-19 Pandemic.International braz j urol46 (2020), 120–124
2020
-
[10]
Fiza Gulzar Hussain, Muhammad Wasim, Seemab Hameed, Abdur Rehman, Muhammad Nabeel Asim, and Andreas Dengel. 2025. Fake news detection landscape: Datasets, data modalities, AI approaches, their challenges, and future perspectives.IEEE Access(2025)
2025
-
[11]
Korir Nancy Jeptoo and Chengjie Sun. 2024. Enhancing fake news detection with large language models through multi-agent debates. InCCF International Conference on Natural Language Processing and Chinese Computing. Springer, 474–486
2024
-
[12]
Ye Jiang, Taihang Wang, Xiaoman Xu, Yimin Wang, Xingyi Song, and Diana Maynard. 2025. Cross-modal augmentation for few-shot multimodal fake news detection.Engineering Applications of Artificial Intelligence142 (2025), 109931
2025
-
[13]
Betul Keles, Niall McCrae, and Annmarie Grealish. 2020. A systematic review: the influence of social media on depression, anxiety and psychological distress in adolescents.International journal of adolescence and youth25, 1 (2020), 79–93
2020
-
[14]
Dhruv Khattar, Jaipal Singh Goud, Manish Gupta, and Vasudeva Varma. 2019. Mvae: Multimodal variational autoencoder for fake news detection. InThe world wide web conference. 2915–2921
2019
-
[15]
Sejeong Kwon, Meeyoung Cha, Kyomin Jung, Wei Chen, and Yajun Wang. 2013. Prominent features of rumor propagation in online social media. In2013 IEEE 13th international conference on data mining. IEEE, 1103–1108
2013
- [16]
-
[17]
Chenhui Li and Weihai Lu. 2026. Decoding the Market’s Pulse: Context-Enriched Agentic Retrieval Augmented Generation for Predicting Post-Earnings Price Shocks. InProceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). 3055–3073
2026
-
[18]
Junnan Li, Dongxu Li, Silvio Savarese, and Steven Hoi. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. InInternational conference on machine learning. PMLR, 19730–19742
2023
-
[19]
Jinjin LI, Guoming SANG, and Yijia ZHANG. 2024. APK-CNN and Transformer- enhanced multi-domain fake news detection model.Journal of Computer Appli- cations(2024), 1930–1939
2024
- [20]
-
[21]
Yuyuan Li, Chaochao Chen, Yizhao Zhang, Weiming Liu, Lingjuan Lyu, Xiaolin Zheng, Dan Meng, and Jun Wang. 2023. Ultrare: Enhancing receraser for recom- mendation unlearning via error decomposition.Advances in Neural Information Processing Systems36 (2023), 12611–12625
2023
-
[22]
Yuyuan Li, Yizhao Zhang, Weiming Liu, Xiaohua Feng, Zhongxuan Han, Chaochao Chen, and Chenggang Yan. 2025. Multi-Objective Unlearning in Rec- ommender Systems via Preference Guided Pareto Exploration.IEEE Transactions on Services Computing(2025)
2025
-
[23]
Shijie Liu, Ruixin Ding, Weihai Lu, Jun Wang, Mo Yu, Xiaoming Shi, and Wei Zhang. 2025. Coherency improved explainable recommendation via large lan- guage model. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 12201–12209
2025
-
[24]
Wayne Lu and Xiaoxi Cui. 2026. DEALT: LLM-driven Diversity-Enhanced Data Augmentation for Long-Tail Text Classification. InProceedings of the AAAI Con- ference on Artificial Intelligence, Vol. 40. 32338–32346
2026
-
[25]
Wayne Lu and Yiheng Li. 2026. From Blind Transfer to Wise Selection: Prototype- Driven Neighbor-Domain Adaptation for Fake News Detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 40. 818–826
2026
-
[26]
Weihai Lu, Yu Tong, and Zhiqiu Ye. 2025. Dammfnd: Domain-aware multimodal multi-view fake news detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 559–567
2025
-
[27]
Weihai Lu and Li Yin. 2025. Dmmd4sr: Diffusion model-based multi-level multi- modal denoising for sequential recommendation. InProceedings of the 33rd ACM International Conference on Multimedia. 6363–6372. Conference acronym ’XX, June 03–05, 2018, Woodstock, NY Trovato et al
2025
-
[28]
Xiaoxiao Ma, Yuchen Zhang, Kaize Ding, Jian Yang, Jia Wu, and Hao Fan. 2024. On fake news detection with LLM enhanced semantics mining. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. 508–521
2024
-
[29]
Despoina Mouratidis, Andreas Kanavos, and Katia Kermanidis. 2025. From misinformation to insight: machine learning strategies for fake news detection. Information16, 3 (2025), 189
2025
-
[30]
Qiong Nan, Juan Cao, Yongchun Zhu, Yanyan Wang, and Jintao Li. 2021. MD- FEND: Multi-domain fake news detection. InProceedings of the 30th ACM inter- national conference on information & knowledge management. 3343–3347
2021
-
[31]
Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding.arXiv preprint arXiv:1807.03748(2018)
work page internal anchor Pith review Pith/arXiv arXiv 2018
-
[32]
Eleftheria Papageorgiou, Christos Chronis, Iraklis Varlamis, and Yassine Himeur
-
[33]
A survey on the use of large language models (llms) in fake news.Future Internet16, 8 (2024), 298
2024
-
[34]
Eleftheria Papageorgiou, Iraklis Varlamis, and Christos Chronis. 2025. Harnessing Large Language Models and deep neural networks for fake news detection. Information16, 4 (2025), 297
2025
-
[35]
Zitao Peng, Nankai Lin, Yongmei Zhou, Dong Zhou, and Aimin Yang. 2023. Prompt learning for low-resource multi-domain fake news detection. In2023 international conference on asian Language Processing (IALP). IEEE, 314–319
2023
-
[36]
Gordon Pennycook and David G Rand. 2021. The psychology of fake news.Trends in cognitive sciences25, 5 (2021), 388–402
2021
-
[37]
Ben Poole, Sherjil Ozair, Aaron Van Den Oord, Alex Alemi, and George Tucker
-
[38]
InInternational conference on machine learning
On variational bounds of mutual information. InInternational conference on machine learning. PMLR, 5171–5180
-
[39]
Peng Qi, Zehong Yan, Wynne Hsu, and Mong Li Lee. 2024. Sniffer: Multimodal large language model for explainable out-of-context misinformation detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 13052–13062
2024
-
[40]
Shaina Raza, Drai Paulen-Patterson, and Chen Ding. 2025. Fake news detection: comparative evaluation of BERT-like models and large language models with generative AI-annotated data.Knowledge and Information Systems67, 4 (2025), 3267–3292
2025
-
[41]
Amila Silva, Ling Luo, Shanika Karunasekera, and Christopher Leckie. 2021. Em- bracing domain differences in fake news: Cross-domain fake news detection using multi-modal data. InProceedings of the AAAI conference on artificial intelligence, Vol. 35. 557–565
2021
-
[42]
Chenguang Song, Nianwen Ning, Yunlei Zhang, and Bin Wu. 2021. Knowl- edge augmented transformer for adversarial multidomain multiclassification multimodal fake news detection.Neurocomputing462 (2021), 88–100
2021
-
[43]
Malliga Subramanian, B Premjith, Kogilavani Shanmugavadivel, Santhiya Pandiyan, Balasubramanian Palani, and Bharathi Raja Chakravarthi. 2025. Overview of the shared task on fake news detection in dravidian languages- DravidianLangTech@ NAACL 2025. InProceedings of the fifth workshop on speech, vision, and language technologies for dravidian languages. 759–767
2025
-
[44]
Naftali Tishby, Fernando C Pereira, and William Bialek. 2000. The information bottleneck method.arXiv preprint physics/0004057(2000)
work page Pith review arXiv 2000
-
[45]
Yu Tong, Weihai Lu, Xiaoxi Cui, Yifan Mao, and Zhejun Zhao. 2025. Dapt: Domain- aware prompt-tuning for multimodal fake news detection. InProceedings of the 33rd ACM International Conference on Multimedia. 7902–7911
2025
-
[46]
Yu Tong, Weihai Lu, Zhe Zhao, Song Lai, and Tong Shi. 2024. Mmdfnd: Multi- modal multi-domain fake news detection. InProceedings of the 32nd ACM Inter- national Conference on Multimedia. 1178–1186
2024
-
[47]
Zhao Tong, Yimeng Gu, Huidong Liu, Qiang Liu, Shu Wu, Haichao Shi, and Xiao-Yu Zhang. 2025. Generate first, then sample: Enhancing fake news detection with LLM-augmented reinforced sampling. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 24276–24290
2025
-
[48]
Bo Wang, Jing Ma, Hongzhan Lin, Zhiwei Yang, Ruichao Yang, Yuan Tian, and Yi Chang. 2024. Explainable fake news detection with large language model via defense among competing wisdom. InProceedings of the ACM web conference
2024
-
[49]
Yaqing Wang, Fenglong Ma, Zhiwei Jin, Ye Yuan, Guangxu Xun, Kishlay Jha, Lu Su, and Jing Gao. 2018. Eann: Event adversarial neural networks for multi-modal fake news detection. InProceedings of the 24th acm sigkdd international conference on knowledge discovery & data mining. 849–857
2018
-
[50]
Xiaolong Wei, Yuehu Dong, Xingliang Wang, Xingyu Zhang, Zhejun Zhao, Dong- dong Shen, Long Xia, and Dawei Yin. 2026. Beyond react: A planner-centric framework for complex tool-augmented llm reasoning. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 40. 33845–33853
2026
-
[51]
Xiaolong Wei, Bo Lu, Xingyu Zhang, Zhejun Zhao, Dongdong Shen, Long Xia, and Dawei Yin. 2025. Igniting creative writing in small language models: Llm-as- a-judge versus multi-agent refined rewards. InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing. 17171–17197
2025
-
[52]
Shitao Xiao, Zheng Liu, Peitian Zhang, Niklas Muennighoff, Defu Lian, and Jian-Yun Nie. 2024. C-pack: Packed resources for general chinese embeddings. InProceedings of the 47th international ACM SIGIR conference on research and development in information retrieval. 641–649
2024
-
[53]
Aolin Xu and Maxim Raginsky. 2017. Information-theoretic analysis of gen- eralization capability of learning algorithms.Advances in neural information processing systems30 (2017)
2017
-
[54]
Junxiao Xue, Yabo Wang, Yichen Tian, Yafei Li, Lei Shi, and Lin Wei. 2021. De- tecting fake news by exploring the consistency of multimodal data.Information Processing & Management58, 5 (2021), 102610
2021
-
[55]
Kaiying Yan, Moyang Liu, Yukun Liu, Ruibo Fu, Zhengqi Wen, Jianhua Tao, Xuefei Liu, and Guanjun Li. 2025. MTPareto: A MultiModal Targeted Pareto Framework for Fake News Detection. InICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 1–5
2025
-
[56]
Xuankai Yang, Yan Wang, Xiuzhen Zhang, Shoujin Wang, Huaxiong Wang, and Kwok Yan Lam. 2025. A macro-and micro-hierarchical transfer learning framework for cross-domain fake news detection. InProceedings of the ACM on Web Conference 2025. 5297–5307
2025
-
[57]
Jingyuan Yi, Zeqiu Xu, Tianyi Huang, and Peiyang Yu. 2025. Challenges and innovations in llm-powered fake news detection: A synthesis of approaches and future directions. InProceedings of the 2025 2nd international conference on generative artificial intelligence and information security. 87–93
2025
-
[58]
Xinquan Yu, Ziqi Sheng, Wei Lu, Xiangyang Luo, and Jiantao Zhou. 2025. Racmc: Residual-aware compensation network with multi-granularity constraints for fake news detection. InProceedings of the AAAI Conference on Artificial Intelli- gence, Vol. 39. 986–994
2025
-
[59]
Zhi Zeng, Minnan Luo, Xiangzheng Kong, Huan Liu, Hao Guo, Hao Yang, Zihan Ma, and Xiang Zhao. 2024. Mitigating World Biases: A Multimodal Multi-View Debiasing Framework for Fake News Video Detection. InProceedings of the 32nd ACM International Conference on Multimedia. 6492–6500
2024
-
[60]
Zhi Zeng, Jiaying Wu, Minnan Luo, Xiangzheng Kong, Zihan Ma, Guang Dai, and Qinghua Zheng. 2025. Understand, Refine and Summarize: Multi-View Knowledge Progressive Enhancement Learning for Fake News Video Detection. InProceedings of the 33rd ACM International Conference on Multimedia. 9216– 9225
2025
-
[61]
Zhi Zeng, Jiaying Wu, Minnan Luo, Herun Wan, Xiangzheng Kong, Zihan Ma, Guang Dai, and Qinghua Zheng. 2025. Imol: Incomplete-modality-tolerant learn- ing for multi-domain fake news video detection. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 30921–30933
2025
- [62]
-
[63]
Chaowei Zhang, Zongling Feng, Zewei Zhang, Jipeng Qiang, Guandong Xu, and Yun Li. 2025. Is llms hallucination usable? llm-based negative reasoning for fake news detection. InProceedings of the AAAI Conference on Artificial Intelligence, Vol. 39. 1031–1039
2025
-
[64]
Xing Zhou, Juan Cao, Zhiwei Jin, Fei Xie, Yu Su, Dafeng Chu, Xuehui Cao, and Junqiang Zhang. 2015. Real-time news cer tification system on sina weibo. In Proceedings of the 24th international conference on world wide web. 983–988
2015
- [65]
-
[66]
Yongchun Zhu, Qiang Sheng, Juan Cao, Shuokai Li, Danding Wang, and Fuzhen Zhuang. 2022. Generalizing to the future: Mitigating entity bias in fake news detection. InProceedings of the 45th international ACM SIGIR conference on research and development in information retrieval. 2120–2125
2022
-
[67]
Explosive footage! A poll worker in XX County was caught on a hidden camera shredding pro-government ballots to rig the election. The proof is undeniable!
Yongchun Zhu, Qiang Sheng, Juan Cao, Qiong Nan, Kai Shu, Minghui Wu, Jindong Wang, and Fuzhen Zhuang. 2022. Memory-guided multi-view multi-domain fake news detection.IEEE Transactions on Knowledge and Data Engineering35, 7 (2022), 7178–7191. Retrieval-Augmented Multimodal Model for Fake News Detection Conference acronym ’XX, June 03–05, 2018, Woodstock, N...
2022
-
[68]
Let the information content of the repre- sentations be decomposed ash 𝑢 ≜𝐶∪𝑈 andh + 𝑢 ≜𝐶∪𝑈 +, where 𝐶 is the common narrative and 𝑈 , 𝑈+ are unique components
Limitations of Standard Contrastive Learning (CL).The stan- dard contrastive objective, such as InfoNCE, maximizes a variational lower bound on the mutual information (MI) between the encoded views: LCL ∝ −𝐼(𝑓 𝜃 (h𝑢 );𝑓 𝜃 (h+ 𝑢 ))(22) where 𝑓𝜃 is the encoder. Let the information content of the repre- sentations be decomposed ash 𝑢 ≜𝐶∪𝑈 andh + 𝑢 ≜𝐶∪𝑈 +, wh...
-
[69]
Limitations of the Canonical Information Bottleneck (IB).The canonical IB principle [41] is formulated as: max z 𝐼(z;y) −𝛽𝐼(z;x)(24) Adapting this to our self-supervised setting by settingx ≡ h𝑢 and y≡h + 𝑢 , the objective becomes: max z𝑢 𝐼(z 𝑢 ;h + 𝑢 ) −𝛽𝐼(z 𝑢 ;h 𝑢 )(25) This formulation is inherently flawed. The compression term, −𝛽𝐼( z𝑢 ;h 𝑢 ), penaliz...
-
[70]
The CIBL as a Variational Information-Theoretic Objective.CIBL resolves the aforementioned limitations by jointly optimizing three information-theoretic desiderata. The full objective is a variational upper bound on the following functional: F (𝜙, 𝜓)=−𝜆 1𝐼(z 𝑢 ;h + 𝑢 ) | {z } Alignment + (−𝜆2E𝑞𝜙 [log𝑝 𝜓 (h𝑢 |z𝑢 )]) | {z } Reconstruction +𝜆 3KL(𝑞𝜙 (z𝑢 |h𝑢,...
2018
-
[71]
Minimizing Lrecon prevents the loss of unique, essential information fromh 𝑢, directly addressing the limitation of standard CL
This term maximizes a lower bound on the MI required to reconstruct the original input: 𝐼(z 𝑢 ;h 𝑢 ) ≥ H (h 𝑢 ) −E z𝑢 ∼𝑞𝜙 [−log𝑝 𝜓 (h𝑢 |z𝑢 )]=H (h 𝑢 ) − L recon (28) where H (·) is the differential entropy. Minimizing Lrecon prevents the loss of unique, essential information fromh 𝑢, directly addressing the limitation of standard CL. • Compression: Lcompr...
-
[72]
Generalization Bound Perspective.The superiority of CIBL is further substantiated by learning theory. From a PAC-Bayesian standpoint, the generalization error Egen is bounded by the empiri- cal error Eemp plus a complexity term related to the KL divergence between the posterior and prior distributions over model param- eters [50]. This complexity term can...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.