Multi-Modal Fake News Detection via Bridging the Gap between Modals

Multi-modal fake news detection aims to identify fake information through text and corresponding images. The current methods purely combine images and text scenarios by a vanilla attention module but there exists a semantic gap between different scenarios. To address this issue, we introduce an imag...

Full description

Saved in:
Bibliographic Details
Published inEntropy (Basel, Switzerland) Vol. 25; no. 4; p. 614
Main Authors Liu, Peng, Qian, Wenhua, Xu, Dan, Ren, Bingling, Cao, Jinde
Format Journal Article
LanguageEnglish
Published Switzerland MDPI AG 04.04.2023
MDPI
Subjects
Online AccessGet full text
ISSN1099-4300
1099-4300
DOI10.3390/e25040614

Cover

More Information
Summary:Multi-modal fake news detection aims to identify fake information through text and corresponding images. The current methods purely combine images and text scenarios by a vanilla attention module but there exists a semantic gap between different scenarios. To address this issue, we introduce an image caption-based method to enhance the model’s ability to capture semantic information from images. Formally, we integrate image description information into the text to bridge the semantic gap between text and images. Moreover, to optimize image utilization and enhance the semantic interaction between images and text, we combine global and object features from the images for the final representation. Finally, we leverage a transformer to fuse the above multi-modal content. We carried out extensive experiments on two publicly available datasets, and the results show that our proposed method significantly improves performance compared to other existing methods.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:1099-4300
1099-4300
DOI:10.3390/e25040614