CSpace
Multi-modality frequency-aware cross attention network for fake news detection
Cui, Wei1,2; Zhang, Xuerui2; Shang, Mingsheng2
2024
摘要An increasing number of fake news combining text, images and other forms of multimedia are spreading rapidly across social platforms, leading to misinformation and negative impacts. Therefore, the automatic identification of multimodal fake news has become an important research hotspot in academia and industry. The key to multimedia fake news detection is to accurately extract features of both text and visual information, as well as to mine the correlation between them. However, most of the existing methods merely fuse the features of different modal information without fully extracting intra- and inter-modal connections and complementary information. In this work, we learn physical tampered cues for images in the frequency domain to supplement information in the image space domain, and propose a novel multimodal frequency-aware cross-attention network (MFCAN) that fuses the representations of text and image by jointly modelling intra- and inter-modal relationships between text and visual information whin a unified deep framework. In addition, we devise a new cross-modal fusion block based on the cross-attention mechanism that can leverage inter-modal relationships as well as intra-modal relationships to complement and enhance the features matching of text and image for fake news detection. We evaluated our approach on two publicly available datasets and the experimental results show that our proposed model outperforms existing baseline methods.
关键词Fake news detection multimoal cross attention frequency domain
DOI10.3233/JIFS-233193
发表期刊JOURNAL OF INTELLIGENT & FUZZY SYSTEMS
ISSN1064-1246
卷号46期号:1页码:433-455
通讯作者Zhang, Xuerui(zxr@cigit.ac.cn) ; Shang, Mingsheng(msshang@cigit.ac.cn)
收录类别SCI
WOS记录号WOS:001163267400030
语种英语