Content area
Abstract
Fake news with multimedia data is ubiquitous on the Internet nowadays, and it is difficult for users to distinguish them. Therefore, it is necessary to design automatic multi-modal fake news detectors. However, the existing works make poor utilization of visual information, and do not fully consider the semantic interaction of multi-modal data. In this paper, we propose the multi-modal transformer using two-level visual features (MTTV) for fake news detection. First, we model texts and images from news uniformly as sequences that can be processed by transformer, and two-level visual features, i.e. global feature and entity-level feature, are used to improve the utilization of news images. Second, we extend the transformer model for natural language processing to multi-modal transformer which can make multi-modal data interact fully and capture the semantic relationships between them. In addition, we propose a scalable classifier to improve the classification balance of fine-grained fake news detection with the problem of class imbalance. Extensive experiments on two public datasets demonstrate that our method achieved significant performance improvement compared to the state-of-the-art methods. The source code is available at https://github.com/cqu-wb/MTTV.
Details
; Xiong, Xian-cai 2 ; Wang, Yong-heng 3 ; Qiang, Bao-hua 4 1 Chongqing University, College of Computer Science, Chongqing, China (GRID:grid.190737.b) (ISNI:0000 0001 0154 0904)
2 Key Laboratory of Monitoring, Evaluation and Early Warning of Territorial Spatial Planning Implementation, Ministry of Natural Resources, Chongqing, China (GRID:grid.453137.7) (ISNI:0000 0004 0406 0561); Chongqing Institute of Planning and Natural Resources Investigation and Monitoring, Chongqing, China (GRID:grid.453137.7)
3 8# of Zhejiang Lab, Hangzhou, China (GRID:grid.510538.a) (ISNI:0000 0004 8156 0818)
4 Guilin University of Electronic Technology, Guangxi Key Laboratory of Trusted Software, Guilin, China (GRID:grid.440723.6) (ISNI:0000 0001 0807 124X)





