Next Article in Journal
Multi-Level Fusion Model for Person Re-Identification by Attribute Awareness
Next Article in Special Issue
Automated Pixel-Level Deep Crack Segmentation on Historical Surfaces Using U-Net Models
Previous Article in Journal
Boosting Iris Recognition by Margin-Based Loss Functions
Previous Article in Special Issue
Ensemble Machine Learning Model to Predict the Waterborne Syndrome
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction

School of Cyber Science and Engineering, College of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Algorithms 2022, 15(4), 119; https://doi.org/10.3390/a15040119
Submission received: 2 March 2022 / Revised: 22 March 2022 / Accepted: 28 March 2022 / Published: 29 March 2022
(This article belongs to the Special Issue Algorithms for Feature Selection)

Abstract

:
In the existing false information detection methods, the quality of the extracted single-modality features is low, the information between different modalities cannot be fully fused, and the original information will be lost when the information of different modalities is fused. This paper proposes a false information detection via multimodal feature fusion and multi-classifier hybrid prediction. In this method, first, bidirectional encoder representations for transformers are used to extract the text features, and S win-transformer is used to extract the picture features, and then, the trained deep autoencoder is used as an early fusion method of multimodal features to fuse text features and visual features, and the low-dimensional features are taken as the joint features of the multimodalities. The original features of each modality are concatenated into the joint features to reduce the loss of original information. Finally, the text features, image features and joint features are processed by three classifiers to obtain three probability distributions, and the three probability distributions are added proportionally to obtain the final prediction result. Compared with the attention-based multimodal factorized bilinear pooling, the model achieves 4.3% and 1.2% improvement in accuracy on Weibo dataset and Twitter dataset. The experimental results show that the proposed model can effectively integrate multimodal information and improve the accuracy of false information detection.

1. Introduction

With the development of social networks, the speed of information dissemination is increasing rapidly, which not only facilitates people to socialize but also facilitates the spread of false information [1,2]. False information seriously damages the credibility of the media, infringes the public’s right to know, hinders the public’s right to participate and supervise, and in severe cases can disrupt social order, cause people’s property damage, cause public panic, disrupt social order and have serious adverse effects on society. For example, in 2011, the rumor that iodized salt would prevent nuclear radiation caused by the Fukushima nuclear accident in Japan caused everyone to frantically buy salt, which not only caused a waste of resources but also disturbed social stability. Therefore, how to detect false information at an early stage has become a recent research hotspot. Early information is in the form of plain text, so the early methods mainly detect false information by extracting text features from text content or extracting corresponding single-modality features from other single-modality data [3,4]. With the development of social media, the form of information has changed from the form of plain text to the form of multimedia [5,6,7]. Most of the existing information is in the form of multimedia. At the same time, research found that the fusion of features of different modalities can effectively improve the performance of the model. Therefore, the recent research in the field of false information detection is mainly based on multimodal methods. However, the existing multimodal methods have shortcomings. First, most of the existing multimodal methods use early fusion methods or late fusion methods. However, there are many limitations in using only one method to fuse multimodal information, and it is difficult to fully fuse the information of different modalities. Second, for text feature extraction, most of them rely on the concatenated output of bi-gated recurrent unit (Bi-GRU) at each time step. However, the feature extraction process lacks the participation of corresponding factual knowledge. Such methods have limited ability to understand named entities in the text, and thus, it is difficult to fully capture the clues at the semantic level of false information [8]. The extraction of image features mostly relies on VGG19 [9] to obtain features. This method requires a large number of parameters and computing resources during training. The quality of features extracted by these two models is lower than that of some existing models. Finally, the problem of original information loss is not considered when performing information fusion between modalities.
In this paper, the above-mentioned issues are investigated separately:
(1)
This paper studies the problem of how to mitigate the influence of irrelevant features and how to enhance fusion between multimodalities. In this paper, the deep autoencoder is used to fuse the information of different modalities early, and the noise removal function of the deep autoencoder is used to reduce the impact of noise on the model performance. The multi-classifier hybrid prediction method is used to fuse the multimodal information in the later stage to enhance the fusion effect between different modalities.
(2)
For the problem of low quality of single-modality features, this paper uses bidirectional encoder representations for transformers (BERT) [10] to replace Bi-GRU in the text feature extraction. BERT has proven its effectiveness in many fields. Shifted window transformer (SWTR) [11] is used to replace VGG19 in image feature extraction. SWTR has achieved good performance in multiple tasks, such as object detection, instance segmentation and semantic segmentation since its release. In this paper, SWTR is applied to multimodal false information detection tasks as an image feature extraction module. The experimental results show that SWTR also has a good performance in the field of false information detection.
(3)
For the problem that the deep autoencoder will lose some information while removing noise, after the joint feature is obtained by the deep autoencoder, the text feature and image feature are spliced with the feature to reduce the loss of original information.
(4)
Comparative experiments and ablation experiments are performed on the Chinese and English datasets, which prove the effectiveness of the model and its various modules in false information detection.

2. Related Work

False information is defined as unsubstantiated or intentionally fabricated stories or statements [12]. Multimodal data refers to the description of different angles of the same object; the description of each angle is a modality [13]. This paper divides the current research into two categories according to the number of modalities used: methods based on single modality and methods based on multimodality [14].

2.1. Single-Modality-Based Approach

Content features refer to features that can be extracted directly from published information. Content features mainly include text features and visual features [15].
In the early research, people’s processing of text content is mainly through manual extraction of text features and the use of machine-learning classification algorithms to detect false information. The method based on text features detects posts by extracting features such as text length, word frequency, part-of-speech and word vectors, morphological and syntactic features from the text information of a post. Verónica [16] et al. built a fake news detector based on linguistic features. The authors extract N grams, punctuation, psycholinguistic features, features indicating text intelligibility and grammar into five sets of linguistic features from the text and finally use the SVM classifier and five cross-validations for detection. Kwon [17] used structural information and linguistic features to capture the multimodal phenomenon of rumor propagation using three classification models: support vector machine, random forest classifier and decision tree. In the early days of the multimedia age, fake news began to use multimedia content with pictures to attract readers to spread false information. Jin [18] et al. demonstrated the importance of images in fake news detection. Due to lack of expertise, it is difficult for traditional machine-learning-based false information detection models to obtain handcrafted features.
With the widespread application of deep learning, Ma [19] et al. used a recurrent neural network to obtain the hidden features of text, which were used to detect rumors. Qi [20] et al. used an attention mechanism to dynamically fuse the features of the image frequency domain and pixel domain for false information detection. Although deep learning can automatically extract features, it is often affected by noise, which degrades its model performance. Inspired by generative adversarial networks (GAN), Ma [21] et al. proposed a model based on GAN network, which can remove noise and irrelevant features through adversarial training to obtain more discriminative features.
Although the detection of false information using text features or visual features alone has been proved to be effective, with the advent of the multimedia era, the existing form of information has changed from the form of plain text in the past to the form that contains text, pictures, videos and other multimodal data at the same time, so if false information is detected only from the perspective of text or pictures, not only are the information utilization ratio and detection accuracy low, but it is also difficult to adapt to the multimedia era that had arrived.

2.2. Multimodal-Based Approach

Most of the existing multimodal-based methods use two kinds of modal information, text and picture, to detect false information. Singhal [22] et al. used BERT to extract text features and VGG19 to extract image features, which were concatenated as joint features for false information detection. Kumari [23] et al. proposed a multimodal fusion model based on multimodal decomposition bilinear pooling. This method can solve the problem that the simple concatenating of different features cannot determine the feature boundary, and the correlation between image feature and text feature cannot be found.
Although the above studies have proposed different ideas in the fusion module, the input is only two modalities of text and image, ignoring the role of the other modalities. Giachanou [24] et al. extract sentiment from text while extracting features from text, extract image labels when extracting image features and perform similarity calculation between image labels and text features. Through the processing of the above method, more information can be obtained from the text and images, and the utilization ratio of the information can be increased. However, it only performs feature fusion through simple concatenating, which fails to fully consider the correlation between different modalities. YUAN [8] et al. proposed semantically enhanced multimodal false information detection, which obtains image features, graph labels and text in the graph through a convolutional network and updates the text information by adding the text extracted from the picture to the text information. The graph labels, original text information and updated text information are used to extract features through enhanced representation from knowledge integration (ERNIE) [25]. The graph labels and graph feature vectors are updated through an attention mechanism using features extracted from the original text. Finally, the updated text features, the updated graph label features and the updated image features are concatenated as the final joint feature for detection.
Meng et al. [14] proposed a multimodal fusion model based on attention mechanism. The model first uses Bi-GRU to extract semantic features and uses convolutional recurrent neural network to extract the features of different levels of the picture and then uses the inter-modality attention and the intra-modality attention to fuse the multimodal information to obtain joint features. Finally, the text features, image features and joint features are fused through the attention mechanism to enhance the role of the original information. The above research works use simple splicing or attention methods to fuse text features and image features, and they only perform early fusion of the two features, ignoring the effect of later fusion [26]. At the same time, in the early fusion, the influence of noise on the fusion results is not considered [27]. In this paper, the above problems are studied, and a deep-autoencoder-based false information detection method is proposed.

3. False Information Detection Method

Problem Definition: Suppose P = { p 1 , p 2 ,   p m } is a dataset of multimodal posts in social networks where p i is the ith post. T = { t 1 , t 2 , t m } is the text set, t i is the text content in the ith post. V = { v 1 , v 2 , , v m } is an image set, v i is the image included in the ith post. L = { l 1 , l 2 , l m } is the tag set, l i is the tag of the ith post. p i = { t i , v i , l i } , The task of false information detection can be described as learning a function f ( T , V ) = Y , Y = { y 1 , y 2 , , y m } , y i is the predicted label value of the ith post, y { 0 , 1 } , 0 represents true information, and 1 represents false information.
The model in this paper is mainly composed of four parts: text feature extractor, image feature extractor, early fusion module and late fusion module. Figure 1 shows our proposed false information detection method based on multimodal feature fusion and multi-classifier hybrid prediction.
The model first uses the text feature extractor and the image feature extractor to extract the text and image features, and then input the two obtained features into the early fusion module to obtain joint features, and the text features, image features and joint features are concatenated to strengthen the joint features. Finally, the text features, the image features and the enhanced joint features are used to obtain the prediction results through the later fusion module. The early fusion module realizes early fusion between different modal features through the deep autoencoder, and the late fusion module uses the method of multi-classifier mixed prediction to fuse the prediction probabilities of different modalities, The pseudocode is shown in Table 1.

3.1. Text Feature Extraction

The text is the main part of the post. It contains the main information that the publisher wants to express, and it can also show the publisher’s emotions and other information. We adopt the BERT pre-training model with strong modeling ability to deal with this important part.
BERT is a pre-training model based on Transformer [28]. Its special unsupervised task enables it to learn contextual information, and it only uses self-attention mechanism instead of RNNs so that it can be parallelized to speed up the training process, and because it has enough parameters, the model can learn more knowledge through learning on large-scale pre-training corpus. It learns some syntactic knowledge and common sense. In some recent similar tasks, usually, BERT-based models will perform better than other networks built on RNNs and CNNs. These experiments show that higher-quality text features can be extracted using BERT. Therefore, this paper uses BERT to extract text features. BERT is used to input text in the form of ‘[CLS]’ + sentence + ’[SEP]’, and the output is an embedded representation of a set of words. The text features are calculated as follows:
T i = B e r t ( t i )
where T i is a collection of text, T i R d i is the feature vector of the text, d i is the dimension of the feature vector.
Since BERT uses the self-attention mechanism internally, if a word with a specific meaning is used to represent the entire sentence, the word vector will be affected by the word itself, and it is difficult to objectively represent the characteristics of the entire sentence. Since the [CLS] vector is a marker bit and does not have semantic information, it will not be affected by itself, so it can be better used as a feature of the entire text. Therefore, this paper selects the [CLS] vector of each sentence to represent the features of the entire text.

3.2. Image Feature Extraction

Picture can supplement the text information and increase the credibility of the content. Processing image parts can better understand the semantics of multimodal posts and thus better detect false information. We use SWTR to process the image part.
SWTR is also an image feature extraction model based on Transformer. This model expands the applicability of Transformer, transfers the high performance of Transformer to the visual neighborhood, solves the shortage of CNN for global information feature extraction, and at the same time, with its unique window mechanism, it greatly reduces the computational complexity of self-attention and solves the problem of fixed token scale, which has become the general backbone of computer vision. Since it was proposed, it has achieved relatively good results in tasks, such as image classification and segmentation tasks. Therefore, this paper uses SWTR as the image feature extractor. The input of SWTR is a picture, and the output is the feature vector of the picture. The specific calculation process is as follows:
V i = S W T R ( I i )
where I i is a picture contained in the ith post, V i R d i is the feature vector of the picture, and d i is the dimension of the feature vector.

3.3. Early Fusion

We obtained text features T i and image features V i . In order to obtain the interactive relationship between the text and the image, this paper uses a deep autoencoder to interact with text and image features to obtain multimodal joint features. The deep autoencoder will convert the high-dimensional feature vector into a low-dimensional feature vector through a function. The low-dimensional feature vector will contain most of the main features of the data, and some tiny features may be lost, so as to eliminate irrelevant features in the features. This module firstly inputs the features of text and images into the encoder for encoding. After the features of text and images enter the encoder, they will be compressed into low-dimensional feature vectors through two linear layers, and then add the encoded two feature vectors by dimension. Finally, the added feature vector is input into the two decoders to obtain two outputs. The specific calculation process is as follows:
T e n c o d e r 1 = L i n e a r ( T i )
T e n c o d e r 2 = L i n e a r ( T e n c o d e r 1 )
T e n c o d e r 3 = L i n e a r ( T e n c o d e r 2 )
V e n c o d e r 1 = L i n e a r ( V i )
V e n c o d e r 2 = L i n e a r ( V e n c o d e r 1 )
V e n c o d e r 3 = L i n e a r ( V e n c o d e r 2 )
F c o = a d d ( T e n c o d e r , V e n c o d e r )
T d e c o d e r 3 = L i n e a r ( F c o )
T d e c o d e r 2 = L i n e a r ( T d e c o d e r 3 )
T d e c o d e r 1 = L i n e a r ( T d e c o d e r 2 )
V d e c o d e r 3 = L i n e a r ( F c o )
V d e c o d e r 2 = L i n e a r ( V d e c o d e r 3 )
V d e c o d e r 1 = L i n e a r ( V d e c o d e r 2 )
L i n e a r ( ) is a linear transformation function, which transforms the dimension of the feature vector, and a d d ( ) is an addition operation, which adds the two input feature vectors according to the corresponding dimensions. T e n c o d e r R e i is the feature of the text feature T i processed by the encoder, and V e n c o d e r R e i is the feature of the image feature Vi processed by the encoder. T e n c o d e r , V e n c o d e r , F c o R e i , e i is the dimension of the feature vector processed by the encoder. T d e c o d e r , V d e c o d e r R d i is the feature vector decoded by the T e n c o d e r , and V e n c o d e r , di is the dimension of the feature vector.
The T e n c o d e r and V e n c o d e r are only used when training this module. This module obtains the prediction error by comparing the output of the decoder with the input of the encoder, and then backpropagation is performed to train the deep autoencoder and gradually improve the accuracy of autoencoding, and finally, the trained model is used as the early fusion module of the model in this paper. In this paper, the F c o after the addition of T e n c o d e r and V e n c o d e r is taken as the joint features of multimodality.

3.4. Late Fusion

Because the deep autoencoder changes the feature vector from high dimensional to low dimensional, it not only removes some irrelevant features but also loses some original information of text and images. Therefore, this paper does not use the joint features F c o extracted from the early fusion module to train classifier 2 but uses the joint features F f i n a l enhanced by the original information to train it. The enhanced joint features F f i n a l is obtained by concatenating the text features T i and image feature V i with the multimodal joint feature F c o .
F f i n a l = T i F C O V i
is the concatenating operation, F f i n a l = T i , F c o , V i .
The three classifiers, F C 1 , F C 2 and F C 3 , are all fully connected layers with activation function s o f t m a x . F C 1 classifies the true and false information through text information, F C 2 classifies the true and false information through the fusion information of text and images, and F C 3 classifies the true and false information through image information. The text feature T i , the enhanced joint features F f i n a l and the image feature V i are projected to the binary target space to obtain the probability distributions P t e x t , P c o and P i m a g e .
P t e x t = s o f t max ( W t T i + b t )
P c o = s o f t max ( W c F f i n a l + b c )
P i m a g e = s o f t max ( W v V i + b v )
where W t , W c and W v represent weight parameters, and b t , b c and b v represent bias terms.
Then, the obtained three probability distributions are fused in proportion to obtain the final predicted probability.
P f i n a l = w 1 p c o + w 2 p t e x t + w e p i m a g e
w 1 + w 2 + w 3 = 1
where w 1 , w 2 and w 3 , respectively, represent the degree of influence of different parts of the model on the final detection results.

3.5. False Information Detection

After normalizing the probability distribution of the later fusion by the function, take out the one with the highest probability as the final result.
P = s o f t max ( P f i n a l )
y i = arg max ( P )
y i is the predicted tag value of our model for ith post.
The loss function is defined as the cross-entropy loss function between the predicted probability distribution and the true label:
L = i = 1 m [ l i log p i + ( 1 l i ) log ( 1 p i ) ]
where m is the number of posts, which l i 0 , 1 is the true label value, 1 represents false information, 0 represents true information, and p i represents the probability of being predicted to be false information.

4. Experimental Results and Analysis

4.1. Experiments

Machine configuration and environment for this experiment: CPU: Intel Xeon E5-2630L v3, 62 G memory, 8cores, GPU: NVIDIA GeForce RTX 3090, PyTorch(1.7.1), Python(3.8), Cuda(10.2). In our experiments, we use two publicly available datasets, namely Twitter and Weibo. The selected Twitter dataset is a public dataset: https://github.com/MKLab-ITI/image-verification-corpus (accessed on 18 October 2021). The selected Weibo dataset is a public dataset: https://drive.google.com/file/d/14VQ7EWPiFeGzxp3XC2DeEHiBEisDINn/view?usp=sharing (accessed on 18 October 2021). The Twitter dataset contains text, related images and contextual information. The information in the training and test sets is collected from different events, so there is no temporal overlap between the training and test sets. This training set contains 14,483 tweets divided into three categories: false, true and humorous. A total of 6841 of them are fake, 5009 are true, and 2633 are humorous. The test set contains 3781 tweets. Since the test set does not contain humor class instances, we ignore these instances in our experiments. The Twitter dataset contains 360 training set images and 50 test set images. We use 20% of the training set as the validation set [18]. The Weibo dataset contains a total of 9563 news items. Of these, 4784 were false and 4779 were true. The test dataset includes 7567 messages. We use 20% as the validation dataset, and the test dataset includes 1996 messages, Table 2 shows the complete data distribution of the two data sets. All the hyper-parameters used for training the proposed model are listed in Table 3.

4.2. Evaluation Index

For binary classification problems, we generally use accuracy, precision, recall and F1 value to evaluate the performance of the model. The accuracy rate is the proportion of all predictions that are correct to the total. The accuracy rate is the proportion of all positive predictions that are correctly predicted. The recall rate is the proportion of correct predictions that are positive to all actual positives. The F1 value is an overall evaluation of precision and recall. The calculation formulas of Accuracy, Precision, Recall and F1 value are described as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 T P 2 T P + F P + F N
The specific meanings of TP, TN, FP and FN are shown in Table 4. TP is a positive sample predicted by the model as a positive class. TN is a negative sample predicted by the model to be a negative class. FP is a negative sample predicted by the model as a positive class. FN is a positive sample predicted by the model as a negative class.

4.3. Baseline Model

We implement some baseline models based on single-modality and multimodal methods to verify the effectiveness of our proposed model.

4.3.1. Single-Modality Model

(1)
Text: First input the text content into the trained BERT pre-training model to obtain the word embedding representation of the word. Then, extract the [CLS] vector output by BERT, and then, input it into a fully connected layer whose activation function is s o f t m a x for classification.
(2)
Visual: Input the image into the trained SWTR model to obtain the feature vector of the image, and then, input it into a fully connected layer whose activation function is s o f t m a x for classification.

4.3.2. Multimodal Model

(1)
att-RNN [29]: att-RNN is to use RNN with attention mechanism to fuse text, image and social context features to achieve the task of rumor detection.
(2)
EANN [30]: Even adversarial neural network (EANN) has three main components: multimodal feature extractor, false information detector, event discriminator. In the model, the feature extractor will extract the textual and visual features of the event with the help of the event discriminator and then concatenate the obtained textual and visual features and finally use the false information detector to detect the authenticity of news posts.
(3)
MVAE [31]: Multimodal variational autoencoder (MVAE) trains three sub-networks to detect fake news. Here, a variational autoencoder is trained for better representation of textual and visual features. The shared latent features are further used for classification.
(4)
AMFB [23]: Attention-based multimodal factorized bilinear (AMFB). The network consists of three parts, a feature extraction module, a feature fusion module and a false information detection module. Attention-based multimodal factorization bilinear pooling to fuse text and image features.
(5)
OUR1: The false information detection model proposed in this paper consists of a feature extractor, an image feature extractor, an early fusion module and a late fusion module. The later fusion module of the model adopts the method of adding the probability distributions obtained by the three classifiers in proportion to obtain the final probability distribution and determine the result by the added probability distribution.
(6)
OUR2: This model adopts the same method as OUR1 in text feature extractor, image feature extractor and early fusion module, but it adopts a voting method in the later fusion module to obtain the result. Each classifier can cast one vote, and the one with the most votes is the final result.
(7)
OUR3: This model is the same as OUR2, only the late fusion module was changed. This model takes into account that the proportion of real information in real life is much larger than false information, so it adopts the ‘one-vote veto’ method in the later fusion module to obtain the results. The method first checks the classification results output by the three classifiers, and if one of them is true, the information is considered to be true.

4.4. Analysis of Results

The experimental results of the baseline models and our model on the two datasets are shown in Table 5. We report the accuracy, precision, recall and F1-score of our model. The experimental results show that the accuracy of our model is better than the baseline models on the Weibo dataset and the Twitter dataset.
By comparing the experimental results of the models using different late fusion methods, it can be found that the model using the proportional addition method for late fusion has higher accuracy, which shows that the effect of proportional addition is better than the two fusion strategies of voting and ‘one-vote veto’, so this paper uses OUR1 as the final false information detection model.
From the perspective of single modality and multimodality, the accuracy of the multimodal model is higher than that of the single-modality model, indicating that multimodal models can obtain more informative features and improve the accuracy of model detection through complementary learning between modalities.
The accuracy of our model on the Twitter dataset is 1.2% higher than that of AMFB, 15% higher than that of MVAE and 15.4% higher than that of EANN. On the Weibo dataset, the accuracy of our model is 4.3% higher than that of AMFB, 5.1% higher than that of MVAE and 8.4% higher than that of EANN.

4.5. Ablation Experiment

In this paper, the effectiveness of each module is verified by comparing the original model with the model after deleting different modules. The results are shown in Table 6 and Table 7:
(1)
O U R : contains all modules.
(2)
D e l e t e _ t : The late fusion module is removed, and the text features in the enhanced joint features are removed.
(3)
D e l e t e _ v : The late fusion module is removed, and the image features in the enhanced joint features are removed.
(4)
D e l e t e _ c : The late fusion module is removed, and the original joint features in the enhanced joint features are removed.
(5)
D e l e t e _ b : Deleting the late fusion module.
We use 5-fold cross validation experiment to demonstrate the validity of our results. Due to the imbalanced distribution of the Twitter dataset, cross-validation cannot be performed well, so we only conduct experiments on the Weibo dataset; the results are shown in Table 6. The accuracy of each model in Table 6 is the average value after 5-fold cross-validation, and the standard deviation of each method is calculated. As can be seen from Table 6 and Table 7, the performance of the model decreases when any module is removed. Compared with the original model, the D e l e t e _ b model has a 0.3% reduction in the accuracy of the Weibo dataset and Twitter dataset, which indicates that the contributions of different modal features to the final result are not the same, and the multimodal features are only passed through the early stage, which is insufficient, and the information between the modalities can be fully fused by using the early fusion module and the late fusion module at the same time. Compared with the original model, the D e l e t e _ c model has lower accuracy on the Weibo dataset and Twitter dataset, indicating that the use of deep autoencoder for early fusion is effective. The accuracy rates of D e l e t e _ t and D e l e t e _ v also decrease on both datasets, indicating that in the process of using the deep autoencoder for early feature fusion, the unique denoising function of the deep autoencoder can not only remove irrelevant features but also lose some useful features. Both text information and image information are partially lost. This part of the loss can be reduced by splicing the original features. The loss of this part can be reduced by concatenating the original features. Comparing the performances of D e l e t e _ t and D e l e t e _ v on the two datasets, it can be found that deleting text features has a greater impact on model accuracy than deleting image features, indicating that text features will have more losses than image features during the fusion process.

4.6. Hypothetical Test

The model proposed in this paper is improved on the basis of the AMFB model, and the AMFB model is the best performing model among the above-mentioned benchmark models. Therefore, the t-test between the model proposed in this paper and the AMFB model is performed to detect whether the difference between them is significant. First, we hypothesize that the AMFB model is not significantly different from the model proposed in this paper. As can be seen from Table 8, the p-value is less than 0.05, so the null hypothesis is rejected, indicating that the AMFB model is significantly different from the model proposed in this paper. As a test, t-values are also looked up based on degrees of freedom and confidence levels to verify that the conclusions are correct. By querying the t-distribution table, when the degree of freedom is 18 and the confidence level is 95%, the t-value is 1.7341, and the t-value in Table 8 is greater than this value, proving that the conclusion is correct.

5. Conclusions

The detection model proposed in this paper has the following advantages: (1) It can interact with the data of different modalities in the early stage to exploit its correlation; (2) It can remove the influence of noise during the fusion process; (3) It can reduce the loss of information in the fusion process; (4) The post-fusion module enables the model to handle the asynchrony of the data, allowing different modalities to use their most appropriate analysis method. At the same time, it contains several disadvantages: (1) The remaining information, other than text and images, cannot be used; (2) The extraction of features is difficult. The features of different modalities must have the same format. It is difficult to express the time synchronization between multimodal features, and it is difficult to obtain the cross-correlation between features as the number of features increases; (3) The model structure is complex and the training is difficult; (4) Only the overall fusion is performed on the data of different modalities, and the local fusion is not performed.
This paper proposes a model for false information detection using multimodal information. The model uses a hybrid fusion method to fuse multimodal information, firstly using a deep autoencoder as an early fusion model to obtain multimodal joint features, and the loss of original information in the fusion process is reduced by concatenating text features and image features on the joint features. Finally, the multimodal data are further fused by the late fusion method. The experimental results on Weibo dataset and Twitter dataset show that the detection accuracy of our model is better than the baseline model. At the same time, the model proposed in this paper has limitations: (1) For a post containing multiple pictures, only one of the pictures can be used to detect it, and all the pictures cannot be used at the same time; (2) The complex structure of the model has a large number of parameters and cannot be run on small devices. The following issues should be addressed in future work: (1) Reduce the complexity of the model so that it can be applied to small devices; (2) Use all the information in the post to detect it; (3) Detect false information that has begun to spread on social platforms.

Author Contributions

Conceptualization, Y.L.; methodology, Y.L.; software, T.T. and A.H.; validation, Y.L., T.T. and A.H.; formal analysis, Y.L.; investigation, Y.L. and T.T.; resources, T.T.; data curation, T.T.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L. and T.T.; visualization, Y.L.; supervision, T.T.; project administration, T.T. and A.H.; funding acquisition, T.T. and A.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by the National Natural Science Foundation of China (62166042, U2003207), Natural Science Foundation of Xinjiang, China (2021D01C076), and Strengthening Plan of National Defense Science and Technology Foundation of China (2021-JCJQ-JJ-0059).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nasir, J.A.; Khan, O.S.; Varlamis, I. Fake news detection: A hybrid CNN-RNN based deep learning approach. Int. J. Inf. Manag. Data Insights 2021, 1, 100007. [Google Scholar] [CrossRef]
  2. Song, C.; Ning, N.; Zhang, Y.; Wu, B. A multimodal fake news detection model based on crossmodal attention residual and multichannel convolutional neural networks. Inf. Process. Manag. 2021, 58, 102437. [Google Scholar] [CrossRef]
  3. Rashkin, H.; Choi, E.; Jang, J.Y.; Volkova, S.; Choi, Y. Truth of varying shades: Analyzing language in fake news and political fact-checking. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017; pp. 2931–2937. [Google Scholar]
  4. Popat, K.; Mukherjee, S.; Strötgen, J.; Weikum, G. Credibility assessment of textual claims on the web. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, New York, NY, USA, 24–28 October 2016; pp. 2173–2178. [Google Scholar]
  5. Alonso-Bartolome, S.; Segura-Bedmar, I. Multimodal Fake News Detection. arXiv 2021, arXiv:2112.04831. [Google Scholar]
  6. Peng, X.; Xintong, B. An effective strategy for multi-modal fake news detection. Multimed. Tools Appl. 2022. [Google Scholar] [CrossRef]
  7. Choi, H.; Ko, Y. Effective Fake News Video Detection Using Domain Knowledge and Multimodal Data Fusion on YouTube. Pattern Recognit. Lett. 2022, 154, 44–52. [Google Scholar] [CrossRef]
  8. Yuan, P.; Cao, J.; Sheng, Q. Semantics-enhanced multi-modal fake news detection. J. Comput. Res. Dev. 2021, 58, 1456. [Google Scholar]
  9. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  10. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  11. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. arXiv 2021, arXiv:2103.14030. [Google Scholar]
  12. Gupta, M.; Zhao, P.; Han, J. Evaluating event credibility on twitter. In Proceedings of the 2012 SIAM International Conference on Data Mining, Anaheim, CA, USA, 26–28 April 2012; pp. 153–164. [Google Scholar]
  13. Zhao, L. Research on Multimodal Data Fusion Methods. Ph.D. Thesis, Dalian University of Technology, Dalian, China, 2018. [Google Scholar]
  14. Meng, J.; Wang, L.; Yang, Y.; Lian, B. Multi-model deep fusion for false information detection. J. Comput. Appl. 2021, 42, 419–425. [Google Scholar]
  15. Wang, J.; Wang, Y.C.; Huang, M.J. False Information in Social Networks: Definition, Detection and Control. Comput. Sci. 2021, 48, 263–277. [Google Scholar]
  16. Pérez-Rosas, V.; Kleinberg, B.; Lefevre, A.; Mihalcea, R. Automatic detection of fake news. arXiv 2017, arXiv:1708.07104. [Google Scholar]
  17. Kwon, S.; Cha, M.; Jung, K.; Chen, W.; Wang, Y. Prominent features of rumor propagation in online social media. In Proceedings of the 2013 IEEE 13th International Conference on Data Mining, Dallas, TX, USA, 7–10 December 2013; pp. 1103–1108. [Google Scholar]
  18. Jin, Z.; Cao, J.; Zhang, Y.; Zhou, J.; Tian, Q. Novel visual and statistical image features for microblogs news verification. IEEE Trans. Multimed. 2016, 19, 598–608. [Google Scholar] [CrossRef]
  19. Ma, J.; Gao, W.; Mitra, P.; Kwon, S.; Jansen, B.J.; Wong, K.F.; Cha, M. Detecting rumors from microblogs with recurrent neural networks. In Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI 2016), New York, NY, USA, 9–15 July 2016. [Google Scholar]
  20. Qi, P.; Cao, J.; Yang, T.; Guo, J.; Li, J. Exploiting multi-domain visual information for fake news detection. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 518–527. [Google Scholar]
  21. Ma, J.; Gao, W.; Wong, K.F. Detect rumors on twitter by promoting information campaigns with generative adversarial learning. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 17 May 2019; pp. 3049–3055. [Google Scholar]
  22. Singhal, S.; Shah, R.R.; Chakraborty, T.; Kumaraguru, P.; Satoh, S.I. Spotfake: A multi-modal framework for fake news detection. In Proceedings of the 2019 IEEE Fifth International Conference on Multimedia Big Data (BigMM), Singapore, 11–13 September 2019; pp. 39–47. [Google Scholar]
  23. Kumari, R.; Ekbal, A. AMFB: Attention based multimodal Factorized Bilinear Pooling for multimodal Fake News Detection. Expert Syst. Appl. 2021, 184, 115412. [Google Scholar] [CrossRef]
  24. Giachanou, A.; Zhang, G.; Rosso, P. Multimodal fake news detection with textual, visual and semantic information. In International Conference on Text, Speech, and Dialogue; Springer: Cham, Switzerland, 2020; pp. 30–38. [Google Scholar]
  25. Sun, Y.; Wang, S.; Li, Y.; Feng, S.; Chen, X.; Zhang, H.; Tian, X.; Zhu, D.; Tian, H.; Wu, H. Ernie: Enhanced representation through knowledge integration. arXiv 2019, arXiv:1904.09223. [Google Scholar]
  26. Ren, Z.Y.; Wang, Z.C.; Ke, Z.W.; Li, Z.; Wushour, S. Survey of Multimodal Data Fusion. Comput. Eng. Appl. 2021, 57, 49–64. [Google Scholar]
  27. Bourlard, H.; Kamp, Y. Auto-association by multilayer perceptrons and singular value decomposition. Biol. Cybern. 1988, 59, 291–294. [Google Scholar] [CrossRef] [PubMed]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  29. Jin, Z.; Cao, J.; Guo, H.; Zhang, Y.; Luo, J. Multimodal fusion with recurrent neural networks for rumor detection on microblogs. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 27 October 2017; pp. 795–816. [Google Scholar]
  30. Wang, Y.; Ma, F.; Jin, Z.; Yuan, Y.; Xun, G.; Jha, K.; Su, L.; Gao, J. Eann: Event adversarial neural networks for multi-modal fake news detection. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 23 August 2018; pp. 849–857. [Google Scholar]
  31. Khattar, D.; Goud, J.S.; Gupta, M.; Varma, V. Mvae: Multimodal variational autoencoder for fake news detection. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 17 May 2019; pp. 2915–2921. [Google Scholar]
Figure 1. Our false information detection model.
Figure 1. Our false information detection model.
Algorithms 15 00119 g001
Table 1. Whole pseudocode.
Table 1. Whole pseudocode.
Input :   t i , v i # t i , v i represents text and image content in a message
Output :   y i # y i is the prediction result of the model for this information
T i = B E R T t i R 1 768 # T i is text feature
V i = S w i n T r a n s f o r m e r v i R 1 768 # V i is image feature
F c o = D e e p _ a u t o e n c o d e r T i , V i R 1 64 # F c o is fused feature
F f i n a l = c o n c a t e T i , V i , F c o R 1 1600 # concatenation   T i , V i   and   F c o
P t e x t = F C 1 T I R 1 2 # P t e x t   is   the   predicted   probability   through   the   text   feature   T i
P c o = F C 2 F f i n a l R 1 2 # P c o   is   the   predicted   probability   through   the   fused   feature   F f i n a l
P i m a g e = F C 3 V i R 1 2 # P i m a g e is the predicted probability through the image feature 𝑉𝑖
P f i n a l = w 1 p t e x t + w 2 p c o + w 3 p i m a g e R 1 2 # P f i n a l is the final predicted probability
y i = a r g m a x P f i n a l # y i is the predicted result
Table 2. The dataset used in the experiment.
Table 2. The dataset used in the experiment.
DataTrainTestImage
FakeRealFakeReal
Twitter6841500925641217410
Delete_b37843783100099613,274
Table 3. Parameter settings.
Table 3. Parameter settings.
ParametersTwitterWeibo
Text length3295
Image size(2,242,243)(2,242,243)
Batch size12832
OptimizerBertAdam (lr = 5 × 10−5)BertAdam (lr = 5 × 10−5)
Epochs4040
Dropout0.60.6
w t e x t 0.530.48
w c o 0.030.07
w i m a g e 0.440.45
Table 4. Meaning of TP, FP, TN, FN.
Table 4. Meaning of TP, FP, TN, FN.
RealityPredict Result
Positive ExampleNegative Example
Positive exampleTPFN
Negative exampleFPTN
Table 5. Experimental results on two datasets.
Table 5. Experimental results on two datasets.
DatasetModelAccuracyFake NewsReal News
PrecisionRecallF-ScorePrecisionRecallF-Score
TwitterText0.8200.8560.8580.8570.7580.7560.757
Visual0.8390.8500.9030.8760.8160.7290.770
att-RNN0.6640.7490.6150.6760.5890.7280.651
EANN0.7410.6900.5500.6100.7600.8500.810
MVAE0.7450.8010.7190.7580.6890.7770.730
AMFB0.8830.8900.9500.9200.8700.7600.810
OUR10.8950.9380.8920.9140.8310.9000.864
OUR20.8410.9080.8320.8680.7510.8560.800
OUR30.7560.7280.9770.8340.9070.3820.538
WeiboText0.8510.8290.8990.8630.8800.8000.838
Visual0.7050.7510.6470.6950.6680.7680.715
att-RNN0.7720.7970.7130.6920.6840.8400.754
EANN0.7910.8400.7200.7800.7600.8600.800
MVAE0.8240.8540.7690.8090.8020.8750.837
AMFB0.8320.8200.8600.8400.8500.8100.830
OUR10.8750.8500.9210.8840.9060.8250.863
OUR20.7730.7390.8720.8000.8280.6670.739
OUR30.5970.5640.9940.7190.9650.1670.284
Table 6. Ablation experiments on Weibo dataset.
Table 6. Ablation experiments on Weibo dataset.
ModelAccuracyFake NewsReal NewsSTDEV
PrecisionRecallF-ScorePrecisionRecallF-Score
OUR0.8750.8500.9210.8840.9060.8250.8630.011
Delete_t0.8290.8600.8000.8290.7990.8590.8280.012
Delete_v0.8530.8290.9040.8650.8850.7980.8390.001
Delete_c0.8480.8500.9130.8800.8980.8260.8600.008
Delete_b0.8720.8540.9110.8810.8960.8310.8620.007
Table 7. Ablation experiments on Twitter dataset.
Table 7. Ablation experiments on Twitter dataset.
ModelAccuracyFake NewsReal News
PrecisionRecallF-ScorePrecisionRecallF-Score
OUR0.8950.9380.8920.9140.8310.9000.864
Delete_t0.8380.8500.9010.8750.8130.7290.769
Delete_v0.8670.8490.9590.9010.9110.7110.799
Delete_c0.8650.8830.9050.8940.8310.7980.814
Delete_b0.8920.9300.8960.9130.8340.8860.859
Table 8. The results of t-test.
Table 8. The results of t-test.
Compare ModelsTwitterWeibo
OUR and AMFBt = 6.937, p = 5.452 × 10−12t = 8.437, p = 9.314 × 10−17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liang, Y.; Tohti, T.; Hamdulla, A. False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction. Algorithms 2022, 15, 119. https://doi.org/10.3390/a15040119

AMA Style

Liang Y, Tohti T, Hamdulla A. False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction. Algorithms. 2022; 15(4):119. https://doi.org/10.3390/a15040119

Chicago/Turabian Style

Liang, Yi, Turdi Tohti, and Askar Hamdulla. 2022. "False Information Detection via Multimodal Feature Fusion and Multi-Classifier Hybrid Prediction" Algorithms 15, no. 4: 119. https://doi.org/10.3390/a15040119

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop