Next Article in Journal
A Secure Personal Health Record Sharing System with Key Aggregate Dynamic Searchable Encryption
Next Article in Special Issue
LASNet: A Light-Weight Asymmetric Spatial Feature Network for Real-Time Semantic Segmentation
Previous Article in Journal
Integrating Teachers’ TPACK Levels and Students’ Learning Motivation, Technology Innovativeness, and Optimism in an IoT Acceptance Model
Previous Article in Special Issue
Specific Emitter Identification Based on Self-Supervised Contrast Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Rumor Detection Method Based on Multimodal Feature Fusion by a Joining Aggregation Structure

1
Department of Computer and Information Security, Zhejiang Police College, Hangzhou 310053, China
2
Big Data and Cyber Security Research Institute, Zhejiang Police College, Hangzhou 310053, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(19), 3200; https://doi.org/10.3390/electronics11193200
Submission received: 4 September 2022 / Revised: 2 October 2022 / Accepted: 3 October 2022 / Published: 6 October 2022

Abstract

:
Online rumors spread rapidly through social media, which is a great threat to public safety. Existing solutions are mainly based on content features or propagation structures for rumor detection. However, due to the variety of strategies in creating rumors, only considering certain features cannot achieve good enough detection results. In addition, existing works only consider the rumor propagation structure and ignore the aggregation structures of rumors, which cannot provide enough discriminative features (especially in the early days of rumors, when the structure of the propagation is incomplete). To solve these problems, this paper proposes a rumor detection method with multimodal feature fusion and enhances the feature representation of the rumor propagation network by adding aggregation features. More specifically, we built a graph model of the propagation structure as well as the aggregation structure. Next, by utilizing the BERT pre-training model and the bidirectional graph convolutional network, we captured the features of text content, propagation structure, and aggregation structure, respectively. Finally, the multimodal features were aggregated based on the attention mechanism, and the final result was obtained through the MLP classifier. Experiments on real-world datasets show that our model outperforms state-of-the-art approaches.

1. Introduction

Social media platforms are typical places for misinformation to spread. Social media platforms are frequently used by rumor-makers to spread false information. Rumors are spread by rumor mongers to manipulate public events, which can lead to negative consequences. For instance, in politics, rumors have influenced opinions on crucial matters, such as Brexit [1] and the 2016 US presidential election [2]. The “information pandemic” [3] brought on by rumors (in the context of the recent new crown outbreak) has resulted in significant opposition to the suppression of the disease. Therefore, it is crucial to identify and control rumors.
Many research studies have invested in rumor detection. The most direct and efficient approaches involve fact-checking, which involves using known facts to validate the veracity of the news. Known facts come from domain experts, authoritative media, popular science websites, etc. By building a library of known facts, the news to be predicted is searched for in the library, and if there is no similar content in the library, it is considered a rumor. However, there are huge costs involved in building a library of known facts, so the coverage of the library of known facts is often extremely limited [4].
The two primary categories of automated rumor detection techniques, aside from fact-checking-based techniques, are content feature-based methods and propagation structure-based methods. Most content feature-based methods take advantage of textual content, such as user retweets and source tweets [5,6]. To extract text features, the majority of them employ pre-trained models, such as word2vec [7], Glove [8], BERT [9], etc. BERT (bidirectional encoder representation from transformers) is the encoder of a bidirectional transformer. The model uses two methods, masked LM and next sentence prediction, to capture the representation of words and sentences, respectively. However, methods solely based on text content cannot fit all rumor detection methods, mainly due to the following reasons. Firstly, some rumors are created by imitating the sentence patterns and word patterns of normal text. Secondly, adding real information to rumors makes it more difficult for the classifier to judge.
Recently, researchers found that the propagation structures of rumors and truth are significantly different. According to Vosoughi et al. [10,11], rumors spread more quickly, deeply, and widely than the truth. As a result, recent research studies [12,13,14,15] have presented high-level representations from propagation structures to identify rumors. With regard to encoders, deep learning algorithms are widely adopted due to their excellent performances [16,17,18].
Methods based on propagation structures usually represent the spread of rumors as graph or tree structures, with tweets acting as nodes and the interactions formed by retweets and comments acting as edges. To extract features from the graph structure, existing research studies are mainly divided into two types—RNN-based methods and CNN-based methods. RNN-based techniques can capture sequential features from rumor propagation structures [13], including long short-term memory (LSTM), gated recurrent unit (GRU), and recurrent neural network (RvNN). However, RNN-based approaches ignore the correlation between rumors and instead primarily concentrate on the sequential propagation patterns of rumors. As a result, some researchers use convolutional neural networks (CNNs) to extract the connection features of rumors [15]. However, the global propagation structural relationships of graphs cannot be handled by CNN-based algorithms, although they can capture relevant features of local neighbors [19].
Therefore, some researchers [20,21,22] have attempted to extract high-dimensional feature representations of rumor propagation structures using graph convolutional networks (GCNs). A GCN is a graph-data version of CNN that can successfully capture global graph features by aggregating node neighborhood data.
However, existing propagation-structure-based methods typically only consider the top-to-bottom propagation of rumors when constructing a graph model, ignoring the bottom-to-top aggregation of public opinions during the rumor-spreading process. Starting with the source tweet, a top-to-bottom propagation network is built between tweets through retweets or comments, as seen in the left part of Figure 1. On the other hand, all comments or retweets are the public’s opinions on the previous tweet, and opinions are continuously gathered from the bottom to the top, as shown in the right picture of Figure 1. Existing methods often only consider the top-to-bottom propagation of rumors, which is not comprehensive. By considering the aggregation structure of a rumor, we can capture more discriminative rumor features. Considering the aggregation structure is helpful for the enhancement of rumor features, especially in the early stage of rumor propagation, when there are few tweets in the propagation network.
Additionally, the methods based on the propagation structure cannot effectively utilize the text features of the rumors, even though some existing methods use the text features of tweets as the initialization features of the nodes and then use the graph convolutional network method to extract the propagation structure features. However, the initialized text features gradually fade away during the training process as the iterations increase.
This paper suggests a rumor detection method based on multimodal information fusion that fully considers rumor text features, propagation structure features, and aggregate structure features. First, we used the BERT pre-training model to extract the source tweet text features of the event. Next, we built two graphs for the propagation structure and aggregation structure, and the bidirectional graph convolutional network (Bi-GCN) [21] to extract the propagation and aggregation structure features. Finally, we chose a multilayer perceptron (MLP) to obtain the final result. MLP is a forward-structured artificial neural network. The main contributions of this paper are as follows:
  • We introduce an attention mechanism and suggest a multimodal information fusion-based rumor detection method that fully integrates the text features and propagation–aggregation structure features of rumors.
  • We improved the rumor propagation network by including the aggregation structure, which can more accurately distinguish rumors from the truth, particularly in the early stages of rumors with an incomplete propagation structure.
  • Two real-world Twitter datasets were used in the experiments, and the results show that our method can identify rumors more accurately.
This paper is a continuation of previous work [22] presented at the Fifth ICET conference. The previous version did not discuss the early detection of rumors. This paper supplements the experiments of early detections of rumors, adds related work chapters, and elaborates on the methodology and experimental process.

2. Related Work

Rumor detection methods are mainly divided into two categories—content feature-based methods and propagation structure-based methods. The methods based on content features mainly use the text content, picture content, video content, etc., of the original tweet as the model input to detect fake news.
The main method is based on text content. For example, Ma et al. [12] applied deep learning techniques to fake news detection. This method inputs each sentence of the text into the recurrent neural network, uses the hidden layer vector of the recurrent neural network to represent the news information, and inputs the hidden layer features into the classifier to obtain the classification result. Yu et al. [14] used a convolutional neural network to extract text features, and input the obtained embedding vector into the classifier to obtain the final classification result. Vaibhav et al. [23] modeled news articles as graphs with sentences as nodes and inter-sentence similarities as edges; they transformed the fake news detection problem into a graph classification problem. Cheng et al. [24] used a variational autoencoder (VAE) to self-encode text information to obtain the embedding representation of news text, and multi-task learning of the obtained news vector to improve the effectiveness of the model.
News or tweets contain textual as well as visual information, such as pictures and videos. Traditional statistics-based methods use the number of attached images, image popularity, and image type to detect fake news. However, these statistics-based features cannot describe the semantic features of images.
With the rise of deep learning, a large number of works [25,26,27,28] are using convolutional neural networks, such as VGG [29] or ResNet [30], to extract features from pictures; researchers are using the extracted features to detect fake news. However, the existing image forgery technology can change the semantic information of the image. The traditional CNN-based model can only extract the pixel-level information of the image, and cannot identify whether the image has been forged.
However, existing rumor makers often write fake news in the same way as real news. Therefore, it is not enough to distinguish fake news based on its content. Sociological studies [10,11] have shown that the propagation of real news and fake news in social networks is often different. Therefore, more researchers are using the propagation structure of news to detect fake news.
For example, Liu et al. [31] regarded the spread of rumors and comment information as a time series; they used RNN and CNN to model the sequence, splice two latent vectors together, and input them into the classification layer to obtain the classification result. Ma et al. [13] modeled the propagation process of rumors as a tree structure. This work constructed a bottom-up propagation tree and a top-down propagation tree, and used recurrent neural networks to extract node features in the tree to classify fake news. Song et al. [32] modeled the news propagation graph as a dynamic graph. Considering the dynamic changes of the news propagation process, the dynamic graph embedding vector was obtained by using the dynamic graph neural network, which was input into the classifier to obtain the classification result.
However, the existing methods did not consider the aggregation features of rumors, which will lead to the loss of information. In addition, the existing methods cannot aggregate the multimodal rumor features well.

3. Problem Statement

The definition of the dataset for rumor detection is C = { c 1 , c 2 , , c m } , where c i is the i -th event and m is the total number of events. c i = { r i , w 1 i , w 2 i , , w n i 1 i , G i } , where n i is the number of posts in event c i , r i is the source post, w j i is the j -th relevant responsive post, and G i = ( V i , E i ) is defined as a graph, where V i = { r i , w 1 i , w 2 i , , w n 1 i } represents the set of nodes, and E i = { e s t i | s , t = 0 , n i 1 } represents a forwarding or replying relationship between two nodes.
For instance, there is a directed edge w 1 i w 2 i , if w 2 i is a retweet of w 1 i , which is represented as e 12 i . The adjacency matrix is defined as A i { 0 , 1 } n i * n i . e s t i represents the corresponding element value of the s -th row and t -th column of matrix A i , i.e.:
a s t i = { 1 ,   if   e s t i E i 0 , otherwise
We define X i = [ x 0 i , x 1 i , , x n i 1 i ] as the feature matrix composed of all tweets in event c i , where x 0 i is the feature vector of the source tweet r i and x j i is the feature vector of other responded tweets w j i . In this paper, we used the BERT pre-trained model to extract the feature vector of each tweet’s content. Furthermore, a ground-truth label y i was connected to each event c i . In this paper, y i { N , F , T , U } , stands for non-rumor, false rumor, true rumor, and unverified rumor, respectively.
The objective of rumor detection is to learn a classifier from the dataset, i.e.:
f : C Y
where C and Y, respectively, stand for the sets of events and labels.

4. Materials and Methods

Figure 2 depicts the general workflow of the model employed in this article. We first extracted the text features from tweets using the BERT pre-training model, then we extracted the propagation and aggregation features from rumors using Bi-GCN, and finally we utilized a fully-connected neural network to obtain the result.
We will go into more detail about the model in this section. We exclude the subscript i from the following part in order to better illustrate our methodology.

4.1. Text Feature Extraction

To extract the content features of rumor texts, the BERT [9] pre-training model was used in this part. BERT can more effectively address the issue of polysemy, i.e., the output of the model for the same word in different contexts is also different (in contrast to the standard word2vec, Glove, and other approaches).
Figure 3 shows the pre-training model architecture of BERT. Each word’s embedding, which is divided into three parts by the input layer (token embedding, segment embedding, and position embedding) is one of them. Token embedding is a traditional word-embedding method, such as the one-hot method. There are two embedded special tokens, [CLS] and [SEP], at the start and end of the sentence, respectively. The sentence number to which the word belongs to is labeled using segmentation embeddings. Positional embeddings are used to represent the input sequence’s sequential features. The bidirectional transformer allows the BERT pre-training model to learn how the context word affects the current word, which improves the extraction of semantic deep features.
We utilized the BERT pre-training model to extract the content features of each tweet text in each event, which is represented as x 0 . We then used an additional two layers of fully-connected layers to convert it into a v 2 -dimensional vector S T E X T , in order to keep its dimensions consistent with the dimensions of the propagation structure feature and aggregation structure feature, which will be introduced later.

4.2. Structure Feature Extraction

By using a bidirectional graph convolutional network (Bi-GCN), we extracted event propagation and aggregation structure features. Compared with traditional GCN, Bi-GCN can capture bidirectional neighbor features in the graph model. Based on the retweet and respond relationship, we built a propagation structure graph for each event. Meanwhile, we denoted the adjacency matrix and feature matrix as A { 0 , 1 } n * n and X, respectively. A and X were the inputs of the model.
The Bi-GCN was made up of two components: a bottom-to-up graph convolutional network (BU-GCN) and a top-to-down graph convolutional network (TD-GCN). Even though their model structures were relatively similar, they had different adjacency matrices. The adjacency matrix in TD-GCN is represented as A T D = A , whereas in BU-GCN, it is represented as A B U = A T , which is the transposition of A. The feature matrix X is the same for TD-GCN and BU-GCN.
The features of an event’s propagation structure and aggregation structure were extracted based on TD-GCN and BU-GCN, respectively. To extract features in TD-GCN, we utilized two graph convolutional layers (GCL); the calculation formula is as follows:
H 1 T D = σ ( A T D X W 0 T D )
H 2 T D = σ ( A T D H 1 T D W 1 T D )
where H 1 T D n * v 1 , H 2 T D n * v 2 is the output of the first layer and second layer GCL of TD-GCN, known as the hidden states; the total number of nodes is n , the first layer’s output dimension is v 1 , the second layer’s output dimension is v 2 . W 0 T D d * v 1 and W 1 T D v 1 * v 2 are parameter matrices in TD-GCN. For the activation function, σ ( ) , we employed the ReLU function. As shown in (3) and (4), the calculations of H 1 B U and H 2 B U in BU-GCN are the same as H 1 T D , H 2 T D .
H 1 B U = σ ( A B U X W 0 B U )
H 2 B U = σ ( A B U H 1 B U W 1 B U )
The pooling layer’s input was made up of H 2 T D and H 2 B U . The propagation structure feature and aggregation structures were read-out in the pooling layer using the average pooling approach, as illustrated in (5) and (6):
S T D = m e a n p o o l i n g ( H 2 T D )
S B U = m e a n p o o l i n g ( H 2 B U )
The operation of m e a n p o o l i n g in (5) and (6) is shown in Figure 4, where the feature matrix was H 2 T D or H 2 B U , a sliding window of size n × 1 was set (shown by the dotted line in Figure 4). The average of the elements in the window was calculated from the first column, and the window was moved backward in turn; finally, a pooling result of size 1 × v 2 was made, i.e., the propagation structure feature S T D and the aggregation structure feature S B U .

4.3. Attention Mechanism

The attention mechanism allows the model to focus on the key points, rather than treating them equally. Many sequence-based [33] or image-based [34] tasks have shown the effectiveness of attention mechanisms. In this paper, we faced the multimodal features of text content, propagation structure, and aggregation structure, and we integrated them by introducing an attention mechanism. In the learning procedure, we estimated the attention coefficients adaptively for various types of information. We have:
S = t t y p e s o f t m a x ( α t ) · S t
where t y p e { T E X T ,   T D , B U } is a set of the feature type; the elements in the set represent the text, propagation structure, and aggregation structure, respectively. Moreover, s o f t m a x ( α t ) = exp ( α t ) i exp ( α i ) , where α t is a free parameter needed to be estimated. The rumor’s final feature is S .

4.4. Rumor Classification and Training

Compared with other classifiers, MLP classifiers have better recognition rates and faster classification speeds. So, the MLP classifier and a SoftMax layer were used to calculate the event’s predicted label y ^ :
y ^ = S o f t m a x ( F C ( S ) )
y ^ is a four-dimensional vector, with each element’s value representing the likelihood that the event belongs to the corresponding class.
During model training, we minimized the cross-entropy between predictions and ground truth for all events. Additionally, to prevent the overfitting issue, we added the L 2 regular term throughout the training process. The definition of the loss function L is:
L = | C | i { 0 , 1 , 2 , 3 } y i l o g y i ^ + β L 2
where β is the coefficient of the regular term and | C | is the total number of events. The value y ^ i and y i are each element of y ^ and y , representing the probability that the event belonged to the corresponding class.

5. Experiment Results

In this section, we show the experimental results of our approaches on two real-world datasets.

5.1. Datasets

On two real-world datasets, Twitter15 and Twitter16 [35], we evaluated our suggested methodology. The datasets came from Twitter, one of the most popular social media platforms in the world. We created a propagation graph for each event in both datasets, using nodes to represent tweets and edges to represent their relationships with retweets or comments. The labels of each event in Twitter15 and Twitter16 were annotated according to the authenticity labels of articles on rumor-debunking websites (such as snopes.com, Emer-gent.info, and so on), whereas Ma et al. [35] refined the labels from binary classes. They have been expanded to quaternary classes: non-rumor (N), false rumor (F), true rumor (T), and unverified rumor (U). Table 1 shows the comprehensive statistics for the two datasets.

5.2. Experimental Settings

This paper chooses the following baseline models for the comparative experiments:
DTC [36]: A rumor detection approach that employs decision tree as a classifier and obtains credibility based on handcrafted features.
SVM-RBF [37]: A rumor detection method that uses post statistics as features and an SVM model with an RBF kernel for classification.
SVM-TS [38]: A linear SVM classifier used to generate time series models with hand-crafted features.
SVM-TK [35]: A rumor propagation structure-based SVM classifier with a propagation tree kernel.
RvNN [13]: A rumor detection approach that uses a tree-structured recurrent neural network with a GRU unit to learn rumor representation via the rumor propagation structure.
PPC_RNN+CNN [31]: A rumor detection model that uses rumor propagation paths to learn rumor representations by combining RNN and CNN.
The features used by each model are quite different; we describe them in detail in Table 2.
We separated the two datasets into five parts at random and performed a five-fold cross-validation. The parameters of the model suggested in this study were updated using stochastic gradient descent, and the model was optimized using the Adam method with a learning rate of 0.002. Each node in Bi-GCN had an output dimension of 196 dimensions, which was also the output dimensions of the event text features, propagation structure features, and aggregation structure features. The training process was repeated 200 times, and the training ended when the validation loss stops reduced after 10 iterations.

5.3. Confusion Matrix and Metrics

The experiment in this paper is a four-classification problem, and its confusion matrix is shown in Table 3.
Each value in Table 3 indicates the number of events in a category that are predicted to be a certain value. For example, A12 indicates the number of events with true labels as non-rumors (N), but were predicted to be false rumors (F).
We compare models using accuracy and F1 values of each class. The accuracy refers to the closeness of the predicted value to its “true value”, which is a commonly used and an effective evaluation indicator. The calculation formula is as follows:
Accuracy = A 11 + A 22 + A 33 + A 44 N 1 + N 2 + N 3 + N 4
The F1 value is an indicator used in statistics to measure the performance of a classification model. It takes into account both the precision and recall of the classification model.
In the binary classification, the formula for calculating the F1 value is:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 * ( P r e c i s i o n * R e c a l l ) P r e c i s i o n + R e c a l l = T P T P + F N + F P 2
where TP is true positive, FP is false positive, and FN is false negative. In the multi-classification problem in this paper, the formula used for calculating the F1 value of each category can be obtained by a simple derivation. The final F1 value calculation formula is as follows:
N F 1 = A 11 A 11 + ( M 1 A 11 ) + ( N 1 A 11 ) 2
F F 1 = A 22 A 22 + ( M 2 A 22 ) + ( N 2 A 22 ) 2
T F 1 = A 33 A 33 + ( M 3 A 33 ) + ( N 3 A 33 ) 2
U F 1 = A 44 A 44 + ( M 4 A 44 ) + ( N 4 A 44 ) 2

5.4. Comparative Experiment

All approaches were tested in comparative experiments on the Twitter15 and Twitter16 datasets, and the findings are given in Table 4 and Table 5. The experimental results demonstrate that the approach in this work is superior to other baseline methods, with accuracies of 83.6% and 85.1%, respectively, and the influence on the F1 value is also superior among these models. It shows that our method can effectively distinguish rumor and truth by aggregating multimodal features.
Furthermore, we can see that models SVM-TK and RvNN, which considered event propagation structural features, performed significantly better than other methods that ignored propagation structures, indicating that the propagation structures of rumor events are indeed different from those of ordinary events and are sufficient in the process of rumor detection. Considering the propagation structures of rumors helps to improve detection accuracy.
In addition to traditional text content features and propagation structure features, our method also extracts aggregation structure features. The experimental results show that adding aggregation structure features can detect rumors more accurately.

5.5. Ablation Study

We conducted an ablation study to see whether each module contributed to the model and which module contribute more. In other words, we demonstrate that text features, propagation structure features, and aggregation structure features can all benefit from rumor identification. Our proposed model’s main modules are BERT, TD-GCN, and BU-GCN, which stand for extract text features, propagation structure features, and aggregation structure features, respectively. Based on the whole model, we systematically removed the above modules and compared their changes in accuracy and the F1 value for each category. Figure 5 and Figure 6 show the experiment results.
The following conclusions can be drawn from the results of the ablation experiments shown in Figure 5 and Figure 6. After removing any module, all indicators decreased, demonstrating that each module in our model has a beneficial effect on rumor recognition results, suggesting that each feature is essential. The indicators declined the most after removing the TD-GCN module, showing that the propagation structure feature had the biggest impact on rumor detection.
The contribution ranking of the three modules in the model for the rumor detection results was: TD-GCN, BERT, and BU-GCN. That is to say, the contributions of different types of features to the results of rumor recognition are ranked from high to low as text feature, propagation structure feature, and aggregation structure feature.

5.6. Early Detection

We found that the early detection of rumors is significantly improved by fusing aggregated structures. Fake news is easily forwarded and spread by a large number of users on social media, causing serious impacts in a short period of time. Therefore, the earlier it is detected, the more the negative impacts can be avoided. However, in the early days of rumor spreading, the available spread of data is limited. In order to verify that the method in this paper improved the effect of the early detection of rumors, we used a small number of early rumor data, as shown in Figure 7, to compare the two methods: (1) The method using only the propagation structure (PS); (2) the method to fuse the propagation and aggregation structures (PS+AS). The results are shown in Table 6 and Table 7. The ratio column in the table indicates the ratio of the propagation structure used during training.
From the experimental results of early detection, it can be seen that the method considering the aggregated structure is significantly better than the method that only considers the propagation structure. Moreover, the lower the ratio, the more obvious the performance differences. This shows that considering the aggregation structure can effectively enhance rumor features and improve the distinction between real information and rumor features.

5.7. Classifier

In this paper, a fully-connected neural network with four layers was chosen as the classifier, and the model structure was 192 × 96 × 48 × 4. In this section, we evaluate the effects of different model topologies on the performance of rumor detection through experiments. Layers 2, 3, 5, and 6 had the following model structures, as illustrated in Figure 8: 192 × 4, 192 × 96 × 4, 192 × 96 × 48 × 24 × 4, 192 × 96 × 48 × 24 × 12 × 4. Each evaluation indicator was the most effective when the model structure had four layers. This is because when the model structure is too simple, the model under-fits, and when the model structure is too complex, the model slightly over-fits. All of them contributed to a decline in the model performance.

6. Conclusions

In this paper, we propose a new rumor detection method and compared it with existing methods; we fully considered the aggregation structures of rumors when extracting rumor features. Moreover, through the attention mechanism, the content features of rumor text, propagation structure, and aggregation structure were integrated. More discriminative rumor features were obtained by this method. Among them, we used the BERT pre-training model to extract text features, TD-GCN to extract propagation structural features, and BU-GCN to extract aggregation structural features. The comparative experimental results show that the proposed method can detect rumors more accurately than the existing methods. The ablation study results show that each module of our method has a positive effect on the detection results. The experiments of early detection show that the method considering the aggregation structure can deal with the early detection of rumors more effectively.
Rumor detection is currently a research hotspot in academia, and there are still many problems that need to be solved, for example, (1) how to detect false information in visual information (considering that social media tweets contain a large number of visual information); (2) how to accurately detect rumors in the early stages of a rumor (so as to cut off the spread of the rumor as soon as possible); (3) how to establish an effective mechanism so that the model can provide reasons for making judgments while giving the rumor detection results (so that the detection results are more convincing). In future work, we will conduct more in-depth research on the above issues.

Author Contributions

Methodology, W.D. and G.Z.; software, N.Z.; validation, N.Z. and J.Z.; writing—original draft preparation, N.Z.; writing—review and editing, N.Z. and G.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (no. U1509219), Basic Research Project of the Ministry of Public Security (No. 2019GABJC36), and the Youth Foundation of Zhejiang Police College (No. 2021QNY002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data included in this study are available upon request by contacting the first author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bastos, M.T.; Mercea, D. The Brexit botnet and user-generated hyperpartisan news. Soc. Sci. Comput. Rev. 2019, 37, 38–54. [Google Scholar] [CrossRef] [Green Version]
  2. Allcott, H.; Gentzkow, M. Social media and fake news in the 2016 election. J. Econ. Perspect. 2017, 31, 211–236. [Google Scholar] [CrossRef] [Green Version]
  3. Zarocostas, J. How to fight an infodemic. Lancet 2020, 395, 676. [Google Scholar] [CrossRef]
  4. Hassan, N.; Zhang, G.; Arslan, F.; Caraballo, J.; Jimenez, D.; Gawsane, S.; Hasan, S.; Joseph, M.; Kulkarni, A.; Tremayne, M.; et al. Claimbuster: The first-ever end-to-end fact-checking system. Proc. VLDB Endow. 2017, 10, 1945–1948. [Google Scholar] [CrossRef]
  5. Chen, Y.; Hu, L.; Sui, J. Text-based fusion neural network for rumor detection. In International Conference on Knowledge Science, Engineering and Management; Springer: Cham, Switzerland, 2019; pp. 105–109. [Google Scholar]
  6. Zeng, H.; Wang, R.; Huang, Y.; Cui, X.; Jiang, Q. Scientific Rumors Detection in Short Online Texts. In Proceedings of the 2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Melbourne, Australia, 17–20 October 2021; IEEE: New York, NY, USA, 2021; pp. 1233–1240. [Google Scholar]
  7. Xue, B.; Fu, C.; Shaobin, Z. A study on sentiment computing and classification of sina weibo with word2vec. In Proceedings of the 2014 IEEE International Congress on Big Data, Anchorage, AK, USA, 27 June–2 July 2014; IEEE: New York, NY, USA, 2014; pp. 358–363. [Google Scholar]
  8. Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  9. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  10. Vosoughi, S. Automatic detection and verification of rumors on Twitter. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2015. [Google Scholar]
  11. Vosoughi, S.; Roy, D.; Aral, S. The spread of true and false news online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef] [PubMed]
  12. Ma, J.; Gao, W.; Mitra, P.; Kwon, S.; Jansen, B.J.; Wong, K.F.; Cha, M. Detecting Rumors from Microblogs with Recurrent Neural Networks; AAAI Press: New York, NY, USA, 2016. [Google Scholar]
  13. Ma, J.; Gao, W.; Wong, K.F. Rumor Detection on Twitter with Tree-Structured Recursive Neural Networks; Association for Computational Linguistics: Melbourne, Australia, 2018. [Google Scholar]
  14. Yu, F.; Liu, Q.; Wu, S.; Wang, L.; Tan, T. A Convolutional Approach for Misinformation Identification. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI), Melbourne, Australia, 19–25 August 2017; pp. 3901–3907. [Google Scholar]
  15. Yu, F.; Liu, Q.; Wu, S.; Wang, L.; Tan, T. Attention-based convolutional approach for misinformation identification from massive and noisy microblog posts. Comput. Secur. 2019, 83, 106–121. [Google Scholar] [CrossRef]
  16. Li, S.; Xue, K.; Zhu, B.; Gao, X.; Wei, D.S.L.; Wan, T.; Wu, F. Falcon: A fourier transform based approach for fast and secure convolutional neural network predictions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8705–8714. [Google Scholar]
  17. Eisenschlos, J.M.; Gor, M.; Müller, T.; Cohen, W.W. MATE: Multi-view Attention for Table Transformer Efficiency. arXiv 2021, arXiv:2109.04312. [Google Scholar]
  18. Song, Y.; Sun, W. Pc-mlp: Model-based reinforcement learning with policy cover guided exploration. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 18–24 July 2021; pp. 9801–9811. [Google Scholar]
  19. Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral networks and locally connected networks on graphs. arXiv 2013, arXiv:1312.6203. [Google Scholar]
  20. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  21. Bian, T.; Xiao, X.; Xu, T.; Zhao, P.; Huang, W.; Rong, Y.; Huang, J. Rumor detection on social media with bi-directional graph convolutional networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 549–556. [Google Scholar]
  22. Nanjiang, Z.; Guomin, Z.; Weijie, D.; Jiawen, Z. A Rumor Detection Method Based on Multimodal Information Fusion. In Proceedings of the 2022 IEEE 5th International Conference on Electronics Technology (ICET), Chengdu, China, 13–16 May 2022; IEEE: New York, NY, USA, 2022; pp. 1032–1037. [Google Scholar]
  23. Vaibhav, V.; Annasamy, R.M.; Hovy, E. Do sentence interactions matter? Leveraging sentence level representations for fake news classification. arXiv 2019, arXiv:1910.12203. [Google Scholar]
  24. Cheng, M.; Nazarian, S.; Bogdan, P. Vroc: Variational autoencoder-aided multi-task rumor classifier based on text. In Proceedings of the Web Conference, Taipei, Taiwan, 20–24 April 2020; pp. 2892–2898. [Google Scholar]
  25. Jin, Z.; Cao, J.; Guo, H.; Zhang, Y.; Luo, J. Multimodal fusion with recurrent neural networks for rumor detection on microblogs. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 795–816. [Google Scholar]
  26. Wang, Y.; Ma, F.; Jin, Z.; Yuan, Y.; Xun, G.; Jha, K.; Su, L.; Gao, J. Eann: Event adversarial neural networks for multi-modal fake news detection. In Proceedings of the 24th Acm Sigkdd International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 849–857. [Google Scholar]
  27. Khattar, D.; Goud, J.S.; Gupta, M.; Varma, V. Mvae: Multimodal variational autoencoder for fake news detection. In Proceedings of the The World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2915–2921. [Google Scholar]
  28. Qi, P.; Cao, J.; Yang, T.; Guo, J.; Li, J. Exploiting multi-domain visual information for fake news detection. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; IEEE: New York, NY, USA, 2019; pp. 518–527. [Google Scholar]
  29. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  31. Liu, Y.; Wu, Y.F.B. Early detection of fake news on social media through propagation path classification with recurrent and convolutional networks. In Proceedings of the Thirty-second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  32. Song, C.; Shu, K.; Wu, B. Temporally evolving graph neural network for fake news detection. Inf. Processing Manag. 2021, 58, 102712. [Google Scholar] [CrossRef]
  33. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  34. Desimone, R.; Duncan, J. Neural mechanisms of selective visual attention. Annu. Rev. Neurosci. 1995, 18, 193–222. [Google Scholar] [CrossRef] [PubMed]
  35. Ma, J.; Gao, W.; Wong, K.F. Detect Rumors in Microblog Posts Using Propagation Structure Via Kernel Learning; Association for Computational Linguistics: Melbourne, Australia, 2017. [Google Scholar]
  36. Castillo, C.; Mendoza, M.; Poblete, B. Information credibility on twitter. In Proceedings of the 20th International Conference On World Wide Web, Hyderabad, India, 28 March–1 April 2011; pp. 675–684. [Google Scholar]
  37. Yang, F.; Liu, Y.; Yu, X.; Yang, M. Automatic detection of rumor on sina weibo. In Proceedings of the ACM SIGKDD Workshop on Mining Data Semantics, Beijing China, 12–16 August 2012; pp. 1–7. [Google Scholar]
  38. Ma, J.; Gao, W.; Wei, Z.; Lu, Y.; Wong, K.F. Detect rumors using time series of social context information on microblogging websites. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 18–23 October 2015; pp. 1751–1754. [Google Scholar]
Figure 1. The propagation and aggregation of rumors. The orange nodes represent tweets, and the root node is the source tweet. The left part shows the top-to-bottom propagation structures of rumors, and the right part shows the bottom-to-top aggregation structure of public opinions during the spread of rumors.
Figure 1. The propagation and aggregation of rumors. The orange nodes represent tweets, and the root node is the source tweet. The left part shows the top-to-bottom propagation structures of rumors, and the right part shows the bottom-to-top aggregation structure of public opinions during the spread of rumors.
Electronics 11 03200 g001
Figure 2. Our rumor detection model. We used the BERT pre-training model to extract text content features, extracted the rumor propagation structure and aggregation structure through Bi-GCN, and fused multimodal features through the attention mechanism; finally, the results were obtained by the fully-connected neural network classifier.
Figure 2. Our rumor detection model. We used the BERT pre-training model to extract text content features, extracted the rumor propagation structure and aggregation structure through Bi-GCN, and fused multimodal features through the attention mechanism; finally, the results were obtained by the fully-connected neural network classifier.
Electronics 11 03200 g002
Figure 3. BERT model structure.
Figure 3. BERT model structure.
Electronics 11 03200 g003
Figure 4. The operation of mean-pooling.
Figure 4. The operation of mean-pooling.
Electronics 11 03200 g004
Figure 5. Results of the ablation experiments on Twitter15.
Figure 5. Results of the ablation experiments on Twitter15.
Electronics 11 03200 g005
Figure 6. Results of the ablation experiments on Twitter16.
Figure 6. Results of the ablation experiments on Twitter16.
Electronics 11 03200 g006
Figure 7. Comparison of the early and complete propagation structures of rumors.
Figure 7. Comparison of the early and complete propagation structures of rumors.
Electronics 11 03200 g007
Figure 8. The effects of different classifier structures on the results.
Figure 8. The effects of different classifier structures on the results.
Electronics 11 03200 g008
Table 1. Statistics of the datasets.
Table 1. Statistics of the datasets.
StatisticTwitter15Twitter16
# of posts331,612204,820
# of events1490818
# of True rumors374205
# of False rumors370205
# of Unverified rumors374203
# of Non-rumors372205
Avg. # of posts/event223251
Max # of posts/event17682765
Min # of posts/event5581
Table 2. Comparison of the features of all methods.
Table 2. Comparison of the features of all methods.
MethodHandcrafted FeatureText ContentPropagation FeatureAggregation Feature
DTC---
SVM-RBF---
SVM-TS---
SVM-TK---
RvNN--
PPC_RNN+CNN--
Ours-
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Ground TruthPredicted Value
NFTUTotal
NA11A12A13A14N1
FA21A22A23A24N2
TA31A31A33A34N3
UA41A42A43A44N4
TotalM1M2M3M4-
Table 4. Rumor detection results on Twitter15.
Table 4. Rumor detection results on Twitter15.
Twitter15
MethodACC.NFTU
F1F1F1F1
DTC0.4520.4140.3650.7150.332
SVM-RBF0.3010.2140.1580.4810.239
SVM-TS0.5370.7810.4760.3950.462
SVM-TK0.7450.7940.6510.7500.742
RvNN0.7410.6910.7730.8170.630
PPC_RNN+CNN0.4850.3570.5130.3720.712
Ours0.8360.8400.8050.9030.831
Table 5. Rumor detection results on Twitter16.
Table 5. Rumor detection results on Twitter16.
Twitter16
MethodACC.NFTU
F1F1F1F1
DTC0.4910.3160.2090.4600.591
SVM-RBF0.6330.6800.1060.1840.381
SVM-TS0.6150.7670.4370.5600.544
SVM-TK0.7380.7180.7750.8560.725
RvNN0.7280.6340.6920.8200.731
PPC_RNN+CNN0.5310.5760.5360.3850.683
Ours0.8510.8100.8440.9170.856
Table 6. Early detection results of rumors on Twitter15.
Table 6. Early detection results of rumors on Twitter15.
RatioMethodACC.NFTU
F1F1F1F1
10%PS0.6160.6190.5980.6840.607
PS+AS0.6510.6640.6380.7390.658
30%PS0.6820.7050.6410.7690.706
PS+AS0.7150.7300.6850.7920.722
50%PS0.7510.7690.7230.8160.721
PS+AS0.7800.8020.7310.8580.774
60%PS0.7730.7820.7450.8600.765
PS+AS0.7910.8120.7670.8760.799
80%PS0.8040.8170.7530.8820.786
PS+AS0.8220.8380.7740.8910.815
Table 7. Early detection results of rumors on Twitter16.
Table 7. Early detection results of rumors on Twitter16.
RatioMethodACC.NFTU
F1F1F1F1
10%PS0.6200.6060.6140.6910.628
PS+AS0.6740.6510.6730.7470.663
30%PS0.7080.6620.6810.7630.674
PS+AS0.7530.7360.7280.8020.721
50%PS0.7710.7310.7450.8070.726
PS+AS0.8020.7430.7600.8370.758
60%PS0.7950.7430.7590.8580.766
PS+AS0.8200.7620.7960.8700.790
80%PS0.8370.7850.8200.8900.819
PS+AS0.8470.8060.8310.9100.831
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhong, N.; Zhou, G.; Ding, W.; Zhang, J. A Rumor Detection Method Based on Multimodal Feature Fusion by a Joining Aggregation Structure. Electronics 2022, 11, 3200. https://doi.org/10.3390/electronics11193200

AMA Style

Zhong N, Zhou G, Ding W, Zhang J. A Rumor Detection Method Based on Multimodal Feature Fusion by a Joining Aggregation Structure. Electronics. 2022; 11(19):3200. https://doi.org/10.3390/electronics11193200

Chicago/Turabian Style

Zhong, Nanjiang, Guomin Zhou, Weijie Ding, and Jiawen Zhang. 2022. "A Rumor Detection Method Based on Multimodal Feature Fusion by a Joining Aggregation Structure" Electronics 11, no. 19: 3200. https://doi.org/10.3390/electronics11193200

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop