Next Article in Journal
Startups and Consumer Purchase Behavior: Application of Support Vector Machine Algorithm
Previous Article in Journal
Service Oriented R-ANN Knowledge Model for Social Internet of Things
Previous Article in Special Issue
Estimating Causal Effects When the Treatment Affects All Subjects Simultaneously: An Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RoBERTaEns: Deep Bidirectional Encoder Ensemble Model for Fact Verification

by
Muchammad Naseer
,
Jauzak Hussaini Windiatmaja
,
Muhamad Asvial
and
Riri Fitri Sari
*
Department of Electrical Engineering, Faculty of Engineering, Universitas Indonesia, Depok 16424, Indonesia
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2022, 6(2), 33; https://doi.org/10.3390/bdcc6020033
Submission received: 31 December 2021 / Revised: 2 March 2022 / Accepted: 8 March 2022 / Published: 22 March 2022
(This article belongs to the Special Issue Big Data Analytics for Social Services)

Abstract

:
The application of the bidirectional encoder model to detect fake news has been widely applied because of its ability to provide factual verification with good results. Good fact verification requires the most optimal model and has the best evaluation to make news readers trust the reliable and accurate verification results. In this study, we evaluated the application of a homogeneous ensemble (HE) on RoBERTa to improve the accuracy of a model. We improve the HE method using a bagging ensemble from three types of RoBERTa models. Then, each prediction is combined to build a new model called RoBERTaEns. The FEVER dataset is used to train and test our model. The experimental results showed that the proposed method, RoBERTaEns, obtained a higher accuracy value with an F1-Score of 84.2% compared to the other RoBERTa models. In addition, RoBERTaEns has a smaller margin of error compared to the other models. Thus, it proves that the application of the HE functions increases the accuracy of a model and produces better values in handling various types of fact input in each fold.

Graphical Abstract

1. Introduction

Nowadays, fact verification models have been developed and implemented using machine learning [1,2,3,4,5]. These developments have improved the research aimed at detecting fake news in recent years [6,7,8,9,10]. Generally, a bidirectional encoder-based model has been developed to detect fake news to verify a fact. For example, the Bidirectional Encoder Representations from Transformers (BERT) [11] model has been successfully implemented to detect fake news [12,13] and fact verification [14,15]. Another model developed from BERT to detect fake news, called RoBERTa [16], has been successfully implemented [17,18]. RoBERTa has been implemented to verify a fact [19]. Some researchers found a new challenge in machine learning models to optimize and improve the performance of better methods, including applying ensemble models [20]. An ensemble is a classifier mechanism running concurrently that combines the predicted results of the components. Several cases of ensemble research carried out homogeneously have increased accuracy values in a model [21,22].
Ensemble deployments are becoming popular due to the advantages of using a single model across multiple datasets and machine learning tasks. In a simple case, by using an ensemble, a network is trained using a large dataset to check the result of the accuracy, loss, etc., in the validation set [23]. It is conducted to improve accuracy as a parameter to measure the success of the ensemble application for a model development [20]. Good fact verification requires the most optimal and has the best evaluation model. The goal is to make the news readers trust the reliable and accurate verification results.
One study [24] stated that RoBERTa has a higher accuracy value than the other models in verifying facts. Based on several previous studies, it has been proven that the homogeneous ensemble (HE) can improve model performance. This paper combined HE with the RoBERTa model to claim a fact. We evaluated the application of the HE on RoBERTa to prove that the application of an ensemble can be used to improve the accuracy of a model. To obtain optimal results, we modified the algorithm structure of RoBERTa in the training and testing stages. Modifications made in the training phase can be used to set three samples on RoBERTa to improve model performance. Improved model performance using three samples has been proven [25]. Afterwards, the evaluation results of each model were summed.
Model accuracy and F1-Score were used to measure the performance of the method. In the experimental process, we combined three models of RoBERTa. We found no research related to fact verification by combining ensembles with the model, and we were interested in contributing to this topic to improve model performance in fact verification. Therefore, the results of the fact verification were expected to have better verification.
In summary, our contributions in this paper are as follows:
  • Improving a RoBERTa model by implementing the HE from three best parameters settings in RoBERTa, we call it RoBERTaEns.
  • Implementing one of the HE methods, i.e., bagging ensemble to RoBERTa, the kind of bagging ensemble used was called Sum.
  • Testing and comparing the performance of BERT, RoBERTa, XL-Net, and XLM models with RoBERTaEns to obtain the best model performance.

2. Related Works

The impact of fake news that is widespread in society becomes a challenge in the world of journalism when one must label a fact from the results of claims of whether news or information is true or false based on evidence. Nowadays, fact verification has become a very challenging task in the realm of Natural Language Processing (NLP) [26,27], and the research community has paid much attention to research issues related to fact verification [28,29,30,31]. There are three types of labelling a fact [32]: (1) If there is an evidence supporting a claim, the facts are labelled with SUPPORTS; (2) If there is an evidence refuting a claim, the facts are labelled with REFUTES; (3) If there is no evidence to support or refute the claim, the facts are labelled with NOT ENOUGH INFO.
There have been many studies related to fact verification for making a model and dataset [31,33] to test various types of models [34] for fact verification [35]. Many studies have tested various datasets to verify facts. For example, a study comparing two datasets [36], i.e., the Fake News Challenge (FNC) dataset and the FEVER dataset using the Recognizing Textual Entailment (RTE) model. One of the results of the research has been reviewed by using the FEVER dataset on the train and evaluation domains, and the accuracy of the FEVER dataset was better than the FNC. The FEVER dataset had an accuracy of 83.43% on the test results. Other studies have also implemented the FEVER dataset for fact verification [15,34]. The FEVER dataset has the subtasks of document retrieval, sentence retrieval, and claim verification [37].
In addition to comparing datasets, a comparison of models to obtain the best accuracy in verifying facts has been carried out by several researchers [14] who compared three models, such as (1) BERT [11]; (2) Enhanced LSTM (ESIM) [38]; and (3) the Neural Semantic Matching Network (NSMN) [39]. The research results prove that BERT has the best results, with an accuracy of 61%, for fact verification. Furthermore, a different study [16] developed the BERT model by changing the BERT training structure, resulting in a new model called RoBERTa.
The results of the research were conducted by comparing three models, such as (1) BERT; (2) Generalized Autoregressive Pretraining for Language Understanding (XL-Net) [40]; and (3) RoBERTa. This study shows that the result of RoBERTa has a higher accuracy than BERT, with an accuracy of 92.5%. These results attracted researchers to develop the RoBERTa model in order to produce novel or better classifications, one of which is a study implementing Natural Language Inference on the RoBERTa model [41]. The model architecture that has been developed is given in Figure 1.
The application of HE in several studies proves an improvement in model performance [25,42]. HE has a model architecture built with multiple base learners and is generated from a single learning algorithm, in contrast to a heterogeneous ensemble that is built with different algorithms. Many researchers in their research more commonly use HE compared to the heterogeneous one [43]. For example, recently, several studies have proven HE models to improve model performance. The study applied a tree-based HE model with feature selection to predict diabetic retinopathy. The results showed that HE was established to enhance the performance of the Support Vector Machine (SVM) model [42]. Another study [25] also developed a model by applying the HE to improve the evaluation performance of the Decision Trees, Logistics Model Tree, and Bagged Multi-layer Perceptron algorithms. As a result, an effective HE gave satisfactory results, with an average accuracy of 98.1%. The existence of various evidence of improving model performance by applying the HE makes us interested in examining the application of the HE on the RoBERTa model. The bagging ensemble method used in this study was Sum [44].
Many researchers have carried out the implementation of RoBERTa by combining datasets with the aim of verifying facts. A study concerning fact verification by combining the FEVER dataset and RoBERTa as a model for fact verification has been carried out [45]. This combination obtained an accuracy of 95.1% by using the F1-Score calculation. F1-Score is the average harmonic value obtained from the precision model. The majority of F1-Scores are used to evaluate the machine learning [46,47,48] model, especially for NLP [49]. With the application of the HE, the model applied for fact verification is expected to improve the performance of the RoBERTa model. One of the interesting results of the related works that we studied refers to the RoBERTa model and has an updated RoBERTa model explicitly by applying the HE. Our HE application is called RoBERTaEns. The application of HE in the transformer family has been applied [50,51] to detect hate speech. Meanwhile, the HE applied in our research focused on fact verification.
RoBERTaEns is formed from three RoBERTa-based models with different epoch settings and batch sizes. RoBERTa-m1 with epoch 5 and batch size 32, RoBERTa-m2 with epoch 30 and batch size 20, and RoBERTa-m3 with epoch 50 and batch size 32. The best epoch and batch size settings are from the research conducted in [52,53,54]. We chose RoBERTa because it has a better accuracy results based on several previous studies. We need the highest accuracy of the claim results to verify the facts to impact the news readers’ trust in the news.

3. Methods

This section presents the proposed HE modification. As mentioned in Section 1, the main idea of applying the HE to RoBERTa is to improve the performance of the model. The more optimal the model, the more effective fact verification results for better fact verification. Algorithm modification in the RoBERTa model was applied at the training and prediction stages. At the training phase, we applied the HE to obtain the model predictions from the datasets. At the testing phase, we calculated the performance of the obtained model using accuracy, recall, precision, and F1-Score.
Furthermore, to calculate the significant value of the model, we used the 95% Confidence Interval (CI) [55]. CI indicates the probability that a parameter is below the mean value and measures the level of uncertainty or certainty in the sampling method. Cross-validation 10-fold times [56] were used for classification performance measurement in machine learning to obtain the CI values.

3.1. System Overview

RoBERTaEns reads input in sentences and encodes the sentences from the input. The sentences that have been encoded are matched with the data in the FEVER dataset. For example, “The chairwoman of Lockheed Martin and the President are currently the same person” which has the label “Supports” on the dataset. Each model with predefined parameters was combined with the HE using the bagging ensemble method. The bagging ensemble method that we used for classifying predictions was Sum. The bagging ensemble method is one of the HE methods that serves to create a better classifier by combining the respective predictions to form the final prediction. Figure 2 shows the process of applying the HE to RoBERTa to improve model performance.
Combined models created a higher learning rate for exploring parameters, and there is a potential for higher accuracy rates. A higher learning rate impacts the performance of the model in the sense that it takes less time due to the combination of the models [43]. It means that the application of the HE to the model potentially improves performance by requiring a shorter time than the time before HE was applied to the model. The results of the fact verification prediction are issued after the predicted value of the ensemble model is obtained. Each RoBERTa model was combined at the verification and prediction phases.
This section describes a method of improving our classification by applying the HE to RoBERTa. The first phase of the model read tokenized input sentences with a [CLS] token at the beginning and a [SEP] token at the end of the sentences. [CLS] stands for classification and represents the sentence-level classification of news samples, [CLS] performed sentence representation during pretraining tokens used to decide sentence order. Furthermore, [SEP] separated sentences and made it easier for the model to know that the input consisted of several sample news sentences. The implementation of RoBERTaEns is conducted by combining the results of each RoBERTa with different epoch and batch size settings. The merging process was carried out using the bagging ensemble method. Each model produced model accuracy and labeled the verification of facts obtained from the classification. News classification was obtained from the output value of the neurons generated by the model. There are two output neurons generated: refutes and supports. The output neuron with the highest value indicates that the classification is in the highest neuron value class. If the resulting neuron output is the same value, it suggests that the news being tested is in another class, i.e., Not Enough Info (NEI). After obtaining the results for each model, the bagging ensemble works by combining the prediction results in each model and forming a new fact verification label. Figure 3 describes the flow diagram of the RoBERTaEns algorithm.

3.2. Training

At this stage, the model architecture was combined, and the equation of RoBERTa was taken from a series of two segments (sequences of tokens), i.e., Question ( x ) and Answer ( y ) in BERT is x 1 , x N a , y 1 , y M a . Usually, each segment consists of more than one natural sentence. Both segments are presented sequentially from a single input to BERT with tokens: C L S ,   x 1 , x N ,   S E P   and   y 1 , y M , E O S . The RoBERTa model can be calculated as follows:
R o B E R T a s e t u p = M a + N a < T a
where Ta is the parameter controlling the maximum sequence length during training, Ma is RoBERTa model, and Na is a sequence of Ma. Ma and Na are restricted. Ma is the RoBERTa model, and Na is the sequence of Ma. Ma reads the input token that is read by the RoBERTa model. Then, Ma is set to the epoch value and batch size in the model to produce a different number of tokens. Ma and Na are limited by Ta, so that the total Ma + Na is not more than Ta, where Ta is the parameter controlling the maximum sequence length of the original RoBERTa during the training process.
Furthermore, the input value of feature X, label Y, model (Ma), epoch (E) were tokenized using RobertaTokenizerFast. In the next step, we initialized and built the model. After the model was successfully created, it was run with the X and Y settings on the epochs that have been set. Algorithm 1 shows the training process in RoBERTa.
Algorithm 1. Pseudocode RoBERTa
  • Input: Training feature X, training label Y, RoBERTa model Ma, epoch E
  • Output: Trained RoBERTa models
  • X ← tokenize X with RoBERTaTokenizerFast
  • Y ← tokenize Y with RoBERTaTokenizerFast
  • Initiate RoBERTa model Ma argument
  • Construct RoBERTa model Ma architecture
  • Compile RoBERTa model Ma
  • Train RoBERTa model Na with X and Y on E epoch
The combination of the model was carried out on the training model by implementing the HE into the RoBERTa model. We called it RoBERTaEns. The regulatory architecture in RoBERTaEns can be calculated as follows:
R o B E R T a E n s s e t u p = M e n s l + N e n s l < T e n s l
where Tensl is the total parameter controlling the maximum sequence length during training, Mensl is the total model, and Nensl is the total sequence of Mensl. Mensl and Nensl are restricted. Mensl is a model of RoBERTaEns, and Nensl is an order of Mensl. Mensl reads the total input tokens from each RoBERTa model and sets the epoch and batch size values for each model to produce a different number of tokens. Mensl and Nensl are limited by Tensl, so the total of Mensl + Nensl is not more than Tensl, where Tensl is the parameter controlling the maximum sequence length of RoBERTaEns during training.
RoBERTaEns read the total of that tokenized using RobertaTokenizerFast. Afterwards, we initialized the model and run the model with the X and Y settings on M-repeated epochs on the RoBERTaEns with the M = 3. The results of each model in RoBERTa were summed for being processed in the HE, which aimed to improve model performance. Algorithm 2 shows the training process in RoBERTaEns.
Algorithm 2. Pseudocode RoBERTaEns
  • Input: Training feature X, training label Y, RoBERTaEns models Mensl, epoch E
  • Output: Trained RoBERTa models
  • X ← tokenize X with RoBERTaEns TokenizerFast
  • Y ← tokenize Y with RoBERTaEns TokenizerFast
  • For i = 1 to M
  • Initiate RoBERTaEns model Mensl argument
  • Construct RoBERTaEns model Mensl architecture
  • Compile RoBERTaEns model Mensl
  • Train RoBERTaEns model Nensl with X and Y on E epoch
  • End

3.3. Parameters

We tested two parameters for setting the model: epoch and batch size. Epoch carried out the learning process on NLP and batch size iterated on each training process. As previously discussed, in RoBERTaEns the three models were applied, i.e., RoBERTa-m1, RoBERTa-m2, and RoBERTa-m3. For each model, the different epochs and batch sizes were set to prove the best setting for fact verification and chose the best setting results of epochs and batch sizes for each model tested in previous studies [52,53,54]. The parameter settings of each model are summarized in Table 1.

3.4. Prediction

At this phase, the parameters in the model prediction were modified to adapt and implement the ensemble model. The predictions in the RoBERTa model were determined by setting one parameter that had been executed, while RoBERTaEns combined the prediction results from each run model. The sequence architecture of the RoBERTa and RoBERTaEns algorithms are shown in Algorithms 3 and 4, respectively.
Algorithm 3. RoBERTa predictions
  • Input: Test feature X, test label Y, trained RoBERTa model Ma
  • Output: Result
  • X ← tokenize X with RobertaTokenizerFast
  • Y ← tokenize Y with RobertaTokenizerFast
  • results < -predict Na with X and Y
Algorithm 4. RoBERTaEns predictions
  • Input: Test feature X, test label Y, trained RoBERTa models Mensl
  • Output: Final_result
  • Declare Array results
  • Declare Array final_result
  • X ← tokenize X with RobertaTokenizerFast
  • Y ← tokenize Y with RobertaTokenizerFast
  • For i = 1 to length (Nensl)
  •  results ← append predict Nensl with X and Y
  • End
  • final_result ← sum of results
The proposed model was evaluated using accuracy, precision, recall, and F1-Score. F1-Score performs a weighted average comparison between precision and recall. True Positive (TP) is positive news data detected correctly, True Negative (TN) is negative news data detected correctly, False Positive (FP) is positive news data classified as false by the system, and False Negative (FN) is negative news data classified as false by the system. The accuracy can be calculated as follows:
Accuracy = TP + TN TP + FP + TN + FN
Precision is used to identify positive cases with high false-positive values, which can be calculated as follows:
Precision = TP TP +   FP
In contrast to precision, recall serves to identify positive cases with a high false-negative value that can be calculated as follows:
Recall = TP TP +   FN
The F1-Score provides a harmonic average of precision and recall, which can be measured as follows:
F 1 Score = 2 × Precision   ×   Recall Precision + Recall

3.5. Dataset

Our study used the FEVER dataset [31,57] because it yielded good accuracy for fact verification. In contrast to the other studies [40], which used the SQuaD, MNLI, SST, and RACE datasets to implement the RoBERTa model. The FEVER dataset is usually used for research based on fact verification [57,58]. The FEVER dataset has been proven to have the best performance and it is used in Claim Extraction and Fact Verification competitions [32,59]. The formation process for collecting the datasets has been carried out using a semi-automatic technique. Data labeling is done manually at the final stage of news labeling. The data labels of the claimed-results dataset are Supports, Refutes, and Not Enough Information. The data collected in the dataset are 10,000 data for the training phase and 10,000 data for the testing phase.

4. Result and Discussion

4.1. Model Performance Results

The results of RoBERTaEns proved that the application of the HE to RoBERTa improved model performance based on the results of accuracy and F1-Score that were obtained. RoBERTaEns combines RoBERTa-m1, RoBERTa-m2, and RoBERTa-m3 models to produce the HE. By testing several RoBERTa models and other models (BERT, XL-Net, XLM), the HE accuracy and F1-Score results are higher than the previously applied ensembles. In RoBERTa-m1, with an epoch setting of 5 and a batch size of 32, the best accuracy, precision, and F1-Score results were obtained by comparing RoBERTa-m2 and RoBERTa-m3. RoBERTa-m1 had an accuracy value 1.1% higher than RoBERTa-m2, a precision value 2.3% higher than RoBERTa-m2, and a F1-Score of 0.4%, which is higher than RoBERTa-m2 with an epoch setting of 30 and a batch size of 20. In addition, RoBERTa-m1 also produced the most of the measurements higher than RoBERTa-m3, with an accuracy value of 0.5% higher than RoBERTa-m3, a precision value of 1.2% higher than RoBERTa-m3, and F1-Score value of 0.2% higher than RoBERTa-m3. The results are summarized in Table 2.
In this test, RoBERTa-m3 produced a recall value of 0.9% higher than RoBERTa-m1. RoBERTa-m3 resulted a recall value of 0.1% higher than RoBERTa-m2. Overall, RoBERTa-m1 was more dominant in obtaining accuracy, precision, and F1-Score. However, the highest recall was in the RoBERTa-m1 setting with 5 epochs and 32 batch sizes. Combining all models using the HE on RoBERTa improved the performance of the model for fact verification, i.e., it achieved an accuracy of 78.4%, a precision of 82%, a recall of 86.5%, and an F1-Score of 84.2%. The experimental results proved that combining several models can improve accuracy.
Figure 4 visualizes the comparison of the best RoBERTa model with RoBERTaEns (after the RoBERTa model was applied with HE). In our experiment, the performance of the RoBERTaEns model has higher accuracy and F1-Score values than other models. It proved that the implementation of the HE has succeeded in improving the performance of the model. However, the accuracy and F1-Score of the RoBERTaEns model is higher than RoBERTa-m1, RoBERTa-m2, and RoBERTa-m3.
In addition, we compared it with the other models, i.e., BERT, XL-Net, and XLM models. Each epoch and batch size setting in the model followed the setting with the best results in the comparison process of RoBERTa-m1, RoBERTa-m2, and RoBERTa-m3. The best epoch and batch size settings are 5 and 32, respectively. As a result, using the same dataset as RoBERTaEns, i.e., the FEVER dataset, RoBERTaEns has higher accuracy and better F1-Score than BERT, XL-Net, and XLM. The results of our comparison models are summarized in Table 3.
RoBERTaEns indicates that the correct positive prediction results are slightly lower than the comparison algorithm. However, the accuracy and F1-Score results are still higher than in BERT, XL-Net, and XLM. These results proved that RoBERTaEns has a measurement of the quantity that is closer to the actual value and its robustness is better than the three other models. Additional evidence of the HE’s application on the model has been shown to improve the performance of the model we developed, i.e., RoBERTaEns. By applying bagging ensemble, RoBERTaEns improved its performance to produce a better classification. Figure 5 shows comparison results of the performance metrics.

4.2. Fact Verification Result

We have proven that RoBERTaEns produces a good model based on the model test performance and that it compares the results with the other models. We documented the results of the fact verification to find out whether it is claimed to be “Supports”, “Refutes”, or “NEI”. The output neuron value generated by RoBERTaEns determined the label owned by the news to be claimed. RoBERTaEns required evidence sourced from datasets. Figure 6 shows an example in our experiment that required scattered evidence from multi-structured textual sources to conclude. Considering that the RoBERTaEns performance produces the best performance compared to the other models, the fact verification results generated by the RoBERTaEns model have a greater chance of producing a more precise fact verification.
We compared the performance of RoBERTa with RoBERTaEns after each datum in our dataset was grouped according to the following categories: Disease, Health, Political, Government, and Industry. The results of the performance comparison experiment are shown in Table 4. The F1-Score values for the “Government” and “Disease” categories in RoBERTaEns obtained the highest scores, with 98.90% and 92.27% respectively. However, the F1-Score value in the “Health” category had the lowest accuracy, indicating that this category is difficult to identify in the FEVER dataset. Various reasons for the low value obtained, among others, the amount of data in this category has a small sample which can cause errors of fact verification. Overall, we observed that RoBERTaEns has a higher accuracy and F1-Score values than RoBERTa-m1, except for the “Industry” category. In this category, the F1-Score obtained by RoBERTaEns has a lower accuracy of 2.28% and a lower accuracy of 1.75% compared to RoBERTa-m1. We used RoBERTa-m1 to compare RoBERTaEns because RoBERTa-m1 is better than RoBERTa-m2 and RoBERTa-m3. The complete results of accuracy (A), precision (P), recall (R), and F1-Score (F) for each topic category in the RoBERTa-m1 and RoBERTaEns models are shown in Table 4.

4.3. Error Analysis

The 95% confidence level is based on the CI shown in Table 5 with n (sample data) of 10,000 using 10-fold cross-validations. All the RoBERTaEns results show the lowest CI value compared to the other models. It validated our hypothesis that RoBERTaEns consistently outperformed the other models. These results also indicate that RoBERTaEns is more robust in handling various types of fact input in each fold.

5. Conclusions

Our research focuses on developing a new model based on three types of RoBERTa models for fact verification. We applied the Homogeneous Ensemble (HE) or a system based on a single classification approach for fact verification on the RoBERTa models with a FEVER dataset of 10,000 data. Our method is called RoBERTaEns. In this study, we found that the application of the HE on RoBERTa resulted in better accuracy and a higher F1-Score than RoBERTa-m1, RoBERTa-m2, RoBERTa-m3, BERT, XL-NET, and XLM. The experimental results showed that RoBERTaEns improves model accuracy by 1.1% compared to RoBERTa-m1, 2% compared to RoBERTa-m2, and 1.6% compared to RoBERTa-m3. The F1-Score results on RoBERTaEns also showed an increased percentage of 1.1% compared to RoBERTa-m1, 1.5% compared to RoBERTa-m2, and 1.3% compared to RoBERTa-m3. Regarding the stability of prediction accuracy, the ensemble application has a better fact verification than the other models. In addition, RoBERTaEns has a smaller margin of error compared to the other models, and it proves that RoBERTaEns produces more robust values in dealing with different types of fact input in each fold.
In future research, we suggest applying another ensemble method to the RoBERTa model, i.e., the heterogeneous ensemble, and comparing it with the HE. This exploration could be carried out on different datasets in future studies.

Author Contributions

Conceptualization, M.N., J.H.W., M.A. and R.F.S.; methodology, M.N., M.A. and R.F.S.; software, M.N. and J.H.W.; validation, M.N., J.H.W., M.A. and R.F.S.; formal analysis, M.N., J.H.W., M.A. and R.F.S.; investigation, M.N., M.A. and R.F.S.; resources, M.N., J.H.W., M.A. and R.F.S.; data curation, M.N., J.H.W., M.A. and R.F.S.; writing—original draft preparation, M.N.; writing—review and editing, M.N., M.A. and R.F.S.; visualization, M.N.; supervision, M.A. and R.F.S.; project administration, M.N. and R.F.S.; funding acquisition, R.F.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research and the APC were funded by Universitas Indonesia under PUTI Doktor 2020, Grant number NKB-664/UN2.RST/HKP.05.00/2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is publicly available at https://fever.ai, accessed on 1 November 2021.

Acknowledgments

We thank the Universitas Indonesia for financial support for this research. The authors would like to express their deep gratitude to the reviewers for their valuable suggestions and important comments that have greatly helped to improve the presentation of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ramadhani, R.; Suswanta, S.; Shaddiq, S. E-Marketing of village tourism development strategy (Case study in the tourist village puncak sosok). J. Robot. Control 2021, 2, 72–77. [Google Scholar] [CrossRef]
  2. Humaira, S.P.; Nursuprianah, I.; Darwan, D. Forecasting of the Number of Schizophrenia Disorder by using the Box-Jenkins of Time Series Analysis. J. Robot. Control (JRC) 2020, 1, 213–219. [Google Scholar] [CrossRef]
  3. Ikhsanudin, M.; Irawan, Y. Application for Selection of Student Final Project Supervisors Based on the Selected Category and Expertise of Lecturers Using the Naive Bayes Classifier Method. J. Robot. Control (JRC) 2021, 2, 316–321. [Google Scholar] [CrossRef]
  4. Oyewola, D.O.; Augustine, A.F.; Dada, E.G.; Ibrahim, A. Predicting impact of COVID-19 on crude oil price image with directed acyclic graph deep convolutional neural network. J. Robot. Control 2021, 2, 103–109. [Google Scholar]
  5. Purwono, P.; Ma’arif, A.; Negara, I.S.M.; Rahmaniar, W.; Rahmaniar, J.; Rahmaniar, W. A linkage detection of features that cause stroke using feyn qlattice machine learning model. J. Ilm. Tek. Elektro Komput. Inform. 2021, 7, 423–432. [Google Scholar] [CrossRef]
  6. Damayantia, D.; Kartinia, S.D.A.; Artanti, A.R. Empowerment in business development through marketing expansion and business legalityin batik Anugrah Art Rembang MSMEs. J. Pengabdi. Pemberdaya. Masy. Indones. 2022, 2, 26–33. [Google Scholar]
  7. Nurdina, A.; Khasanah, S.I.; Istiyantoa, J.; Anggrainia, S.R.; Patonah, L. Optimization of digital marketing as an effort to expand the reach of promotion of MSME products assist by village enterprises Ngerangan Jaya. J. Pengabdi. Pemberdaya. Masy. Indones. 2021, 1, 479–484. [Google Scholar]
  8. Salamah, U.; Wiharto; Saptono, R. Desimination of technology for increasing the security of community-based citizens of Gawanan village, Colomadu using i-siskamling. J. Pengabdi. Pemberdaya. Masy. Indones. 2021, 1, 405–412. [Google Scholar]
  9. Faidati, N.; Fitriahadi, E. Desimination of technology for utilizing online learning media during the Covid-19 pandemic at SD Muhammadiyah Mlangi. J. Pengabdi. Pemberdaya. Masy. Indones. 2021, 1, 303–310. [Google Scholar]
  10. Mutmainah, N.F.; Mahfida, S.L. Utilization of online learning media during the Covid-19 pandemic in teaching and learning activities by teachers at MTs Muhammadiyah Karangkajen. J. Pengabdi. Pemberdaya. Masy. Indones. 2021, 1, 103–110. [Google Scholar]
  11. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  12. Zellers, R.; Holtzman, A.; Rashkin, H.; Bisk, Y.; Farhadi, A.; Roesner, F.; Choi, Y. Defending against neural fake news. arXiv 2019, arXiv:1905.12616. [Google Scholar]
  13. Jwa, H.; Oh, D.; Park, K.; Kang, J.M.; Lim, H. exBAKE: Automatic Fake News Detection Model Based on Bidirectional Encoder Representations from Transformers (BERT). Appl. Sci. 2019, 9, 4062. [Google Scholar] [CrossRef] [Green Version]
  14. Schuster, T.; Shah, D.; Yeo, Y.J.S.; Ortiz, D.R.F.; Santus, E.; Barzilay, R. Towards Debiasing Fact Verification Models. arXiv 2019, arXiv:1908.05267. [Google Scholar]
  15. Soleimani, A.; Monz, C.; Worring, M. BERT for Evidence Retrieval and Claim Verification. In Advances in Information Retrieval; Springer: Cham, Switzerland, 2020; Volume 12036, pp. 359–366. [Google Scholar] [CrossRef] [Green Version]
  16. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A robustly optimized BERT pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  17. Umer, M.; Imtiaz, Z.; Ullah, S.; Mehmood, A.; Choi, G.S.; On, B.-W. Fake News Stance Detection Using Deep Learning Architecture (CNN-LSTM). IEEE Access 2020, 8, 156695–156706. [Google Scholar] [CrossRef]
  18. Slovikovskaya, V.; Attardi, G. Transfer learning from transformers to fake news challenge stance detection (FNC-1) task. In Proceedings of the Language Resources and Evaluation Conference, Marseille, France, 13–15 May 2020; European Language Resources Association: Luxembourg, Luxembourg, 2020; Volume 1, pp. 1211–1218. [Google Scholar]
  19. Barrón-Cedeño, A.; Elsayed, T.; Nakov, P.; Martino, G.D.S.; Hasanain, M.; Suwaileh, R.; Haouari, F.; Babulkov, N.; Hamdan, B.; Nikolov, A.; et al. Overview of CheckThat! 2020: Automatic Identification and Verification of Claims in Social Media. In Proceedings of the International Conference of the Cross-Language Evaluation Forum for European Languages, Thessaloniki, Greece, 22–25 September 2020; Springer: Cham, Switzerland, 2020; Volume 12260 LNCS, pp. 215–236. [Google Scholar]
  20. Chen, H.; Lundberg, S.; Lee, S.-I. Checkpoint ensembles: Ensemble methods from a single training process. arXiv 2017, arXiv:1710.03282. [Google Scholar]
  21. Gan, K.S.; Chin, K.O.; Anthony, P.; Chang, S.V. Homogeneous ensemble feedforward neural network in CIMB stock price forecasting. In Proceedings of the 2018 IEEE International Conference on Artificial Intelligence in Engineering and Technology (IICAIET), Kota Kinabalu, Malaysia, 8 November 2018; IEEE: Piscataway, NJ, USA, 2019; pp. 111–116. [Google Scholar]
  22. Douibi, K.; Settouti, N.; Chikh, M.E.A. The Homogeneous Ensemble Methods for MLknn Algorithm. In Proceedings of the International Conference on Compute and Data Analysis, Lakeland, FL, USA, 19–23 May 2017; Association for Computing Machinery: New York, NY, USA, 2017; Volume Part F1302, pp. 197–201. [Google Scholar]
  23. Muguro, J.K.; Laksono, P.W.; Rahmaniar, W.; Njeri, W.; Sasatake, Y.; Suhaimi, M.S.A.b.; Matsushita, K.; Sasaki, M.; Sulowicz, M.; Caesarendra, W. Development of Surface EMG Game Control Interface for Persons with Upper Limb Functional Impairments. Signals 2021, 2, 48. [Google Scholar] [CrossRef]
  24. Khan, J.Y.; Khondaker, T.I.; Afroz, S.; Uddin, G.; Iqbal, A. A benchmark study of machine learning models for online fake news detection. Mach. Learn. Appl. 2021, 4, 100032. [Google Scholar] [CrossRef]
  25. Handhika, T.; Fahrurozi, A.; Sari, I.; Lestari, D.P.; Zen, R.I. Hybrid Method for Sentiment Analysis Using Homogeneous Ensemble Classifier. In Proceedings of the 2019 2nd International Conference of Computer and Informatics Engineering (IC2IE), Banyuwangi, Indonesia, 10–11 September 2019; pp. 232–236. [Google Scholar]
  26. Thorne, J.; Vlachos, A. Evidence-based factual error correction. arXiv 2020, arXiv:2012.15788. [Google Scholar]
  27. Vijjali, R.; Potluri, P.; Kumar, S.; Teki, S. Two stage transformer model for COVID-19 fake news detection and fact checking. arXiv 2020, arXiv:2011.13253. [Google Scholar]
  28. Liu, Z.; Xiong, C.; Dai, Z.; Sun, S.; Sun, M.; Liu, Z. Adapting Open Domain Fact Extraction and Verification to COVID-FACT through In-Domain Language Modeling. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, Online, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 2395–2400. [Google Scholar] [CrossRef]
  29. Wang, Y.; Xia, C.; Si, C.; Yao, B.; Wang, T. Robust Reasoning Over Heterogeneous Textual Information for Fact Verification. IEEE Access 2020, 8, 157140–157150. [Google Scholar] [CrossRef]
  30. Liu, Z.; Xiong, C.; Sun, M. Fine-grained fact verification with kernel graph attention network. arXiv 2020, arXiv:1910.09796. [Google Scholar]
  31. Wang, Y.; Xia, C.; Si, C.; Zhang, C.; Wang, T. The Graph Reasoning Approach Based on the Dynamic Knowledge Auxiliary for Complex Fact Verification. Electronics 2020, 9, 1472. [Google Scholar] [CrossRef]
  32. Thorne, J.; Vlachos, A.; Christodoulopoulos, C.; Mittal, A. FEVER: A Large-scale Dataset for Fact Extraction and VERification. arXiv 2018, arXiv:1803.05355. [Google Scholar]
  33. Chen, W.; Wang, H.; Chen, J.; Zhang, Y.; Wang, H.; Li, S.; Zhou, X.; Wang, W.Y. Tabfact: A large-scale dataset for table-based fact verification. arXiv 2019, arXiv:1909.02164. [Google Scholar]
  34. Reddy, A.J.; Rocha, G.; Esteves, D. Defactonlp: Fact verification using entity recognition, tfidf vector comparison and decomposable attention. arXiv 2018, arXiv:1809.00509. [Google Scholar]
  35. Bekoulis, G.; Papagiannopoulou, C.; Deligiannis, N. A Review on Fact Extraction and Verification. ACM Comput. Surv. 2021, 55, 1–35. [Google Scholar] [CrossRef]
  36. Suntwal, S.; Paul, M.; Sharp, R.; Surdeanu, M. On the Importance of Delexicalization for Fact Verification. arXiv 2019, arXiv:1909.09868. [Google Scholar] [CrossRef]
  37. Bekoulis, G.; Papagiannopoulou, C.; Deligiannis, N. Understanding the Impact of Evidence-Aware Sentence Selection for Fact Checking. In Fourth Workshop on NLP for Internet Freedom: Censorship, Disinformation, and Propaganda; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 23–28. [Google Scholar]
  38. Chen, Q.; Zhu, X.; Ling, Z.-H.; Wei, S.; Jiang, H.; Inkpen, D. Enhanced LSTM for natural language inference. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, Canada, 30 July–4 August 2017; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; Volume 141, pp. 1657–1668. [Google Scholar]
  39. Nie, Y.; Chen, H.; Bansal, M. Combining Fact Extraction and Verification with Neural Semantic Matching Networks. Proc. Conf. AAAI Artif. Intell. 2019, 33, 6859–6866. [Google Scholar] [CrossRef] [Green Version]
  40. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.; Le, Q.V. XLNet: Generalized autoregressive pretraining for language understanding. arXiv 2019, arXiv:1906.08237. [Google Scholar]
  41. Saha, S.; Nie, Y.; Bansal, M. ConjNLI: Natural Language Inference Over Conjunctive Sentences. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Virtual, 16–20 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 8240–8252. [Google Scholar]
  42. Dagogo-George, T.E.; Mojeed, H.A.; Balogun, A.O.; Mabayoje, M.A.; Salihu, S.A. Tree-based homogeneous ensemble model with feature selection for diabetic retinopathy prediction. J. Teknol. dan Sist. Komput. 2020, 8, 297–303. [Google Scholar] [CrossRef]
  43. Bian, S.; Wang, W. On diversity and accuracy of homogeneous and heterogeneous ensembles. Int. J. Hybrid Intell. Syst. 2007, 4, 103–128. [Google Scholar] [CrossRef]
  44. Nannia, L.; Ghidoni, S.; Brahnam, S. Ensemble of convolutional neural networks for bioimage classification. Appl. Comput. Informatics 2020, 17, 19–35. [Google Scholar] [CrossRef]
  45. Pan, L.; Chen, W.; Xiong, W.; Kan, M.-Y.; Wang, W.Y. Zero-shot Fact Verification by Claim Generation. arXiv 2021, arXiv:2105.14682. [Google Scholar]
  46. Rahmaniar, W.; Hernawan, A. Real-time human detection using deep learning on embedded platforms: A review. J. Robot. Control 2021, 2, 462–468. [Google Scholar]
  47. Rahmaniar, W.; Wang, W.; Chiu, C.-W.E.; Hakim, N.L.L. Real-time bi-directional people counting using an RGB-D camera. Sens. Rev. 2021, 41, 341–349. [Google Scholar] [CrossRef]
  48. Banadaki, Y.; Razaviarab, N.; Fekrmandi, H.; Sharifi, S. Toward enabling a reliable quality monitoring system for additive manufacturing process using deep convolutional neural networks. arXiv 2020, arXiv:2003.08749. [Google Scholar]
  49. Yang, S.M.; Yoo, S.Y.; Jeong, O.R. DeNERT-KG: Named entity and relation extraction model using DQN, knowledge graph, and BERT. Appl. Sci. 2020, 10, 6429. [Google Scholar] [CrossRef]
  50. Das, K.A.; Baruah, A.; Barbhuiya, F.A.; Dey, K. KAFK at SemEval-2020 Task 12: Checkpoint Ensemble of Transformers for Hate Speech Classification. In Proceedings of the Fourteenth Workshop on Semantic Evaluation, Barcelona, Spain (Online), 12–13 December 2020; International Committee for Computational Linguistics: New York, NY, USA,, 2020; pp. 2023–2029. [Google Scholar]
  51. Li, J.; Zhou, X.; Zhang, Z. Lee at SemEval-2020 Task 12: A BERT Model Based on the Maximum Self-ensemble Strategy for Identifying Offensive Language. In Proceedings of the Fourteenth Workshop on Semantic Evaluation, Barcelona, Spain (Online), 12–13 December 2020; International Committee for Computational Linguistics: New York, NY, USA, 2020; pp. 2067–2072. [Google Scholar] [CrossRef]
  52. Liu, X.; He, P.; Chen, W.; Gao, J. Multi-Task Deep Neural Networks for Natural Language Understanding. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, Florence, Italy, 28 July–2 August 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 4487–4496. [Google Scholar]
  53. Narin, A.; Pamuk, Z. Effect of Different Batch Size Parameters on Predicting of COVID19 Cases. arXiv 2020, arXiv:2012.05534. [Google Scholar] [CrossRef]
  54. Kandel, I.; Castelli, M. The effect of batch size on the generalizability of the convolutional neural networks on a histopathology dataset. ICT Express 2020, 6, 312–315. [Google Scholar] [CrossRef]
  55. Clemente, F.M.; Ramirez-Campillo, R.; Afonso, J.; Sarmento, H. Effects of Small-Sided Games vs. Running-Based High-Intensity Interval Training on Physical Performance in Soccer Players: A Meta-Analytical Comparison. Front. Physiol. 2021, 12, 642703. [Google Scholar] [CrossRef] [PubMed]
  56. Browarczyk, J.; Kurowski, A.; Kostek, B. Analyzing the Effectiveness of the Brain–Computer Interface for Task Discerning Based on Machine Learning. Sensors 2020, 8, 2403. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Li, Q.; Zhou, W. Connecting the dots between fact verification and fake news detection. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain (Online), 12–13 December 2020; International Committee for Computational Linguistics: New York, NY, USA, 2020; pp. 1820–1825. [Google Scholar]
  58. Aly, R.; Guo, Z.; Schlichtkrull, M.; Thorne, J.; Vlachos, A.; Christodoulopoulos, C.; Cocarascu, O.; Mittal, A. Feverous: Fact extraction and verification over unstructured and structured information. arXiv 2021, arXiv:2106.05707. [Google Scholar]
  59. Thorne, J.; Vlachos, A.; Cocarascu, O.; Christodoulopoulos, C.; Mittal, A. The Fact Extraction and VERification (FEVER) Shared Task. arXiv 2018, arXiv:1811.10971. [Google Scholar]
Figure 1. Architecture diagram of predicate-aware RoBERTa model.
Figure 1. Architecture diagram of predicate-aware RoBERTa model.
Bdcc 06 00033 g001
Figure 2. Block diagram of RoBERTaEns model.
Figure 2. Block diagram of RoBERTaEns model.
Bdcc 06 00033 g002
Figure 3. RoBERTaEns Flowchart.
Figure 3. RoBERTaEns Flowchart.
Bdcc 06 00033 g003
Figure 4. Performance of RoBERTaEns and various RoBERTa models.
Figure 4. Performance of RoBERTaEns and various RoBERTa models.
Bdcc 06 00033 g004
Figure 5. Performance of RoBERTaEns and other Transformer models.
Figure 5. Performance of RoBERTaEns and other Transformer models.
Bdcc 06 00033 g005
Figure 6. Fact verification result.
Figure 6. Fact verification result.
Bdcc 06 00033 g006
Table 1. RoBERTa model parameter settings for ensemble.
Table 1. RoBERTa model parameter settings for ensemble.
ModelEpochBatch Size
RoBERTa-m1 [52]532
RoBERTa-m2 [53]3020
RoBERTa-m3 [54]5032
Table 2. Performance of various RoBERTa model.
Table 2. Performance of various RoBERTa model.
ModelEpochBatch SizeAccPrecisionRecallF1-Score
RoBERTa-m153277.3%82.3%84.0%83.1%
RoBERTa-m2302076.4%80.0%84.8%82.7%
RoBERTa-m3503276.8%81.1%84.9%82.9%
RoBERTaEns78.4%82.0%86.5%84.2%
Table 3. RoBERTaEns model accuracy results compared to BERT, XL-NET, and XLM.
Table 3. RoBERTaEns model accuracy results compared to BERT, XL-NET, and XLM.
ModelAccuracyPrecisionRecallF1-Score
RoBERTaEns78.4%82.0%86.5%84.2%
BERT77.5%80.4%87.5%83.8%
XL-Net77.2%79.3%89.0%83.8%
XLM72.5%71.9%96.1%82.3%
Table 4. Comparison of RoBERTa-m1 and RoBERTaEns by the News Topic Category.
Table 4. Comparison of RoBERTa-m1 and RoBERTaEns by the News Topic Category.
Topic CategoryParametersRoBERTa-m1RoBERTaEns
DiseaseA79.05%92.17%
P78.48%93.27%
R84.09%92.27%
F81.19%92.77%
HealthA72.46%80.66%
P74.26%81.29%
R80.92%80.52%
F79.51%80.92%
PoliticalA87.20%84.00%
P88.98%82.60%
R92.90%90.71%
F90.26%88.90%
GovernmentA82.25%90.66%
P88.90%91.86%
R84.65%95.43%
F86.71%98.90%
IndustryA74.35%72.60%
P78.27%77.97%
R80.54%78.53%
F79.18%76.90%
Table 5. Margin of Error comparison using 95% CI.
Table 5. Margin of Error comparison using 95% CI.
RoBERTaEnsR-m1R-m2R-m3BERTXL-NetXLM
A0.01%0.02%0.01%0.01%0.02%0.06%0.10%
P0.00%0.04%0.01%0.02%0.05%0.12%0.02%
R0.01%0.10%0.01%0.01%0.14%0.04%0.09%
F0.00%0.02%0.01%0.00%0.05%0.13%0.05%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Naseer, M.; Windiatmaja, J.H.; Asvial, M.; Sari, R.F. RoBERTaEns: Deep Bidirectional Encoder Ensemble Model for Fact Verification. Big Data Cogn. Comput. 2022, 6, 33. https://doi.org/10.3390/bdcc6020033

AMA Style

Naseer M, Windiatmaja JH, Asvial M, Sari RF. RoBERTaEns: Deep Bidirectional Encoder Ensemble Model for Fact Verification. Big Data and Cognitive Computing. 2022; 6(2):33. https://doi.org/10.3390/bdcc6020033

Chicago/Turabian Style

Naseer, Muchammad, Jauzak Hussaini Windiatmaja, Muhamad Asvial, and Riri Fitri Sari. 2022. "RoBERTaEns: Deep Bidirectional Encoder Ensemble Model for Fact Verification" Big Data and Cognitive Computing 6, no. 2: 33. https://doi.org/10.3390/bdcc6020033

Article Metrics

Back to TopTop