Next Article in Journal
Optimal Thresholding for Multi-Window Computed Tomography (CT) to Predict Lung Cancer
Previous Article in Journal
A Transformer-Based Approach to Authorship Attribution in Classical Arabic Texts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Machine Learning Method with Hybrid Feature Selection for Improved Credit Card Fraud Detection

by
Ibomoiye Domor Mienye
*,† and
Yanxia Sun
Department of Electrical and Electronic Engineering Science, University of Johannesburg, Johannesburg 2006, South Africa
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(12), 7254; https://doi.org/10.3390/app13127254
Submission received: 18 March 2023 / Revised: 9 June 2023 / Accepted: 13 June 2023 / Published: 18 June 2023

Abstract

:
With the rapid developments in electronic commerce and digital payment technologies, credit card transactions have increased significantly. Machine learning (ML) has been vital in analyzing customer data to detect and prevent fraud. However, the presence of redundant and irrelevant features in most real-world credit card data degrades the performance of ML classifiers. This study proposes a hybrid feature-selection technique consisting of filter and wrapper feature-selection steps to ensure that only the most relevant features are used for machine learning. The proposed method uses the information gain (IG) technique to rank the features, and the top-ranked features are fed to a genetic algorithm (GA) wrapper, which uses the extreme learning machine (ELM) as the learning algorithm. Meanwhile, the proposed GA wrapper is optimized for imbalanced classification using the geometric mean (G-mean) as the fitness function instead of the conventional accuracy metric. The proposed approach achieved a sensitivity and specificity of 0.997 and 0.994, respectively, outperforming other baseline techniques and methods in the recent literature.

1. Introduction

Over the years, electronic payments (e-payments) have been the most common payment option due to technological advancements and the development of several electronic funding methods [1]. E-payment systems are essential to the present competitive financial sector and are mostly performed using credit cards [2]. The introduction of credit cards has resulted in convenient and seamless e-payments. A recent study stated that in the second quarter of 2021, Mastercard and Visa issued 1131 million and 1156 million cards, respectively [3]. However, the rise of credit card usage globally has increased the fraud rate, affecting consumers and merchants [4]. For instance, a report stated that financial losses due to credit and debit cards are among the leading causes of losses in the financial sector [3]. Therefore, developing efficient credit card fraud-detection systems is necessary to reduce such losses.
Machine learning algorithms have been widely employed to detect credit card fraud [5,6,7]. Meanwhile, there have been enormous datasets with very high dimensions due to the advent of big data and the Internet of Things (IoT) [8,9]. Furthermore, some features in these datasets might be redundant or less significant to the response variable. Using such features for machine learning could increase the complexity of the model and lead to overfitting [10]. Therefore, to handle the high dimensionality issue, an approach containing dimensionality reduction, such as feature selection, is necessary to obtain valuable insights and make accurate predictions [11].
Feature-selection techniques aim to identify the most important attributes needed to develop a well-performing machine learning model [12,13], ensuring improved classification performance and reduced computational complexity by removing irrelevant and redundant features. Feature selection techniques are usually grouped into three methodological groups: filters, wrappers, and embedded methods [10,14]. The internal workings and configuration of the various feature-selection methods make them suitable for different applications. Filter methods employ attribute ranking to determine the most informative features. Features that attain scores above a given threshold are selected, and those below the threshold are discarded. After identifying the most important features, they can be fed as input to the learning algorithm. Filter methods vary from wrapper and embedded methods as they are not dependent on a classifier and are, therefore, independent of the classifier’s bias [15].
However, wrapper methods use an ML classifier’s performance as the evaluation metric in selecting the most relevant feature set. Wrapper methods usually lead to better classification performance than filter techniques because the feature-selection procedure is optimized for the chosen classification algorithm [16,17]. Generally, wrapper methods employ a search strategy to identify the candidate subsets. The classifier’s performance on the various feature subsets is measured, and the subset that leads to the highest performance is selected as the most informative subset. Examples of wrapper-based feature selection techniques include the Boruta algorithm, forward selection, backward elimination, and the genetic algorithm. Embedded methods select the features that enhance the model’s performance during training. The feature selection is incorporated into the learning procedure [13]. Unlike wrapper methods, this type of feature selection aims to reduce the time used in training different subsets. Embedded methods include random forest, decision tree, gradient boosting, elastic net, and LASSO [10].
Meanwhile, the GA wrapper is an effective method for feature selection, with applications in diverse domains, including natural language processing (NLP) [18], fraud detection [19], sentiment analysis [20], and medical diagnosis [21]. This study proposes a hybrid feature-selection approach, combining the IG-based filter and GA-based wrapper techniques. The main contributions and objectives of the work include the following:
  • Using the information gain technique for initial feature selection to rank the features in the credit card dataset, only the top-ranked features are fed into the GA wrapper to reduce the search space and enhance the classification performance.
  • Secondly, the GA wrapper is employed to select the best feature subset that results in optimal classification performance, and the ELM is employed as the learning algorithm in the GA wrapper.
  • Additionally, this study employs the G-mean as the fitness function in the GA wrapper instead of the conventional accuracy evaluation criterion, ensuring the recognition rate of the minority samples is considered and improved.
The rationale behind this approach is that the initial IG-based feature selection and ELM’s ability to produce promising performance while converging faster than traditional neural networks could reduce the computational complexity of the GA and improve the classification performance. The ELM is chosen as the learning algorithm in the GA wrapper because it converges far more rapidly and achieves higher generalization performance than conventional neural networks. At the same time, its learning process is thousands of times quicker than neural networks trained via backpropagation [22]. Furthermore, for convenience, the proposed hybrid approach is called IG-GAW. It would be compared with the conventional ELM classifier, an ELM classifier with IG-based feature selection (IG-ELM), the GA wrapper (GAW), and well-performing methods in related literature.
The rest of this paper is structured as follows: Section 2 presents related works, focusing on feature-selection methods in the literature. Section 3 discusses the dataset and algorithms used in this study. The proposed credit card fraud-prediction approach is introduced in Section 4. Section 5 presents the results, while Section 6 concludes the paper and provides appropriate future research directions.

2. Related Works

Recently, ML algorithms have been widely applied for credit card fraud detection [23,24,25]. Researchers have used both traditional ML and deep learning (DL) algorithms to predict credit card fraud efficiently. For example, Alarfaj et al. [26] conducted a study using ML and DL techniques for detecting credit card fraud, while Van Belle et al. [27] employed inductive graph representation learning, Esenogho et al. [28] used a neural network ensemble, and Zhang et al. [29] employed an ensemble classifier based on isolation forest and adaptive boosting.
Some problems encountered when dealing with credit card datasets include high dimensionality and imbalance class [30,31], making it difficult for ML classifiers to learn and make accurate predictions. In addition, high dimensional data often make the learning process complex and computationally expensive, resulting in models with poor generalization ability [32]. Therefore, feature selection is essential in such datasets to reduce the computational burden and enhance the model’s generalization ability. For example, Chaquet-Ulldemolins et al. [33] recorded an increase in the classification performance of ML classifiers after introducing feature selection. Generally, feature-selection methods are useful in applications where the number of features affects the classifier’s performance.
The wrapper feature-selection methods have been widely applied in numerous applications [34,35]. They compute the importance of each feature based on its usefulness when training the ML model. The primary components of a wrapper method are the learning classifier and search strategy. The wrapper technique exists as a wrapper around the learning classifier and uses the same classifier to select the most relevant features. Therefore, a robust learning classifier could enhance the wrapper-based feature selection. Furthermore, the search strategy employed in the wrapper could affect the feature selection, and using the right search strategy for a given application is crucial in obtaining good performance.
Evolutionary search techniques such as genetic algorithms can avoid becoming stuck in local optima. Unlike deterministic algorithms, they can identify reduced feature sets that can effectively represent the original feature set [36]. The GA-based wrapper can easily identify feature redundancy and correlations. In addition, selecting a suitable classifier is vital in developing robust GA wrapper models since the wrapper procedure is tied to the selected classifier’s performance. However, there are specific issues to consider when selecting the classifier. Firstly, the classifier should be able to achieve good classification performance and have excellent generalization ability. Secondly, since the classifier would be used to train numerous subsets, it should have good training speed. Thirdly, the number of features in the various subsets might differ. Therefore, using the same model parameters might not be enough to obtain good performance in all the subsets [37]. Hence, it would be preferred to use a classifier that automatically updates the model parameters for every feature subset to achieve good performance.
Other recent methods for credit card fraud detection include a signal processing framework [38], signal processing on graphs [39], and a deep learning ensemble [40]. In addition, in the literature, several learning algorithms (such as decision tree [41], naïve Bayes [42], SVM [43], and random forest [44]) have been used as the classifier in the GA wrapper. However, these classifiers are not able to consider the issues mentioned above. Therefore, a hybrid wrapper approach that considers all the above-mentioned issues is proposed. The proposed approach employs the IG-based filter feature selection to rank the attributes, and only the top-ranked features would be used as input into the GA wrapper. Meanwhile, the GA wrapper employs the ELM as the learning classifier. The ELM can achieve excellent classification performance and generalization ability with an extremely fast learning speed compared to conventional training methods. Furthermore, unlike traditional neural networks based on backpropagation algorithms, the ELM’s training process is entirely automatic and does not require it to be tuned iteratively.

3. Materials and Methods

3.1. Credit Card Dataset

The European cardholders dataset [45] is used in this study. It is publicly available and comprises 284807 transactions made by European cardholders in September 2013. The dataset has been widely used in different credit card fraud-detection studies [1,6,46]. It contains 492 fraudulent transactions, and the rest are legitimate transactions, i.e., only 0.17% of the dataset belongs to the minority class, and 99.83% belongs to the majority class; hence, the dataset is highly skewed and it is challenging for conventional ML algorithms to learn from the dataset.
Due to privacy concerns, the features in the dataset were anonymized as V 1 , V 2 , …, and V 28 , except for the “Time” and “Amount” features. The “Time” attribute indicates the seconds elapsed between a transaction and the first transaction in the dataset, whereas the “Amount” indicates the value of the transaction. Meanwhile, the “Class” attribute is the response variable, representing legitimate and fraudulent transactions and having values 0 and 1, respectively.

3.2. Information Gain

The information gain technique, or mutual information, is one of the most used filter criteria. The IG criterion is modeled after the concept of entropy in information theory [47]. The entropy measures the impurity or uncertainty in a group of observations, while information gain computes the decrease in entropy before and after adding an attribute. An attribute with a high IG value is usually preferred to those with low IG values. Assuming X and Y are features in a dataset, the information gain of X, given Y is represented mathematically, is:
G ( X | Y ) = H ( X ) H ( X | Y )
In (1), H ( X ) denotes the entropy of X and H ( X | Y ) is the conditional entropy for X given Y [48]. Meanwhile, H ( X ) and H ( X | Y ) can be represented as:
H ( X ) = x X P ( x ) l o g 2 ( x )
H ( X | Y ) = x X P ( x ) y Y P ( x | y ) l o g 2 ( P ( x | y ) )
Therefore, given two features X and Z, a response variable Y is more correlated to X than Z if I G ( X | Y ) > I G ( Z | Y ) [49]. Lastly, the information gain technique considers each feature separately, computes the IG value, and outputs its importance to the response variable.

3.3. Genetic Algorithm

The genetic algorithm, inspired by genetics in biological systems, can perform well in high-dimensional feature-selection problems because of its robustness, making it suitable for credit card fraud detection. Given a set of candidate features (called the population), the GA finds the optimal solution via a series of iterative genetic operations. It is superior to most traditional search methods in three main areas: firstly, the GA conducts a parallel search all over the population of solutions; instead of optimizing its parameters, the GA uses chromosomes, an encoded form of a possible solution, to achieve faster convergence. Lastly, the GA employs a fitness value to identify a potential solution. The flowchart of the GA is shown in Figure 1.
The genetic operators include crossover, mutation, and selection [50]. The crossover operator determines how the chromosomes are combined to obtain a new chromosome for the next generation [51]. In the GA, a population of candidate solutions (also known as individuals) to the optimization problem is evolved to obtain the optimal solution. The various candidate solutions have certain attributes called chromosomes that can be mutated; meanwhile, these solutions are represented in 0 and 1 binary strings, indicating whether the corresponding attribute has been selected or not. The mutation operator ensures some bits of the chromosomes are flipped randomly based on probability. The selection step involves choosing chromosomes based on their fitness score for further processing.
A common termination criterion is running the algorithm for a specified number of times [52]. Therefore, the algorithm ends after the specified number of iterations, outputting the optimal solution identified after going through all the generations.

3.4. Extreme Learning Machine

The ELM was developed by Huang et al. [50] to fix the slow learning speed of feedforward neural networks. The authors attributed the slow learning speed to the use of gradient descent-based learning algorithms for training neural networks and how such algorithms iteratively tune the neural network parameters [22]. The ELM has excellent generalization ability with extremely fast learning.
Unlike traditional neural networks, the hidden layer parameters of the ELM are randomly generated without being iterative tuned [37], thereby reducing the learning procedure to just estimating the optimal output weights β . For a given dataset ( x j , t j ) ( j = 1 ) N , where N is the number of instances and the SLFN has L hidden nodes, the activations function g ( x ) can be mathematically represented as:
i = 1 L β i g ( w i . x j + b i ) = t j
where t j is the output of the network, w i = [ w i 1 , , w i n ] T represent the input weight, b i denotes the bias of the i t h hidden node, β i = [ β i 1 , , β i m ] T denotes the weight vector linking the i t h hidden node with the output nodes, and w i . x j represents the inner product of w i and x j [53]. Equation (4) can be rewritten in the compact matrix form as
H β = T
where H represents the output matrix of the hidden layer and its mathematical formulation, presented by Huang et al. [22], is expressed as
H ( w 1 , , w N , b 1 , , b N , x 1 , , x N ) = g ( w i . x i + b i ) g ( w N . x i + b i ) g ( w 1 . x N + b 1 ) g ( w N . x N + b N ) N × N
β = β 1 T β N T N × m a n d t 1 T t N T N × m

4. Proposed Credit Card Fraud-Detection Approach

This study employs a hybrid feature-selection method, combining the IG-based filter and GA-based wrapper methods. Firstly, the IG technique ranks the attributes to identify the most significant attributes, and the threshold value is obtained by computing the standard deviation [1] of the IG values. Therefore, features with IG values greater or equal to the threshold are selected, while those below the threshold are discarded. The standard deviation has been widely employed in the literature to achieve excellent feature-selection thresholds [54,55,56]. Secondly, the top-ranked features are fed into the GAW, which uses the ELM as the learning algorithm.
The GA aims to identify the best feature subsets from a series of combinations known as generations [57]. Furthermore, after the ELM classifies the chromosomes, they are assigned a fitness value. Generally, the accuracy metric has been used as the standard fitness function [58]. However, it gives equal importance to samples in both majority and minority classes; hence, it is unsuitable for imbalanced classification problems [59].
Since the majority class samples outnumber the minority class, the fitness function will be biased toward the former. Furthermore, since there are more majority class samples, the accuracy metric will return high accuracy values, which could be misleading. Therefore, this study employs the G-mean metric obtained by the chromosomes as the fitness value to handle the imbalanced class. G-mean is a vital metric for imbalanced classification problems, and it considers the classifier’s performance for the majority and minority classes. The G-mean can be represented mathematically as follows:
G - m e a n = T P T P + F N × T N T N + F P
where true positive (TP) indicates a fraudulent transaction that is correctly predicted by the model, false positive (FP) indicates a legitimate transaction that the model wrongly predicts as fraud, true negative (TN) indicates a legitimate transaction that is correctly identified, and false negative (FN) indicates a fraudulent transaction that is predicted as legitimate. The proposed approach is outlined in Algorithm 1.
Algorithm 1 Proposed IG-GAW
  • Compute the information gain of the features in the dataset
  • Rank features according to their importance: F = ( f 1 > f 2 > f 3 , , f n ) .
  • GAW input: Top-ranked features from Step 2 and class variable C, population size n, elitism rate e, number of iterations k.
  • Begin
  • Initialize population with n random solutions
  • Compute the fitness values for each random solution.
  • for  i = 1 , , k :
  • Select the best individuals with respect to e
  • Generate new offspring based on the GA operators
  • Examine the fitness values of the new individuals.
  • Discard least-fit population individuals in the population
  • end for
  • GAW output: Optimal solution S
The GAW approach starts with randomly initializing the GA’s population, where every candidate feature subset is encoded as a chromosome. The next step involves training different ELM neural networks based on each chromosome, and the fitness value for each feature subset is computed. Thirdly, a new population is generated using genetic operators. The procedure continues until the stopping criterion is obtained, i.e., the maximum number of generations. This stopping criterion has been used extensively in the literature to obtain excellent GA performance [60,61,62]. Meanwhile, the rationale behind this approach is that the GA wrapper would select the best feature subset that would lead to enhanced prediction performance.

5. Results and Discussion

The proposed method’s classification performance and other baseline classifiers are presented and discussed in the section. The machine learning models were implemented using scikit-learn [63], a widely used library for machine learning in Python. Meanwhile, the proposed method is compared with the following baseline classifiers: AdaBoost [64], logistic regression (LR) [65], random forest (RF) [66], SVM [67], and decision tree [68]. Furthermore, the stratified 10-fold cross-validation method is employed to measure the performance of the prediction models. The stratified k-fold technique ensures that the proportion of fraudulent and legitimate instances in the dataset is preserved in each fold, and it is usually more suitable for imbalance classification problems than the k-fold cross-validation method [69].
The following metrics are used to evaluate the performance of the models: sensitivity, specificity, the receiver operating characteristic (ROC) curve, and the area under the ROC curve (AUC). Sensitivity refers to the model’s ability to predict a fraud transaction as fraud. Usually, a highly sensitive model is preferred in fraud detection, as it implies there are no or few false negative predictions [70,71]. Meanwhile, the specificity of a model is its ability to predict non-fraudulent transactions as legitimate. A model with high specificity means there are hardly any false positive predictions [72]. Sensitivity and specificity can be computed as follows:
Sensitivity = T P T P + F N
Specificity = T N T N + F P
The ROC curve is used to visualize the performance of binary classifiers. It plots the true positive rate against the false positive rate at various classification thresholds [73]. In contrast, the AUC summarizes the ROC curve, and its value ranges from 0 to 1. An AUC of 0 implies the model’s predictions are all wrong, and a value of 1 implies all the model predictions are correct. The AUC is a crucial metric in imbalance classification problems, such as fraud detection, as it indicates the classifier’s ability to differentiate between the fraud and non-fraud classes.
Furthermore, in line with existing literature, this study estimates the best GA parameters by conducting several trials using different combinations [74,75,76]. The final GA parameters used in this study are outlined in Table 1.

5.1. Performance of the ELM Classifier with Filter, Wrapper, and Hybrid Feature Selection Methods

The ELM’s performance without feature selection is compared with instances where the ELM classifier is coupled with the filter, wrapper, and hybrid feature-selection techniques. Firstly, the performance of the ELM without feature selection is recorded. Secondly, the credit card features and their IG values are ranked by the IG-based filter technique.
The standard deviation [54] of the IG values is calculated and used as the threshold value to select the most informative features. From Table 2, the standard deviation is 0.145. Hence, information gain values above 0.145 are chosen as the essential features employed for training the machine learning model. The features with IG values below 0.145 are removed. Therefore, the top 21 attributes are chosen by the filter approach as the most important features, while the following attributes are removed: V 8 , V 19 , V 24 , V 23 , V 26 , V 13 , V 25 , V 15 , and V 22 .
Thirdly, the GA wrapper coupled with the ELM classifier (GAW) is trained, and its performance is recorded. Lastly, the top 21 features selected by the IG technique are used as input to the GAW. The performance from the four scenarios is tabulated in Table 2, i.e., ELM without feature selection, filter-based IG-ELM, wrapper-based GAW, and the hybrid IG-GAW method. In addition, Figure 2 shows the ROC curves and AUC of the various models. Meanwhile, the complete feature set and the features selected by the various feature-selection methods are tabulated in Table 3.
Table 2 and Figure 2 show that the proposed hybrid IG-GAW obtained the highest sensitivity, specificity, and AUC values of 0.997, 0.994, and 0.990, respectively, outperforming the ELM, the IG-ELM, and IG-GAW.

5.2. Performance Comparison with Baseline Classifiers and Recent Literature

In this section, the proposed hybrid approach is benchmarked with other ML classifiers and methods in the literature. The baseline classifiers were trained with the complete feature set. The performance of the classifiers is tabulated in Table 4 and visualized in Figure 3. The models obtained relatively high specificity compared to the sensitivity, which implies the model correctly predicted more non-fraud transactions (majority class) than fraud transactions (minority class).
Meanwhile, in credit card fraud detection, like every imbalance classification task, it is more important to predict the minority class samples correctly. However, the proposed method achieved excellent sensitivity and specificity, indicating its robustness in predicting the minority and majority class samples. This enhanced performance could be attributed to using the G-mean as the fitness function rather than the widely used accuracy criterion, ensuring the model’s detection rate on the minority class is enhanced.
Furthermore, Table 5 shows the performance of some state-of-the-art methods in the literature. The methods include a weighted extreme learning machine (weighted ELM) [77], a deep neural network (DNN)-based classifier [78], a cost-sensitive neural network ensemble (CS-NNE) [79], a random forest-based genetic algorithm wrapper method (GA-RF) [19], a method that sequentially combines C4.5 and the naïve Bayes model (DT-NB) [80], a model developed using the random forest algorithm with the SMOTE technique (RF-SMOTE) [81], a stochastic ensemble model [82], an XGBoost-SMOTE model [83], a support vector machine (SVM)-based GA wrapper [84], an ensemble model optimized using the particle swarm optimization (PSO) technique [85], a metaheuristic based feature selection algorithm based on rock hyrax swarm optimization (RHSO) [86], and a deep residual network with shuffled shepherd optimization (DRN-SSPO) [87].
Table 5 shows that the proposed IG-GAW demonstrated excellent results compared to the state-of-the-art methods, indicating the proposed method’s robustness. Furthermore, the proposed method is applied to other credit card datasets to show how the method performs in different scenarios. The datasets include the German credit card dataset [88] and the Taiwan credit card dataset [89]. The German dataset has 20 attributes and contains 1000 samples, of which 700 are classified as good and 300 as bad customers, i.e., 70% of the dataset belongs to the majority class, while 30% belongs to the minority class; hence, the dataset is imbalanced. The features in the German dataset and those selected by the various feature selection techniques are shown in Table 6, while Table 7 and Figure 4 show the performance of the various models.
Meanwhile, the Taiwan dataset has 25 attributes and 30,000 samples, of which 23,364 are classified as good and 6636 as bad customers, i.e., 77.88% of the dataset belongs to the majority class and 22.12% belongs to the minority class. Therefore, the Taiwan dataset is also imbalanced. The features in the datasets and the selected feature sets are shown in Table 8. Meanwhile, Table 9 and Figure 5 show the performance of the various methods.
Using the German dataset, the proposed IG-GAW achieved a sensitivity of 0.904, specificity of 0.945, and AUC of 0.910, as shown in Table 7. Meanwhile, the experimental results for the Taiwan dataset are shown in Table 9, and it shows that the proposed method obtained a sensitivity of 0.945, specificity of 0.961, and AUC of 0.940.

5.3. Discussions

Credit card fraud is a huge burden for financial institutions. The advances in e-commerce and digital payment platforms have made credit card fraud more common. This study aimed to utilize machine learning to detect credit card fraud effectively. Meanwhile, since most credit card datasets contain irrelevant attributes that degrade the performance of machine learning algorithms, this study proposed a robust hybrid feature-selection approach comprising filter (information gain technique) and wrapper (genetic algorithm) feature-selection steps, ensuring only the most significant attributes are used for machine learning. This study uses the well-known European Credit Card dataset. The proposed IG-GAW, which uses the ELM as the learning algorithm in the GA wrapper, obtained scores for sensitivity, specificity, AUC, and G-mean of 0.997, 0.994, 0.990, and 0.994, respectively. Additionally, two popular credit risk datasets (German and Taiwan credit datasets) were used to further validate the proposed method’s improved performance.
The proposed IG-GAW outperformed the selected classifiers when trained with the German and Taiwan datasets. The results also showed that the proposed hybrid approach, IG-GAW, achieved superior performance compared to the filter and wrapper methods, i.e., IG-ELM and GAW. Therefore, integrating filter and wrapper techniques in a hybrid setting is a robust approach to detecting credit card fraud. Lastly, the results also showed that introducing feature selection enhanced the ELM’s performance, indicating the importance of effective feature selection.

6. Conclusions

Detecting fraudulent credit card transactions is challenging, and researchers have developed different methods to handle this problem. This study proposed a hybrid approach to enhance the detection rate. The hybrid approach takes advantage of the strength of different feature-selection and ML methods, including information gain, genetic algorithms, and extreme learning machines. The IG technique was employed for initial feature selection, and the top-ranked features served as input to the GA wrapper. Meanwhile, the ELM was used as the learning algorithm in the GA wrapper. The proposed approach outperformed other baseline classifiers and methods in recent literature. Furthermore, the proposed method was applied to two more credit card datasets to validate its performance, and it achieved excellent performance in both datasets, demonstrating its robustness. Therefore, it can be concluded that the proposed hybrid approach is an effective credit card fraud-detection method. Future research work would employ more combinations of evolutionary algorithms and ML-based feature selection methods to enhance new aspects of credit card fraud detection. In addition, future research would explore the potential of obtaining more recent datasets to train ML models.

Author Contributions

Conceptualization, I.D.M. and Y.S.; methodology, I.D.M. and Y.S.; software, I.D.M.; validation, I.D.M. and Y.S.; formal analysis, I.D.M. and Y.S.; investigation, I.D.M. and Y.S.; resources, Y.S.; data curation, I.D.M. and Y.S.; writing—original draft preparation, I.D.M.; writing—review and editing, I.D.M. and Y.S.; visualization, I.D.M. and Y.S.; supervision, Y.S.; funding acquisition, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the South African National Research Foundation under Grant 120106 and Grant 132797 and in part by the South African National Research Foundation Incentive under Grant 132159.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Femila Roseline, J.; Naidu, G.; Samuthira Pandi, V.; Alamelu alias Rajasree, S.; Mageswari, N. Autonomous credit card fraud detection using machine learning approach. Comput. Electr. Eng. 2022, 102, 108132. [Google Scholar] [CrossRef]
  2. Alharbi, A.; Alshammari, M.; Okon, O.D.; Alabrah, A.; Rauf, H.T.; Alyami, H.; Meraj, T. A Novel text2IMG Mechanism of Credit Card Fraud Detection: A Deep Learning Approach. Electronics 2022, 11, 756. [Google Scholar] [CrossRef]
  3. Bin Sulaiman, R.; Schetinin, V.; Sant, P. Review of Machine Learning Approach on Credit Card Fraud Detection. Hum.-Centric Intell. Syst. 2022, 2, 55–68. [Google Scholar] [CrossRef]
  4. Wang, D.; Chen, B.; Chen, J. Credit card fraud detection strategies with consumer incentives. Omega 2019, 88, 179–195. [Google Scholar] [CrossRef]
  5. Nandi, A.K.; Randhawa, K.K.; Chua, H.S.; Seera, M.; Lim, C.P. Credit card fraud detection using a hierarchical behavior-knowledge space model. PLoS ONE 2022, 17, e0260579. [Google Scholar] [CrossRef]
  6. Ileberi, E.; Sun, Y.; Wang, Z. Performance Evaluation of Machine Learning Methods for Credit Card Fraud Detection Using SMOTE and AdaBoost. IEEE Access 2021, 9, 165286–165294. [Google Scholar] [CrossRef]
  7. Rtayli, N.; Enneya, N. Enhanced credit card fraud detection based on SVM-recursive feature elimination and hyper-parameters optimization. J. Inf. Secur. Appl. 2020, 55, 102596. [Google Scholar] [CrossRef]
  8. Oo, M.C.M.; Thein, T. An efficient predictive analytics system for high dimensional big data. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 1521–1532. [Google Scholar] [CrossRef]
  9. Huebner, J.; Fleisch, E.; Ilic, A. Assisting mental accounting using smartphones: Increasing the salience of credit card transactions helps consumer reduce their spending. Comput. Hum. Behav. 2020, 113, 106504. [Google Scholar] [CrossRef]
  10. Pudjihartono, N.; Fadason, T.; Kempa-Liehr, A.W.; O’Sullivan, J.M. A Review of Feature Selection Methods for Machine Learning-Based Disease Risk Prediction. Front. Bioinform. 2022, 2, 927312. [Google Scholar] [CrossRef]
  11. de-la-Bandera, I.; Palacios, D.; Mendoza, J.; Barco, R. Feature Extraction for Dimensionality Reduction in Cellular Networks Performance Analysis. Sensors 2020, 20, 6944. [Google Scholar] [CrossRef]
  12. Bouaguel, W. A New Approach for Wrapper Feature Selection Using Genetic Algorithm for Big Data. In Intelligent and Evolutionary Systems; Springer: Cham, Switzerland, 2016; pp. 75–83. [Google Scholar] [CrossRef]
  13. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  14. Bashir, S.; Khattak, I.U.; Khan, A.; Khan, F.H.; Gani, A.; Shiraz, M. A Novel Feature Selection Method for Classification of Medical Data Using Filters, Wrappers, and Embedded Approaches. Complexity 2022, 2022, e8190814. [Google Scholar] [CrossRef]
  15. Kumar, A.; Bhatia, M.P.S.; Sangwan, S.R. Rumour detection using deep learning and filter-wrapper feature selection in benchmark twitter dataset. Multimed. Tools Appl. 2022, 81, 34615–34632. [Google Scholar] [CrossRef]
  16. Wang, F.; Lu, X.; Chang, X.; Cao, X.; Yan, S.; Li, K.; Duić, N.; Shafie-khah, M.; Catalão, J.P. Household profile identification for behavioral demand response: A semi-supervised learning approach using smart meter data. Energy 2022, 238, 121728. [Google Scholar] [CrossRef]
  17. Wang, Z.; Gao, S.; Zhou, M.; Sato, S.; Cheng, J.; Wang, J. Information-Theory-based Nondominated Sorting Ant Colony Optimization for Multiobjective Feature Selection in Classification. IEEE Trans. Cybern. 2022, 1–14. [Google Scholar] [CrossRef]
  18. Rasool, A.; Tao, R.; Kamyab, M.; Hayat, S. GAWA–A Feature Selection Method for Hybrid Sentiment Classification. IEEE Access 2020, 8, 191850–191861. [Google Scholar] [CrossRef]
  19. Ileberi, E.; Sun, Y.; Wang, Z. A machine learning based credit card fraud detection using the GA algorithm for feature selection. J. Big Data 2022, 9, 24. [Google Scholar] [CrossRef]
  20. Al-Ahmad, B.; Al-Zoubi, A.M.; Abu Khurma, R.; Aljarah, I. An Evolutionary Fake News Detection Method for COVID-19 Pandemic Information. Symmetry 2021, 13, 1091. [Google Scholar] [CrossRef]
  21. Soumaya, Z.; Drissi Taoufiq, B.; Benayad, N.; Yunus, K.; Abdelkrim, A. The detection of Parkinson disease using the genetic algorithm and SVM classifier. Appl. Acoust. 2021, 171, 107528. [Google Scholar] [CrossRef]
  22. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar] [CrossRef]
  23. Han, S.; Zhu, K.; Zhou, M.; Cai, X. Competition-Driven Multimodal Multiobjective Optimization and Its Application to Feature Selection for Credit Card Fraud Detection. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 7845–7857. [Google Scholar] [CrossRef]
  24. Malik, E.F.; Khaw, K.W.; Belaton, B.; Wong, W.P.; Chew, X. Credit Card Fraud Detection Using a New Hybrid Machine Learning Architecture. Mathematics 2022, 10, 1480. [Google Scholar] [CrossRef]
  25. Zioviris, G.; Kolomvatsos, K.; Stamoulis, G. Credit card fraud detection using a deep learning multistage model. J. Supercomput. 2022, 78, 14571–14596. [Google Scholar] [CrossRef]
  26. Alarfaj, F.K.; Malik, I.; Khan, H.U.; Almusallam, N.; Ramzan, M.; Ahmed, M. Credit Card Fraud Detection Using State-of-the-Art Machine Learning and Deep Learning Algorithms. IEEE Access 2022, 10, 39700–39715. [Google Scholar] [CrossRef]
  27. Van Belle, R.; Van Damme, C.; Tytgat, H.; De Weerdt, J. Inductive Graph Representation Learning for fraud detection. Expert Syst. Appl. 2022, 193, 116463. [Google Scholar] [CrossRef]
  28. Esenogho, E.; Mienye, I.D.; Swart, T.G.; Aruleba, K.; Obaido, G. A Neural Network Ensemble with Feature Engineering for Improved Credit Card Fraud Detection. IEEE Access 2022, 10, 16400–16407. [Google Scholar] [CrossRef]
  29. Zhang, Y.-F.; Lu, H.-L.; Lin, H.-F.; Qiao, X.-C.; Zheng, H. The Optimized Anomaly Detection Models Based on an Approach of Dealing with Imbalanced Dataset for Credit Card Fraud Detection. Mob. Inf. Syst. 2022, 2022, e8027903. [Google Scholar] [CrossRef]
  30. Ala’raj, M.; Abbod, M.F.; Majdalawieh, M.; Jum’a, L. A deep learning model for behavioural credit scoring in banks. Neural Comput. Appl. 2022, 34, 5839–5866. [Google Scholar] [CrossRef]
  31. Zhang, X.; Yu, L.; Yin, H.; Lai, K.K. Integrating data augmentation and hybrid feature selection for small sample credit risk assessment with high dimensionality. Comput. Oper. Res. 2022, 146, 105937. [Google Scholar] [CrossRef]
  32. Yang, Y.; Fan, C.; Chen, L.; Xiong, H. IPMOD: An efficient outlier detection model for high-dimensional medical data streams. Expert Syst. Appl. 2022, 191, 116212. [Google Scholar] [CrossRef]
  33. Chaquet-Ulldemolins, J.; Gimeno-Blanes, F.-J.; Moral-Rubio, S.; Muñoz-Romero, S.; Rojo Álvarez, J.-L. On the Black-Box Challenge for Fraud Detection Using Machine Learning (I): Linear Models and Informative Feature Selection. Appl. Sci. 2022, 12, 3328. [Google Scholar] [CrossRef]
  34. Al-Yaseen, W.L.; Idrees, A.K.; Almasoudy, F.H. Wrapper feature selection method based differential evolution and extreme learning machine for intrusion detection system. Pattern Recognit. 2022, 132, 108912. [Google Scholar] [CrossRef]
  35. Beheshti, Z. BMPA-TVSinV: A Binary Marine Predators Algorithm using time-varying sine and V-shaped transfer functions for wrapper-based feature selection. Knowl.-Based Syst. 2022, 252, 109446. [Google Scholar] [CrossRef]
  36. Prashanth, S.K.; Shitharth, S.; Praveen Kumar, B.; Subedha, V.; Sangeetha, K. Optimal Feature Selection Based on Evolutionary Algorithm for Intrusion Detection. SN Comput. Sci. 2022, 3, 439. [Google Scholar] [CrossRef]
  37. Xue, X.; Yao, M.; Wu, Z. A novel ensemble-based wrapper method for feature selection using extreme learning machine and genetic algorithm. Knowl. Inf. Syst. 2018, 57, 389–412. [Google Scholar] [CrossRef]
  38. Salazar, A.; Safont, G.; Rodriguez, A.; Vergara, L. Combination of multiple detectors for credit card fraud detection. In Proceedings of the 2016 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), Limassol, Cyprus, 12–14 December 2016; pp. 138–143. [Google Scholar] [CrossRef]
  39. Vergara, L.; Salazar, A.; Belda, J.; Safont, G.; Moral, S.; Iglesias, S. Signal processing on graphs for improving automatic credit card fraud detection. In Proceedings of the 2017 International Carnahan Conference on Security Technology (ICCST), Madrid, Spain, 23–26 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
  40. Mienye, I.D.; Sun, Y. A Deep Learning Ensemble With Data Resampling for Credit Card Fraud Detection. IEEE Access 2023, 11, 30628–30638. [Google Scholar] [CrossRef]
  41. Gkikas, D.C.; Theodoridis, P.K.; Beligiannis, G.N. Enhanced Marketing Decision Making for Consumer Behaviour Classification Using Binary Decision Trees and a Genetic Algorithm Wrapper. Informatics 2022, 9, 45. [Google Scholar] [CrossRef]
  42. Mabdeh, A.N.; Al-Fugara, A.; Ahmadlou, M.; Al-Adamat, R.; Al Shabeeb, A.R. GIS-based landslide susceptibility assessment and mapping in Ajloun and Jerash governorates in Jordan using genetic algorithm-based ensemble models. Acta Geophys. 2022, 70, 1253–1267. [Google Scholar] [CrossRef]
  43. Tao, P.; Sun, Z.; Sun, Z. An Improved Intrusion Detection Algorithm Based on GA and SVM. IEEE Access 2018, 6, 13624–13631. [Google Scholar] [CrossRef]
  44. Kasongo, S.M. An Advanced Intrusion Detection System for IIoT Based on GA and Tree Based Algorithms. IEEE Access 2021, 9, 113199–113212. [Google Scholar] [CrossRef]
  45. Credit Card Fraud Detection. Available online: https://kaggle.com/mlg-ulb/creditcardfraud (accessed on 26 October 2021).
  46. Lin, T.-H.; Jiang, J.-R. Credit Card Fraud Detection with Autoencoder and Probabilistic Random Forest. Mathematics 2021, 9, 2683. [Google Scholar] [CrossRef]
  47. Mienye, I.D.; Obaido, G.; Aruleba, K.; Dada, O.A. Enhanced Prediction of Chronic Kidney Disease Using Feature Selection and Boosted Classifiers. In Intelligent Systems Design and Applications; Springer: Cham, Switzerland, 2022; pp. 527–537. [Google Scholar] [CrossRef]
  48. Alhaj, T.A.; Siraj, M.M.; Zainal, A.; Elshoush, H.T.; Elhaj, F. Feature Selection Using Information Gain for Improved Structural-Based Alert Correlation. PLoS ONE 2016, 11, e0166017. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Ebiaredoh-Mienye, S.A.; Swart, T.G.; Esenogho, E.; Mienye, I.D. A Machine Learning Method with Filter-Based Feature Selection for Improved Prediction of Chronic Kidney Disease. Bioengineering 2022, 9, 350. [Google Scholar] [CrossRef]
  50. Katoch, S.; Chauhan, S.S.; Kumar, V. A review on genetic algorithm: Past, present, and future. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef]
  51. Schulte, R.V.; Prinsen, E.C.; Hermens, H.J.; Buurke, J.H. Genetic Algorithm for Feature Selection in Lower Limb Pattern Recognition. Front. Robot. AI 2021, 8, 710806. Available online: https://www.frontiersin.org/articles/10.3389/frobt.2021.710806 (accessed on 23 November 2022). [CrossRef] [PubMed]
  52. Kalita, K.; Dey, P.; Haldar, S.; Gao, X.-Z. Optimizing frequencies of skew composite laminates with metaheuristic algorithms. Eng. Comput. 2020, 36, 741–761. [Google Scholar] [CrossRef]
  53. Jovanovic, D.; Antonijevic, M.; Stankovic, M.; Zivkovic, M.; Tanaskovic, M.; Bacanin, N. Tuning Machine Learning Models Using a Group Search Firefly Algorithm for Credit Card Fraud Detection. Mathematics 2022, 10, 2272. [Google Scholar] [CrossRef]
  54. Prasetiyowati, M.I.; Maulidevi, N.U.; Surendro, K. Determining threshold value on information gain feature selection to increase speed and prediction accuracy of random forest. J. Big Data 2021, 8, 84. [Google Scholar] [CrossRef]
  55. Xie, J.; Wang, M.; Xu, S.; Huang, Z.; Grant, P.W. The Unsupervised Feature Selection Algorithms Based on Standard Deviation and Cosine Similarity for Genomic Data Analysis. Front. Genet. 2021, 12, 684100. Available online: https://www.frontiersin.org/article/10.3389/fgene.2021.684100 (accessed on 15 January 2022). [CrossRef]
  56. Van Hulse, J.; Khoshgoftaar, T.M.; Napolitano, A.; Wald, R. Threshold-based feature selection techniques for high-dimensional bioinformatics data. Netw. Model. Anal. Health Inform. Bioinform. 2012, 1, 47–61. [Google Scholar] [CrossRef] [Green Version]
  57. Theodoridis, P.K.; Gkikas, D.C. Optimal Feature Selection for Decision Trees Induction Using a Genetic Algorithm Wrapper—A Model Approach. In Strategic Innovative Marketing and Tourism; Springer: Cham, Switzerland, 2020; pp. 583–591. [Google Scholar] [CrossRef]
  58. Kumar, A.; Sinha, N.; Bhardwaj, A. A novel fitness function in genetic programming for medical data classification. J. Biomed. Inform. 2020, 112, 103623. [Google Scholar] [CrossRef]
  59. Mienye, I.D.; Sun, Y. Effective Feature Selection for Improved Prediction of Heart Disease. In Pan-African Artificial Intelligence and Smart Systems; Springer: Cham, Switzerland, 2022; pp. 94–107. [Google Scholar] [CrossRef]
  60. Costa-Carrapiço, I.; Raslan, R.; González, J.N. A systematic review of genetic algorithm-based multi-objective optimisation for building retrofitting strategies towards energy efficiency. Energy Build. 2020, 210, 109690. [Google Scholar] [CrossRef]
  61. Maghawry, A.; Hodhod, R.; Omar, Y.; Kholief, M. An approach for optimizing multi-objective problems using hybrid genetic algorithms. Soft Comput. 2021, 25, 389–405. [Google Scholar] [CrossRef]
  62. Blank, J.; Deb, K. A Running Performance Metric and Termination Criterion for Evaluating Evolutionary Multi- and Many-objective Optimization Algorithms. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar] [CrossRef]
  63. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  64. Schapire, R.E. A brief introduction to boosting. IJCAI 1999, 99, 1401–1406. [Google Scholar]
  65. Cramer, J.S. The Origins of Logistic Regression. In Social Science Research Network; SSRN Scholarly Paper ID 360300; SSRN: Rochester, NY, USA, 2002. [Google Scholar] [CrossRef] [Green Version]
  66. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  67. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  68. Krzywinski, M.; Altman, N. Classification and regression trees. Nat. Methods 2017, 14, 8. [Google Scholar] [CrossRef]
  69. Prusty, S.; Patnaik, S.; Dash, S.K. SKCV: Stratified K-fold cross-validation on ML classifiers for predicting cervical cancer. Front. Nanotechnol. 2022, 4, 972421. Available online: https://www.frontiersin.org/articles/10.3389/fnano.2022.972421 (accessed on 8 November 2022). [CrossRef]
  70. Trevethan, R. Sensitivity, Specificity, and Predictive Values: Foundations, Pliabilities, and Pitfalls in Research and Practice. Front. Public Health 2017, 5, 307. Available online: https://www.frontiersin.org/article/10.3389/fpubh.2017.00307 (accessed on 25 January 2022). [CrossRef]
  71. Mienye, I.D.; Sun, Y. A Survey of Ensemble Learning: Concepts, Algorithms, Applications, and Prospects. IEEE Access 2022, 10, 99129–99149. [Google Scholar] [CrossRef]
  72. Obaido, G.; Ogbuokiri, B.; Swart, T.G.; Ayawei, N.; Kasongo, S.M.; Aruleba, K.; Mienye, I.D.; Aruleba, I.; Chukwu, W.; Osaye, F.; et al. An Interpretable Machine Learning Approach for Hepatitis B Diagnosis. Appl. Sci. 2022, 12, 11127. [Google Scholar] [CrossRef]
  73. Mienye, I.D.; Sun, Y.; Wang, Z. Improved Predictive Sparse Decomposition Method with Densenet for Prediction of Lung Cancer. Int. J. Comput. 2020, 1, 533–541. [Google Scholar] [CrossRef]
  74. Zain, A.M.; Haron, H.; Sharif, S. Application of GA to optimize cutting conditions for minimizing surface roughness in end milling machining process. Expert Syst. Appl. 2010, 37, 4650–4659. [Google Scholar] [CrossRef]
  75. Mirjalili, S. Genetic Algorithm. In Evolutionary Algorithms and Neural Networks: Theory and Applications; Mirjalili, S., Ed.; Springer International Publishing: Cham, Switzerland, 2019; pp. 43–55. [Google Scholar] [CrossRef]
  76. Mienye, I.D.; Kenneth Ainah, P.; Emmanuel, I.D.; Esenogho, E. Sparse noise minimization in image classification using Genetic Algorithm and DenseNet. In Proceedings of the 2021 Conference on Information Communications Technology and Society (ICTAS), Durban, South Africa, 10–11 March 2021; pp. 103–108. [Google Scholar] [CrossRef]
  77. Zhu, H.; Liu, G.; Zhou, M.; Xie, Y.; Abusorrah, A.; Kang, Q. Optimizing Weighted Extreme Learning Machines for imbalanced classification and application to credit card fraud detection. Neurocomputing 2020, 407, 50–62. [Google Scholar] [CrossRef]
  78. Alkhatib, K.I.; Al-Aiad, A.I.; Almahmoud, M.H.; Elayan, O.N. Credit Card Fraud Detection Based on Deep Neural Network Approach. In Proceedings of the 2021 12th International Conference on Information and Communication Systems (ICICS), Valencia, Spain, 24–26 May 2021; pp. 153–156. [Google Scholar] [CrossRef]
  79. Yotsawat, W.; Wattuya, P.; Srivihok, A. A Novel Method for Credit Scoring Based on Cost-Sensitive Neural Network Ensemble. IEEE Access 2021, 9, 78521–78537. [Google Scholar] [CrossRef]
  80. Kalid, S.N.; Ng, K.-H.; Tong, G.-K.; Khor, K.-C. A Multiple Classifiers System for Anomaly Detection in Credit Card Data With Unbalanced and Overlapped Classes. IEEE Access 2020, 8, 28210–28221. [Google Scholar] [CrossRef]
  81. Mrozek, P.; Panneerselvam, J.; Bagdasar, O. Efficient Resampling for Fraud Detection During Anonymised Credit Card Transactions with Unbalanced Datasets. In Proceedings of the 2020 IEEE/ACM 13th International Conference on Utility and Cloud Computing (UCC), Leicester, UK, 7–10 December 2020; pp. 426–433. [Google Scholar] [CrossRef]
  82. Carta, S.; Ferreira, A.; Reforgiato Recupero, D.; Saia, R. Credit scoring by leveraging an ensemble stochastic criterion in a transformed feature space. Prog. Artif. Intell. 2021, 10, 417–432. [Google Scholar] [CrossRef]
  83. Xie, Y.; Li, A.; Gao, L.; Liu, Z. A Heterogeneous Ensemble Learning Model Based on Data Distribution for Credit Card Fraud Detection. Wirel. Commun. Mob. Comput. 2021, 2021, e2531210. [Google Scholar] [CrossRef]
  84. Saheed, Y.K.; Hambali, M.A.; Arowolo, M.O.; Olasupo, Y.A. Application of GA Feature Selection on Naive Bayes, Random Forest and SVM for Credit Card Fraud Detection. In Proceedings of the 2020 International Conference on Decision Aid Sciences and Application (DASA), Sakheer, Bahrain, 8–9 November 2020; pp. 1091–1097. [Google Scholar] [CrossRef]
  85. Verma, B.P.; Verma, V.; Badholia, A. Hyper-Tuned Ensemble Machine Learning Model for Credit Card Fraud Detection. In Proceedings of the 2022 International Conference on Inventive Computation Technologies (ICICT), Lalitpur, Nepal, 20–22 July 2022; pp. 320–327. [Google Scholar] [CrossRef]
  86. Padhi, B.K.; Chakravarty, S.; Naik, B.; Pattanayak, R.M.; Das, H. RHSOFS: Feature Selection Using the Rock Hyrax Swarm Optimization Algorithm for Credit Card Fraud Detection System. Sensors 2022, 22, 9321. [Google Scholar] [CrossRef]
  87. Ganji, V.R.; Chaparala, A.; Sajja, R. Shuffled shepherd political optimization-based deep learning method for credit card fraud detection. Concurr. Comput. Pract. Exp. 2023, 35, e7666. [Google Scholar] [CrossRef]
  88. UCI Machine Learning Repository: Statlog (German Credit Data) Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/statlog+(german+credit+data) (accessed on 5 December 2022).
  89. UCI Machine Learning Repository: Default of credit card clients Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/default+of+credit+card+clients (accessed on 5 December 2022).
Figure 1. Genetic algorithm flowchart.
Figure 1. Genetic algorithm flowchart.
Applsci 13 07254 g001
Figure 2. ROC curve of the ELM classifier, filter, wrapper, and hybrid feature-selection methods.
Figure 2. ROC curve of the ELM classifier, filter, wrapper, and hybrid feature-selection methods.
Applsci 13 07254 g002
Figure 3. Comparative analysis of the various methods using the credit card dataset.
Figure 3. Comparative analysis of the various methods using the credit card dataset.
Applsci 13 07254 g003
Figure 4. Comparative analysis of the various methods using the German dataset.
Figure 4. Comparative analysis of the various methods using the German dataset.
Applsci 13 07254 g004
Figure 5. Comparative analysis of the various methods using the Taiwan dataset.
Figure 5. Comparative analysis of the various methods using the Taiwan dataset.
Applsci 13 07254 g005
Table 1. GA parameters.
Table 1. GA parameters.
ParameterValue
Population size50
Number of generations100
Crossover rate0.6
Mutation rate0.01
Fitness functionG-mean
Stopping criteriaMax number of generations
Type of mutationUniform mutation
Type of crossoverSingle point
Parent selection methodTournament selection
Tournament size2
Table 2. Performance of the ELM classifier, filter, wrapper, and hybrid feature-selection methods.
Table 2. Performance of the ELM classifier, filter, wrapper, and hybrid feature-selection methods.
ClassifierSensitivitySpecificityAUCG-Mean
ELM0.8810.9040.9000.892
IG-ELM0.9360.9600.9400.947
GAW0.9490.9620.9500.955
IG-GAW0.9970.9940.9900.994
Table 3. Feature sets from the European cardholders dataset.
Table 3. Feature sets from the European cardholders dataset.
Feature-Selection MethodFeatures
Complete feature set V 1 , V 2 , V 3 , V 4 , V 5 , V 6 , V 7 , V 8 , V 9 , V 10 , V 11 ,
V 12 , V 13 , V 14 , V 15 , V 16 , V 17 , V 18 , V 19 ,
V 20 , V 21 , V 22 , V 23 , V 24 , V 25 , V 26 ,
V 27 , V 28 , Time, Amount, Class
IG V 1 , V 2 , V 3 , V 4 , V 5 , V 6 , V 7 , V 9 , V 10 , V 11 ,
V 12 , V 14 , V 16 , V 17 , V 18 , V 20 ,
V 21 , V 27 , V 28 , Time, Amount
GAW V 1 , V 2 , V 5 , V 6 , V 7 , V 9 , V 10 , V 11 , V 12 , V 16 , V 17 ,
V 18 , V 20 , V 21 , V 27 , V 28 , Time, Amount
IG-GAW V 1 , V 2 , V 5 , V 6 , V 7 , V 9 , V 11 , V 12 , V 16 , V 17 , V 18 ,
V 20 , V 21 , V 27 , V 28
Table 4. Performance comparison with other baseline classifiers.
Table 4. Performance comparison with other baseline classifiers.
ClassifierSensitivitySpecificityAUCG-Mean
AdaBoost0.8890.9180.9000.903
LR0.7520.9160.8100.829
RF0.8690.9400.8900.904
SVM0.5850.8270.6600.695
DT0.5900.8010.6900.688
Proposed IG-GAW0.9970.9940.9900.994
Table 5. Performance comparison with other well-performing methods in the literature.
Table 5. Performance comparison with other well-performing methods in the literature.
ReferenceAlgorithmSensitivitySpecificityAUC
Zhu et al. [77]Weighted ELM0.982-0.978
Alkhatib et al. [78]DNN0.955-0.990
Yotsawat et al. [79]CS-NNE-0.9360.980
Ileberi et al. [19]GA-RF72.56-0.950
Kalid et al. [80]DT-NB0.8721.000-
Mrozek et al. [81]Random forest-SMOTE0.829-0.910
Carta et al. [82]Stochastic ensemble0.915-0.876
Xie et al. [83]XGBoost-SMOTE0.988-0.970
Saheed et al. [84]GA-SVM0.9630.963-
Verma et al. [85]PSO-based Ensemble model0.97--
Padhi et al. [86]RHSO0.951--
Ganji et al. [87]DRN-SSPO0.9120.902-
This paperProposed IG-GAW0.9970.9940.990
Table 6. Feature sets from the German dataset.
Table 6. Feature sets from the German dataset.
Feature-Selection MethodFeatures
Complete feature setStatus of existing checking account, duration in month, credit history, purpose, credit amount, savings account, present employment since, installment rate as a percentage of disposable income, personal status and sex, other debtors, present residence since, property, age, other installment plans, housing, number of existing credits at this bank, job, number of dependents, telephone, foreign worker
IGStatus of existing checking account, duration in month, credit history, purpose, credit amount, savings account, present employment since, installment rate as a percentage of disposable income, personal status and sex, other debtors, property, age, other installment plans, housing, number of dependents, foreign worker
GAWStatus of existing checking account, duration in month, credit history, purpose, credit amount, savings account, present employment since, property, age, other installment plans, housing, number of dependents, foreign worker
IG-GAWCredit amount, status of existing checking account, duration in months, age, credit history, purpose, property, present employment since, and housing
Table 7. Performance comparison using the German dataset.
Table 7. Performance comparison using the German dataset.
ClassifierSensitivitySpecificityAUCG-Mean
AdaBoost0.7850.8920.8100.837
LR0.6880.8130.7000.748
RF0.7960.9040.8300.850
SVM0.6490.7920.6500.716
DT0.6300.7870.6400.704
ELM0.7040.8300.7100.763
IG-ELM0.7960.9030.8100.847
GAW0.8200.9250.8600.871
Proposed IG-GAW0.9040.9460.9100.925
Table 8. Feature sets from the Taiwan dataset.
Table 8. Feature sets from the Taiwan dataset.
Feature Selection MethodFeatures
Complete feature setID, LIMIT_BAL, SEX, EDUCATION,
MARRIAGE, AGE, PAY_0, PAY_2, PAY_3,
PAY_4, PAY_5, PAY_6, BILL_AMT1, BILL_AMT2,
BILL_AMT3, BILL_AMT4, BILL_AMT5,
BILL_AMT6, PAY_AMT1, PAY_AMT2,
PAY_AMT3, PAY_AMT4, PAY_AMT5, PAY_AMT6
IGSEX, PAY_0, PAY_2, PAY_3, PAY_4, PAY_5,
PAY_6, BILL_AMT1, BILL_AMT2, BILL_AMT3,
BILL_AMT4, BILL_AMT5, BILL_AMT6,
PAY_AMT1, PAY_AMT2, PAY_AMT3, PAY_AMT4,
PAY_AMT5, PAY_AMT6
GAWPAY_0, PAY_2, PAY_4, PAY_5, BILL_AMT1,
BILL_AMT2, BILL_AMT3, BILL_AMT4,
BILL_AMT5, BILL_AMT6, PAY_AMT1, PAY_AMT2,
PAY_AMT3, PAY_AMT4, PAY_AMT5, PAY_AMT6
IG-GAWBILL_AMT1, BILL_AMT2, BILL_AMT3,
BILL_AMT4, BILL_AMT5, BILL_AMT6, PAY_AMT1,
PAY_AMT2, PAY_AMT3, PAY_AMT6, PAY_AMT4,
PAY_AMT5, PAY_0, and PAY_2
Table 9. Performance comparison using the Taiwan dataset.
Table 9. Performance comparison using the Taiwan dataset.
ClassifierSensitivitySpecificityAUCG-Mean
AdaBoost0.8700.8900.8700.880
LR0.6250.8370.6400.723
RF0.8290.9140.8400.870
SVM0.6260.8190.6500.716
DT0.5740.7730.6100.666
ELM0.7100.8850.7300.793
IG-ELM0.8740.9110.8900.892
GAW0.8990.9200.9000.909
Proposed IG-GAW0.9450.9610.9400.952
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mienye, I.D.; Sun, Y. A Machine Learning Method with Hybrid Feature Selection for Improved Credit Card Fraud Detection. Appl. Sci. 2023, 13, 7254. https://doi.org/10.3390/app13127254

AMA Style

Mienye ID, Sun Y. A Machine Learning Method with Hybrid Feature Selection for Improved Credit Card Fraud Detection. Applied Sciences. 2023; 13(12):7254. https://doi.org/10.3390/app13127254

Chicago/Turabian Style

Mienye, Ibomoiye Domor, and Yanxia Sun. 2023. "A Machine Learning Method with Hybrid Feature Selection for Improved Credit Card Fraud Detection" Applied Sciences 13, no. 12: 7254. https://doi.org/10.3390/app13127254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop