Figures
Abstract
Credit card fraud identification is an important issue in risk prevention and control for banks and financial institutions. In order to establish an efficient credit card fraud identification model, this article studied the relevant factors that affect fraud identification. A credit card fraud identification model based on neural networks was constructed, and in-depth discussions and research were conducted. First, the layers of neural networks were deepened to improve the prediction accuracy of the model; second, this paper increase the hidden layer width of the neural network to improve the prediction accuracy of the model. This article proposes a new fusion neural network model by combining deep neural networks and wide neural networks, and applies the model to credit card fraud identification. The characteristic of this model is that the accuracy of prediction and F1 score are relatively high. Finally, use the random gradient descent method to train the model. On the test set, the proposed method has an accuracy of 96.44% and an F1 value of 96.17%, demonstrating good fraud recognition performance. After comparison, the method proposed in this paper is superior to machine learning models, ensemble learning models, and deep learning models.
Citation: Jiang S, Liao X, Feng Y, Gao Z, Onasanya BO (2024) A new fusion neural network model and credit card fraud identification. PLoS ONE 19(10): e0311987. https://doi.org/10.1371/journal.pone.0311987
Editor: Maleika Heenaye- Mamode Khan, University of Mauritius, MAURITIUS
Received: December 21, 2023; Accepted: September 29, 2024; Published: October 28, 2024
Copyright: © 2024 Jiang et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data underlying the results presented in the study are available from https://www.heywhale.com/mw/dataset/5b56a592fc7e9000103c0442/content.
Funding: This work is sponsored by the Natural Science Foundation of Chongqing, P.R. China (Grant No. 2024NSCQ-LZX0121, 2024NSCQ-LZX0120, 2023TIAD-ZXX0017, CSTB2023NSCQ-LZX0135) awarded to YF, the Scientific and Technological Research Program of Chongqing Municipal Education Commission, P.R. China (KJZD-K202301023) awarded to ZG, the Scientific and Technological Research Program of Wanzhou District, P.R. China (WZSTC-20230309) awarded to ZG, the National Natural Science Foundation of China (12201086) and Program of Chongqing Municipal Education Commission, P.R. China (KJQN202201209, 233356) awarded to YF. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
In recent years, China’s credit cards have made significant progress. Credit cards play a huge role in the market economy and national economy. With the rapid development of credit cards, the fraudulent behavior of credit cards is also increasing. Credit card fraud [1] refers to fraudulent behavior committed with the purpose of illegally occupying resources, in the process of purchasing, using, and consuming credit cards, in violation of relevant bank regulations or national laws related to credit card issuance.
Common credit card fraud behaviors [2] include: (1) Using forged or invalidated credit cards for consumption, overdraft, or fraud. (2) Pretending to be someone else’s credit card for consumption, overdraft, or fraud. (3) Intentionally or maliciously overdrawing a credit card and long-term denial. The fraudulent behavior of credit cards poses a huge threat to the safe and stable development of banks and the financial industry. Therefore, researching and constructing a machine learning model and method that can accurately identify credit card fraudulent transactions is of great significance for the scientific management and rapid and healthy development of credit card business.
In order to avoid the huge risks brought by credit card fraud [3], the banking and financial industries urgently need a credit card fraud identification system that can effectively identify whether there is fraudulent behavior in transaction records.
The current credit card fraud identification method still has the problem that accuracy and F1 score is not high enough. At the same time, there is a great demand for a method with low time complexity and low space complexity. At present, neural networks are a very popular direction, and deep feedforward networks and wide neural networks can optimize credit card fraud detection models to a certain extent, improving their detection accuracy and correctness. The application of these two networks in credit card fraud detection is still relatively small. The method proposed in this paper not only provides a new model for this problem, but also improves the recognition accuracy.
This article will propose a fusion neural network that combines deep neural networks and width neural networks, and apply it to credit card fraud identification. Due to the excellent performance of neural networks [4] in binary classification problems, this paper will construct a credit card fraud recognition model based on neural networks. The first method is to deepen the layers of the neural network to improve the prediction accuracy of the model. The second method is to increase the hidden layer width of the neural network to improve the prediction accuracy of the model. Finally, by combining deep neural networks and wide neural networks, a new fusion neural network model is proposed and applied to credit card fraud identification. The model in this article can effectively improve the accuracy of credit card fraud identification and provide a model with good time and space complexity
In the following chapters, this article first introduces the current research status of credit card fraud identification. In the method section of the article, this paper will introduce the Independent variable and dependent variable respectively, deep neural network, width neural network, fusion neural network, model training, model evaluation criteria. In the numerical experiment section, this paper will discuss data set, analysis of influencing factors, experimental results, comparison with other models, time and space complexity analysis. The last part is a conclusion.
Related works
At present, credit card recognition technology mainly relies on machine learning [5,6] methods. Machine learning algorithms can be divided into two categories: Supervised learning [7] and Semi-Supervised Learning. The methods of machine learning mainly include: Logistic regression [8], K Nearest Neighbor (KNN) [9], Decision tree [10], Bayesian network [11], support vector machine (SVM) [12].
At present, many scholars have conducted in-depth research on credit card fraud detection based on machine learning methods, including: algorithm based on decision tree and boolean logic [13], cost sensitive decision tree algorithm [14], risk induced cost sensitive minimal Bayesian algorithm [15], parallel fuzzy neural network [16], a framework for detecting potential fraudulent transactions in credit card transactions mining based on CNN [2], support vector machine model with Spark [17]. Some scholars have introduced semi supervised learning [18] and ensemble learning [19] into the problem of credit card fraud identification.
Some scholars have improved the credit card scoring system itself to obtain a better credit card scoring model to increase the success rate of fraud detection. They used a combination of logistic regression and weighted evidence to construct a hybrid credit scoring model, which improved the accuracy and predictive power of credit scoring and reduced the risk of credit fraud [20]. In the realm of deep learning, scholars have used supervised learning algorithms such as artificial neural networks (ANN), support vector machines (SVM), and deep neural networks (DNN) to predict fraud risks, achieving high recognition accuracy [21,22]. In order to discover the patterns of credit card fraud transactions, fuzzy neural networks are used, and it is pointed out that fuzzy neural networks can discover fraud patterns in a short period of time. Some scholars use self-encoder neural networks to detect credit card fraud [23].
At present, some scholars have conducted in-depth research on credit card fraud detection using ensemble learning methods [24,25]. Through supervised learning on the dataset using random forest, logistic regression, and AdaBoost classifiers, the detection accuracy was greatly improved, and a research idea was proposed to combine LSTM prediction with random forest based on HMM [26,27]. Some scholars have designed a credit card fraud prediction model based on clustering analysis and ensemble support vector machines, using the idea of ensemble learning to further address the imbalance of data and improve the recognition of minority classes by classifiers. Finally, testing is conducted [28].
In the field of heuristic learning, many authors have done relevant work. For example, some scholars have proposed tuning machine learning models using a group search firefly algorithm for fraud detection [29]. Some scholars have proposed metaheuristics based hyperparameters optimization for credit card fraud detection [30]. Some scholars have used firefly metaheuristics to optimize the adaboost algorithm [31]. Some scholars have optimized the feature selection for credit card fraud identification based on the Oppositional Cat Swarm Optimization method [32]. Some scholars optimize the XGBoost model based on heuristic optimization methods to identify fraud [33]. Some scholars have constructed the adaboost algorithm based on the SNS heuristic optimization method [34]. In general, the current predictive model of heuristic optimization is a popular trend. These methods can optimize fraud detection to a certain extent.
In the research on credit card fraud identification based on neural networks, the current methods mainly include: Neural Network Ensemble with Feature Engineering [35], Stacked Sparse Autoencoder Approach [36], Method based on deep convolutional neural network [37], Graph Neural Network-based Method [38], Method based on competitive graph neural network [39]. These methods can promote the application of neural networks in credit card fraud identification.
Fraud identification methods
In this article, a credit card fraud recognition model based on deep neural networks and wide neural networks is proposed. In order to establish this model, this paper must determine the dependent and independent variables of the credit card fraud identification problem, build a deep neural network model, build a width neural network model, study the basic principle and structure of the fusion neural network model that combines the deep neural network and the width neural network, then train the these model, and finally use the test set to test and evaluate. The flowchart for modeling is shown in Fig 1.
Independent variable and dependent variable
The content of this article is credit card fraud identification. Credit card fraud identification is to predict whether a credit card is fraudulent based on several attributes and transaction amounts of each transaction. In this article, the j-th attribute of the i-th transaction is defined as:
(1)
This article uses all 28 variables and transaction amount in the data set as input variables to predict whether there is fraudulent behavior. The model we want to establish is as follows:
(2)
where value of yi is 0 or 1, if yi = 1, it indicates that the transaction is fraudulent, and if yi = 0, it indicates that the transaction is not fraudulent. f is the model we want to establish, where vi,j represent the values of the i-th transaction and the j-th attribute. ai is the amount of the i-th transaction. In order to construct the training data set for the model, this paper record the output variables of the model training set as:
(3)
where Y is an m-dimensional column vector, this paper note the input variable of the model training set as:
(4)
This paper use Y and V to obtain this model parameters, and then use this model to predict whether there is fraudulent behavior in credit card transactions, and test the accuracy of the model on the test set.
Deep neural network
The credit card fraud recognition in this article is a typical nonlinear binary classification problem, which is applicable to the nonlinear classification problem in the input space. The input variables for this problem are the 28 attributes and transaction amount of the credit card, with outputs of 0 or 1, where 0 represents normal transactions and 1 represents fraudulent transactions. Neural network models can effectively handle nonlinear classification problems because they can characterize any nonlinear function. In order to achieve better fitting results, this article deepens the layers of the neural network to improve the model’s generalization ability. The neural network in this article consists of 6 hidden layers and 6 relu layers [20]. Finally, it is output as a category variable through the softmax layer. The forward calculation formula for deep neural networks is as follows:
(5)
where ω(1) is the coefficient of the first layer model, ω(2) is the coefficient of the second layer model, ω(6) is the 6-th layer model coefficient. n1 is the number of neurons in the first layer, n2 is the number of neurons in the second layer, n6 is the number of neurons in the 6-th layer, v is the input variable, a is the amount of each transaction, vi is the data in the training set. x(1) is the output of the first hidden layer, x(2) is the output of the second hidden layer, x(6) is the output of the 6-th hidden layer, f is the nonlinear transformation function of the hidden layer, L is the number of categories (Take L as 2 in this article), b is the bias term. Fig 2 is the structural diagram of the deep neural network. The relu is a nonlinear activation function. Its definition is as follows:
(6)
Width neural network
The neural network model can effectively handle nonlinear classification problems, and deepening the neural network can improve the model’s generalization ability. In order to achieve better fitting results, the second solution in this article is to improve the width of the neural network to improve the prediction accuracy of the model. The model consists of 2 hidden layers and 2 relu layers, which are then output as category variables through the softmax layer. The calculation formula for a wide neural network is as follows:
(7)
where n1 is the number of neurons in the first layer. In the width neural network, this paper specify that its value is at least greater than 1000 dimensions, n2 is the number of neurons in the second layer. This is a binary classification problem, therefore n2 = 2. ω(1) is the coefficient of the first layer model, ω(2) is the coefficient of the second layer model. v is the input variable, a is the amount of each transaction, vi is the data in the training set, x(1) is the output of the first hidden layer, x(2) is the output of the second hidden layer, f is the nonlinear transformation function of the hidden layer, L is the number of categories (Take L as 2 in this article), b is the bias term. The relu is a nonlinear activation function. Fig 3 is the structural diagram of the width neural network in this article.
Fusion neural network
In order to further improve the recognition accuracy of the model, this paper combine deep neural networks with width neural networks. First, after normalization, the input is processed through a deep neural network and a width neural network, respectively. Second, the outputs of the deep and width neural networks are merged into a vector, which is processed through a hidden layer and a softmax layer to output as the class variables of the neural network. this paper refer to it as a fusion neural network, and the structure of the fusion neural network is as Fig 4.
The calculation formula of the deep neural network in the fusion neural network is as follows:
(8)
The calculation formula of width neural network in fusion neural network is as follows:
(9)
The calculation formula of neural network fusion is as follows:
(10)
where ⊕ stands for splicing two vectors, d stands for deep neural network, w stands for width neural network and f stands for fusion neural network. Softmax function is the category output function of neural network, and its definition is as follows:
(11)
Model training
The training of the model in this paper is equivalent to minimize the cross entropy loss, as follows:
(12)
where pi,j is the probability of taking the j-th category from the i-th sample (Calculated forward by the neural network model). yi,j is the sample label, which represents whether the i-th sample belongs to class j. If the value is 1, it represents belonging to that class, and if it is 0, it represents not belonging to that class.
In neural network learning, the Loss function is used to guide model learning. Generally, the smaller its value is, the better the performance of the model is. This means that the fewer errors the model produces, the higher the accuracy of classification. Therefore, when training neural networks, goal of this paper is often to minimize the loss function as far as possible. In order to achieve this goal, this paper need to use some optimization algorithms to adjust the parameters of the model. Here, this paper use the rmsprop algorithm to solve the optimization problem. Rmsprop algorithm [40] is a common variant of gradient descent algorithm. its basic idea is to adjust the learning rate according to the square root of gradient to better adapt to different data inputs. The rmsprop algorithm is very common in deep learning, and this paper will not elaborate on its specific details here. In summary, by adopting this optimization algorithm, this paper can better optimize the performance of the model and obtain more accurate and refined prediction results.
Model evaluation criteria
For the binary classification problem with balanced data sets, accuracy is a useful evaluation metric.At the same time, for any binary classification problem, the F1 score is also a useful evaluation metric. Therefore, two evaluation indicators are used to evaluate the performance of the prediction model in this article, namely accuracy and F1 score [41].
The calculation formula for the accuracy of the classification model is as follows:
(13)
The formula for calculating the F1 score of the classification model is:
(14)
The precision score calculation formula for the classification model is:
(15)
The formula for calculating the recall score of the classification model is:
(16)
where TP is the number of true cases, TN is the number of true negative cases, FP is the number of false positive cases, and FN is the number of false negative cases.
Numerical experiment
Data set
The data set [42] used in this article contains transaction information from European cardholders using credit cards in September 2013. This data set shows transactions that occurred within two days, with 492 fraud cases out of 284807 transactions. The data set is highly imbalanced, with fraud accounting for 0.172% of all transactions (see Fig 5). The original dataset has undergone desensitization and PCA processing. 28 anonymous variables are the principal components obtained by PCA, and the only variables that have not been processed by PCA are time and amount. Time is the interval between each transaction and the first transaction in the data set, measured in seconds; amount is the transaction amount, class is a Categorical variable, which is 1 in case of fraud, and 0 otherwise. We first randomly downsample the non-fraud data in the dataset, as the dataset is imbalanced and the non-fraud data greatly outweighs the fraud data. After downsampling, the categories of the dataset in this article are in a balanced state. Subsequently, 80% of the data was randomly sampled as the training set, and 20% as the testing set. Table 1 is an example of this data. When inputting data into a neural network for calculation, each dimension of the data is subjected to corresponding normalization processing.
(a) Proportion of normal transactions and fraudulent transactions of raw data. (b) The proportion of normal transactions and fraudulent transactions of the data after down sampling.
Fig 6 shows the feature distribution on the training set of this data set, and different colors represent different values. On the far right of the picture are category variables, yellow represents fraudulent transactions, and green represents normal transactions. The horizontal axis is the attributes and fraud categories of each transaction. This paper plot each feature of each transaction record according to its value. The closer the color is to green, the smaller the value, and the closer the color is to yellow, the larger the value. The vertical axis of the graph is the transaction id, and the upper part of the graph is normal transactions, while the lower part is fraudulent transactions. From the graph, it can be seen that there are some differences in the left features between the data in the upper part and the data in the lower part. Therefore, these features can be used as effective features for fraud detection.
This paper extract the first, second, and fifth attributes and group the data based on the label category. Draw their scatter plots and histograms to partially showcase the characteristics of this data, as shown in Fig 7. As can be seen from the figure, there are different distributions of the three variables v1, v2, and v5 for normal data and fraudulent data. These three variables can be used to predict whether there is fraudulent behavior on credit cards.
Analysis of influencing factors
This article first studies the correlation between each credit card attribute and whether it is fraudulent. The calculation formula for the correlation coefficient is as follows:
(17)
where Cov represents the covariance of two variables, and Var represents the variance of the variable [43]. This paper visually display the calculation results in Fig 8, and it can be seen that some attributes (v1, v2, v5, v6, v7, v8, v10, v20, v23, v25) have a certain correlation with whether fraud occurs. These attributes highly correlated with the label can be used to detect whether there is fraud on the credit card transaction.
This paper take the 10-th attribute of a credit card as an example to analyze the relationship between various variables and output variables in the data set. We draw a distribution of data labeled as fraudulent and a distribution of data labeled as normal, as shown in Fig 9. As can be seen from the figure, fraudulent transactions and normal transactions have different distributions on the 10-th variable. Therefore, this feature can be considered as an effective variable for fraud identification.
The upper part is the distribution of the tenth variable of normal data, and the lower part is the distribution of the tenth variable of fraudulent data.
To further understand the impact of each input on the final result, this article uses interpretable AI technology to analyze the data. This article uses Logstic regression to fit the model [44]. The parameters of each variable in the model can reflect the importance of each variable to a certain extent. The greater the weight, the higher the correlation between the variable and the label. This article shows the weights of each input, as shown in the Table 2.
Experimental results
This paragraph mainly describes the hyperparameters and experimental results of deep neural networks. When inputting data into a neural network, normalization is the first step. Set the dimensions of the all hiden layers to 64, and the dimensions of the sixth layer to 2. Finally, through the softmax layer, the output is a category variable. The nonlinear activation function uses the relu function. The maximum number of iterations is set to 100, and the batch size is set to 64. In this paper, numerical experiments are carried out, and the average accuracy is 94.92% and the average F1 score is 94.44%, the recall score is 90.43%, the precision score is 98.83%.
This paragraph mainly describes the hyperparameters and experimental results of the width neural network.The first layer dimension of the width neural network is set to 5000 dimensions, and the second layer dimension is set to 2. After passing through the softmax layer, the output is a category variable. When inputting data into a neural network, normalization is the first step. The maximum number of iterations is set to 1000, and the batch size is set to 64. Using a wide neural network and conducting 10 calculations repeatedly. The average accuracy of experiments is 95.43%, and the average F1 score is 95.03%, the recall score is 91.49%, the precision score is 98.85%.
This paragraph mainly describes the hyperparameters and experimental results of the fusion network. Before entering the deep neural network and the wide neural network for computation, the data needs to be normalized first. Set the number of layers for the deep neural network to 6, with an output length of 64 for all hidden layers. The last level output dimension of the deep branch is 2. The first layer dimension of the width neural network is 5000, and the second layer length is 2. Splice the output of the width neural network and the depth neural network together, pass through a hidden layer with a length of 16, and then pass through a dense and softmax layer to output as a category variable. This paper set the batch size as 64.
Some of the hyperparameters in this article, such as the number of hidden units per layer in deep networks and the number of hidden units in wide networks, are mainly optimized based on 5-fold cross-validation on the training set. Regarding the choice of activation function, we have chosen the relu function for all hidden layers because it can effectively prevent the problems of vanishing or exploding gradients. The number of layers and batch size of the neural network are mainly selected with reference to relevant paper and experience on neural networks [45].
On the test set, the experimental results of this method are shown in Fig 10.
Table 3 shows the statistical analysis of 10 experiments on the fusion neural network. The average accuracy and F1 score were 96.45% and 96.17%, respectively. The standard deviations of the two indicators in each experiment were 0.39% and 0.34%.
It can be seen that the average accuracy of the fusion neural network is 96.44%, and the average F1 score is 96.17%, the recall score is 93.62%, the precision score is 98.88%.
Comparison with other models
This article compares the proposed method with machine learning algorithms. Firstly, the accuracy of the KNN algorithm [9] is compared as shown in Table 4. The number of neighbors in the KNN algorithm is set to 5, and the distance measurement is based on Euclidean distance. The weight is weighted using the inverse proportion of distance. Secondly, this paper compared the accuracy of the decision tree [14]. The maximum depth of the decision tree is set to 15, the splitting index is used as the Gini index, and the maximum number of features is set to 50. Finally, this paper compared the accuracy of Bayesian algorithm [15].
From the above analysis of results, it can be concluded that the method (deep neural network, width neural network, fusion neural network) proposed in this paper is better than machine learning algorithms such as KNN and decision trees, Bayes, etc. In addition, it is worth noting that compared with the width neural network and the depth neural network, the fusion neural network proposed in this paper has made some progress to a certain extent.
To further verify the performance of the fusion network, this paper compares the model with deep learning models and ensemble models. The ensemble models this paper want to compare include: random forest and adaboost [25]. The deep learning models this paper want to compare include: long short-term memory network (LSTM) and convolutional neural network (CNN) [46]. The number of weak classifiers in the random forest is set to 100, the segmentation evaluation metric is set to gini, the maximum depth is set to 3, and the maximum number of features is set to 5. The base learner of adaboost is set to decision tree, the number of base learners is set to 100, the learning rate is set to 1, and the training algorithm is SAMME. The number of hidden layer units in the LSTM model is set to 32. This paper treat each feature as a time step and input it into a recurrent neural network. The hidden layer activation function uses relu, and the last layer activation function uses softmax. The CNN model has 4 one-dimensional convolutional layers, each with 8 channels. The activation function is relu, and the last layer is a fully connected layer with softmax activation function. The comparison results are shown in Table 5. From Table 5, it can be seen that the detection F1 score of this model is slightly higher than that of deep learning models and ensemble learning models.
Time and space complexity analysis
The time complexity of forward calculation of the model is an important indicator, because the time of forward calculation has a great impact on the successful deployment of the model. For a fair comparison, both models were implemented on Huawei computers, with a CPU frequency of 1.9GHz and 16GB of memory. Table 6 shows the Time complexity comparison between the model in this paper and several models to be compared.
(Calculation time for 100 forward inferences).
The wide neural network model in this article consumes 0.039 seconds for forward inference of 100 data points, while the deep neural network model consumes 0.041 seconds. The calculation time for 100 data points in the fusion neural network inference is 0.039 seconds. Generally speaking, a neural network model with 100 forward inferences less than 1 second can meet the actual deployment requirements [47]. Therefore, the time complexity of the model in this paper can meet the needs of fast calculation.
This paper measures the space complexity of the model by the size of the model file exported from the software. Table 7 shows the test results of space complexity of this model and space complexity of other models. The deep neural network model consumes 273 KB of memory, and the wide neural network model consumes 1894 KB of memory, the memory consumed by the fusion neural neural network is 2170KB. Space complexity of the model in this paper is very small and can be ignored compared with the current computers and servers. Therefore, the space complexity of this model can meet the requirements.
Conclusions
This paper proposes a fusion network model combining deep neural networks and wide neural networks, and applies it to credit card fraud identification. The width neural network proposed in this article achieved an accuracy of 95.43% and an F1 value of 95.03% on the test set. The constructed deep neural network achieved an accuracy of 94.92% and an F1 score of 95.03% on the test set. The fusion neural network constructed in this article achieved an average accuracy of 96.44% and an F1 value of 96.17%. This article compares the method with several traditional machine learning models, including decision trees, KNN, and Bayesian networks. At the same time, this article also compares the model with the ensemble model and deep learning model. From the experimental results, method of this paper has achieved good results. From a time complexity analysis, it takes 0.359 seconds to perform 100 inferences. In terms of spatial complexity, the size of model of this paper can be negligible compared to current computers and servers.
This article combines deep neural networks and wide neural networks to improve detection accuracy. However, there is still a problem that the binding mechanism is not deep enough. In future research, we will delve deeper into the integration mechanism of deep neural networks and wide neural networks.
Currently, the transformer model is a very popular deep learning model. In our future work, we will try to propose a similar credit card fraud identification model to further improve the accuracy of identification. At the same time, we will also conduct experiments on more features and larger-scale data sets.
References
- 1. Ravi S., Kumar J. T., Joseph Dr. L., Kunjeti S. R., Saniboina N. V., and Endela S. R. An unique methodology for credit card fraud detection based on convolutional neural network. Alinteri Journal of Agriculture Sciences, 2021, 36(1), 277–280. https://doi.org/10.47059/alinteri/v36i1/ajas21041.
- 2. Fu K., Cheng D., Tu Y., and Zhang L. Credit card fraud detection using convolutional neural networks. In Neural Information Processing, Lecture Notes in Computer Science. 2016, 483–490. https://doi.org/10.1007/978-3-319-46675-0_53.
- 3. Van-Vlasselaer V., Bravo C., Caelen O., Eliassi-Rad T., Akoglu L., Snoeck M., and Baesens B. APATE: A novel approach for automated credit card transaction fraud detection using network-based extensions. Decision Support Systems, 2015, 75(7), 38–48.https://doi.org/10.1016/j.dss.2015.04.013.
- 4. Tong C., Li J., and Zhu F. A convolutional neural network based method for event classification in event-driven multi-sensor network. Computers and Electrical Engineering, 2017, 60(1), 90–99.https://doi.org/10.1016/j.compeleceng.2017.01.005.
- 5. Li L., Su X., Zhang Y., Lin Y., and Li Z. Trend modeling for traffic time series analysis: an integrated study. IEEE Transactions on Intelligent Transportation Systems, 2015, 16(6), 3430–3439. https://doi.org/10.1109/tits.2015.2457240.
- 6. Shi Q. X., Li Z. L, Chen X. Q. A traffic breakdown model based on queueing theory. Networks and spatial economics, 2014, 14(3), 485–504. https://doi.org/10.1007/s11067 -014-9246-6.
- 7. Ke J., Zheng H., Yang H., and Chen X. (Michael). Short-term forecasting of passenger demand under on-demand ride services: A spatio-temporal deep learning approach. Transportation Research Part C: Emerging Technologies, 2017, 85(12),591–608.https://doi.org/10.1109/ictis.2017.8047810.
- 8. Hauser R. P., and Booth D. Predicting Bankruptcy with Robust Logistic Regression. Journal of Data Science, 2021, 9(4), 565–584. https://doi.org/10.6339/jds.201110_09(4).0006.
- 9. Tran T. N., Wehrens R., and Buydens L. M. C. KNN-kernel density-based clustering for high-dimensional multivariate data. Computational Statistics and Data Analysis, 2006, 51(2), 513–525. https://doi.org/10.1016/j.csda.2005.10.001.
- 10. Munmun F. A., and Khatun S. A hybrid method: hierarchical agglomerative clustering algorithm with classification techniques for effective heart disease prediction. International Journal of Research and Innovation in Applied Science, 2022, 7(7), 56–60.https://doi.org/10.51584/ijrias.2022.7704.
- 11. Udelnow A., Leinung S., Grochola , Lukasz F., Henne-Bruns D., and Wfcrl P. Prediction of outcome in multiorgan resections for cancer using a Bayes-network. Hepato-Gastroenterology, 2013, 60(125), 1009–1013.https://doi.org/10.5754/hge10510.
- 12. Tong S., and Koller D. Support vector machine active learning with applications to text classification. The Journal of Machine Learning Research, 2002, 2(1), 999–1006. https://doi.org/10.1162/153244302760185243.
- 13.
Kokkinaki, A. I. On atypical database transactions: identification of probable frauds using machine learning for user profiling. Proceedings 1997 IEEE Knowledge and Data Engineering Exchange Workshop, Newport Beach, CA, USA, 2002, 107–113. https://doi.org/10.1109/kdex.1997.629848.
- 14. Sahin Y., Bulkan S., and Duman E. A cost-sensitive decision tree approach for fraud detection. Expert Systems with Applications, 2013, 40(15), 5916–5923. https://doi.org/10.1016/j.eswa.2013.05.021.
- 15.
Bahnsen, A. C., Stojanovic, A., Aouada, D., and Ottersten, B. Cost sensitive credit card fraud detection using Bayes minimum risk. 2013 12th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA. 2013, 333–338. https://doi.org/10.1109/icmla.2013.68.
- 16.
Syeda, M., Zhang, Y. Q., and Pan, Y. Parallel granular neural networks for fast credit card fraud detection. IEEE International Conference on Fuzzy Systems. IEEE, 2002, 572–577. https://doi.org/10.1109/fuzz.2002.1005055.
- 17.
Kumar, S., Gunjan, V. K., Ansari, M. D., and Pathak, R. Credit card fraud detection using support vector machine. In Proceedings of the 2nd International Conference on Recent Trends in Machine Learning, IoT, Smart Cities and Applications,Lecture Notes in Networks and Systems. 2022, 27–37. https://doi.org/10.1007/978-981-16-6407-6_3.
- 18. Carcillo F., Le Borgne Y. A., Caelen O., Kessaci Y., Oblé F., and Bontempi G. Combining unsupervised and supervised learning in credit card fraud detection. Information Sciences, 2021, 557, 317–331. https://doi.org/10.1016/j.ins.2019.05.042.
- 19. Zareapoor M., and Shamsolmoali P. Application of credit card fraud detection: based on bagging ensemble classifier. Procedia Computer Science, 2015, 48, 679–685.https://doi.org/10.1016/j.procs.2015.04.201.
- 20. Chen K., Yadav A., Khan A., and Zhu K. Credit fraud detection based on hybrid credit scoring model. Procedia Computer Science, 2020, 167, 2–8. https://doi.org/10.1016/ jprocs.2020.03.176.
- 21. Dharwadkar N. V., and Patil P. S. Customer retention and credit risk analysis using ANN, SVM and DNN. International Journal of Society Systems Science, 2018, 10(4), 316–332. https://doi.org/10.1504/IJSSS.2018.095601.
- 22. Bhattacharyya S., Jha S., Tharakunnel K., and Westland J. C. Data mining for credit card fraud: A comparative study. Decision support systems, 2011, 50(3), 602–613.https://doi.org/10.1016/j.dss.2010.08.008.
- 23.
Chen, J., Shen, Y., and Ali, R. Credit card fraud detection using sparse autoencoder and generative adversarial network. In 2018 IEEE 9th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON) 2018,1054–1059. IEEE. https://doi.org/10.1109/IEMCON.2018.8614815.
- 24.
Xuan, S., Liu, G., Li, Z., Zheng, L., Wang, S., and Jiang, C. Random forest for credit card fraud detection. In 2018 IEEE 15th international conference on networking, sensing and control (ICNSC), 2018, 1–6. IEEE. https://doi.org/10.1109/ICNSC.2018.8361343.
- 25. Randhawa K., Loo C. K., Seera M., Lim C. P., and Nandi A. K. Credit card fraud detection using AdaBoost and majority voting. IEEE access, 2018, 6, 14277–14284. https://doi.org/.101109/ACCESS.2018.2806420.
- 26. Benchaji I., Douzi S., and El Ouahidi B. Credit card fraud detection model based on LSTM recurrent neural networks. Journal of Advances in Information Technology, 2021, 12(2), 113–118.
- 27. Srivastava A., Kundu A., Sural S., and Majumdar A. Credit card fraud detection using hidden Markov model. IEEE Transactions on dependable and secure computing, 2008, 5(1), 37–48. https://doi.org/10.1109/TDSC.2007.70228.
- 28. Wang C., & Han D. Credit card fraud forecasting model based on clustering analysis and integrated support vector machine. Cluster Computing, 2019, 22(Suppl 6), 13861–13866.
- 29. Jovanovic D., Antonijevic M., Stankovic M., Zivkovic M., Tanaskovic M., and Bacanin N. Tuning machine learning models using a group search firefly algorithm for credit card fraud detection. Mathematics, 2022, 10(13), 2272.
- 30. Tayebi M., and El Kafhali S. Performance analysis of metaheuristics based hyperparameters optimization for fraud transactions detection. Evolutionary intelligence, 2024, 17(2),921–939.
- 31.
Petrovic, A., Bacanin, N., Zivkovic, M., Marjanovic, M., Antonijevic, M., and Strumberger, I., The adaboost approach tuned by firefly metaheuristics for fraud detection. In 2022 IEEE world conference on applied intelligence and computing (AIC). 2022, pp. 834–839). IEEE.
- 32. Prabhakaran N., and Nedunchelian R. Oppositional Cat Swarm Optimization‐Based Feature Selection Approach for Credit Card Fraud Detection. Computational Intelligence and Neuroscience, 2023, 2023(1), 2693022. pmid:36688222
- 33.
Petrovic, A., Antonijevic, M., Strumberger, I., Jovanovic, L., Savanovic, N., and Janicijevic, S. The xgboost approach tuned by tlb metaheuristics for fraud detection. In Proceedings of the 1st international conference on innovation in information technology and business (ICIITB 2022) (Vol. 104, p. 219). Springer Nature.
- 34.
Djuric, M., Jovanovic, L., Zivkovic, M., Bacanin, N., Antonijevic, M., & Sarac, M., February). The AdaBoost approach tuned by SNS metaheuristics for fraud detection. In Proceedings of the International Conference on Paradigms of Computing, Communication and Data Sciences: PCCDS 2022. 2023, (pp. 115–128). Singapore: Springer Nature Singapore.
- 35. Esenogho E., Mienye I. D., Swart T. G., Aruleba K., and Obaido G. A neural network ensemble with feature engineering for improved credit card fraud detection. IEEE Access, 2022, 10, 16400–16407.
- 36. Ebiaredoh-Mienye S. A., Esenogho E., and Swart T. G. Artificial neural network technique for improving prediction of credit card default: A stacked sparse autoencoder approach. International Journal of Electrical and Computer Engineering, 2021, 11(5), 4392.
- 37. Chen J. I. Z., and Lai K. L. Deep convolution neural network model for credit-card fraud detection and alert. Journal of Artificial Intelligence, 2021, 3(02), 101–112.
- 38. Liu Z., Dou Y., Yu P. S., Deng Y., & Peng H. Alleviating the inconsistency problem of applying graph neural network to fraud detection. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, 2020, (pp. 1569–1572).
- 39. Zhang G., Li Z., Huang J., Wu J., Zhou C., Yang J., and Gao J. efraudcom: An e-commerce fraud detection system via competitive graph neural networks. ACM Transactions on Information Systems (TOIS), 2022, 40(3), 1–29.
- 40. Hanif M. S. Similarity learning for person reidentification using projected gradient method. Journal of electronic imaging, 2019, 28(3), 1–9. https://doi.org/10.1117/1.JEI.28.3.033017.
- 41. Pereira F., Mitchell T., and Botvinick M. Machine learning classifiers and fmri: a tutorial overview. Neuroimage, 2009, 45(1), 199–209. pmid:19070668
- 42. worldline and the machine learning group, credit card fraud detection dataset, https://www.heywhale.com/mw/dataset/5b56a592fc7e9000103c0442/content.
- 43. Temizhan E., Mirtagioglu H., and Mendes M. Which correlation coefficient should be used for investigating relations between quantitative variables. Acad. Sci. Res. J. Eng. Technol. Sci, 2022, 85, 265–277.
- 44.
Das A. Logistic regression. In Encyclopedia of Quality of Life and Well-Being Research. 2024, (pp. 3985–3986). Cham: Springer International Publishing.
- 45. Choi R. Y., Coyner A. S., Kalpathy-Cramer J., Chiang M. F., and Campbell J. P. Introduction to machine learning, neural networks, and deep learning. Translational vision science & technology, 2020, 9(2), 14–14. pmid:32704420
- 46. Muppalaneni N. B., Ma M., Gurumoorthy S., Vardhani P. R., Priyadarshini Y. I., and Narasimhulu Y. CNN data mining algorithm for detecting credit card fraud. Soft computing and medical bioinformatics, 2019, 85–93.
- 47. Alzubaidi L., Zhang J., Humaidi A. J., Al-Dujaili A., Duan Y., Al-Shamma O., and Farhan L. Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. Journal of Big Data. 2021, 8(1), 1–74. https://doi.org/10.1186/s40537-021-00444–8.