Credit Risk Assessment Models in Financial Technology: A Review

Modelos para la evaluación de riego crediticio en el ámbito de la tecnología financiera: una revisión

DOI 10.22430/22565337.2679 Logotecnologicas PDF Table Figure

Received: March 24, 2023
Accepted: October 30, 2023
Available: December 20, 2023

 

How to cite / Cómo citar
F. E. Tadeo Espinoza and M. A. Coral Ygnacio, “Credit Risk Assessment Models in Financial TecnoLógicas, vol. 26, nro. 58, eXXX, 2023. https://doi.org/10.22430/22565337.2679

 


Highlights

  • Demographic and socioeconomic factors can improve the results of Credit Risk Assessment (CRA).
  • CRA systems can improve credit evaluation processes for loan granting.
  • Default is one of the biggest problems faced by financial institutions.
  • Highlights

  • Los factores demográficos y socioeconómicos permiten mejorar los resultados obtenidos en la evaluación de riesgo crediticio.
  • El uso de los sistemas de riesgo crediticio permite mejorar los procesos para el otorgamiento de crédito.
  • La morosidad es una de las mayores problemáticas a las que se enfrentan las entidades financieras.
  • Los sistemas de riesgo crediticio deben desarrollarse considerando la solución que mejor se adecue al contexto.

  • Abstract

    This review analyzes a selection of scientific articles on the implementation of Credit Risk Assessment (CRA) systems to identify existing solutions, the most accurate ones, and limitations and problems in their development. The PRISMA statement was adopted as follows: the research questions were formulated, the inclusion criteria were defined, the keywords were selected, and the search string was designed. Finally, several descriptive statistics of the selected articles were calculated. Thirty-one solutions were identified in the selected studies; they include methods, models, and algorithms. Some of the most widely used models are based on Artificial Intelligence (AI) techniques, especially Neural Networks and Random Forest. It was concluded that Neural Networks are the most efficient solutions, with average accuracies above 90 %, but their development can have limitations. These solutions should be implemented considering the context in which they will be employed.

    Keywords: credit assessment, credit risk, technology solutions, machine learning, algorithms

    Resumen

    Esta revisión analiza una selección de artículos científicos sobre la implantación de sistemas de evaluación del riesgo de crédito para identificar las soluciones existentes, las más acertadas y las limitaciones y problemas en su desarrollo. Se adoptó la declaración PRISMA del siguiente modo: se formularon las preguntas de investigación, se definieron los criterios de inclusión, se seleccionaron las palabras clave y se diseñó la cadena de búsqueda. Por último, se calcularon varios estadísticos descriptivos de los artículos seleccionados. En los estudios seleccionados se identificaron 31 soluciones, entre métodos, modelos y algoritmos. Algunos de los modelos más utilizados se basan en técnicas de Inteligencia Artificial (IA), especialmente Redes Neuronales y Bosques Aleatorios. Se concluyó que las Redes Neuronales son las soluciones más eficientes, con precisiones medias superiores al 90 %, pero su desarrollo puede tener limitaciones. Estas soluciones deben implementarse teniendo en cuenta el contexto en el que se van a emplear.

    Palabras clave: evaluación crediticia, riesgo de crédito, soluciones tecnológicas, aprendizaje automático, algoritmos

    1. INTRODUCTION

    Credit risk is defined as the classification or evaluation that is assigned to someone after having applied for a loan from a bank. The main task of Credit Risk Assessment (CAR) systems is to solve the problem of classifying customers based on their behavior or credit history [1]. Nevertheless, their results vary depending on the variables that are considered in the process, such as unpaid bills, balance in other loans, characteristics of personal accounts, monthly income, and demographic data (e.g., age and marital status) [2]. This problem has been present for a long time. However, recently, and due to the COVID-19 pandemic, it has become worse because the number of loan applications has increased significantly, and many borrowers had a poor credit history or requested amounts higher than they could pay back. As a result, their default risk assessments were deficient, and a massive number of loan defaulters resulted in financial losses for banking institutions. In spite of the gradual return to normality, the number of defaulters keeps rising. Therefore, different and more efficient solutions should be developed to minimize these losses [3].

    This is a brief description of the problem in CRA. Hence, new and affordable technologies should be adopted in CRA to support the loan granting decisions of banking institutions [4]. This is where CRA systems emerge as a viable alternative to improve credit assessment processes for loan granting. Different systems of this kind have been increasingly used, applying many current technologies. Several banks are developing their own systems based on their own selection criteria to reduce the probability of losses in loans requested by customers or firms [5]. Some of the most widely used CRA systems are based on Artificial Intelligence (AI) and Machine Learning (ML) techniques that analyze and pinpoint trends in potential debtors [6]. These solutions have proven to be efficient, but their performance depends on different factors, such as required data and standards. Many banks, especially microfinance institutions, do not have access to these data, and the latter do not include demographic characteristics, which would improve the system performance. In addition, currently, there are no standards in place to define this type of data treatment as a “solution.” Nevertheless, these solutions aim to improve the process of CRA and thus reduce financial losses for financial institutions [7].

    CRA can use technology-based methods, models, and algorithms to make predictions and select borrowers [8]. In this type of assessment, banks create computer models to represent the process of selecting customers for loan granting so that the approach is clearer and produces better results [9]. The operation of these tools is based on the application of several statistical models depending on the complexity of the assessment.

    Among the statistical models most commonly used for data treatment in CRA systems, the most popular is the Synthetic Minority Over-sampling Technique Evaluation (SMOTE) [10]-[12]. The SMOTE applies algorithms and mathematical processes to artificially generate synthetic samples, that is, artificial samples are created based on the characteristics of original ones in the classes under evaluation. Therefore, it can generate more general and balanced samples to be treated [13]. Besides this model, the efficiency of each kind of statistical model depends on the approach and the type of model adopted by the authors, who define their own efficiency measurements to evaluate the aspects that they deem convenient [14]. In addition, CRA can change depending on the financial institution and the financial model they use. This is because the set of variables that institutions use to measure risk can vary depending on the approach they need [15].

    In CRA systems, AI techniques have been especially used to process and classify data-substituting human analysis to improve the accuracy and speed of the assessment. Based on predictions, these systems classify the risk level of a loan by evaluating different variables: economic, social, demographic, financial factors, etc. [16].

    CRA systems appeared at the beginning of the 2000s, with the development of new technologies and financial institutions’ need for information systems that could conduct CRA. The first CRA systems implemented algorithms and statistical techniques based on financial models to determine potential debtors [17]. Later, ML-based models appeared and evolved into AI models that use big data techniques to process big databases, which is necessary to improve the results of the assessment [18].

    CRA systems also involve the use of different financial models, which should be compared and validated with suitable theoretical foundations to provide reliable solutions. In addition, these systems include statistical models to treat vast amounts of data, which is necessary to continuously improve them [19].

    In these systems, the literature recommends the implementation of ML techniques, especially algorithms based on Random Forest, thanks to their accuracy considering several factors. It also recommends AI techniques. Among them, the most popular are solutions based on Neural Networks, which combine statistical algorithms and algorithmic processes. The literature also suggests to take into consideration the hardware that is used for CRA, which should be adequate to run the algorithms [20].

    CRA systems need models that determine the variables to be assessed. However, only a few studies analyzed in this literature review describe the logic behind their solutions for CRA. In addition, there is little information on the limitations of CRA systems.

    This Systematic Literature Review (SLR) proposes four research questions, which will be detailed later. In particular, this SLR aims to determine what methods, models, and algorithms have been employed in CRA systems; their individual efficiency; the logic behind their assessment models; and their limitations.

    This paper describes and analyzes the most efficient solutions used in CRA systems, as well as their limitations and problems.

    This SLR followed the PRISMA statement, which offers guidelines to orderly and systematically review published documents about a research topic [21].

    As a result, this SLR found 41 common solutions in CRA systems. Ten of them are techniques based on financial models. This review also determined three limitations or problems that could arise when CRA systems are implemented—one of them is particularly significant and should be considered after the implementation process.

    This paper is structured as follows. Section 1 introduces the study and presents the state of the art. Section 2 details the methodology of this SLR and reports the results of the statistical analysis. Section 3 provides answers to the research questions. Finally, Section 4 draws the conclusions of this work.

    2. METHODOLOGY

    This SLR implemented the PRISMA statement to investigate the available knowledge and studies about the topic addressed here. This statement provides researchers with a framework to adequately carry out SLRs that are accepted by the entire scientific community. It also details the necessary steps for an SLR, including establishing goals, eligibility criteria, results, and conclusions [21]. The protocol of this SLR followed the 17 steps in the PRISMA statement.

    In addition, the research objectives of this SLR were clearly established in order to cover and obtain all the necessary information. Table 1 details the four research objectives and four research questions in this SLR based on the PICO (Problem, Intervention, Comparison, and Outcome) model provided by the PRISMA statement. The PICO model is used to formulate research questions [21]. The first research objective in this SLR is quantitative because it aims to identify and count the total number of methods, models, or algorithms employed in CRA systems. The other three objectives are qualitative because they intend to determine the most efficient models, methods, and algorithms, as well as the problems or limitations that may arise during the implementation of CRA systems. The four questions were formulated to achieve the four research objectives, respectively.

    Table 1. Research objectives of this SLR
    Source: Own work.
    Research objectives Research questions Approach
    To determine the methods, models, and algorithms employed in the implementation of CRA systems. What methods, models, and algorithms are employed in CRA systems? Quantitative
    To determine the most efficient methods, models, and algorithms employed in the implementation of CRA systems. What are the most efficient methods, models, and algorithms in CRA systems? Qualitative
    To identify the credit risk models used by banking institutions. What credit risk models are used by banking institutions? Qualitative
    To determine the limitations or problems that may arise in the implementation of CRA systems. What problems or limitations may arise in the implementation of CRA systems? Qualitative

    Subsequently, a search string was defined considering three groups of terms that were identified. The first group contained terms related to the implementation of CRA systems, i.e., methods, models, algorithms, and technologies. The second group included terms related to the topic of this review, i.e., credit risk or credit risk assessment. The last group referred to solutions: system, implementation, and software. Considering these three groups of terms, the following search string was formulated: (methods OR models OR algorithms OR technologies) AND (“credit risk” OR “credit risk assessment”) AND (system OR implementation OR software).

    In addition, six inclusion criteria were established. First, this SLR included only documents in English as it is the most widely accepted language in this scientific area. Second, to study recent literature, the articles should have been published between 2018 and 2022. Third, to ensure the legitimacy of the material, the articles should be final published versions. Fourth, they should be indexed in research databases. Fifth, the articles should be about Credit Risk Assessment (CRA) systems. Sixth, articles about the implementation of said systems were also included. Table 2 details these criteria.

    Table 2. Inclusion and exclusion criteria
    Source: Own work.
    Inclusion criteria Exclusion criteria
    Articles in English Articles published before 2018
    Articles published between 2018 and 2022 Articles in languages other than English
    Final published versions Literature reviews, meta-analyses, or incomplete papers
    Articles indexed in research databases Articles about topics remotely related to the implementation of CRA systems.
    Articles about CRA systems Articles without a Digital Object Identifier (DOI)
    Articles about the implementation of CRA systems

    After the search string and the inclusion and exclusion criteria were defined, the search was conducted in three databases: Scopus, IEEExplore, and ResearchGate. These three databases were selected because we had access to the articles indexed in them. The search was performed on September 13, 2022, and only articles indexed in those databases were selected.

    A preliminary search found 2,285 articles in these three databases. Among them, 88 were excluded because they were duplicates. Then, 1,860 records were excluded as they did not meet the pre-established inclusion criteria. At that point, there were 337 articles: 82 in Scopus, 117 in IEEExplore, and 38 in ResearchGate. Afterward, the titles and abstracts of the documents were analyzed to determine if they latter were about topics closely related to the implementation of CRA systems. After this process, there were 104 articles: 71 in Scopus, 17 in IEEExplore, and 16 in ResearchGate. Subsequently, the full text of these articles was examined to evaluate their connection to the research topic addressed of this study. Thus, other 35 records were excluded, for 69 articles in the final sample. This process is described in detail in Figure 1.

    PRISMA flow diagram.
    Figure 1. PRISMA flow diagram. Source: Own work.

    In this methodology, due to the criteria established and the limited access to articles behind paywalls, the results could present some bias. Because of funding reasons, only articles we had access to were reviewed. In addition, time and language limitations influenced the course of this study. The list of all the articles that were selected for this review was used to generate a series of figures to illustrate the characteristics of these documents.

    Figure 2 details the number of documents published every year in the selected period. The year 2021 presented the highest number of publications (26), followed by 2022 (15). This figure shows the publication trend in this field during that period.

    Number of articles published every year in the selected period and classified by quartile of their journal.
    Figure 2. Number of articles published every year in the selected period and classified by quartile of their journal. Source: Own work.

    It also includes the distribution of those articles into the quartiles of the journals where they were published. The website Scimago was used to classify the journals into quartiles (from Q1 to Q4). The results show that 33 of the articles were published in Q1 journals; 19, in Q2; 13, in Q3; and 4, in Q4. This chart provides information about the quality and impact of the sources where the articles were published.

    In Figure 3, the articles are classified based on the nationality of their main author. Most of them are from China (36), followed by India (5), the US (4), and the UK (4).

    Authors classified by country.
    Figure 3. Authors classified by country. Source: Own work.

    Table 3 lists the top 15 articles in number of citations, showing the most influential papers in this research field.

    Table 3. Top 15 articles by number of citations
    Source: Own work.
    Reference Title Citations
    [22] Forecasting SMEs’ credit risk in supply chain finance with an enhanced hybrid ensemble machine learning approach 111
    [23] Application of new deep genetic cascade ensemble of SVM classifiers to predict the Australian credit scoring 95
    [2] Predicting mortgage default using convolutional neural networks 67
    [24] Enterprise credit risk evaluation based on neural network algorithm 48
    [7] A dynamic credit risk assessment model with data mining techniques: evidence from Iranian banks 43
    [25] Loan evaluation in P2P lending based on Random Forest optimized by genetic algorithm with profit score 41
    [26] SecureBoost: A Lossless Federated Learning Framework 40
    [27] Credit risk modeling using Bayesian network with a latent variable 35
    [28] Multi-view ensemble learning based on distance-to-model and adaptive clustering for imbalanced credit risk assessment in P2P lending 33
    [10] Big data analytics on enterprise credit risk evaluation of e-Business platform 32
    [29] A novel classifier ensemble approach for financial distress prediction 31
    [30] Automatic feature weighting for improving financial Decision Support Systems 22
    [31] Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology: Application to Eurobonds Credit Risk Evaluation 22
    [32] An economic order quantity model under two-level partial trade credit for time varying deteriorating items 21
    [6] Utilizing historical data for corporate credit rating assessment 19
    Total 660

    In turn, Figure 4 shows the countries where the articles included in this SLR were published. Again, most of them were published in China (33), followed by India (5).

    Number of articles by country
    Figure 4. Number of articles by country. Source: Own work.

    Table 4 details all the keywords retrieved from the selected documents. They were grouped into three themes: technologies, supply chain, and financial processes. The first theme includes 32 keywords related to technologies that are employed to implement CRA systems. The second one features three terms associated with the supply chain. The third theme has 22 keywords related to financial processes in CRA.

    Table 4. Keywords classified by theme
    Source: Own work.
    Theme Keywords
    Technologies Machine learning, fuzzy random forest, Genetic algorithm, logistic regression, python web framework2, Predictive intelligence, Deep multiple kernel classifier, Deep learning, Optimal model, Data mining, EM algorithm, Deep neural network, Mixture model, Z-Score Model, Internet of Things, RS-MultiBoosting, improved neural network, Random Forest, explainable algorithms, Deep forest, fuzzy clustering, Artificial Neural Networks, random forest, support vector machine;, Firefly algorithm, SAW Method, Oriented Fuzzy Number, artificial intelligence, Machine learning algorithms, Prediction algorithms, XGBoost, CatBoost algorithm, Data mining, EM algorithm, Deep neural network, Mixture model, Z-Score Model, Internet of Things, RS-MultiBoosting, improved neural network, Random Forest, explainable algorithms.
    Supply chain management Supply chain management, supply chain finance, supply chain financing.
    Financial processes Credit risk scoring, Credit risk, Credit scoring, Bankruptcy prediction, Banknote authentication, Credit scoring, Optimal model, e-Business, Credit risk assessment, Intelligent Manufacturing, Risk analysis, P2P lending, Peer-to-peer lending, Mortgage default model, financial credit evaluation assessment, Internet finance, Federated Learning, systemic risk, soft information, Credit scoring model, feature selection, Feature selection.

    Following the PRISMA statement, the next step was to extract key information from the selected documents. This included methods, results obtained, and any other variables that might have been relevant for the objectives of this review. The data was extracted systematically and following a pre-established protocol.

    Subsequently, the results of this data extraction were summarized to identify patterns and trends and draw conclusions about the reviewed literature. This enabled us to address the research objectives and answer the research questions.

    Simultaneously, the quality of the studies was assessed. The PRISMA statement encourages researchers to conduct a critical assessment of the methodological quality of the selected studies. This includes factors such as study design, sample size, validity of the results, and presence of possible bias. The quality of the selected studies was taken into account in the interpretation of the results. When the selected studies presented comparable and sufficiently homogenous results, a meta-analysis was conducted to combine the data and obtain global estimations of the effects. This contributed to a more accurate assessment of the available information. Finally, the heterogeneity and bias in the included studies were continuously evaluated. Thanks to this continuous evaluation, it was possible to identify possible sources of heterogeneity and bias as the review progressed, which contributed to a more accurate interpretation of the results.

    3. RESULTS AND DISCUSSION

    3.1 Research question 1: What solutions are used in CRA systems?

    In different studies, diverse computational solutions have been implemented in CRA systems. These solutions can be classified into three types: methods, models, and algorithms. Authors in this field adopt a particular solution in accordance with the topic they are investigating or their purpose. Table 5 details the solutions identified in the reviewed studies (31 references).

    Table 5. Solutions used in the implementation of CRA systems
    Source: Own work.
    Reference Solution Description
    [18] SMOTE-PSO-LSSVM This method consists in the application of a synthetic minority oversampling technique (SMOTE) and the least square support vector machine (LSSVM) algorithm optimized by particle swarm optimization (PSO).
    [33] CNN Model This multi-layer hierarchical model performs several transformations using a set of convolutional kernels. The convolution procedure helps to extract valuable characteristics from data points that are spatially connected.
    [34] Deep Multiple Kernel Classifier This method is an optimized version of the Deep Learning structure. It uses multiple processing kernels instead of just one to increase richness in the representation of the features.
    [35] RB Genetic Algorithm This algorithm combines the adaptive random balance (RB) method and the XGBoost algorithm to construct a credit risk assessment model.
    [36], [37] Genetic Algorithm This paper describes an adaptive sequential-filtering learning system for credit risk modeling using genetic algorithms for preselected patterns.
    [10] SMOTE algorithm The SMOTE algorithm is an oversampling technique in which synthetic samples are generated for the minority class. This algorithm can solve the overfitting problem posed by random oversampling.
    [38] Dynamic mutation Particle Swarm Optimization (PSO) algorithm The algorithm in this article was designed to avoid the problem of particles falling into local optima in the optimization process. It is used to optimize the SVM parameters and the integration of AdaBoost as a weak classifier to build an integrated model with good performance in many aspects.
    [2], [39] Convolutional Neural Network (CNN) method In this method, a series of networks are created inspired by the way a human brain works. These networks can learn on different levels of abstraction.
    [27] Bayesian Network (BN) with a latent variable In this model, while the BN structure models the probabilistic relationships between the factors that lead to credit default payment, the latent variable can be used to represent different classes of probability distributions.
    [40] Deep Neural Networks The main purpose of this method is to receive a set of inputs, make progressively more complex calculations on them, and provide an output to solve real-life problems, such as classification. It is limited to feeding the neural networks.
    [41] Random forest algorithm The weighted random forest algorithm is used to classify the financial credit risk data, construct the evaluation index system, and use the analytic hierarchy process to evaluate the financial credit risk level.
    [42] RS-MultiBoosting This approach consists of two classical ensemble ML approaches, i.e., random subspace (RS) and MultiBoosting, to improve the forecasting performance of SMEs’ credit risk. The forecasting result shows that RS-MultiBoosting has good performance in dealing with a small sample size.
    [42] AHP-LSTM This model consists of the analytic hierarchy process (AHP) and the long short-term memory (LSTM) model. Firstly, the characteristic information is extracted, and the financial credit risk assessment index system structure is established. The data are input into the AHP-LSTM neural network, and the index data are fused with the AHP so as to obtain the risk level and serve as the expected output of the LSTM neural network.
    [43] Bagging algorithm for Random Forest The random forest algorithm is an example of a bagging ensemble. It consists in training a series of decision trees so that later they (1) do a majority voting in classifications or (2) obtain the mean in regression problems.
    [44] Multi-layer gradient boosting decision tree (GBDT) In this model, gradient boosting decision trees are used to optimize the prediction results by means of ML processes.
    [28] Distance-to-model and adaptive clustering-based multi-view ensemble classification method In this method, multi-view ensemble learning and a method based on adaptive clustering are used to produce a set of different ensembles composed of gradient boosting decision trees.
    [45] Back Propagation Neural Network (BPNN) In this article, BPNN methods are applied to design a default prediction model for online loans. It was demonstrated that said model (based on BPNN) presents high applicability and prediction accuracy compared to other models.
    [46] ReG-Rules This study describes the general framework of a rule-based ensemble classification system which consists of 5 stages with several operations: (1) Diversity Generation, (2) Base Classifiers Inductions, (3) Models Selection, (4) Rule Merging, and (5) Combination and Prediction.
    [47] Multi-grained augmented gradient boosting decision tree (mg-GBDT) A multi-grained scanning is introduced for feature augmentation, which enriches the input feature of GBDT; the GBDT-based step-wisely optimization mechanism ensures low-deviation of credit scoring; besides, the proposed method inherits the good interpretability of tree-based structure, which provides intuitive reference results for policy-makers.
    [48] Support vector machines This approach includes three components: (1) a novel noise filtering scheme that avoids noisy examples based on fuzzy clustering; (2) the principal component analysis algorithm, which was used to eliminate noise in attributes and classes to produce an optimal clean dataset; and (3) the classifiers of the support vector machine, based on the improved particle swarm optimization algorithm.
    [8]-[14] Machine Learning These two papers compare several ML-based models that can be employed in the development of a credit risk assessment system.
    [6] Parallel Artificial Neural Networks (PANNs) The proposed Parallel ANNs model consists of three stages; the first stage is to create the neural network classifiers; the second stage is to integrate multiple classifiers into an ensemble output; the third stage is to perform the learning process of the PANNs model.
    [49] Firefly algorithm This article proposes an improvement to the firefly algorithm to solve optimization problems and feature selection.
    [43] Imperialist competitive algorithm with modified fuzzy min–max classifier (ICA-MFMCN) This algorithm was designed to identify an optimal subset of features and increased through accuracy classification and scalability through assessment of credit risk.
    [50] Fuzzy Neural Network In this study, a credit risk assessment system called SC-IR2FNN is developed based on fuzzy neural networks.
    [7] Adaptive network-based fuzzy inference system (ANFIS) This adaptive network-based system was implemented to design an adaptive credit risk assessment model that removes human judgement completely.
    [51] Support vector machine (SVM) The concept of SVM is to find optimal hyper plane with maximum margin to linearly separate the data set into two classes.
    [52] CatBoost algorithm Swindle implements CatBoost algorithm is used for predicting loan defaults along with a document verification module using Tesseract and Camelot and also a personalized loan module, thereby mitigating the risk of the financial institutes in issuing loans to defaulters and unauthorized customers.

    3.2 Research question 2: What are the most efficient solutions to implement a CRA system?

    The efficiency of the solutions found in the selected articles was evaluated based on the accuracy percentage of the systems developed and described by the authors. There are three efficiency levels: inefficient, efficient, and very efficient. Table 6 details the solutions and their efficiency levels.

    Table 6. Most efficient solutions
    Source: Own work.
    Reference Solution Efficiency level Advantages/Disadvantages Conditions
    [18] SMOTE-PSO-LSSVM Very efficient It improves accuracy as it addresses data imbalance, but it can be computationally expensive. Ideal for imbalanced datasets.
    [33] CNN Model Very efficient Effective in the extraction of complex data features, but it requires a large volume of data to be trained. Recommended for large amounts of data.
    [34] Deep Multiple Kernel Classifier Efficient It can deal with nonlinear features, but it may require adjustments for sensitive hyper-parameters. Ideal for nonlinear relationships between features.
    [35] RB-XGBoost algorithm Efficient It strikes a good balance between speed and performance, but it may be susceptible to overfitting. Suitable for medium-size and big datasets with mixed features.
    [36] Genetic Algorithm Efficient It can find global solutions, but it may have a higher computational cost. Useful to find optimal solutions in complex search spaces.
    [10] SMOTE algorithm Very efficient Its performance improves in minority classes, but it may produce synthetic samples superimposed on existing ones. Recommended to deal with class imbalances.
    [38] Dynamic mutation particle swarm optimization algorithm (AdaBoost-DPSO-SVM) Very efficient It combines multiple algorithms to have better accuracy, but it is computationally demanding. Suitable when the goal is high performance and multiple algorithms can be used.
    [39] Convolutional Neural Network (CNN) method Very efficient Effective in the extraction of complex data features, but it requires a large volume of data to be trained. Recommended for large amounts of data.
    [27] Bayesian Network (BN) with a latent variable Very efficient It can handle uncertainty effectively, but its operation is more complex. Ideal for modeling uncertainty when probabilistic information is available.
    [40] Deep Neural Networks Efficient They can learn complex data representations but require large volumes of training data. Suitable for problems with complex and scattered data.
    [24] PNN Efficient It can classify patterns fast, but it becomes less efficient as the number of parameters increases. Useful for quick classifications with a few parameters.
    [41] Random forest algorithm Very efficient It is effective with noisy data and remotely related features. Recommended if the data structure or volume are unknown.
    [22] RS-MultiBoosting Efficient It is efficient in multi-class classification, but its accuracy is low in the face of noisy data. Useful for multi-class classifications with clean datasets.
    [42] AHP-LSTM Very efficient It incorporates multi-criteria analysis in the learning process, which involves more complexity. Adequate when multiple criteria should be considered in decision-making.
    [37] Random Forest optimized by genetic algorithm with profit score Efficient It offers better optimization, although with more complexity. Ideal for maximizing classification results.
    [43] Bagging algorithm for Random Forest Very efficient It can reduce variance and overfitting, but it may not work so well with highly imbalanced data. Suitable for classification problems with high variance.
    [44] Multi-layer gradient boosting decision tree (GBDT) Efficient It can deal with different types of features better, although it requires more parameter adjustments. Ideal when the data to be processed have mixed features.
    [28] Distance-to-model and adaptive clustering-based multi-view ensemble (DM–ACME) classification method Efficient It can efficiently process multi-view data, but it requires a careful selection of features. Suitable for multi-view learning.
    [45] Back Propagation Neural Network (BPNN) Very efficient Suitable for regression and classification problems, but it is less efficient with big datasets. Useful for limited datasets because it is less efficient as the dataset is larger.
    [2] Convolutional Neural Network (CNN) method Very efficient Suitable for rule-based problems, but it might be less accurate than complex models. Recommended for large amounts of data.
    [46] ReG-Rules Very efficient Suitable for rule-based problems, but it might be less accurate than complex models. Ideal for interpretability and clear rules.
    [47] Multi-grained augmented gradient boosting decision tree (mg-GBDT) Inefficient It works well in problems with multiple granularity levels, but it is slow compared to other algorithms. Useful to model data granularity in test models.
    [53] EN-AdaPSVM model Very efficient It combines multiple algorithms to improve its performance, but it requires more computational power. Suitable when the goal is high performance and multiple algorithms can be used.
    [8] Machine Learning Efficient It can be more flexible depending on the algorithm employed, although it requires more adjustments. Ideal for a final classification of different kinds of data.
    [[6] Parallel Artificial Neural Networks (PANNs) Efficient They exploit parallel computing for efficient data processing, but that increases the complexity of the algorithm. Useful to improve the processing speed for specific data volumes.
    [49] Firefly algorithm Efficient It can be used for a global optimization of the implementation, but it may require parameter adjustments. Useful to find optimal solutions in complex search spaces.
    [5] Imperialist competitive algorithm with modified fuzzy min–max classifier (ICA-MFMCN) Efficient It can deal with uncertainty and classifications with high numbers of parameters. Recommended to deal with uncertainty in credit assessment.
    [5] Fuzzy Neural Network Efficient It is suitable for problems with high uncertainty and data that have not been normalized yet. Useful to model relationships between data to determine common features.
    [7] Adaptive network-based fuzzy inference system (ANFIS) Efficient It combines fuzzy logic and machine learning to obtain better results, but it may require parameter adjustments. Adequate to determine relationships between features and when training data are available.
    [51] Support vector machine (SVM) Very efficient It can efficiently separate linear and nonlinear classes, but it is largely dependent on kernel selection. Recommended for high-performance classification problems.
    [14] Machine Learning Very efficient It can be more flexible depending on the algorithm employed, although it requires more adjustments. Recommended for a general solution with different types of data.
    [1] PCA-GA-FS model Very efficient It combines feature selection and dimensionality reduction techniques, although it may require parameter adjustments. Useful to obtain a more compact representation of the data.
    [52] CatBoost algorithm Efficient It was optimized for sets of categorical data, although it may be sensitive to hyper-parameter selection. Suitable for datasets with many categorical features.

    3.3 Research question 3: What credit risk models are used by banking institutions?

    This review also identified financial models that banking institutions use for CRA. A credit assessment model can assign weights to qualitative and quantitative variables to evaluate customer credit quality [54]. These models enable banking institutions to measure (based on their own variables and specifications) the risk level that represents granting a loan to a borrower. Table 7 details the credit risk models identified in this review.

    Table 7. Financial models for CRA
    Source: Own work.
    Reference Model Comment
    [13] Comprehensive model for credit scoring The objective of this model is to support decision-making by collecting data and applying statistics, artificial intelligence, and other techniques.
    [32] Economic order quantity model under two-level partial trade credit This article proposed a two-level credit model for time varying deteriorating items. Due to the granting credit terms, we applied default risk. Moreover, we assume that the supplier/retailer offers the partial trade credit to the retailer/customer.
    [17] Optimal backward elimination model and forward regression method The purpose of this paper is to verify whether there is a relationship between credit risk, main threat to the banks, and the demographic, marital, cultural, and socio-economic characteristics of a sample of 40 credit applicants, by using the optimal backward elimination model and the forward regression method.
    [55] Credit evaluation adjustment method According to the measurement requirement of counterparty credit risk in Basel Accord, counterparty credit risk-weighted asset is the summation of the default risk-weighted asset of counterparty credit risk and the credit valuation adjustment risk-weighted asset.
    [56] Additive generalized models It outlines the implementation via frequentist and Bayesian MCMC methods. We apply them to a large portfolio of credit card accounts, and show how GAMs can be used to improve not only the application, behavioral and macro-economic components of survival models for credit risk data at individual account level, but also the accuracy of predictions.
    [12] Logistic regression model According to the definition of whether the dependent variable borrower defaults in Logistic regression, the borrower is divided into default and non-default. Among them, non-default borrowers are marked as 0, and default borrowers are marked as 1.
    [11] Binary spatial regression model This study applied a binary spatial regression model to measure contagion effects in credit risk arising from bank failures. To derive interconnectedness measures, we use the World Input-Output Trade (WIOT) statistics between economic sectors.
    [19] B2B supply chain Based on the B2B e-commerce platform, it is used for online transactions and transactions between companies and companies. Information, integrating logistics, business flow, information flow, and capital flow for data analysis and processing.
    [31 Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology The T1OWA operator methodology is used to assess the creditworthiness of European bonds based on real credit risk ratings of individual Eurozone member states modeled as unbalanced fuzzy linguistic labels.
    [57] Moral hazard model Moral hazard is a problem mainly caused by ex-post information asymmetry in a contract especially after signing the contract. This problem arises due to the inability of agents to be able to observe the actions of other agents.

    A total of 10 models were found in the selected documents. They define the necessary variables for CRA and are adapted according to specific requirements. Many of these models include the demographic characteristics of the applicants to improve the accuracy of the assessment.

    3.4. Research question 4: What problems or limitations may arise in the implementation of CRA systems?

    Finally, in the selected articles, this review found three limitations or problems that may arise in the implementation of CRA systems. A limitation or problem is a factor that can negatively influence research, creating obstacles for authors. Table 8 details the problems and limitations found in the selected research papers. It is concluded that, nevertheless, these limitations or problems do not significantly affect the implementation of CRA systems.

    Table 8. Limitations and problems
    Source: Own work.
    Reference Model Comment
    [13] Comprehensive model for credit scoring The objective of this model is to support decision-making by collecting data and applying statistics, artificial intelligence, and other techniques.
    [22] Economic order quantity model under two-level partial trade credit This article proposed a two-level credit model for time varying deteriorating items. Due to the granting credit terms, we applied default risk. Moreover, we assume that the supplier/retailer offers the partial trade credit to the retailer/customer.
    [17] Optimal backward elimination model and forward regression method The purpose of this paper is to verify whether there is a relationship between credit risk, main threat to the banks, and the demographic, marital, cultural, and socio-economic characteristics of a sample of 40 credit applicants, by using the optimal backward elimination model and the forward regression method.
    [55] Credit evaluation adjustment method According to the measurement requirement of counterparty credit risk in Basel Accord, counterparty credit risk-weighted asset is the summation of the default risk-weighted asset of counterparty credit risk and the credit valuation adjustment risk-weighted asset.
    [56] Additive generalized models It outlines the implementation via frequentist and Bayesian MCMC methods. We apply them to a large portfolio of credit card accounts, and show how GAMs can be used to improve not only the application, behavioral and macro-economic components of survival models for credit risk data at individual account level, but also the accuracy of predictions.
    [12] Logistic regression model According to the definition of whether the dependent variable borrower defaults in Logistic regression, the borrower is divided into default and non-default. Among them, non-default borrowers are marked as 0, and default borrowers are marked as 1.
    [11] Binary spatial regression model This study applied a binary spatial regression model to measure contagion effects in credit risk arising from bank failures. To derive interconnectedness measures, we use the World Input-Output Trade (WIOT) statistics between economic sectors.
    [19] B2B supply chain Based on the B2B e-commerce platform, it is used for online transactions and transactions between companies and companies. Information, integrating logistics, business flow, information flow, and capital flow for data analysis and processing.
    [31] Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology The T1OWA operator methodology is used to assess the creditworthiness of European bonds based on real credit risk ratings of individual Eurozone member states modeled as unbalanced fuzzy linguistic labels.
    [57] Moral hazard model Moral hazard is a problem mainly caused by ex-post information asymmetry in a contract especially after signing the contract. This problem arises due to the inability of agents to be able to observe the actions of other agents.

    Three limitations were found in this SLR. First, the fact that logistic regression is the standard approach in the sector (due to explainability regulations) means that there should be a balance between innovation and transparency in credit decision-making. Second, the “No Free Lunch” theorem indicates that algorithms should be adapted to variations in credit data, acknowledging that there is no “one-size-fits-all” approach. Third, default prediction performance in imbalanced classification situations shows that we should not only focus on absolute high accuracy, but on achieving good global performance and generalization ability.

    The limitations identified here do not significantly hinder the implementation of CRA systems. However, these challenges should be strategically addressed to further improve the quality and efficiency of CRA in a constantly evolving environment.

    This review found CRA solutions in 31 of the selected documents. AI-based algorithms and techniques are present in most of the implementations reported in the articles. Another widely used solution is Neural Networks. Many authors claim that the latter can process data efficiently, be adapted, and combined with other algorithms to improve their performance. Recently, decision trees have become more popular due to their predictive power based on features. Although these solutions have produced good results in implementations of CRA systems, many of them have been validated and tested using simulated databases based on information collected by several banks around the world. These simulated databases can be used to demonstrate the functionality of the systems, but their results do not represent the reality of some organizations.

    Among the 31 articles that described solutions for CRA, 16 of them (i.e., 52 % in this sample) reported accuracies above 90 %. Therefore, their solutions were classified as “Very efficient.” In particular, AI techniques showed great efficiency. Although these studies stress the high performance of their systems, some of them may be biased as they reported 100 % accuracies, which is very unlikely.

    CRA systems need to have models that determine the variables to be assessed. However, only a few studies analyzed in this review describe the logic behind their solutions for CRA.

    Finally, only three articles in this review described the limitations or problems that should be considered in the implementation of their CRA solutions.

     

    4. CONCLUSIONS

    Currently, CRA systems are developed all over the world, mainly in China, India, and the US. These countries are particularly interested in constantly innovating in the implementation of CRA systems.

    The methods, models, and algorithms reviewed here make a great contribution to these systems. However, their functionality should be analyzed using real (not simulated) databases to ensure that the results represent the reality of many organizations. Among these methods, models, and algorithms, AI was employed in most studies. In particular. Neural Networks and Random Forest were present in most articles. They have been combined with different algorithms to guarantee specific results according to the implementation.

    All the solutions analyzed here have shown outstanding efficiency, with their own particularities. Neural Networks—in particular, deep and convolutional Neural Networks—have demonstrated high efficiency due to their ability to predict outcomes based on feature recognition. Their accuracy in the implemented systems is above average. These networks can identify and thoroughly analyze features and patterns in data, which results in accurate, reliable predictions. Likewise, Random Forest-based algorithms exhibited an outstanding predictive power in CRA. These algorithms exploit the diversity and predictive power of a collection of individual trees to offer reliable and consistent predictions, which results in more robust and generalizable models. Nevertheless, this can mean a possible deficiency in cases of large volumes of data or variables, which can increase the computational complexity and require significant processing and storage resources. Although these technologies presented remarkable efficiency, most of the other solutions achieved satisfactory results as well. Still, as some studies referred to margins of error, the methods they employed should be continuously improved and refined.

    To develop CRA systems, it is important to describe the logic behind the variables that are involved in the implementation. This logic is not described in many of the articles reviewed here. Therefore, future studies should focus on this specific area to further improve CRA systems. Also, they could examine aspects such as the optimization of AI algorithms more thoroughly to achieve ever greater accuracy in CRA. They should use real data (instead of simulations) to identify and address possible challenges and limitations in the practical implementation of CRA systems. Furthermore, they can include demographic factors in CRA, which is a key opportunity. They can explore how to effectively incorporate these factors and how to adapt models to different populations. Finally, they should investigate the impact that current and future research in this field can have on the financial sector and decision-making regarding credit risk. These studies should benefit financial institutions and consumers in general.

    5. ACKNOWLEDGMENTS AND FUNDING

    This study did not receive funding from any public or private institution or organization.

    CONFLICTS OF INTEREST

    The authors declare no conflict of interest.

    AUTHOR CONTRIBUTIONS

    Frank Edward Tadeo Espinoza drafted the manuscript, developed the methodology, reported the results, and drew the conclusions. Marco Antonio Coral Ygnacio drafted and structured the manuscript.

    6. REFERENCES