Received: March 24, 2023
Accepted: October 30, 2023
Available: December 20, 2023
This review analyzes a selection of scientific articles on the implementation of Credit Risk Assessment (CRA) systems to identify existing solutions, the most accurate ones, and limitations and problems in their development. The PRISMA statement was adopted as follows: the research questions were formulated, the inclusion criteria were defined, the keywords were selected, and the search string was designed. Finally, several descriptive statistics of the selected articles were calculated. Thirty-one solutions were identified in the selected studies; they include methods, models, and algorithms. Some of the most widely used models are based on Artificial Intelligence (AI) techniques, especially Neural Networks and Random Forest. It was concluded that Neural Networks are the most efficient solutions, with average accuracies above 90 %, but their development can have limitations. These solutions should be implemented considering the context in which they will be employed.
Keywords: credit assessment, credit risk, technology solutions, machine learning, algorithms
Esta revisión analiza una selección de artículos científicos sobre la implantación de sistemas de evaluación del riesgo de crédito para identificar las soluciones existentes, las más acertadas y las limitaciones y problemas en su desarrollo. Se adoptó la declaración PRISMA del siguiente modo: se formularon las preguntas de investigación, se definieron los criterios de inclusión, se seleccionaron las palabras clave y se diseñó la cadena de búsqueda. Por último, se calcularon varios estadísticos descriptivos de los artículos seleccionados. En los estudios seleccionados se identificaron 31 soluciones, entre métodos, modelos y algoritmos. Algunos de los modelos más utilizados se basan en técnicas de Inteligencia Artificial (IA), especialmente Redes Neuronales y Bosques Aleatorios. Se concluyó que las Redes Neuronales son las soluciones más eficientes, con precisiones medias superiores al 90 %, pero su desarrollo puede tener limitaciones. Estas soluciones deben implementarse teniendo en cuenta el contexto en el que se van a emplear.
Palabras clave: evaluación crediticia, riesgo de crédito, soluciones tecnológicas, aprendizaje automático, algoritmos
Credit risk is defined as the classification or evaluation that is assigned to someone after having applied for a loan from a bank. The main task of Credit Risk Assessment (CAR) systems is to solve the problem of classifying customers based on their behavior or credit history [
This is a brief description of the problem in CRA. Hence, new and affordable technologies should be adopted in CRA to support the loan granting decisions of banking institutions [
CRA can use technology-based methods, models, and algorithms to make predictions and select borrowers [
Among the statistical models most commonly used for data treatment in CRA systems, the most popular is the Synthetic Minority Over-sampling Technique Evaluation (SMOTE) [
In CRA systems, AI techniques have been especially used to process and classify data-substituting human analysis to improve the accuracy and speed of the assessment. Based on predictions, these systems classify the risk level of a loan by evaluating different variables: economic, social, demographic, financial factors, etc. [
CRA systems appeared at the beginning of the 2000s, with the development of new technologies and financial institutions’ need for information systems that could conduct CRA. The first CRA systems implemented algorithms and statistical techniques based on financial models to determine potential debtors [
CRA systems also involve the use of different financial models, which should be compared and validated with suitable theoretical foundations to provide reliable solutions. In addition, these systems include statistical models to treat vast amounts of data, which is necessary to continuously improve them [
In these systems, the literature recommends the implementation of ML techniques, especially algorithms based on Random Forest, thanks to their accuracy considering several factors. It also recommends AI techniques. Among them, the most popular are solutions based on Neural Networks, which combine statistical algorithms and algorithmic processes. The literature also suggests to take into consideration the hardware that is used for CRA, which should be adequate to run the algorithms [
CRA systems need models that determine the variables to be assessed. However, only a few studies analyzed in this literature review describe the logic behind their solutions for CRA. In addition, there is little information on the limitations of CRA systems.
This Systematic Literature Review (SLR) proposes four research questions, which will be detailed later. In particular, this SLR aims to determine what methods, models, and algorithms have been employed in CRA systems; their individual efficiency; the logic behind their assessment models; and their limitations.
This paper describes and analyzes the most efficient solutions used in CRA systems, as well as their limitations and problems.
This SLR followed the PRISMA statement, which offers guidelines to orderly and systematically review published documents about a research topic [
As a result, this SLR found 41 common solutions in CRA systems. Ten of them are techniques based on financial models. This review also determined three limitations or problems that could arise when CRA systems are implemented—one of them is particularly significant and should be considered after the implementation process.
This paper is structured as follows. Section 1 introduces the study and presents the state of the art. Section 2 details the methodology of this SLR and reports the results of the statistical analysis. Section 3 provides answers to the research questions. Finally, Section 4 draws the conclusions of this work.
This SLR implemented the PRISMA statement to investigate the available knowledge and studies about the topic addressed here. This statement provides researchers with a framework to adequately carry out SLRs that are accepted by the entire scientific community. It also details the necessary steps for an SLR, including establishing goals, eligibility criteria, results, and conclusions [
In addition, the research objectives of this SLR were clearly established in order to cover and obtain all the necessary information. Table 1 details the four research objectives and four research questions in this SLR based on the PICO (Problem, Intervention, Comparison, and Outcome) model provided by the PRISMA statement. The PICO model is used to formulate research questions [
Research objectives | Research questions | Approach |
To determine the methods, models, and algorithms employed in the implementation of CRA systems. | What methods, models, and algorithms are employed in CRA systems? | Quantitative |
To determine the most efficient methods, models, and algorithms employed in the implementation of CRA systems. | What are the most efficient methods, models, and algorithms in CRA systems? | Qualitative |
To identify the credit risk models used by banking institutions. | What credit risk models are used by banking institutions? | Qualitative |
To determine the limitations or problems that may arise in the implementation of CRA systems. | What problems or limitations may arise in the implementation of CRA systems? | Qualitative |
Subsequently, a search string was defined considering three groups of terms that were identified. The first group contained terms related to the implementation of CRA systems, i.e., methods, models, algorithms, and technologies. The second group included terms related to the topic of this review, i.e., credit risk or credit risk assessment. The last group referred to solutions: system, implementation, and software. Considering these three groups of terms, the following search string was formulated: (methods OR models OR algorithms OR technologies) AND (“credit risk” OR “credit risk assessment”) AND (system OR implementation OR software).
In addition, six inclusion criteria were established. First, this SLR included only documents in English as it is the most widely accepted language in this scientific area. Second, to study recent literature, the articles should have been published between 2018 and 2022. Third, to ensure the legitimacy of the material, the articles should be final published versions. Fourth, they should be indexed in research databases. Fifth, the articles should be about Credit Risk Assessment (CRA) systems. Sixth, articles about the implementation of said systems were also included. Table 2 details these criteria.
Inclusion criteria | Exclusion criteria |
Articles in English | Articles published before 2018 |
Articles published between 2018 and 2022 | Articles in languages other than English |
Final published versions | Literature reviews, meta-analyses, or incomplete papers |
Articles indexed in research databases | Articles about topics remotely related to the implementation of CRA systems. |
Articles about CRA systems | Articles without a Digital Object Identifier (DOI) |
Articles about the implementation of CRA systems |
After the search string and the inclusion and exclusion criteria were defined, the search was conducted in three databases: Scopus, IEEExplore, and ResearchGate. These three databases were selected because we had access to the articles indexed in them. The search was performed on September 13, 2022, and only articles indexed in those databases were selected.
A preliminary search found 2,285 articles in these three databases. Among them, 88 were excluded because they were duplicates. Then, 1,860 records were excluded as they did not meet the pre-established inclusion criteria. At that point, there were 337 articles: 82 in Scopus, 117 in IEEExplore, and 38 in ResearchGate. Afterward, the titles and abstracts of the documents were analyzed to determine if they latter were about topics closely related to the implementation of CRA systems. After this process, there were 104 articles: 71 in Scopus, 17 in IEEExplore, and 16 in ResearchGate. Subsequently, the full text of these articles was examined to evaluate their connection to the research topic addressed of this study. Thus, other 35 records were excluded, for 69 articles in the final sample. This process is described in detail in Figure 1.
In this methodology, due to the criteria established and the limited access to articles behind paywalls, the results could present some bias. Because of funding reasons, only articles we had access to were reviewed. In addition, time and language limitations influenced the course of this study. The list of all the articles that were selected for this review was used to generate a series of figures to illustrate the characteristics of these documents.
Figure 2 details the number of documents published every year in the selected period. The year 2021 presented the highest number of publications (26), followed by 2022 (15). This figure shows the publication trend in this field during that period.
It also includes the distribution of those articles into the quartiles of the journals where they were published. The website Scimago was used to classify the journals into quartiles (from Q1 to Q4). The results show that 33 of the articles were published in Q1 journals; 19, in Q2; 13, in Q3; and 4, in Q4. This chart provides information about the quality and impact of the sources where the articles were published.
In Figure 3, the articles are classified based on the nationality of their main author. Most of them are from China (36), followed by India (5), the US (4), and the UK (4).
Table 3 lists the top 15 articles in number of citations, showing the most influential papers in this research field.
Reference | Title | Citations |
[ |
Forecasting SMEs’ credit risk in supply chain finance with an enhanced hybrid ensemble machine learning approach | 111 |
[ |
Application of new deep genetic cascade ensemble of SVM classifiers to predict the Australian credit scoring | 95 |
[ |
Predicting mortgage default using convolutional neural networks | 67 |
[ |
Enterprise credit risk evaluation based on neural network algorithm | 48 |
[ |
A dynamic credit risk assessment model with data mining techniques: evidence from Iranian banks | 43 |
[ |
Loan evaluation in P2P lending based on Random Forest optimized by genetic algorithm with profit score | 41 |
[ |
SecureBoost: A Lossless Federated Learning Framework | 40 |
[ |
Credit risk modeling using Bayesian network with a latent variable | 35 |
[ |
Multi-view ensemble learning based on distance-to-model and adaptive clustering for imbalanced credit risk assessment in P2P lending | 33 |
[ |
Big data analytics on enterprise credit risk evaluation of e-Business platform | 32 |
[ |
A novel classifier ensemble approach for financial distress prediction | 31 |
[ |
Automatic feature weighting for improving financial Decision Support Systems | 22 |
[ |
Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology: Application to Eurobonds Credit Risk Evaluation | 22 |
[ |
An economic order quantity model under two-level partial trade credit for time varying deteriorating items | 21 |
[ |
Utilizing historical data for corporate credit rating assessment | 19 |
Total | 660 |
In turn, Figure 4 shows the countries where the articles included in this SLR were published. Again, most of them were published in China (33), followed by India (5).
Table 4 details all the keywords retrieved from the selected documents. They were grouped into three themes: technologies, supply chain, and financial processes. The first theme includes 32 keywords related to technologies that are employed to implement CRA systems. The second one features three terms associated with the supply chain. The third theme has 22 keywords related to financial processes in CRA.
Theme | Keywords |
Technologies | Machine learning, fuzzy random forest, Genetic algorithm, logistic regression, python web framework2, Predictive intelligence, Deep multiple kernel classifier, Deep learning, Optimal model, Data mining, EM algorithm, Deep neural network, Mixture model, Z-Score Model, Internet of Things, RS-MultiBoosting, improved neural network, Random Forest, explainable algorithms, Deep forest, fuzzy clustering, Artificial Neural Networks, random forest, support vector machine;, Firefly algorithm, SAW Method, Oriented Fuzzy Number, artificial intelligence, Machine learning algorithms, Prediction algorithms, XGBoost, CatBoost algorithm, Data mining, EM algorithm, Deep neural network, Mixture model, Z-Score Model, Internet of Things, RS-MultiBoosting, improved neural network, Random Forest, explainable algorithms. |
Supply chain management | Supply chain management, supply chain finance, supply chain financing. |
Financial processes | Credit risk scoring, Credit risk, Credit scoring, Bankruptcy prediction, Banknote authentication, Credit scoring, Optimal model, e-Business, Credit risk assessment, Intelligent Manufacturing, Risk analysis, P2P lending, Peer-to-peer lending, Mortgage default model, financial credit evaluation assessment, Internet finance, Federated Learning, systemic risk, soft information, Credit scoring model, feature selection, Feature selection. |
Following the PRISMA statement, the next step was to extract key information from the selected documents. This included methods, results obtained, and any other variables that might have been relevant for the objectives of this review. The data was extracted systematically and following a pre-established protocol.
Subsequently, the results of this data extraction were summarized to identify patterns and trends and draw conclusions about the reviewed literature. This enabled us to address the research objectives and answer the research questions.
Simultaneously, the quality of the studies was assessed. The PRISMA statement encourages researchers to conduct a critical assessment of the methodological quality of the selected studies. This includes factors such as study design, sample size, validity of the results, and presence of possible bias. The quality of the selected studies was taken into account in the interpretation of the results. When the selected studies presented comparable and sufficiently homogenous results, a meta-analysis was conducted to combine the data and obtain global estimations of the effects. This contributed to a more accurate assessment of the available information. Finally, the heterogeneity and bias in the included studies were continuously evaluated. Thanks to this continuous evaluation, it was possible to identify possible sources of heterogeneity and bias as the review progressed, which contributed to a more accurate interpretation of the results.
In different studies, diverse computational solutions have been implemented in CRA systems. These solutions can be classified into three types: methods, models, and algorithms. Authors in this field adopt a particular solution in accordance with the topic they are investigating or their purpose. Table 5 details the solutions identified in the reviewed studies (31 references).
Reference | Solution | Description |
[ |
SMOTE-PSO-LSSVM | This method consists in the application of a synthetic minority oversampling technique (SMOTE) and the least square support vector machine (LSSVM) algorithm optimized by particle swarm optimization (PSO). |
[ |
CNN Model | This multi-layer hierarchical model performs several transformations using a set of convolutional kernels. The convolution procedure helps to extract valuable characteristics from data points that are spatially connected. |
[ |
Deep Multiple Kernel Classifier | This method is an optimized version of the Deep Learning structure. It uses multiple processing kernels instead of just one to increase richness in the representation of the features. |
[ |
RB Genetic Algorithm | This algorithm combines the adaptive random balance (RB) method and the XGBoost algorithm to construct a credit risk assessment model. |
[ |
Genetic Algorithm | This paper describes an adaptive sequential-filtering learning system for credit risk modeling using genetic algorithms for preselected patterns. |
[ |
SMOTE algorithm | The SMOTE algorithm is an oversampling technique in which synthetic samples are generated for the minority class. This algorithm can solve the overfitting problem posed by random oversampling. |
[ |
Dynamic mutation Particle Swarm Optimization (PSO) algorithm | The algorithm in this article was designed to avoid the problem of particles falling into local optima in the optimization process. It is used to optimize the SVM parameters and the integration of AdaBoost as a weak classifier to build an integrated model with good performance in many aspects. |
[ |
Convolutional Neural Network (CNN) method | In this method, a series of networks are created inspired by the way a human brain works. These networks can learn on different levels of abstraction. |
[ |
Bayesian Network (BN) with a latent variable | In this model, while the BN structure models the probabilistic relationships between the factors that lead to credit default payment, the latent variable can be used to represent different classes of probability distributions. |
[ |
Deep Neural Networks | The main purpose of this method is to receive a set of inputs, make progressively more complex calculations on them, and provide an output to solve real-life problems, such as classification. It is limited to feeding the neural networks. |
[ |
Random forest algorithm | The weighted random forest algorithm is used to classify the financial credit risk data, construct the evaluation index system, and use the analytic hierarchy process to evaluate the financial credit risk level. |
[ |
RS-MultiBoosting | This approach consists of two classical ensemble ML approaches, i.e., random subspace (RS) and MultiBoosting, to improve the forecasting performance of SMEs’ credit risk. The forecasting result shows that RS-MultiBoosting has good performance in dealing with a small sample size. |
[ |
AHP-LSTM | This model consists of the analytic hierarchy process (AHP) and the long short-term memory (LSTM) model. Firstly, the characteristic information is extracted, and the financial credit risk assessment index system structure is established. The data are input into the AHP-LSTM neural network, and the index data are fused with the AHP so as to obtain the risk level and serve as the expected output of the LSTM neural network. |
[ |
Bagging algorithm for Random Forest | The random forest algorithm is an example of a bagging ensemble. It consists in training a series of decision trees so that later they (1) do a majority voting in classifications or (2) obtain the mean in regression problems. |
[ |
Multi-layer gradient boosting decision tree (GBDT) | In this model, gradient boosting decision trees are used to optimize the prediction results by means of ML processes. |
[ |
Distance-to-model and adaptive clustering-based multi-view ensemble classification method | In this method, multi-view ensemble learning and a method based on adaptive clustering are used to produce a set of different ensembles composed of gradient boosting decision trees. |
[ |
Back Propagation Neural Network (BPNN) | In this article, BPNN methods are applied to design a default prediction model for online loans. It was demonstrated that said model (based on BPNN) presents high applicability and prediction accuracy compared to other models. |
[ |
ReG-Rules | This study describes the general framework of a rule-based ensemble classification system which consists of 5 stages with several operations: (1) Diversity Generation, (2) Base Classifiers Inductions, (3) Models Selection, (4) Rule Merging, and (5) Combination and Prediction. |
[ |
Multi-grained augmented gradient boosting decision tree (mg-GBDT) | A multi-grained scanning is introduced for feature augmentation, which enriches the input feature of GBDT; the GBDT-based step-wisely optimization mechanism ensures low-deviation of credit scoring; besides, the proposed method inherits the good interpretability of tree-based structure, which provides intuitive reference results for policy-makers. |
[ |
Support vector machines | This approach includes three components: (1) a novel noise filtering scheme that avoids noisy examples based on fuzzy clustering; (2) the principal component analysis algorithm, which was used to eliminate noise in attributes and classes to produce an optimal clean dataset; and (3) the classifiers of the support vector machine, based on the improved particle swarm optimization algorithm. |
[ |
Machine Learning | These two papers compare several ML-based models that can be employed in the development of a credit risk assessment system. |
[ |
Parallel Artificial Neural Networks (PANNs) | The proposed Parallel ANNs model consists of three stages; the first stage is to create the neural network classifiers; the second stage is to integrate multiple classifiers into an ensemble output; the third stage is to perform the learning process of the PANNs model. |
[ |
Firefly algorithm | This article proposes an improvement to the firefly algorithm to solve optimization problems and feature selection. |
[ |
Imperialist competitive algorithm with modified fuzzy min–max classifier (ICA-MFMCN) | This algorithm was designed to identify an optimal subset of features and increased through accuracy classification and scalability through assessment of credit risk. |
[ |
Fuzzy Neural Network | In this study, a credit risk assessment system called SC-IR2FNN is developed based on fuzzy neural networks. |
[ |
Adaptive network-based fuzzy inference system (ANFIS) | This adaptive network-based system was implemented to design an adaptive credit risk assessment model that removes human judgement completely. |
[ |
Support vector machine (SVM) | The concept of SVM is to find optimal hyper plane with maximum margin to linearly separate the data set into two classes. |
[ |
CatBoost algorithm | Swindle implements CatBoost algorithm is used for predicting loan defaults along with a document verification module using Tesseract and Camelot and also a personalized loan module, thereby mitigating the risk of the financial institutes in issuing loans to defaulters and unauthorized customers. |
3.2 Research question 2: What are the most efficient solutions to implement a CRA system?
The efficiency of the solutions found in the selected articles was evaluated based on the accuracy percentage of the systems developed and described by the authors. There are three efficiency levels: inefficient, efficient, and very efficient. Table 6 details the solutions and their efficiency levels.
Reference | Solution | Efficiency level | Advantages/Disadvantages | Conditions |
[ |
SMOTE-PSO-LSSVM | Very efficient | It improves accuracy as it addresses data imbalance, but it can be computationally expensive. | Ideal for imbalanced datasets. |
[ |
CNN Model | Very efficient | Effective in the extraction of complex data features, but it requires a large volume of data to be trained. | Recommended for large amounts of data. |
[ |
Deep Multiple Kernel Classifier | Efficient | It can deal with nonlinear features, but it may require adjustments for sensitive hyper-parameters. | Ideal for nonlinear relationships between features. |
[ |
RB-XGBoost algorithm | Efficient | It strikes a good balance between speed and performance, but it may be susceptible to overfitting. | Suitable for medium-size and big datasets with mixed features. |
[ |
Genetic Algorithm | Efficient | It can find global solutions, but it may have a higher computational cost. | Useful to find optimal solutions in complex search spaces. |
[ |
SMOTE algorithm | Very efficient | Its performance improves in minority classes, but it may produce synthetic samples superimposed on existing ones. | Recommended to deal with class imbalances. |
[ |
Dynamic mutation particle swarm optimization algorithm (AdaBoost-DPSO-SVM) | Very efficient | It combines multiple algorithms to have better accuracy, but it is computationally demanding. | Suitable when the goal is high performance and multiple algorithms can be used. |
[ |
Convolutional Neural Network (CNN) method | Very efficient | Effective in the extraction of complex data features, but it requires a large volume of data to be trained. | Recommended for large amounts of data. |
[ |
Bayesian Network (BN) with a latent variable | Very efficient | It can handle uncertainty effectively, but its operation is more complex. | Ideal for modeling uncertainty when probabilistic information is available. |
[ |
Deep Neural Networks | Efficient | They can learn complex data representations but require large volumes of training data. | Suitable for problems with complex and scattered data. |
[ |
PNN | Efficient | It can classify patterns fast, but it becomes less efficient as the number of parameters increases. | Useful for quick classifications with a few parameters. |
[ |
Random forest algorithm | Very efficient | It is effective with noisy data and remotely related features. | Recommended if the data structure or volume are unknown. |
[ |
RS-MultiBoosting | Efficient | It is efficient in multi-class classification, but its accuracy is low in the face of noisy data. | Useful for multi-class classifications with clean datasets. |
[42] | AHP-LSTM | Very efficient | It incorporates multi-criteria analysis in the learning process, which involves more complexity. | Adequate when multiple criteria should be considered in decision-making. |
[ |
Random Forest optimized by genetic algorithm with profit score | Efficient | It offers better optimization, although with more complexity. | Ideal for maximizing classification results. |
[ |
Bagging algorithm for Random Forest | Very efficient | It can reduce variance and overfitting, but it may not work so well with highly imbalanced data. | Suitable for classification problems with high variance. |
[ |
Multi-layer gradient boosting decision tree (GBDT) | Efficient | It can deal with different types of features better, although it requires more parameter adjustments. | Ideal when the data to be processed have mixed features. |
[ |
Distance-to-model and adaptive clustering-based multi-view ensemble (DM–ACME) classification method | Efficient | It can efficiently process multi-view data, but it requires a careful selection of features. | Suitable for multi-view learning. |
[ |
Back Propagation Neural Network (BPNN) | Very efficient | Suitable for regression and classification problems, but it is less efficient with big datasets. | Useful for limited datasets because it is less efficient as the dataset is larger. |
[ |
Convolutional Neural Network (CNN) method | Very efficient | Suitable for rule-based problems, but it might be less accurate than complex models. | Recommended for large amounts of data. |
[ |
ReG-Rules | Very efficient | Suitable for rule-based problems, but it might be less accurate than complex models. | Ideal for interpretability and clear rules. |
[ |
Multi-grained augmented gradient boosting decision tree (mg-GBDT) | Inefficient | It works well in problems with multiple granularity levels, but it is slow compared to other algorithms. | Useful to model data granularity in test models. |
[ |
EN-AdaPSVM model | Very efficient | It combines multiple algorithms to improve its performance, but it requires more computational power. | Suitable when the goal is high performance and multiple algorithms can be used. |
[ |
Machine Learning | Efficient | It can be more flexible depending on the algorithm employed, although it requires more adjustments. | Ideal for a final classification of different kinds of data. |
[[ |
Parallel Artificial Neural Networks (PANNs) | Efficient | They exploit parallel computing for efficient data processing, but that increases the complexity of the algorithm. | Useful to improve the processing speed for specific data volumes. |
[ |
Firefly algorithm | Efficient | It can be used for a global optimization of the implementation, but it may require parameter adjustments. | Useful to find optimal solutions in complex search spaces. |
[ |
Imperialist competitive algorithm with modified fuzzy min–max classifier (ICA-MFMCN) | Efficient | It can deal with uncertainty and classifications with high numbers of parameters. | Recommended to deal with uncertainty in credit assessment. |
[ |
Fuzzy Neural Network | Efficient | It is suitable for problems with high uncertainty and data that have not been normalized yet. | Useful to model relationships between data to determine common features. |
[ |
Adaptive network-based fuzzy inference system (ANFIS) | Efficient | It combines fuzzy logic and machine learning to obtain better results, but it may require parameter adjustments. | Adequate to determine relationships between features and when training data are available. |
[ |
Support vector machine (SVM) | Very efficient | It can efficiently separate linear and nonlinear classes, but it is largely dependent on kernel selection. | Recommended for high-performance classification problems. |
[ |
Machine Learning | Very efficient | It can be more flexible depending on the algorithm employed, although it requires more adjustments. | Recommended for a general solution with different types of data. |
[ |
PCA-GA-FS model | Very efficient | It combines feature selection and dimensionality reduction techniques, although it may require parameter adjustments. | Useful to obtain a more compact representation of the data. |
[ |
CatBoost algorithm | Efficient | It was optimized for sets of categorical data, although it may be sensitive to hyper-parameter selection. | Suitable for datasets with many categorical features. |
3.3 Research question 3: What credit risk models are used by banking institutions?
This review also identified financial models that banking institutions use for CRA. A credit assessment model can assign weights to qualitative and quantitative variables to evaluate customer credit quality [54]. These models enable banking institutions to measure (based on their own variables and specifications) the risk level that represents granting a loan to a borrower. Table 7 details the credit risk models identified in this review.
Reference | Model | Comment |
[ |
Comprehensive model for credit scoring | The objective of this model is to support decision-making by collecting data and applying statistics, artificial intelligence, and other techniques. |
[ |
Economic order quantity model under two-level partial trade credit | This article proposed a two-level credit model for time varying deteriorating items. Due to the granting credit terms, we applied default risk. Moreover, we assume that the supplier/retailer offers the partial trade credit to the retailer/customer. |
[ |
Optimal backward elimination model and forward regression method | The purpose of this paper is to verify whether there is a relationship between credit risk, main threat to the banks, and the demographic, marital, cultural, and socio-economic characteristics of a sample of 40 credit applicants, by using the optimal backward elimination model and the forward regression method. |
[ |
Credit evaluation adjustment method | According to the measurement requirement of counterparty credit risk in Basel Accord, counterparty credit risk-weighted asset is the summation of the default risk-weighted asset of counterparty credit risk and the credit valuation adjustment risk-weighted asset. |
[ |
Additive generalized models | It outlines the implementation via frequentist and Bayesian MCMC methods. We apply them to a large portfolio of credit card accounts, and show how GAMs can be used to improve not only the application, behavioral and macro-economic components of survival models for credit risk data at individual account level, but also the accuracy of predictions. |
[ |
Logistic regression model | According to the definition of whether the dependent variable borrower defaults in Logistic regression, the borrower is divided into default and non-default. Among them, non-default borrowers are marked as 0, and default borrowers are marked as 1. |
[ |
Binary spatial regression model | This study applied a binary spatial regression model to measure contagion effects in credit risk arising from bank failures. To derive interconnectedness measures, we use the World Input-Output Trade (WIOT) statistics between economic sectors. |
[ |
B2B supply chain | Based on the B2B e-commerce platform, it is used for online transactions and transactions between companies and companies. Information, integrating logistics, business flow, information flow, and capital flow for data analysis and processing. |
[ |
Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology | The T1OWA operator methodology is used to assess the creditworthiness of European bonds based on real credit risk ratings of individual Eurozone member states modeled as unbalanced fuzzy linguistic labels. |
[ |
Moral hazard model | Moral hazard is a problem mainly caused by ex-post information asymmetry in a contract especially after signing the contract. This problem arises due to the inability of agents to be able to observe the actions of other agents. |
A total of 10 models were found in the selected documents. They define the necessary variables for CRA and are adapted according to specific requirements. Many of these models include the demographic characteristics of the applicants to improve the accuracy of the assessment.
3.4. Research question 4: What problems or limitations may arise in the implementation of CRA systems?
Finally, in the selected articles, this review found three limitations or problems that may arise in the implementation of CRA systems. A limitation or problem is a factor that can negatively influence research, creating obstacles for authors. Table 8 details the problems and limitations found in the selected research papers. It is concluded that, nevertheless, these limitations or problems do not significantly affect the implementation of CRA systems.
Reference | Model | Comment |
[ |
Comprehensive model for credit scoring | The objective of this model is to support decision-making by collecting data and applying statistics, artificial intelligence, and other techniques. |
[ |
Economic order quantity model under two-level partial trade credit | This article proposed a two-level credit model for time varying deteriorating items. Due to the granting credit terms, we applied default risk. Moreover, we assume that the supplier/retailer offers the partial trade credit to the retailer/customer. |
[ |
Optimal backward elimination model and forward regression method | The purpose of this paper is to verify whether there is a relationship between credit risk, main threat to the banks, and the demographic, marital, cultural, and socio-economic characteristics of a sample of 40 credit applicants, by using the optimal backward elimination model and the forward regression method. |
[ |
Credit evaluation adjustment method | According to the measurement requirement of counterparty credit risk in Basel Accord, counterparty credit risk-weighted asset is the summation of the default risk-weighted asset of counterparty credit risk and the credit valuation adjustment risk-weighted asset. |
[ |
Additive generalized models | It outlines the implementation via frequentist and Bayesian MCMC methods. We apply them to a large portfolio of credit card accounts, and show how GAMs can be used to improve not only the application, behavioral and macro-economic components of survival models for credit risk data at individual account level, but also the accuracy of predictions. |
[ |
Logistic regression model | According to the definition of whether the dependent variable borrower defaults in Logistic regression, the borrower is divided into default and non-default. Among them, non-default borrowers are marked as 0, and default borrowers are marked as 1. |
[ |
Binary spatial regression model | This study applied a binary spatial regression model to measure contagion effects in credit risk arising from bank failures. To derive interconnectedness measures, we use the World Input-Output Trade (WIOT) statistics between economic sectors. |
[ |
B2B supply chain | Based on the B2B e-commerce platform, it is used for online transactions and transactions between companies and companies. Information, integrating logistics, business flow, information flow, and capital flow for data analysis and processing. |
[ |
Type-1 OWA Unbalanced Fuzzy Linguistic Aggregation Methodology | The T1OWA operator methodology is used to assess the creditworthiness of European bonds based on real credit risk ratings of individual Eurozone member states modeled as unbalanced fuzzy linguistic labels. |
[ |
Moral hazard model | Moral hazard is a problem mainly caused by ex-post information asymmetry in a contract especially after signing the contract. This problem arises due to the inability of agents to be able to observe the actions of other agents. |
Three limitations were found in this SLR. First, the fact that logistic regression is the standard approach in the sector (due to explainability regulations) means that there should be a balance between innovation and transparency in credit decision-making. Second, the “No Free Lunch” theorem indicates that algorithms should be adapted to variations in credit data, acknowledging that there is no “one-size-fits-all” approach. Third, default prediction performance in imbalanced classification situations shows that we should not only focus on absolute high accuracy, but on achieving good global performance and generalization ability.
The limitations identified here do not significantly hinder the implementation of CRA systems. However, these challenges should be strategically addressed to further improve the quality and efficiency of CRA in a constantly evolving environment.
This review found CRA solutions in 31 of the selected documents. AI-based algorithms and techniques are present in most of the implementations reported in the articles. Another widely used solution is Neural Networks. Many authors claim that the latter can process data efficiently, be adapted, and combined with other algorithms to improve their performance. Recently, decision trees have become more popular due to their predictive power based on features. Although these solutions have produced good results in implementations of CRA systems, many of them have been validated and tested using simulated databases based on information collected by several banks around the world. These simulated databases can be used to demonstrate the functionality of the systems, but their results do not represent the reality of some organizations.
Among the 31 articles that described solutions for CRA, 16 of them (i.e., 52 % in this sample) reported accuracies above 90 %. Therefore, their solutions were classified as “Very efficient.” In particular, AI techniques showed great efficiency. Although these studies stress the high performance of their systems, some of them may be biased as they reported 100 % accuracies, which is very unlikely.
CRA systems need to have models that determine the variables to be assessed. However, only a few studies analyzed in this review describe the logic behind their solutions for CRA.
Finally, only three articles in this review described the limitations or problems that should be considered in the implementation of their CRA solutions.
Currently, CRA systems are developed all over the world, mainly in China, India, and the US. These countries are particularly interested in constantly innovating in the implementation of CRA systems.
The methods, models, and algorithms reviewed here make a great contribution to these systems. However, their functionality should be analyzed using real (not simulated) databases to ensure that the results represent the reality of many organizations. Among these methods, models, and algorithms, AI was employed in most studies. In particular. Neural Networks and Random Forest were present in most articles. They have been combined with different algorithms to guarantee specific results according to the implementation.
All the solutions analyzed here have shown outstanding efficiency, with their own particularities. Neural Networks—in particular, deep and convolutional Neural Networks—have demonstrated high efficiency due to their ability to predict outcomes based on feature recognition. Their accuracy in the implemented systems is above average. These networks can identify and thoroughly analyze features and patterns in data, which results in accurate, reliable predictions. Likewise, Random Forest-based algorithms exhibited an outstanding predictive power in CRA. These algorithms exploit the diversity and predictive power of a collection of individual trees to offer reliable and consistent predictions, which results in more robust and generalizable models. Nevertheless, this can mean a possible deficiency in cases of large volumes of data or variables, which can increase the computational complexity and require significant processing and storage resources. Although these technologies presented remarkable efficiency, most of the other solutions achieved satisfactory results as well. Still, as some studies referred to margins of error, the methods they employed should be continuously improved and refined.
To develop CRA systems, it is important to describe the logic behind the variables that are involved in the implementation. This logic is not described in many of the articles reviewed here. Therefore, future studies should focus on this specific area to further improve CRA systems. Also, they could examine aspects such as the optimization of AI algorithms more thoroughly to achieve ever greater accuracy in CRA. They should use real data (instead of simulations) to identify and address possible challenges and limitations in the practical implementation of CRA systems. Furthermore, they can include demographic factors in CRA, which is a key opportunity. They can explore how to effectively incorporate these factors and how to adapt models to different populations. Finally, they should investigate the impact that current and future research in this field can have on the financial sector and decision-making regarding credit risk. These studies should benefit financial institutions and consumers in general.
This study did not receive funding from any public or private institution or organization.
The authors declare no conflict of interest.
Frank Edward Tadeo Espinoza drafted the manuscript, developed the methodology, reported the results, and drew the conclusions. Marco Antonio Coral Ygnacio drafted and structured the manuscript.