首页    期刊浏览 2024年11月25日 星期一
登录注册

文章基本信息

  • 标题:A comparison of data mining techniques for credit scoring in banking: a managerial perspective.
  • 作者:Ince, Huseyin ; Aktan, Bora
  • 期刊名称:Journal of Business Economics and Management
  • 印刷版ISSN:1611-1699
  • 出版年度:2009
  • 期号:September
  • 语种:English
  • 出版社:Vilnius Gediminas Technical University
  • 关键词:Banking industry;Business economics;Credit and debit card industry;Credit card industry;Credit ratings;Data mining;Managerial economics

A comparison of data mining techniques for credit scoring in banking: a managerial perspective.


Ince, Huseyin ; Aktan, Bora


1. Introduction

One of the main tasks of a bank is to lend money. As a financial intermediary, one of its roles is to reduce lending risks. Bank lending is an art as well as a science. Success depends on techniques used, knowledge and on an aptitude to assess both credit-worthiness of a potential borrower and the merits of the proposition to be financed. In recent years, banks have increasingly used credit-scoring techniques to evaluate the loan applications they receive from consumers (Blochlinger and Leippold 2006; Vojtek and Kocenda 2006; Macerinskiene and Ivaskeviciute 2008; Karan and Arslan 2008). Since severe competition and rapid growth in the consumer credit market, credit scoring models have been extensively used for the credit admission evaluation. Credit scoring is a method of modeling potential risk of credit applications (Vojtek and Kocenda 2006; Zhao 2007; Avery et al. 2004; Bodur and Teker 2005; Crook and Banasik 2004; Jacobson and Roszbach 2003). Credit scoring models have been developed by the financial institution and researchers in order to solve the problems involved during the evaluation process.

In the first beginning, financial institutions always utilized the rules or principles built by the analysts to decide whom to give credit. Since the number of applicants increase tremendously, it is impossible in both economic and man power terms to evaluate the credit applications. Several quantitative methods have been developed for credit admission decision. The credit scoring models are developed to categorize applicants as either accepted or rejected with respect to the applicants' characteristics. The objective of credit scoring models is to assign credit applicants to either a 'good credit' group that is likely to repay financial obligation or a 'bad credit' group whose application will be denied because of its high possibility of defaulting on the financial obligation (Lee et al. 2006). The statistical methods, nonparametric statistical methods, and artificial intelligence approaches have been proposed to support the credit decision (Thomas 2000). Credit scoring problems are basically in the domain of the more general and widely discussed classification problems (Lee et al. 2002).

The classification problems have long played important roles in business related decision making due to its wide applications in decision support, financial forecasting, fraud detection, marketing strategy, process control, and other related fields (Chen et al. 1996; Fayyad et al. 1996; Lee at al. 2006). The classification problem can be solved by using different techniques ranging from statistical methods to artificial intelligence algorithms.

Statistical methods, including regression, linear and nonlinear discriminant analysis, logit and probit models were most commonly applied to construct credit scoring models (Vojtek and Kocenda 2006; Lee et al. 2002; Lee et al. 2006). The most popular methods applied to credit scoring models are linear discriminant analysis, logistic regression and their variations. They are relatively easy to implement and are able to generate straightforward results that can be readily interpreted. However, there are some limitations associated with their applications in credit scoring. First of all, these methods are not effective for problems with high-dimensional inputs and small sample size. Most importantly, these techniques rely on linear separability and normality assumptions. Furthermore, it is difficult to automate the modeling process and design a continuous update flow. According to Yang (2007), the static models usually fail to adapt when environment or population changes over the time. Therefore, these models may need to be rebuilt from scratch.

In addition to these classical methodologies, artificial intelligence techniques have been applied to credit scoring. Practitioners and researchers have developed a variety of techniques for credit scoring, which involve k-nearest neighbor (Henley and Hand 1996), decision trees (Lee et al. 2006), neural networks (Lee et al. 2002; Malhotra, R. and Malhotra, D. K. 2002; West 2000), and genetic programming (Ong et al. 2005), support vector machines models. These techniques can be used as an alternative to discriminant analysis and logistic regression, in situations where the dependent and independent variables exhibit complex nonlinear relationships (Lee et al. 2006).

The purpose of this study is to explore the performance of credit scoring using discriminant analysis, logistic regression, neural networks and classification and regression tree. The rest of the paper is organized as follows: We will briefly review the literature on credit scoring models and a brief outline of statistical methods and artificial intelligence techniques in Section 2. The analytic results of credit scoring models using discriminant analysis, logistic regression, neural networks and classification and regression trees are presented in Section 3. Finally, Section 4 addresses the conclusion.

2. Research methodology and literature review

The credit scoring models investigate the objective and subjective factors that may influence the individuals. In order to predict individual's ability to fulfill his or her financial commitment as expected, credit scoring models have been developed by using quantitative and qualitative analysis. Next, we briefly review the background and related literature on credit scoring models.

2.1. Statistical methods

Two models have been used widely in credit scoring. These are discriminant analysis and logistic regressions. Several variations of these methods have been proposed. Discriminant analysis, proposed by Fisher (1936), involves the linear combination of explanatory variables that differentiate best between a priori defined groups. In order to achieve this, one has to maximize the between-group variance relative to the within group variance. The following equation expresses the discriminant analysis:

Z = [[beta].sub.0] + [[beta].sub.1][X.sub.1] + [[beta].sub.2][X.sub.2] + ... + [[beta].sub.n][X.sub.n] ..., (1)

where Z is the discriminant score, [beta] are the coefficients and X are the independent variables. Discriminant analysis can be used if the dependent variable is categorical and the independent variables are metric. In order to use discriminant analysis, the data has to be independent and normally distributed and covariance matrix is required to comply with the variation homogeneity assumption (Rencher 2002). If the covariance matrices of the given populations are not equal, then the separation surface of the discriminant function is quadratic. Therefore, the quadratic discriminant analysis (QDA) needs to be used. Despite the fact that LDA is only a special case of QDA with stronger assumptions, LDA has been reported to be a more robust method when the theoretical presumptions are violated (Lee and Chen 2005). Discriminant analysis has been used to solve classification problems for finance, business, and marketing research (Lee et al. 1997; Kim et al. 2000; Trevino and Daniels 1995). For credit scoring problems, several researchers have proposed and used the discriminant analysis and its variations (Lee et al. 2002; Lee et al. 2006).

Logistic regression is a widely used statistical modeling technique in which the probability of a dichotomous outcome is related to a set of potential explanatory variables in the form (Hosmer and Lemeshow 1989):

log[p/(1 -p)] = [[beta].sub.0] + [[beta].sub.1][x.sub.1] + [[beta].sub.2][x.sub.2] + ... + [[beta].sub.n][x.sub.n] ..., (2)

where p is the probability of the outcome of interest, [[beta].sub.0] is the intercept term, and [[beta].sub.i] represents the [beta] coefficient associated with the corresponding independent variable [x.sub.i] (i = 1, ..., n). According to Lee et al. (2002), the logistic regression model does not necessarily require the assumptions of discriminant analysis. However, logistic regression can be as efficient and accurate as discriminant analysis even though the assumptions of discriminant analysis are satisfied. An advantage of discriminant analysis is that ordinary least square estimation procedure can be implemented to estimate the coefficients of the linear discriminant function, but maximum likelihood methods are required for the estimation of logistic regression models. Logistic regression models have been widely adopted in many areas ranging from business to engineering (Laitinen, E. K. and Laitinen, T. 2000; Suh et al. 1999; Vellido et al. 1999). Logistic regression has also been explored by several in building credit scoring models for personal loan, business loan, and credit card applications (Lee et al. 2006).

2.2. Artificial intelligence techniques

The artificial intelligence techniques, which have made significant contribution to the field of information science (Chen and Liu 2004) can be adopted to construct the credit scoring models. Several artificial intelligence techniques, which are decision trees, neural networks, genetic programming, k-nearest neighbor models, have been developed by practitioners and researchers for credit scoring (Malhotra, R. and Malhotra, D. K. 2002; West 2000, Ong et al. 2005; Lee et al. 2006; Lee and Chen 2005; Lee et al. 2002). In this study, we will develop and compare credit scoring models based on neural networks, and classification and regression trees.

Neural network (NN), which is an algorithmic procedure for transforming inputs into desired outputs using highly inter-connected networks of relatively simple processing elements (nodes), is a class of nonlinear regression and discrimination models. The neural networks consist of the nodes, the network topology describing the connections between nodes, and the training algorithm used to determine the values of network weights for a particular network. The nodes are connected to one another in the sense that the output from one node can be served as the input to other nodes. Each node transforms an input to an output using a transfer function. Network topology gives the organization of nodes and the types of connections. The nodes are arranged in a series of layers with connections between nodes in different layers. The first layer called input layer receives the inputs. An example of neural networks with one hidden layer is shown in Fig. 1 (Crook et al. 2007). The appropriate network topology (i.e., the number of hidden neurons in hidden layer) can be determined by the following equation:

h = [square root of n + m +a, [alpha] [member of]] [1,10] ..., (3)

where h is the number of the hidden units, n and m is the number of input and output units respectively.

[FIGURE 1 OMITTED]

Neural networks can be classified into different categories such as feedforward and feedback networks. The nodes in feedforward networks can take inputs only from the previous layer and send outputs to the next layer. The multilayer perceptron (MLP) uses back propagation algorithm which is a gradient steepest descent algorithm. In order to find the optimal weight, BP tries to minimize the network error. The step size, called the learning rate, must be specified first. The learning rate is crucial for BPN since smaller values tend to slow down the training process before convergence while larger ones may cause network oscillation and are unable to converge. Several variations of BP algorithm have been proposed to overcome the difficulties such as reaching local minimum, slow convergence and overtraining, detailed information on neural networks can be found in (Haykin 1998).

Decision tree is one of the different approaches to build a classification model by using inductive reasoning. It produces a model of tree-shaped structure representing segmentation of the data that is created by applying a series of simple rules. These rules can be used for prediction through repetitive process of splitting. The decision tree theory is very suitable for credit scoring model and used widely (Lee and Chen 2005). The following decision tree algorithms have been used for prediction and classification: ID3, C4.5, Classification and Regression Trees (CART), and Chi-squared Automatic Interactive Detector (CHAID) models.

ID3 (Iterative Dichotomiser 3) was proposed by Quinlan (1993) to generate decision trees. It is based on theory of information gain. ID3 determines the optimal information gain as an attribute for branching of decision trees so that the tree thus built has a simple structure (Zhao 2007). Information gain is computed by the entropy of the sub-trees produced by a node of a decision tree using a certain attribute, as well as that of the whole data set. The disadvantage of ID3 is that it uses the information gain as a rule to select attributes for branching which result in bias over attributes of higher values. In order to remove this drawback, C4.5, which is an extension and revision of ID3, was proposed (Chang and Chen 2008). C4.5 algorithm uses information gain-ratio to segment attributes. C5 algorithm offering improvements for C4.5 can be used in processing a huge data set because it uses boosting trees to increase modeling accuracy (Chang and Chen 2008). In addition to this, it is much faster in speed and is more efficient than C4.5 in terms of memory usage. According to (Tso and Yau 2007), C5 has the following advantages over C4.5 algorithm: "(1) the branch-merging option for nominal splits is the default; (2) misclassification costs can be specified; (3) boosting and cross-validation are available; and (4) the algorithm for creating rule sets from trees is much improved".

Besides these algorithms, several researchers have proposed other decision trees techniques. One of them is classification and regression trees known as CART, a statistical procedure introduced by Breiman et al. (1984). It is a recursive partitioning method to be used both for regression and classification. It is primarily used as a classification tool to classify an object into two or more populations. It can be used to analyze the continuous data. The CART algorithm can be summarized in three stages as follows (Chang and Chen 2008):

1. In this stage, recursive partitioning technique is used to select variables and split points using a splitting criterion. The best predictor is chosen using a variety of impurity or diversity measures (Gini, twoing, ordered twoing and least-squared deviation). The detailed information about how to compute these impurity measures can be found in (Breiman et al. 1984). The objective is to produce subsets of the data which are as homogeneous as possible with respect to the target variable (Breiman et al. 1984)

2. After identifying a large tree, CART uses the pruning procedure that incorporates a minimal cost complexity. Pruning procedure yields a nested subset of trees starting from the largest tree grown and continuing the process until only one node of the tree remains.

3. In the last stage, the optimal tree is selected by using the lowest cross-validated or testing set error criteria.

Neural networks and decision trees have been widely used to solve several problems related to engineering, science, business, forecasting fields (Vellido et al. 1999; Lee et al. 2002; Lee et al. 2006). Neural networks and decision trees have been used to deal with credit scoring problems (Lee et al. 2006; Lee et al. 2002). Also, decision trees have been used widely in the context of credit scoring models. We will use multilayer perceptron (MLP) networks, CART decision trees algorithm.

3. Empirical study

To verify the feasibility and effectiveness of the credit scoring models using discriminant analysis, logistic regression, decision trees (C5, CART), and neural networks, credit card data set provided by a Turkish bank is used. Each bank customer in the data set contains nine predictor variables, namely, gender, age, marital status, educational level, occupation, job position, income, customer type and credit cards from the other banks. The response variable is the credit status of the customer-good or bad credit. The data set is composed of 1260 customers' records. Among them, 890 data sets with respect to the ratio of good and bad credit were randomly selected as the training sample to estimate the parameters of the corresponding credit scoring model. The remaining 370 will be retained for validation (evaluating the classification capability of the scoring models).

Weka data mining software (Witten and Frank 2005) will be utilized to develop neural networks, decision trees and logistic regression credit scoring models. The discriminant analysis credit scoring models will be implemented by using SPSS 13.0. All the modeling tasks are implemented on an IBM PC with Intel Pentium D 3.0GHz CPU processor with 2 GB of RAM. The detailed credit scoring results using the above-mentioned five modeling techniques can be summarized as follows.

3.1. Discriminant Analysis

The stepwise discriminant approach (Rencher 2002) is adopted in building the discriminant analysis credit scoring model. The final discriminant function has five significant predictor variables, namely income, education, age, occupation, marital status. The credit scoring results of the training and testing sample using the obtained discriminant function are summarized in Table 1. For training and testing sample, the average correct classification rate is 65.23% and 62.00% respectively. For training set, 137 customers with good credit are classified as bad credit customers for training and 169 customers with bad credit are classified as good credit customers. 52 customers with good credit are classified as bad credit customers, and 81 customers with bad credit are classified as good credit customers for testing.

3.2. Logistic Regression

The stepwise logistic regression procedure is used in building the credit scoring model. The variables included in credit scoring model are income, education, customer type. The following Table shows the credit scoring results of the training and testing sample. As it can be seen from Table 2, average correct classification rates of training and testing are 66.37% and 62.33%, respectively. For training set, 113 customers with good credit are classified as bad credit customers for training and 186 customers with bad credit are classified as good credit customers. 81 customers with good credit are classified as bad credit customers, and 58 customers with bad credit are classified as good credit customers for testing.

3.3. Neural Networks

The most widely used algorithm for neural networks is back propagation (BPN) algorithm (Lee et al. 2006). According to Vellido et al. (1999), more than 75% of business applications using neural networks adopted the BPN algorithm. Based on these facts, we will use the BPN algorithm for credit scoring model. In BPN, data set is splitted into two subsets: a training set of 70% (860), a holdout (testing) set of 30% (370) of the total data (1230) respectively.

According to Lee et al. (2002), any complex system can be modeled by one-hidden-layer network. Determining the optimal number of hidden nodes (neurons) is crucial and complicated. The most commonly used way in determining the number of hidden nodes is via experiments or trial-and-error. In addition to this, equation (3) can be used to determine the number of hidden neurons. In this study, we have used the equation (3), to determine the number of hidden neurons. The number of hidden neurons is determined as thirteen. The convergence criteria used for training are a root-meansquared error (RMSE) less than or equal to 0.0001 or a maximum of 5000 iterations.

The prediction results of the neural networks for training and testing sets are summarized in Table 3. From Table 3, average correct classification rates of training and testing are 78.85% and 61.52%, respectively. For training set, 49 customers with good credit are classified as bad credit customers for training and 139 customers with bad credit are classified as good credit customers. 99 customers with good credit are classified as bad credit customers, and 43 customers with bad credit are classified as good credit customers for testing.

3.4. Decision Trees

We use the single classification tree for credit scoring model. We employ the most commonly used decision tree algorithm CART with 1-SE rule in the pruning procedure. CART methods are always preference for the best effective variable to split the node. Therefore, the order of the split node can reflect the important variable in the credit scoring. The variable, income, customer type, and education level are important. Table 4 shows the classification results of training and testing samples. Table 4 shows that average correct classification rates of training and testing are 72.89% and 65.58%, respectively. For training set, 100 customers with good credit are classified as bad credit customers for training and 141 customers with bad credit are classified as good credit customers. 64 customers with good credit are classified as bad credit customers, and 63 customers with bad credit are classified as good credit customers for testing.

3.5. Comparison of the credit scoring models

In order to evaluate the overall credit scoring capability of the designed credit scoring models, predicted results of the credit scoring models and the misclassification costs are used. The predictive results can be determined by the average correct classification rate for the testing set. The following Table 5 shows the predictive accuracy of the four credit scoring models.

It is apparent that the misclassification costs associated with Type I error (a customer with good credit is misclassified as a customer with bad credit) and Type II error (a customer with bad credit is misclassified as a customer with good credit) are significantly different. The misclassification costs associated with Type II errors are much higher than those associated with Type I errors. Since the relative ratio of misclassification costs associated with Type I and Type II errors is 1-5 (West, 2000), special attention should be paid to Type II errors of the four constructed models in order to evaluate the overall credit scoring capability. Table 6 shows the Type I and Type II errors of the four models being discussed.

As the results revealed in Table 6, the neural networks model has the lowest Type II error in comparison with the other three approaches. Therefore, we can conclude that the neural networks can successfully reduce the possible risks of extra losses due to high misclassification costs associated with Type II errors.

4. Conclusions

In this paper, four different techniques have been applied to explore credit scoring and evaluate the bank's credit card policy. Credit scoring has become an important issue as the competition among financial institutions becomes very intense. More and more, financial institutions are seeking better strategies through the help of credit scoring models. Therefore, credit scoring problems are one of the applications that have gained serious attention over the past decades with advances in information technology and modeling techniques. Modeling techniques like traditional statistical analyses and artificial intelligence techniques have been developed in order to successfully attack the credit scoring tasks.

The purpose of this study is to explore the performance of credit scoring using discriminant analysis, logistic regression, neural networks and classification and regression tree. In order to evaluate the feasibility and effectiveness of these techniques, credit-scoring task is performed on one bank credit card data set. Analytic results demonstrate that CART has better average correct classification rate in comparison with discriminant analysis, logistic regression, and neural networks. On the other hand, neural network credit scoring model has lower Type II errors associated with high misclassification costs and therefore has better overall credit scoring capabilities.

Received 20 June 2008; accepted 20 March 2009

References

Avery, R. B.; Calem, P. S; Canner, G. B. 2004. Consumer credit scoring: Do situational circumstances matter? Journal of Banking and Finance 28: 835-856.

Blochlinger, A. and Leippold, M. 2006. Economic benefit of powerful credit scoring, Journal of Banking and Finance 30: 851-873. doi:10.1016/j.jbankfin.2005.07.014

Bodur, C. and Teker, S. 2005. Credit Scoring of Companies: Application to the ISEM Companies, Itujournal/b 2 (1): 25-36.

Breiman, L.; Friedman, J. H.; Olshen, R. A.; Stone, C. J. 1984. Classification and regression trees. Wadsworth and Brooks/Cole.

Chang, C. L.; Chen, C. H. 2008. Applying decision tree and neural network to increase quality of dermatologic diagnosis, Expert Systems with Applications 36(2): 4035-4041. doi:10.1016/j.eswa.2008.03.007

Chen, M. S.; Han, J.; Yu, P. S. 1996. Data mining: an overview from a database perspective, IEEE Trans. Knowledge Data Engineering 8(6): 866-883. doi:10.1109/69.553155

Chen, S. Y.; Liu X. 2004. The contribution of data mining to information science, Journal of Information Science 30(6): 550-558. doi:10.1177/0165551504047928

Crook, J.; Banasik, J. 2004. Does reject inference really improve the performance of application of scoring models? Journal of Banking and Finance 28: 857-874.

Crook, J. N.; Edelman, D. B.; Thomas, L. C. 2007. Recent developments in consumer credit risk assessment, European Journal of Operational Research 183: 1447-1465. doi:10.1016/j.ejor.2006.09.100

Fayyad, U.; Piatetsky-Shapiro, G.; Smyth, P. 1996. The KDD process for extracting useful knowledge from volumes of data, Communications of the ACM 39: 27-34. doi:10.1145/240455.240464

Fisher, R. A. 1936. The use of multiple measurements in taxonomic problems, Annals of Eugenic 7: 179-188.

Haykin, S. 1998. Neural Networks: A Comprehensive Foundation. Prentice Hall.

Henley, W. E.; Hand, D. J. 1996. Ak-nearest neighbor classifier for assessing consumer credit risk, Statistician 44(1): 77-95. doi:10.2307/2348414

Hosmer, D. W.; Lemeshow, S. 1989. Applied Logistic Regression. John Wiley & Sons, New York.

Jacobson, T.; Roszbach, K. 2003. Bank lending policy, credit scoring and value at risk, Journal of Banking and Finance 27: 615-633. doi:10.1016/S0378-4266(01)00254-0

Karan, M. B.; Arslan, O. 2008. Consumer Credit Risk Factors of Turkish Households, Bank and Bank Systems 3(1): 42-57.

Kim, J. C.; Kim, D. H.; Kim, J. J.; Ye, J. S.; Lee, H. S. 2000. Segmenting the Korean housing market using Multiple discriminant analysis, Construction Management and Economics 18: 45-54. doi:10.1080/014461900370942

Laitinen, E. K.; Laitinen, T. 2000. Bankruptcy prediction: Application of the Taylor's expansion in logistic regression, International Review of Financial Analysis 9 (4): 327-349. doi:10.1016/S1057-5219(00)00039-9

Lee, H.; Jo, H.; and Han, I. 1997. Bankruptcy prediction using case-based reasoning, neural networks, and discriminant analysis, Expert Systems with Applications 13: 97-108. doi:10.1016/S0957-4174(97)00011-0

Lee, T. S.; Chiu, C. C.; Lu, C. J.; Chen, I. F. 2002. Credit scoring using the hybrid neural discriminant technique, Expert Systems with Applications 23: 245-254. doi:10.1016/S0957-4174(02)00044-1

Lee, T. S.; Chen, I. F. 2005. A two-stage hybrid credit scoring model using artificial neural networks and multivariate adaptive regression splines, Expert Systems with Applications 28: 743-752. doi:10.1016/j.eswa.2004.12.031

Lee, T. S.; Chiu, C. C.; Chou, Y. C.; Lu, C. J. 2006. Mining the customer credit using classification and regression tree and multivariate adaptive regression splines, Computational Statistics and Data Analysis 50: 1113-1130. doi:10.1016/j.csda.2004.11.006

Macerinskiene, I.; Ivaskeviciute, L. 2008. The evaluation model of a commercial bank loan portfolio, Journal of Business Economics and Management 9: 269-277. doi:10.3846/1611-1699.2008.9.269-277

Malhotra, R.; Malhotra, D. K. 2002. Differentiating between good credits and bad credits using neuro-fuzzy systems, European Journal of Operational Research 136(1): 190-211. doi:10.1016/S0377-2217(01)00052-2

Ong, C-S.; Huang, J-J.; Tzeng, G-H. 2005. Building credit scoring models using genetic programming, Expert Systems with Applications 29(1): 41-47.

Quinlan, J. R. 1993. C4.5: Programs for Machine Learning. San Francisco, CA: Morgan Kaufman.

Rencher, A. C. 2002. Methods of Multivariate Analysis. John Wiley & Sons. doi:10.1002/0471271357

Suh, E. H.; Noh, K. C.; Suh, C. K. 1999. Customer list segmentation using the combined response model, Expert Systems with Applications 17(2): 89-97. doi:10.1016/S0957-4174(99)00026-3

Thomas, L. C. 2000. A survey of credit and behavioral scoring: Forecasting financial risk of lending to consumers, International Journal of Forecasting 16: 149-172. doi:10.1016/S0169-2070(00)00034-0

Trevino, L. J.; Daniels, J. D. 1995. FDI theory and foreign direct investment in the United States: A comparison of investors and noninvestors, International Business Review 4: 177-194. doi:10.1016/0969-5931(95)00004-J

Tso, K. F. G; Yau, K. K. W. 2007. Predicting electricity energy consumption: A comparison of regression analysis, decision tree and neural networks, Energy 32: 1761-1768. doi:10.1016/j.energy.2006.11.010

Vellido, A.; Lisboa, P. J. G.; Vaughan, J. 1999. Neural networks in business: a survey of applications (1992-1998), Expert Systems with Applications 17: 51-70. doi:10.1016/S0957-4174(99)00016-0

Vojtek, M.; Kocenda, E. 2006. Credit Scoring methods, Czech Journal of Economics and Finance 56: 152-167.

West, D. 2000. Neural network credit scoring Models, Computers and Operational Research 27:1131-1152. doi:10.1016/S0305-0548(99)00149-5

Witten, I. H.; Frank, E. 2005. Data Mining: Practical Machine Learning Tools and Techniques. Second edition. San Francisco, CA: Morgan Kaufmann.

Yang, Y. 2007. Adaptive credit scoring with kernel learning methods, European Journal of Operational Research 183(3): 1521-1536. doi:10.1016/j.ejor.2006.10.066

Zhao, H. 2007. A multi-objective genetic programming approach to developing Pareto optimal decision trees, Decision Support Systems 43: 809-826. doi:10.1016/j.dss.2006.12.011

DOI: 10.3846/1611-1699.2009.10.

Huseyin Ince (1), Bora Aktan (2)

(1) Gebze Institute of Technology, Kocaeli, Turkey (2) Yasar University, Izmir, Turkey; University of Primorska, Slovenia E-mails: (1) h.ince@gyte.edu.tr; (2) bora.aktan@yasar.edu.tr
Table 1. Classification results using discriminant
analysis for training and testing samples

 Training Sample Testing Sample

 Predicted Predicted

 Good Bad Good Bad
Actual Credit Credit Credit Credit

Good 356 137 106 52
Credit (72.21%) (27.79%) (67.09%) (32.91%)

Bad 169 218 81 111
Credit (43.67%) (56.33%) (42.19%) (57.81%)

Average correct classification rate for training: 65.23%

Average correct classification rate for testing: 62.00%

Table 2. Classification results using logistic regression
for training and testing samples

 Training Sample Testing Sample

 Predicted Predicted

 Good Bad Good Bad
Actual Credit Credit Credit Credit

Good 348 113 122 81
Credit (75.49%) (24.51%) (60.10%) (39.90%)

Bad 186 242 58 108
Credit (43.46%) (56.54%) (34.94%) (65.06%)

Average correct classification rate for training: 66.37%

Average correct classification rate for testing: 62.33%

Table 3. Classification results using Neural Networks
for training and testing samples

 Training Sample Testing Sample

 Predicted Predicted

 Good Bad Good Bad
Actual Credit Credit Credit Credit

Good 412 49 104 99
Credit (89.37%) (10.63%) (51.23%) (48.77%)

Bad 139 289 43 123
Credit (32.48%) (67.52%) (25.90%) (74.10%)

Average correct classification rate for training: 78.85%

Average correct classification rate for testing: 61.52%

Table 4. Classification results using CART
for training and testing samples

 Training Sample Testing Sample

 Predicted Predicted

 Good Bad Good Bad
Actual Credit Credit Credit Credit

Good 361 100 139 64
Credit (78.31%) (21.69%) (68.47%) (31.53%)

Bad 141 287 63 103
Credit (32.94%) (67.06%) (37.95%) (62.05%)

Average correct classification rate for training: 72.89%

Average correct classification rate for testing: 65.58%

Table 5. Comparison of credit scoring models
(Discriminant Analysis, Logistic regression,

Neural Networks, CART)

 Discriminant Logistic Neural
 Analysis Regression Networks CART

Predicted 62.20% 62.33% 61.52% 65.58%
 Accuracy

Table 6. Type I and Type II errors of four models

 Type I error Type II error

Discriminant Analysis 31.90% 43.32%
Logistic Regression 42.86% 32.22%
Neural Networks 44.59% 29.25%
CART 39.88% 33.01%


联系我们|关于我们|网站声明
国家哲学社会科学文献中心版权所有