Acessibilidade / Reportar erro

CREDIT SCORING MODELING WITH STATE-DEPENDENT SAMPLE SELECTION: A COMPARISON STUDY WITH THE USUAL LOGISTIC MODELING

Abstract

Statistical methods have been widely employed to assess the capabilities of credit scoring classification models in order to reduce the risk of wrong decisions when granting credit facilities to clients. The predictive quality of a classification model can be evaluated based on measures such as sensitivity, specificity, predictive values, accuracy, correlation coefficients and information theoretical measures, such as relative entropy and mutual information. In this paper we analyze the performance of a naive logistic regression model, a logistic regression with state-dependent sample selection model and a bounded logistic regression model via a large simulation study. Also, as a case study, the methodology is illustrated on a data set extracted from a Brazilian retail bank portfolio. Our simulation results so far revealed that there is nostatistically significant difference in terms of predictive capacity among the naive logistic regression models, the logistic regression with state-dependent sample selection models and the bounded logistic regression models. However, there is difference between the distributions of the estimated default probabilities from these three statistical modeling techniques, with the naive logistic regression models and the boundedlogistic regression models always underestimating such probabilities, particularly in the presence of balanced samples. Which are common in practice.

classification models; naive logistic regression; logistic regression with state-dependent sample selection; bounded logistic regression; performance measures; credit scoring


1 INTRODUCTION

The proper classification of applicants is of vital importance for determining the granting of credit facilities. Historically, statistical classification models have been used by financial institutions as a major tool to help on granting credit to clients.

The consolidation of the use of classification models occurred in the 90s, when changes in the world scene, such as deregulation of interest rates and exchange rates, increase in liquidity and in bank competition, made financial institutions more and more worried about credit risk, i.e., the risk they were running when accepting someone as their client. The granting of credit started to be more important in the profitability of companies in the financial sector, becoming one of the main sources of revenue for banks and financial institutions in general. Due to this fact, this sector of the economy realized that it was highly recommended to increase the amount of allocated resources without losing the agility and quality of credits, at which point the contribution of statistical modeling is essential. For more details about the developments in the credit risk measurement literature over the 80s and 90s, see Altman & Saunders (1998)ALTMAN EI & SAUNDERS A. 1998. Credit risk measurement: Developments over the last 20 years. Journal of Banking & Finance, 21: 1721-1742..

Classification models for credit scoring are based on databases of relevant client information, with the financial performance of clients evaluated from the time when the client-company relationship began as a dichotomic classification. The goal of credit scoring models is to classify loan clients to either good credit or bad credit (Lee et al., 200226 LEE T, CHIU C, LU C & CHEN I . 2002. Credit scoring using the hybrid neural discriminant technique. Expert Systems with Applications, 23(3): 245-254.; Scarpel & Milioni, 200233 SCARPEL RA & MILIONI AZ. 2002. Utilização conjunta de modelagem econométrica e otimização em decisões de concessão de crédito. Pesquisa Operacional, 22(1): 61-72.), predicting the bad payers (Lim & Sohn, 200727 LIM MK & SOHN SY. 2007. Cluster-based dynamic scoring model. Expert Systems with Applications, 32(2): 427-431.).

In this context, discriminant analysis, regression trees, logistic regression, logistic regression with state-dependent sample selection, bounded logistic regression and neural networks are among the most widely used classification models. In fact, logistic regression is still very used in building and developing credit scoring models (Desai et al., 199614 DESAI VS, CROOK JN & OVERSTREET GA. 1996. A comparison of neural networks and linear scoring models in the credit union environment. European Journal of Operational Research, 95(1): 24-37.; Hand & Henley, 199720 HAND DJ & HENLEY WE. 1997. Statistical classification methods in consumer credit scoring: A review. Journal of the Royal Statistical Society: Series A (Statistics in Society), 160(3): 523-541.; Sarlija et al., 200432 SARLIJA N, BENSIC M & BOHACEK Z. 2004. Multinomial model in consumer credit scoring. 10th International Conference on Operational Research. Trogir: Croatia.). Generally, the best technique for all data sets does not exist but we can compare a set of methods using some statistical criteria. Therefore, the main thrust of this paper is to investigate and compare the performance of the naive logistic regression (Hosmer & Lemeshow, 198921 HOSMER WD & LEMESHOW S. 1989. Applied Logistic Regression. New York: John Wiley.), the logistic regression with state-dependent sample selection and the bounded logisticregression (Cramer, 200413 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.) using performance measures, in terms of a simulation study. The idea is to analyze the impact of disproportional samples on credit scoring models. Logistic regression with state-dependent sample selection is a statistical modeling technique used in cases where the sample considered to develop a model, i.e. the selected sample, contains only a portion, usually small, of the individuals who make up one of two study groups, in general the most frequent group. In credit scoring, for instance, the group of good payers is expected to be the predominant group. In short, this recent technique takes into account the principle of sample selection and makes a correction in the estimated default probability from a naive logistic regression model. On the other hand, the bounded logistic regression technique is commonly used in rare events studies and modifies the naive logistic regression model by adding a parameter that quantifies an upper limit for the probability of the event of interest.

The first credit scoring models were developed around 1950 and 1960, and the methods appliedin this kind of problem referred to methods of discrimination suggested by Fisher (1936)16 FISHER RA. 1936. The use multiple measurements in taxonomic problems. Annals of Eugenics, 7(7): 179-188.,where the models were based on his discriminant function. As Thomas (2000)36 THOMAS LC. 2000. A survey of credit and Behavioural Scoring; Forecasting financial risk of lending to consumers. University of Edinburgh, Edinburgh, U.K. points out, David Durand, in 1941, was the first one which recognized that the discriminant analysis technique, invented by Fisher in 1936, could be used to separate good credits from the bad ones. According to Kang & Shin (2000)22 KANG S & SHIN K. 2000. Customer credit scoring model using analytic hierarchy process. Informs & Korms, Seoul, 2197-2204. Korea., Durand presented a model which attributed weights for each variable using discriminant analysis. Thus Fisher's approach can be seen as the starting point for developments and modifications of the methodologies used for granting of credit until today, where statistical techniques, such as discriminant analysis, regression analysis, probit analysis, neural networks and logistic regression, have been used and examined (Boyes et al., 198910 BOYES WJ, HOFFMAN DL & LOW SA. 1989. An econometric analysis of the bank credit scoring problem. Journal of Econometrics, 40(1): 3-14.; Greene, 199818 GREENE W. 1998. Sample selection in credit-scoring models. Japan and the World Economy, 10(3): 299-316.; Banasik et al., 2001BANASIK J, CROOK J & THOMAS L. 2001. Scoring by usage. The Journal of the Operational Research Society, 52(9): 997-1006.; Sarlija et al., 200432 SARLIJA N, BENSIC M & BOHACEK Z. 2004. Multinomial model in consumer credit scoring. 10th International Conference on Operational Research. Trogir: Croatia.; Selau & Ribeiro, 201134 SELAU LPR & RIBEIRO JLD. 2011. A systematic approach to construct credit risk forecast models. Pesquisa Operacional, 31(1): 41-56.). Particularly, considering state-dependent sample selection in order to make a correction in the estimated default probability from a credit scoring model; and/or introducing a ceiling or maximum risk in order to provide a better model for the cases where the event of interest is rare (Cramer, 200413 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.). Survival analysis is an expending area of research in credit scoring and can be applied to model the time to default, for instance (Andreeva et al., 2007ANDREEVA G, ANSELL J & CROOK J. 2007. Modelling profitability using survival combination scores. European Journal of Operational Research, 183(3): 1537-1549.; Bellotti & Crook, 2009BELLOTTI T & CROOK J . 2009. Credit Scoring with Macroeconomic Variables Using SurvivalAnalysis. The Journal of the Operational Research Society, 60(12): 1699-1707.).

The predictive quality of a credit scoring model can be evaluated based on measures such as sensitivity, specificity, correlation coefficients and information measures, such as relative entropy and mutual information (Baldi et al., 2000BALDI P, BRUNAK S, CHAUVIN Y, ANDERSEN CAF & NIELSEN H. 2000. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics Review, 16(5): 412-424.).

Generally, there is no overall best statistical technique used in building credit scoring models, so that the choice of a particular technique depends on the details of the problem, such as the data structure, the features used, the extent to which it is possible to segregate the classes by using those features, and the purpose of the classification (Hand & Henley, 199720 HAND DJ & HENLEY WE. 1997. Statistical classification methods in consumer credit scoring: A review. Journal of the Royal Statistical Society: Series A (Statistics in Society), 160(3): 523-541.). Most studies that made a comparison between different techniques tried to discover that the most recent/advanced credit scoring techniques, such as neural networks and fuzzy algorithms are better than the traditional ones. Nevertheless, the more simple classification techniques, such as linear discriminant analysis and naive logistic regression, have a very good performance, which is in majority of the cases not statistically different from other techniques (Baesens et al., 2003BAESENS B, GESTEL TV, VIAENE S, STEPANOVA M, SUYKENS J & VANTHIENEN J. 2003. Benchmarking state-of-the-art classification algorithms for credit scoring. The Journal of the Operational Research Society, 54(6): 627-635.; Hand, 200619 HAND DJ. 2006. Classifier Technology and the Illusion of Progress. Statistical Science, 21(1): 1-14.).

This paper is organized as follows. Section 2 describes the three commonly used statisticaltechniques in building credit scoring models: the naive logistic regression, the logistic regression with state-dependent sample selection and the bounded logistic regression. Section 3 presents some useful measures that are used to analyze the predictive capacity of a classification model. Section 4 describes the details of a simulation study performed in order to compare the techniques of interest. In Section 5 the methodology is illustrated on a real data set from a Brazilian retail bank portfolio. Finally, Section 6 concludes the paper with some final comments.

2 CREDIT SCORING MODELS

In this Section, the three statistical techniques used for building credit scoring are described.

2.1 Naive Logistic Regression

Naive logistic regression is a widely used statistical modeling technique in which the response variable, i.e. the outcome is binary (0,1) and can thus be used to describe the relationship between the occurrence of an event of interest and a set of potential predictor variables. In the context of credit scoring, the outcome corresponds to the credit performance of a client during a given period of time, usually 12 months. A set of individual characteristics, such as marital status, age and income, as well as information about his credit product in use, such as number of parcels, purpose and credit value, are observed at the time the clients apply for the credit.

Let us consider a large sample of observations with predictors x i and binary (0,1) outcomes Y i. Here, the event Y i=1 represents a bad credit, while the complement Y i=0 represents a good credit. The model specifies that the probability of i being a bad credit as a function of the x i is given by,

In the case that (1) is a naive logistic regression model, p i is given by,

(see Hosmer & Lemeshow, 198921 HOSMER WD & LEMESHOW S. 1989. Applied Logistic Regression. New York: John Wiley.). Thus the objective of a naive logistic regression model in credit scoring is to determine the conditional probability of a specific client belonging to a class, for instance, the bad payers class, given the values of the independent variables of that credit applicant (Lee & Chen, 200525 LEE T & CHEN I. 2005. A two-stage hybrid credit scoring model using artificial neural networksand multivariate adaptive regression splines. Expert Systems with Applications, 28(4): 743-752.).

2.2 Logistic Regression with State-Dependent Sample Selection

Now let us consider the situation where the event Y i=1 represents a bad credit but it has a low incidence, while the complement Y i=0 represents a good credit but it is abundant.

Suppose we wish to estimate β from a selected sample, which is obtained by discarding a large part of the abundant zero observations for reasons of convenience. Assume also that the overall sample, hereafter full sample, is a random sample with sampling fraction α and that only a fraction γ of the zero observations, taken at random, is maintained. The probability that the element i has Y i=1 and it is included in the sample is given by αp i, but for Y i=0 it is given by γα(1-p i), where p i is calculated from (2). Then, the probability that an element of the selected sample has Y i=1 is given by,

The sketch of the proof of (3) is given in Appendix A APPENDIX A Sketch of the proof of (3). If the event I i=1 means that element i is included in the sample, from the Bayesian Theorem (Baron, 1994), it is possible to see that, completing the proof. .

2.2.1 Estimation Procedure

The log-likelihood of the observed sample can be written in terms of as follows,

If the selected sample is drawn from a known full sample (as here) γ is always known. Thus the parameters of any specification of p i from (1) can be estimated from the selected sample by standard maximum likelihood methods. In the special case that (1) is a naive logistic regression model, is given by,

Thus the of the selected sample also obey a naive logistic regression model, and besides the intercept the same parameters β apply as in the full sample. If it is needed, the full sample intercept can be easily recovered by adding lnγ to the intercept of the selected sample.

2.3 Bounded Logistic Regression

The bounded logistic regression model is derived from the naive logistic regression model by introducing a parameter ω which quantifies an upper limit, i.e. a ceiling for the probability of the event of interest. The model specifies that the probability of i being a bad credit as a function of the x i is given by,

where 0<ω<1.

Model (6) was proposed by Cramer (2004)13 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380. who fitted the naive logistic regression model, the complementary log-log model and the bounded logistic regression model to the database of a Dutch financial institution. These data showed a low incidence of the event of interest and the Hosmer-Lemeshow test indicated the bounded logistic regression as the best fitting model to describe the observed data. According to Cramer (2004)13 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380., the parameter ω has the ability to absorb the impact of possible significant covariates excluded from the database.

2.3.1 Estimation Procedure

The log-likelihood of the observed sample can be written in terms of as follows,

Therefore, the maximum likelihood estimates of β and ω can be obtained through numerical maximization of (7).

3 MODEL VALIDATION

After building a statistical model, it is important we evaluate it. Particularly, we can say that a good model is one which produces scores that can distinguish good from bad credits, since the objective is to previously identify such groups and treat them differently considering distinct relationship policies (Thomas et al., 200237 THOMAS LC, EDELMAN DB & CROOK JN. 2002. Credit Scoring and Its Applications. Philadelphia: SIAM.).

The evaluation performance of a model can be made by a comparison between its prediction and the real classification of a client. The true condition of a client is generally known and is present as basic information in the database. There are therefore four possible results for each client:a) The result given by the classification model is a true-positive (TP). In other words, the model indicates that the client is a bad payer, which, in fact, he actually is; b) The result given by the classification model is a false-positive (FP). In other words, the model accuses the client ofbeing a bad payer, which, in fact, he actually is not; c) The result given by the classification model is a false-negative (FN). In other words, the model classifies him as a good payer, which, in fact, he actually is not; d) The result given by the classification model is a true-negative (TN). In other words, the model classifies the client as a good payer, which, in fact, he actually is.

These four possible situations above are summarized in Table 1. Two of these results, TP and TN, can be seen as indicating that the proposed model "got it right" and are important measures of financial behavior. Of the other results that can be considered to indicate that the proposed model "got it wrong", the one that gives a FN classification is often the most important and deservesthe most attention. If the classification given by the proposed model is negative, and the client is, in fact, a bad payer, granting to this client the credit requested can end up with the client defaulting, which, added to other cases of defaults by bad payers, can result in very high overall default rates for the company, and a real threat to its financial sustainability. A FP result can also be negative for the company, since it may lose out by not approving credit for a client that would not, in fact, cause any default.

Table 1
Definitions used to validate classification models that produce dichotomized responses.

The predictive capacity of a classification model is related to its performance measures, which can be calculated from Table 1. Among them we can cite sensitivity, specificity, positive and negative predictive values, accuracy, Matthews correlation coefficient, approximate correlation, relative entropy and mutual information.

Sensitivity (SEN) is defined as the probability that the classification model will produce apositive result, given that the client is a defaulter. In other words, sensitivity corresponds to the proportion of bad payers that are correctly classified by the classification model (Mazucheli et al., 2006), and is given by,

Specificity (SPE) is defined as the probability that a classification model will produce a negative result for a client who is not a defaulter. In other words, specificity represents the proportionof good payers that are correctly classified by the classification model (Mazucheli et al., 2006), and is calculated as follows,

Thus, a model with high sensitivity rarely fails to detect clients with the characteristic of interest (default) and a model with high specificity rarely classifies a client without this feature as a bad payer. According to Fleiss (1981)17 FLEISS JL. 1981. Statistical methods for rates and proportions. New York: John Wiley. and Thibodeau (1981)35 THIBODEAU LA. 1981. Evaluating diagnostic tests. Biometrics, 37(4): 801-804., it was J. Yerushalmy who, in 1947,first defined these measures and suggested they be used to evaluate classification models.

The Accuracy (ACC) is defined as the proportion of successes of a model, i.e. the proportion of true-positives and true-negatives in relation to all possible outcomes (Mazucheli et al., 2006). Accuracy is then given by,

The positive predictive value (PPV) of a model is defined as the proportion of bad payers identified as such by the model, while the negative predictive value (NPV) of a model is defined as the proportion of good payers identified as such by the model (Mazucheli et al., 2006).

The positive and negative predictive values can be estimated directly from the sample, using the equations,

and

respectively (Dunn & Everitt, 199515 DUNN G & EVERITT BS. 1995. Clinical biostatistics: an introduction to evidence based medicine. London: Edward Arnold.).

The more sensitive the model, the greater its negative predictive value. That is, the greater is the assurance that a client with a negative result is not a bad payer. Likewise, the more specific the model, the greater its positive predictive value. That is, the greater the guarantee of an individual with a positive result being a bad payer.

The correlation coefficient proposed by Matthews (1975)28 MATTHEWS BW. 1975. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta, 405: 442-451. uses all four values (TP, FP, TN, TN) of a confusion matrix in its calculation. In general, it is regarded as a balanced measure which can be used even if the studied classes are of very different sizes. The Matthews correlation coefficient (MCC) returns a value between 1 and +1. A value of +1 represents a perfect prediction, i.e. total agreement, 0 a completely random prediction and 1 an inverse prediction, i.e. total disagreement. The MCC is given by,

(see Baldi et al., 2000BALDI P, BRUNAK S, CHAUVIN Y, ANDERSEN CAF & NIELSEN H. 2000. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics Review, 16(5): 412-424.).

If any of the four sums in the denominator is zero, the denominator can be arbitrarily set to one; which results in a MCC of zero. There are situations, however, where the MCC is not a reliable performance measure. For instance, the MCC will be relatively high in cases where a classification model gives very few or no false-positives, but at the same time very few true-positives.

Burset & Guigó (1996)12 BURSET M & GUIGÓ R. 1996. Evaluation of gene structure prediction programs. Genomics, 34(3): 353-367. defined an approximate correlation measure to compensate for a declared problem with the MCC. That is, it is not defined if any of the four sums TP + FN, TP + FP, TN + FP, or TN + FN is zero, e.g. when there are no positive predictions. They use the average conditional probability (ACP) which is defined as,

if all the sums are non-zero; otherwise, it is the average over only those conditional probabilities that are defined. Approximate correlation (AC) is a simple transformation of the ACP given by,

It returns a value between 1 and +1 and has the same interpretation as the MCC. Burset & Guigó (1996)12 BURSET M & GUIGÓ R. 1996. Evaluation of gene structure prediction programs. Genomics, 34(3): 353-367. observe that the AC is close to the real correlation value. However, some authors, like Baldi et al. (2000)BALDI P, BRUNAK S, CHAUVIN Y, ANDERSEN CAF & NIELSEN H. 2000. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics Review, 16(5): 412-424., do not encourage its use, since there is no simple geometrical interpretation for AC.

Suppose that D ' = (d 1, ..., d N) is a vector of true conditions of clients and M ' = (m 1, ..., m N) a vector of predictions from a classification model, both binary (0,1). The d is and m is arethen equal to 0 or 1, for instance, 0 indicates a good payer and 1 a bad payer. The mutualinformation between D and M is measured by,

(see Wang, 199438 WANG ZX. 1994. Assessing the accuracy of protein secondary structure. Nat. Struct. Biol., 1(3): 145-146.), where N is the sample size, =(TP+FN)/N, =(TP+FP)/N and

is the usual entropy, whose roots are in information theory (Kullback & Leibler, 195124 KULLBACK S & LEIBLER RA. 1951. On information and sufficiency. Ann. Math. Stat., 22(1):79-86.; Kullback, 195923 KULLBACK S. 1959. Information Theory and Statistics. New York: John Wiley.; Baldi & Brunak, 1998BALDI P & BRUNAK S. 1998. Bioinformatics: the machine learning approach. MIT Press,Cambridge, MA.).

The mutual information always satisfies 0≤I(D,M)≤H(D), where H(D)=- log -(1 - )log(1 - ). Thus in the assessment of performance of a classification model, it is habitual to use the normalized mutual information coefficient (Rost & Sander, 1993; Rost et al., 1994), which is given by,

The normalized mutual information satisfies 0≤IC(D,M)≤1. If IC(D,M)=0, then I(D,M)=0 and the prediction is completely random (D and M are independent); ifIC(D,M)=1, then I(D,M)=H(D)=H(M) and the prediction is perfect (Baldi et al., 2000BALDI P, BRUNAK S, CHAUVIN Y, ANDERSEN CAF & NIELSEN H. 2000. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics Review, 16(5): 412-424.).

4 SIMULATION STUDY

In this Section, we present results of the performed simulation study. The simulation study was conducted to compare the performance of the naive logistic regression, the logistic regression with state-dependent sample selection and the bounded logistic regression. Here, we considered some circumstances that may arise in the development of credit scoring models, and whichinvolve the number of clients in the training sample and its degree of unbalance.

First, we generated a population of clients of a hypothetical credit institution, with the following composition: 10,000,000 good payers and 100,000 bad payers. The data for good payers were generated from a six-dimensional multivariate normal distribution, with mean vector = (0, ...,0) and covariance matrix 4×I 6, where I 6 is the identity matrix of order 6; while the data for bad payers were generated from a six-dimensional multivariate normal distribution, with mean = and covariance matrix I 6 (Breiman, 199811 BREIMAN L. 1998. Arcing classifiers. The Annals of Statistics, 26(3): 801-849.). We categorized the six observed continuous covariates using the quartiles. Afterwards, we took out a stratified random sample (full sample) of 1,000,000 good payers (10% of the population size of this group) and 10,000 bad payers (10% of the population size of this class) of the generated population. Then the selected samples were obtained by keeping all bad payers of the full sample, plus 10,000*K good payers taken out randomly from the full sample. Here, we studied only the cases where K=1,3,9 and 19, which correspond to 10,000, 30,000, 90,000 and 190,000 good payers in the selected samples, respectively. For each K, we performed 100 simulations, i.e. 100 samples were obtained for each K. In each one of them, the good payers were selected from their group in the full sample via simple random sampling without replacement.

For each obtained sample and for each K we applied the following procedures: a naive logistic regression model (Hosmer & Lemeshow, 1989), a logistic regression with state-dependentsample selection model and a bounded logistic regression model (Cramer, 200413 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.) were fitted to the data and their predictive capacity was investigated in the training sample by the calculation of their performance measures.

After simulating for each K, we obtained vectors of length 100, i.e. vectors with 100 records for each of the studied performance measures. Then we could obtain an average estimate for each measure by computing the mean value of each vector.

We calculated the optimal cutoff point for every adjusted model. Finally, we also compared the original probabilities, estimated from the naive logistic regression models, with the adjusted probabilities, predicted from the logistic regression with state-dependent sample selection models, that are the original probabilities corrected by considering the principles of sample selection, and the predicted probabilities from the bounded logistic regression models. Such comparison was made as follows. After performing the simulation study, we obtained 100 vectors of estimated default probabilities, for each K (K=1,3,9 and 19) and for each of the three techniques studied. Then we sorted, in ascending order, each one of the 100 vectors (columns). Thus,the first row of the resulting spreadsheet contained the lowest estimated probabilities in each of the 100 simulations, while the last line comprised the highest estimated probabilities. We got empirical curves for the distributions of estimated probabilities, original, adjusted or predicted from the bounded logistic regression models, by computing the mean value of each row. The simulations were performed using SAS version 9.0. Interested readers can email the authors for the codes.

Tables 2, 3 and 4 show the empirical averages for the performance measures: naive logistic regression results are presented in Table 2, logistic regression with state-dependent sample selection results are presented in Table 3 and bounded logistic regression results are presented in Table 4. The empirical results presented in Table 2 reveal that the naive logistic regression technique produces good results, with high values of positive and negative predictive values among others, only when the sample used for the development of the model is balanced, i.e. when K=1. As the degree of imbalance increases (K=3, K=9 and K=19) positive predictive value decreases considerably, assuming values less than 0.5 when K=9 and K=19, whereas negative predictive value increases, reaching values near 1 for K=9 and K=19. Note that the values of MCC, IC and AC are also decreasing as K increases.

Table 2
Empirical averages for the performance measures, when we use the naive logistic regression technique.

Table 3
Empirical averages for the performance measures, when we use the logistic regression with state-dependent sample selection technique.

Table 4
Empirical averages for the performance measures, when we use the bounded logistic regression technique.

Discussions about the results of the logistic regression with state-dependent sample selection technique and bounded logistic regression technique (Tables 3 and 4 summarize these results, respectively) are analogous to those made previously, when we use the naive logistic regression technique.

Figures 1 and 2 present the empirical curves for the three studied models. It can be observed that regardless of the K value, the estimated probabilities from the bounded logistic regression model and the estimated probabilities without the adjustment to the constant term of the equation are quite similar and lower than those where the adjustment was made. So the bounded logistic regression model and the naive logistic regression model underestimate the probability ofdefault. Note also that the distance among the curves decreases as the degree of sample imbalance increases. The curves are closer when the degree of imbalance is close to the real one present in the full sample, that is, approximately 99% of good payers and 1% of bad payers. For instance, the distance between the curves is largest for K=1 (Fig. 1 upper panel), i.e. for balanced samples, while the curves are very close to each other for K=19 (Fig. 2 bottom panel), i.e. for 190,000 good payers and 10,000 bad payers in the selected samples.

Figure 1
Estimated cumulative probabilities, where —–—– represents the empirical curve for the original probability; −−−− represents the empirical curve for the adjusted probability; and · · · · · · the empirical curve for the predicted probability from bounded logistic regression model. Upper panel: K=1 and lower panel: K=3.

Figure 2
Estimated cumulative probabilities, where —–—– represents the empirical curve for the original probability; −−−− represents the empirical curve for the adjusted probability; and · · · · · · the empirical curve for the predicted probability from bounded logistic regression model. Upper panel: K=9 and lower panel: K=19.

The results presented in Tables 2, 3 and 4 are very close, showing no statistically significant difference between the majorities of corresponding empirical averages.

5 BRAZILIAN RETAIL BANK CREDIT SCORING DATA

In order to illustrate the measures of a model performance discussed so far, let us examine a data set taken from the overall clients from a Brazilian retail bank portfolio. The data classifies 5,912 clients who took out personal loans from this bank into good or bad payers, according to their credit histories. The variables considered are: client type, gender, age, marital status and length of residence. The dataset used for model-fitting, i.e. the training sample, has 4,139 clients (70% of the original sample) and the validation one, i.e. the test sample, has 1,773 clients (the remaining 30% of the original sample), both with roughly the same proportion of bad payers (30%). Here, we considered the training sample as the full sample. Then a balanced (supported by the good results in the simulation study) selected sample was obtained by maintaining all bad payers of the full sample, plus 122 good payers selected randomly from the full sample. Thus, a naive logistic regression model (Hosmer & Lemeshow, 198921 HOSMER WD & LEMESHOW S. 1989. Applied Logistic Regression. New York: John Wiley.), a logistic regression with state-dependent sample selection model and a bounded logistic regression model (Cramer, 200413 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.) were applied to the balanced selected sample and then the test sample was used to analyze the appropriateness of the adjusted models.

Figure 3 shows the ROC curves (panel on the left) together with the cumulative probabilities curves (panel on the right) for each fitting. As can be seen, the ROC curves for the three studied models are very close. With the help of such curves, we selected cutoff points equal to 0.45, 0.69 and 0.46 for the naive logistic regression model, the logistic regression with state-dependent sample selection model and the bounded logistic regression model, respectively. In the Figure 3 right panel we observe the distributions of estimated default probabilities in the test sample from the three studied techniques. Note that the naive logistic regression model, which provides us the original probability of a client being a bad payer, and the bounded logistic regression model underestimate the probability of default. The measures of the predictive capacity of the adjusted models are showed in Table 5 for the balanced selected sample and test sample. Note that the measures of the three models are very close and are indicative of good predictive capacity, which is confirmed by the ROC curves (Fig. 3 left panel).

Figure 3
The ROC curves constructed from the training sample of a bank's actual portfolio (panels on the left) and the corresponding distributions of estimated probabilities in an unbalanced test sample with 70% good payers and 30% bad payers (panels on the right). To the panels on the left, —–—– represents the naive logistic regression model; −−−− represents the logistic regression with state-dependent sample selection model; and · · · · · · the bounded logistic regression model. To the panels on the right, —–—– represents the original probability; −−−− represents the adjusted one; and · · · · · · the estimated probability from the bounded logistic regression model.

Table 5
The evaluation (performance measures) of the adjusted models in the balanced training sample and in the test sample (unbalanced with 70% good payers and 30% bad payers), where NLR is the naive logistic regression model, LRSD is the logistic regression with state-dependent sample selection model and BLR refers to the bounded logistic regression model.

6 FINAL COMMENTS

Credit scoring techniques have become one of the most important tools currently used by financial institutions such as banks, in the measurement and evaluation of the loans credit risk. Furthermore, credit scoring is regarded as one of the basic applications of misclassificationproblems that have attracted most attention during the past decades.

This paper compares, via simulation, three statistical techniques widely used in modeling credit scoring data: the naive logistic regression (Hosmer & Lemeshow, 198921 HOSMER WD & LEMESHOW S. 1989. Applied Logistic Regression. New York: John Wiley.), the logistic regression with state-dependent sample selection and the bounded logistic regression (Cramer, 200413 CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.). A case study was also performed in order to illustrate the presented procedures on a real data set.

Based on the simulation results, we discover that there is no difference between the cumulative distributions of the default estimated probabilities by the use of the naive logistic regression and bounded logistic regression techniques. However, there is difference between the default estimated probabilities from these two models and the predicted probabilities from the logistic regression with state-dependent sample selection model. Particularly, the naive logistic regression models and the bounded logistic regression models underestimate such probabilities. Although there is no difference concerning to the performance of adjusted models from such techniques when we use measures like sensitivity, specificity and accuracy among others, in the evaluation of such models. The simulation study also showed that regardless of which of these three statistical modeling techniques is used, there is a need for working with balanced samples, which ensure models with good measures of positive and negative predictive values and highaccuracy rate.

Thus, it is important always to work with balanced samples using the logistic regression with state-dependent sample selection, since such modeling lead us to the true probability of default and has predictive capacity similar to the naive logistic regression and bounded logistic regression models, while these two models underestimate the probability of default.

ACKNOWLEDGMENTS

The research is partially founded by CNPq, Brazil.

  • 1
    ABREU HJ. 2004. Aplicação de análise de sobrevivência em um problema de credit scoring e comparação com a regressão logística. Dissertação de Mestrado, DEs-UFSCar.
  • 2
    ALTMAN EI & SAUNDERS A. 1998. Credit risk measurement: Developments over the last 20 years. Journal of Banking & Finance, 21: 1721-1742.
  • 3
    ANDREEVA G, ANSELL J & CROOK J. 2007. Modelling profitability using survival combination scores. European Journal of Operational Research, 183(3): 1537-1549.
  • 4
    BAESENS B, GESTEL TV, VIAENE S, STEPANOVA M, SUYKENS J & VANTHIENEN J. 2003. Benchmarking state-of-the-art classification algorithms for credit scoring. The Journal of the Operational Research Society, 54(6): 627-635.
  • 5
    BALDI P & BRUNAK S. 1998. Bioinformatics: the machine learning approach. MIT Press,Cambridge, MA.
  • 6
    BALDI P, BRUNAK S, CHAUVIN Y, ANDERSEN CAF & NIELSEN H. 2000. Assessing the accuracy of prediction algorithms for classification: an overview. Bioinformatics Review, 16(5): 412-424.
  • 7
    BANASIK J, CROOK J & THOMAS L. 2001. Scoring by usage. The Journal of the Operational Research Society, 52(9): 997-1006.
  • 8
    BARON J. 1994. Thinking and Deciding (2 ed.). Oxford University Press, London.
  • 9
    BELLOTTI T & CROOK J . 2009. Credit Scoring with Macroeconomic Variables Using SurvivalAnalysis. The Journal of the Operational Research Society, 60(12): 1699-1707.
  • 10
    BOYES WJ, HOFFMAN DL & LOW SA. 1989. An econometric analysis of the bank credit scoring problem. Journal of Econometrics, 40(1): 3-14.
  • 11
    BREIMAN L. 1998. Arcing classifiers. The Annals of Statistics, 26(3): 801-849.
  • 12
    BURSET M & GUIGÓ R. 1996. Evaluation of gene structure prediction programs. Genomics, 34(3): 353-367.
  • 13
    CRAMER JS. 2004. Scoring bank loans that may go wrong: a case study. Statistica Neerlandica, 58(3): 365-380.
  • 14
    DESAI VS, CROOK JN & OVERSTREET GA. 1996. A comparison of neural networks and linear scoring models in the credit union environment. European Journal of Operational Research, 95(1): 24-37.
  • 15
    DUNN G & EVERITT BS. 1995. Clinical biostatistics: an introduction to evidence based medicine. London: Edward Arnold.
  • 16
    FISHER RA. 1936. The use multiple measurements in taxonomic problems. Annals of Eugenics, 7(7): 179-188.
  • 17
    FLEISS JL. 1981. Statistical methods for rates and proportions. New York: John Wiley.
  • 18
    GREENE W. 1998. Sample selection in credit-scoring models. Japan and the World Economy, 10(3): 299-316.
  • 19
    HAND DJ. 2006. Classifier Technology and the Illusion of Progress. Statistical Science, 21(1): 1-14.
  • 20
    HAND DJ & HENLEY WE. 1997. Statistical classification methods in consumer credit scoring: A review. Journal of the Royal Statistical Society: Series A (Statistics in Society), 160(3): 523-541.
  • 21
    HOSMER WD & LEMESHOW S. 1989. Applied Logistic Regression. New York: John Wiley.
  • 22
    KANG S & SHIN K. 2000. Customer credit scoring model using analytic hierarchy process. Informs & Korms, Seoul, 2197-2204. Korea.
  • 23
    KULLBACK S. 1959. Information Theory and Statistics. New York: John Wiley.
  • 24
    KULLBACK S & LEIBLER RA. 1951. On information and sufficiency. Ann. Math. Stat., 22(1):79-86.
  • 25
    LEE T & CHEN I. 2005. A two-stage hybrid credit scoring model using artificial neural networksand multivariate adaptive regression splines. Expert Systems with Applications, 28(4): 743-752.
  • 26
    LEE T, CHIU C, LU C & CHEN I . 2002. Credit scoring using the hybrid neural discriminant technique. Expert Systems with Applications, 23(3): 245-254.
  • 27
    LIM MK & SOHN SY. 2007. Cluster-based dynamic scoring model. Expert Systems with Applications, 32(2): 427-431.
  • 28
    MATTHEWS BW. 1975. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta, 405: 442-451.
  • 29
    MAZUCHELI JA, LOUZADA-NETO F & MARTINEZ EZ. 2008. Algumas medidas do valor preditivo de um modelo de classificação. Rev. Bras. Biom., 26(2): 83-91.
  • 30
    ROST B & SANDER C. 1993. Prediction of protein secondary structure at better than 70% accuracy. J. Mol. Biol., 232(2): 584-599.
  • 31
    ROST B, SANDER C & SCHNEIDER R. 1994. Redefining the goals of protein secondary structure prediction. J. Mol. Biol., 235(1): 13-26.
  • 32
    SARLIJA N, BENSIC M & BOHACEK Z. 2004. Multinomial model in consumer credit scoring. 10th International Conference on Operational Research. Trogir: Croatia.
  • 33
    SCARPEL RA & MILIONI AZ. 2002. Utilização conjunta de modelagem econométrica e otimização em decisões de concessão de crédito. Pesquisa Operacional, 22(1): 61-72.
  • 34
    SELAU LPR & RIBEIRO JLD. 2011. A systematic approach to construct credit risk forecast models. Pesquisa Operacional, 31(1): 41-56.
  • 35
    THIBODEAU LA. 1981. Evaluating diagnostic tests. Biometrics, 37(4): 801-804.
  • 36
    THOMAS LC. 2000. A survey of credit and Behavioural Scoring; Forecasting financial risk of lending to consumers. University of Edinburgh, Edinburgh, U.K.
  • 37
    THOMAS LC, EDELMAN DB & CROOK JN. 2002. Credit Scoring and Its Applications. Philadelphia: SIAM.
  • 38
    WANG ZX. 1994. Assessing the accuracy of protein secondary structure. Nat. Struct. Biol., 1(3): 145-146.

APPENDIX A

Sketch of the proof of (3). If the event I i=1 means that element i is included in the sample, from the Bayesian Theorem (Baron, 1994BARON J. 1994. Thinking and Deciding (2 ed.). Oxford University Press, London.), it is possible to see that,

completing the proof.

Publication Dates

  • Publication in this collection
    Jan-Apr 2015

History

  • Received
    01 Sept 2012
  • Accepted
    12 Dec 2013
Sociedade Brasileira de Pesquisa Operacional Rua Mayrink Veiga, 32 - sala 601 - Centro, 20090-050 Rio de Janeiro RJ - Brasil, Tel.: +55 21 2263-0499, Fax: +55 21 2263-0501 - Rio de Janeiro - RJ - Brazil
E-mail: sobrapo@sobrapo.org.br