Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2019
…
28 pages
1 file
Presentation given at the Credit Scoring & Credit Control Conference XVI in Edinburgh, UK. <strong>Abstract</strong> Score calibration is the process of empirically determining the relationship between a score and an outcome on some population of interest, and scaling is the process of expressing that relationship in agreed units. Calibration is often treated as a simple matter and attacked with simple tools – typically, either assuming the relationship between score and log-odds is linear and fitting a logistic regression with the score as the only covariate, or dividing the score range into bands and plotting the empirical log-odds as a function of score band. Both approaches ignore some information in the data. The assumption of a linear score to log-odds relationship is too restrictive and score banding ignores the continuity of the scores. While a linear score to log-odds relationship is often an adequate approximation, the reality can be much more interesting, with...
Knowledge-Based Systems, 2017
Financial institutions use credit scorecards for risk management. A scorecard is a data-driven model for predicting default probabilities. Scorecard assessment concentrates on how well a scorecard discriminates good and bad risk. Whether predicted and observed default probabilities agree (i.e., calibration) is an equally important yet often overlooked dimension of scorecard performance. Surprisingly, no attempt has been made to systematically explore different calibration methods and their implications in credit scoring. The goal of the paper is to integrate previous work on probability calibration, to reintroduce available calibration techniques to the credit scoring community, and to empirically examine the extent to which they improve scorecards. More specifically, using real-world credit scoring data, we first develop scorecards using different classifiers, next apply calibration methods to the classifier predictions, and then measure the degree to which they improve calibration. To evaluate performance, we measure the accuracy of predictions in terms of the Brier Score before and after calibration, and employ repeated measures analysis of variance to test for significant differences between group means. Furthermore, we check calibration using reliability plots and decompose the Brier Score to clarify the origin of performance differences across calibrators. The observed results suggest that post-processing scorecard predictions using a calibrator is beneficial. Calibrators improve scorecard calibration while the discriminatory ability remains unaffected. Generalized additive models are particularly suitable for calibrating classifier predictions.
2003
Recently, the consumer credit industry has experienced a sizeable growth, where scoring techniques have grown to outperform the traditional, judgmental manner of assessing credit risk. Using the data of a Belgian direct-mail company offering consumer credit, the authors have shown a clear improvement in comparison with its current credit evaluation system, constructed by an international company specialized in consumer credit scoring, and identify the size of this performance increase due to population drift versus model improvement. Considering the crucial impact of the accuracy of the score on the cost side (lowered credit risk) as well as on the revenue side (increased accepted applications), in this study, we highlight the importance of introducing different performance measures to quantify credit risk performance. Hence, instead of reporting predictive performance on the total sample, we have quantified the predictive performance in more detail into a graphical overview that ca...
Acta Universitatis Lodziensis. Folia Oeconomica, 2019
Granting a credit product has always been at the heart of banking. Simultaneously, banks are obligated to assess the borrower’s credit risk. Apart from creditworthiness, to grant a credit product, banks are using credit scoring more and more often. Scoring models, which are an essential part of credit scoring, are being developed in order to select those clients who will repay their debt. For lenders, high effectiveness of selection based on the scoring model is the primary attribute, so it is crucial to gauge its statistical quality. Several textbooks regarding assessing statistical quality of scoring models are available, there is however no full consistency between names and definitions of particular measures. In this article, the most common statistical measures for assessing quality of scoring models, such as the pseudo Gini index, Kolmogorov‑Smirnov statistic, and concentration curve are reviewed and their statistical characteristics are discussed. Furthermore, the author prop...
2019
I would also like to thank my committee members Dr. Robert Chun and Mr. Raghavendra Keshavamurthy, for their valuable time and suggestions during this project. Last, but not least, I would like to thank my parents, my sister and friends for supporting and believing in me. 2 LITERATURE REVIEW.
SSRN Electronic Journal, 2003
The New Basel Capital Accord will allow the determination of banks' regulatory capital requirements due to default probabilities which are estimated and forecasted from internal ratings. External ratings from rating agencies play fundamental roles in capital and credit markets. Discriminatory power of internal and external ratings is a key requirement for the soundness of a rating system in general and for the acceptation of a bank's internal rating systems under Basel II. Statistics such as the area under a receiver operating characteristic or the accuracy ratio, are widely used in practice as measures for the performance. This note shows that such measures should only be interpreted with caution. Firstly, the outcomes of the measures depend not only on the discrimination power of the rating system but mainly on the structure of the portfolio under consideration. Thus, the absolute values achieved do not measure the performance of a rating system solely. Secondly, comparisons of the outcomes between different portfolios, different time periods or both may be misleading. As a positive result we show that the value achieved by a rating system which predicts all default probabilities correctly can not be beaten.
REPRESENTATIONS, 2014
2013
Starting in the early 1990s credit scoring became widespread and central in credit granting decisions. Credit scores are scalar representations of default risk. They are used, in turn, to price credit, and as a result alter household borrowing and default decisions. We build on recent work on defaultable consumer credit under asymmetric information to develop a quantitative theory of credit scores. We construct and solve a rich and quantitatively-disciplined lifecycle model of consumption in which households have access to defaultable debt, and lenders are asymmetrically informed about household characteristics relevant to predicting default. We then allow lenders to keep record of inferences on the hidden type of a borrower, as well as a binary "flag" indicating a past default. These inferences arise endogenously from a signalling game induced by borrowers' need to obtain loans. We show how lenders' inferences evolve over the lifecycle as a function of household behavior in a way that can be naturally interpreted as "credit scores." In particular, we first show that lenders' assessments that a household has relatively low default risk matter significantly for the interest rates households pay. We then show that such assessments rise most sharply-and interest rates paid by borrowers fall most sharply (on the order of 5 − 6 percentage points)-when the bankruptcy flag is removed, consistent with work of Musto (2005). Lastly, we compare allocations across information regimes to provide a measure of the social value of credit scores, and the dependence of these measures on lenders' ability to observe borrower characteristics.
2007
Accurate credit-granting decisions are crucial to the efficiency of the decentralized capital allocation mechanisms in modern market economies. Credit bureaus and many .nancial institutions have developed and used credit-scoring models to standardize and automate, to the extent possible, credit decisions. We build credit scoring models for bankcard markets using the Office of the Comptroller of the Currency, Risk Analysis Division
Standard Bank, South Africa, currently employs a particular methodology when developing application or behavioural scorecards. One of the processes in this methodology involves model building using logistic regression. A key aspect of building logistic regression models entails variable selection which involves dealing with multicollinearity. The objective of this study was to investigate the impact of using different variance inflation factor (VIF) thresholds on the performance of these models in a predictive and discriminatory context and to study the stability of the estimated coefficients in order to advise the bank. The impact of the choice of VIF thresholds was researched by means of an empirical and simulation study. The empirical study involved analysing two large data sets that represent the typical size encountered in a retail credit scoring context. The first analysis concentrated on fitting the various VIF models and comparing the fitted models in terms of the stability of coefficient estimates and goodness-of-fit statistics while the second analysis focused on evaluating the fitted models' predictive ability over time. The simulation study was used to study the effect of multicollinearity in a controlled setting. All the abovementioned studies indicate that the presence of multicollinearity in large data sets is of much less concern than in small data sets and that the VIF criterion could be relaxed considerably when models are fitted to large data sets. Our recommendations in this regard have been accepted and implemented by Standard Bank.
ORiON, 2015
Standard Bank, South Africa, currently employs a methodology when developing application or behavioural scorecards that involves logistic regression. A key aspect of building logistic regression models entails variable selection which involves dealing with multicollinearity. The objective of this study was to investigate the impact of using different variance inflation factor 1 (VIF) thresholds on the performance of these models in a predictive and discriminatory context and to study the stability of the estimated coefficients in order to advise the bank. The impact of the choice of VIF thresholds was researched by means of an empirical and simulation study. The empirical study involved analysing two large data sets that represent the typical size encountered in a retail credit scoring context. The first analysis concentrated on fitting the various VIF models and comparing the fitted models in terms of the stability of coefficient estimates and goodness-of-fit statistics while the second analysis focused on evaluating the fitted models' predictive ability over time. The simulation study was used to study the effect of multicollinearity in a controlled setting. All the above-mentioned studies indicate that the presence of multicollinearity in large data sets is of much less concern than in small data sets and that the VIF criterion could be relaxed considerably when models are fitted to large data sets. The recommendations in this regard have been accepted and implemented by Standard Bank.
Higher Education, 2010
Os Dembos na Desconstrução das Retóricas Colonais, 2018
2014 9th International Forum on Strategic Technology (IFOST), 2014
Indian Journal of Surgical Oncology, 2020
Journal of Fish Biology, 2006
SVERIGES FORENSISKA PSYKOLOGERS FÖRENING, 2023
Psicología y Psicopedagogía, 2013
Journal of Environmental Quality, 2002