Upcoming SlideShare
×

# Consumer credit-risk3440

818 views
732 views

Published on

Credit Risk Summer

Published in: Education
0 Likes
Statistics
Notes
• Full Name
Comment goes here.

Are you sure you want to Yes No
• Be the first to comment

• Be the first to like this

Views
Total views
818
On SlideShare
0
From Embeds
0
Number of Embeds
4
Actions
Shares
0
37
0
Likes
0
Embeds 0
No embeds

No notes for slide
• will not calculate until another chapter.
• will not calculate until another chapter.
• Revolving credit: e.g. credit card
• Consumer lending exception: CapitalOne , “a high tech firm that happens to be in the credit card industry”, quote of the founder
• will not calculate until another chapter.
• Consumer lending extensively uses credit scoring technique.
• Categorical data. Dataset is usually very big : say 100,000 individuals.
• Categorical data. Dataset is usually very big : say 100,000 individuals.
• 1. Distributional features of data
• The nature of consumer credit lends itself to statistical analysis.
• DA: e.g. Altman’s Z-score model. Normality: law of large numbers for large samples; so important for small samples
• Apparently the further the two distributions are separated, the better the credit score model can discriminate good and bad credits. There are several measures that can be used to gauge the difference between the two distribution, e.g. Wilks ’ lamda, information value, alpha/beta error etc. see pp92 – 99 of Handbook.
• How are the coefficients derived ? – see Altman ’ s paper. Altman selects the 5 variables from a list of 22 variables as doing the best overall job together in predicting corporate bankruptcy. Uses iterative procedures to evaluate different combinations of eligible variables and selects the profile that does the best job among the alternatives. Two sets of sample firms are used: healthy and bankrupt to find the discriminant function .
• heteroscedasticity: use general least square method U taking 2 values: u = y – beta*x and y = 0, or 1; so 2 values for each i.
• Mathematical programming: An objective criterion to optimize: e.g. the proportion of applicants correctly classified Subject to certain discriminating conditions (to discriminate good and bad credits) Recursive partitioning algorithm: A computerized nonparametric technique based on pattern recognition Results: a binary classification tree which assigns objects into selected a priori groups Terminal nodes represent final classification of all objects. From two samples of default and non-default firms, e.g., one can calculate the misclassified numbers. Expert systems : evidence shows the predictive performances are quite poor . Also known as artificial intelligence systems Computer-based decision-support systems A consultation module: asking users questions until enough evidence has been collected to support a final recommendation A knowledge base containing static data, algorithms, and rules that tell the system ‘what to do if’ A knowledge acquisition and learning module: rules from the lending officer and rules of its own
• Classification methods that are easy to understand (such as regression, nearest neighbor and tree-based methods) are much more appealing , than those which are essentially black boxes (such as neural networks) But neural networks have advantages too
• Two group : one is defaulted, the other non-defaulted. Logistic regression is hence more robust than linear models. Normality : for binary data , such as 0 or 1, it’s hard to justify they have normal distribution. This is severe in significance testing.
• ML method: we want to obtain estimates of parameters that make the observations most likely to happen . This is usually done by specifying the exact distributions of error terms, i.e. the likelihood function.
• The inverse of h is called the logit function : g=log[P/(1-P)] Note that function h guarantees P is between 0 and 1 , as required by probability.
• M is the number of 1’s. that is, the number of defaults in the sample. Order l like this: the first are defaulted ones, then the non-defaulted ones. The probability is conditional here, i.e. conditional on X of the sample Likelihood function serves as the (minus) loss function that is to be optimized. The last function is the cumulative logistic distribution function
• Simplification is correct: Pi_1_m (P)*Pi_m+1_n (1-P) = Pi_1_m (P/(1-P))*Pi_1_n (1-P) When do MLE, use the log version of the above formula NEED TO correct
• All the observations are either default or non-default (so P’s are either 1 or 0 for the observations in the sample). Logit = log(odds) Log(odds) translates the odds of default to nondefault to be the opposite of the odds of nondefault to default (e.g. 2 vs. -2) Then the logit function is assumed to be linear . Not solvable by OLS: coz P is wither 1 or 0. coefficients are solved through MLE . If data set is large enough, we can use the sample relative frequency as an estimate of the true probability for each X level, then we have values of the logits, then OLS can be used This website is a logistic calculator: http://members.aol.com/johnp71/logistic.html
• The regression line would be nonlinear None of the observations actually fall on the regression line. They all fall on 0 or 1 .
• 1. Explain thin tail implications: extreme events
• Probit model uses probit function that maps the probability to a numerical value between – inf to inf. The probit function is the inverse of the normal cdf. Assume the observations are independent, one solves for beta ’ s using the ML estimation . See pp100 – 101 Handbook. The other link function often used is the logit function
• Transfer function : that converts the combination of inputs to an output.
• Several Neural network models: multilayer perceptron (MLP): best model for credit scoring purpose. mixture of experts (MOE) radial basis function (RBF) learning vector quantization (LVQ) fuzzy adaptive resonance (FAR) The weighted combination of inputs is called NET
• The overall input g for a neuron is called potential . potential g is a linear combination of weights for the inputs X to the neuron The activation function is also called the transfer function that converts the potential to an output f. W0 is called bias , or the excitation threshold value . It’s like a constant in the regression model. One can set x0=1 and i starts from 0 in the summation sign
• SVM can perform binary classification (pattern recognition) and real valued function approximation (regression estimation) tasks.
• 1. Smiley faces: good credits; stars: bad credits
• Largest margin means the strongest differentiating power/most robust as any additional obligor that falls out of the margin region can be clearly identified; that falls into the middle is hard to label but this would incur no error in labeling. Support vectors: X (the data) associated with the two obligors are called support vectors .
• NW regression: a nonparametric kernel method Fan’s paper is cited in Atiya’s review paper: predicting bankruptcies
• Data get outdated : e.g. income will change; so behavior will change Application scores: the scores computed for applications, i.e. whether to extend facility based on this score Behavior scores: after facility has been granted Probability scores : not only want to know binary result, i.e. 0 and 1, but also the expected probability. This is important , e.g. calculating capitals and expected returns
• will not calculate until another chapter.
• will not calculate until another chapter.
• will not calculate until another chapter.
• will not calculate until another chapter.
• Revolving credit: e.g. credit card
• Revolving credit: e.g. credit card
• will not calculate until another chapter.
• In this example, X has 2 dimensions. W is perpendicular to the line.
• Use the 2 nd equation minus the 1 st one: w’lamda = 2 =&gt; margin = 2/|w|
• Minimizing |w| is equivalent to maximizing margin n refers to the number of obligors . The constraints: two groups must lie on either side of the margin y identifies default or non-default. Note the label here is different from other methods (say 0 and 1) This is linear SVM , the data are assumed to be linearly separable . (the constraint) The constraint is binding only for support vectors!
• If the constraint is binding (i.e. =0), then lamda &gt; 0 ( economic meaning : the shadow price of the constraint) If not binding (&gt;0), then lamda = 0
• Using dual is more convenient. See: http://en.wikipedia.org/wiki/Linear_programming#Duality Intuition: in primal problem, max the constraint to meet objective; in dual problem, min objective to meet constraint (use a graph to show this). Lamda is price, want to max it so more constraints are binding (less slack).
• ### Consumer credit-risk3440

1. 1. MFIN 7011: Credit Risk Summer, 2008 Dragon Tang <ul><li>Lecture 15 </li></ul><ul><li>Consumer Credit Risk </li></ul><ul><li>Thursday, August 7, 2008 </li></ul><ul><li>Readings: </li></ul><ul><li>Niu (2004); Agarwal, Chomsisengphet, Liu, and Souleles (2006) </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
2. 2. Consumer Credit Risk Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Objectives : </li></ul><ul><li>Credit scoring approach for consumer credit risk </li></ul><ul><li>Practice, challenge, and opportunity </li></ul>
3. 3. Consumer Credit Summer 08, MFIN7011, Tang Consumer Credit Risk Default Risk (low in general) Low High Credit Products Fixed Term Revolving Residential Mortgage Retail Finance Personal Loans Overdrafts Credit Cards
4. 4. Consumer Lending <ul><li>Examples: </li></ul><ul><ul><li>Automobile loans </li></ul></ul><ul><ul><li>Home equity loans </li></ul></ul><ul><ul><li>Revolving credit </li></ul></ul><ul><li>There is an exponential growth in consumer credit outstanding in the US, from USD 9.8 billion in 1946 to USD 2411 billion in January 2007 </li></ul><ul><ul><li>\$878 billion revolving; \$1526 billion non-revolving </li></ul></ul><ul><ul><li>Currently interest rate is 13%; interest accessed is 15% </li></ul></ul>Consumer Credit Risk Summer 08, MFIN7011, Tang
5. 5. Consumer vs. Corporate Lending <ul><li>Consumer lending is not as glamorous as corporate lending </li></ul><ul><li>Consumer lending is a volume business, where low cost producers who can manage the credit losses are able to enjoy profitable margins </li></ul><ul><li>Corporate lending is often unprofitable as every bank is chasing the same corporate customers, depressing margins </li></ul>Consumer Credit Risk Summer 08, MFIN7011, Tang
6. 6. Consumer Credit Risk: Art or Science? Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Art: consumers care about reputation </li></ul><ul><ul><li>Value of reputation is hard to model </li></ul></ul><ul><ul><li>Reduced form model may be useful </li></ul></ul><ul><li>Science: creditworthiness can be predicted from financial health </li></ul><ul><ul><li>Using structural models of Merton type </li></ul></ul><ul><li>The answer is probably both! </li></ul><ul><ul><li>Hybrid structural-reduced form model should be most promising </li></ul></ul>
7. 7. Summer 08, MFIN7011, Tang Consumer Credit Risk Never make predictions, especially about the future. — Casey Stengel
8. 8. The credit Decision Scoring vs. Judgmental <ul><li>Both methods </li></ul><ul><ul><li>Assume that the future will resemble the past </li></ul></ul><ul><ul><li>Compare applicants to past experience </li></ul></ul><ul><ul><li>Aim to grant credit only to acceptable risks </li></ul></ul><ul><li>Added value of scoring </li></ul><ul><ul><li>Defines degree of credit risk for each applicant </li></ul></ul><ul><ul><li>Ranks risk relative to other applicants </li></ul></ul><ul><ul><li>Allows decisions based on degree of risk </li></ul></ul><ul><ul><li>Enables tracking of performance over time </li></ul></ul><ul><ul><li>Permits known and measurable adjustments </li></ul></ul><ul><ul><li>Permits decision automation </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
9. 9. Evaluating the credit applicant Summer 08, MFIN7011, Tang Consumer Credit Risk Time at present address Time at present job Residential status Debt ratio Bank reference Age Income # of Recent inquiries % of Balance to avail. lines # of Major derogs. Overall Decision Odds of repayment • • • CHARACTERISTICS + + - + + N / A - - + + + Accept ? • • • JUDGMENT 12 20 5 21 28 15 5 -7 10 35 212 Accept 11:1 • • • CREDIT SCORING
10. 10. Credit Scoring <ul><li>Project </li></ul><ul><ul><li>Input x feature vector </li></ul></ul><ul><ul><li>Label y, default or not </li></ul></ul><ul><ul><li>Data (x i , y i ) </li></ul></ul><ul><ul><li>Target y=f(x) </li></ul></ul><ul><li>Objective </li></ul><ul><ul><li>Given new x, predict y so that probability of error is minimal </li></ul></ul>Consumer Credit Risk Summer 08, MFIN7011, Tang
11. 11. Typical Input Data <ul><li>Time at present address 0-1, 1-2, 3-4, 5+ years </li></ul><ul><li>Home status Owner, tenant, other </li></ul><ul><li>Telephone Yes, no </li></ul><ul><li>Applicant's annual income \$(0-10000), \$(11000-20000), \$(21000+) </li></ul><ul><li>Credit card Yes, no </li></ul><ul><li>Type of bank account Cheque and/or savings, none </li></ul><ul><li>Age 18-25, 26-40, 41-55, 55+ years </li></ul><ul><li>Type of occupation Coded </li></ul><ul><li>Purpose of loan Coded </li></ul><ul><li>Marital status Married, divorced, single, widow </li></ul><ul><li>Time with bank Years </li></ul><ul><li>Time with employer Years </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
12. 12. Input Data: FICO Score Summer 08, MFIN7011, Tang Consumer Credit Risk Not in the score: demographic data
13. 13. Characteristics of Data <ul><li>X: </li></ul><ul><ul><li>Continuous </li></ul></ul><ul><ul><li>Discrete </li></ul></ul><ul><ul><li>Normal distribution? </li></ul></ul><ul><li>Y: </li></ul><ul><ul><li>Binary data: 0 or 1 (=default) </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
14. 14. Scoring Models <ul><li>Statistical Methods </li></ul><ul><ul><li>DA (Discriminant Analysis) </li></ul></ul><ul><ul><li>Linear regression </li></ul></ul><ul><ul><li>Logistic regression </li></ul></ul><ul><ul><li>Probit analysis </li></ul></ul><ul><ul><li>Non-parametric models </li></ul></ul><ul><ul><ul><li>Nearest-neighbor approach </li></ul></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
15. 15. Statistical Methods: Discriminant Analysis <ul><li>Multivariate statistical analysis: several predictors (independent variables) and several groups (categorical dependent variable, e.g. 0 and 1) </li></ul><ul><li>Predictive DA: for a new observation, calculate the discriminant score, then classify it according to the score </li></ul><ul><li>The objective is to maximize the between group to within group sum of squares ratio that results in the best discrimination between the groups (within group variance is solely due to randomness; between group variability is due to the difference of the means) </li></ul><ul><li>Normal distribution for the response variables (dependent variables) is assumed (but normality only becomes important if significance tests are to be taken for small samples) </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
16. 16. Statistical Credit Scoring Summer 08, MFIN7011, Tang Consumer Credit Risk Credit Score #Customers Good Credit Bad Credit Cut-off Score
17. 17. Statistical Credit Scoring <ul><li>Credit scoring systems: </li></ul><ul><li>Altman Z-score model: </li></ul><ul><li>Z = .012 X 1 +.014 X 2 +.033 X 3 +.006 X 4 +1.0 X 5 </li></ul><ul><ul><li>X 1 = working capital/total assets ratio </li></ul></ul><ul><ul><li>X 2 = retained earnings/total assets ratio </li></ul></ul><ul><ul><li>X 3 = earnings before interest and taxes/total assets ratio </li></ul></ul><ul><ul><li>X 4 = market value of equity/book value of total liabilities ratio </li></ul></ul><ul><ul><li>X 5 = sales/total assets ratio </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
18. 18. Statistical Methods: Linear Regression <ul><li>The regression model is like: </li></ul><ul><li>For the true model, u can take only two values as Y; thus u can’t be normally distributed. </li></ul><ul><li>u has heteroskedastic variances, which makes the OLS inefficient </li></ul><ul><li>The estimated probability may well lie outside [0,1]. </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
19. 19. Statistical Methods: Nearest-Neighbor Approach <ul><li>A historical database has been divided into two groups (good and bad) </li></ul><ul><li>When a consumer comes, calculate the distance between the consumer and everyone in the database </li></ul><ul><li>The consumer will be classified in the category which is the same as the nearest one(s) </li></ul><ul><li>Problems: </li></ul><ul><ul><li>The definition of distance and the number of the nearest ones </li></ul></ul><ul><ul><li>Scoring speed: when a new x comes, we need calculate the distance between the new x and all of the historical data; too much calculation! </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
20. 20. Scoring Models <ul><li>Non-statistical Methods </li></ul><ul><ul><li>Mathematical programming </li></ul></ul><ul><ul><li>Recursive partitioning </li></ul></ul><ul><ul><li>Expert systems </li></ul></ul><ul><ul><li>Machine Learning </li></ul></ul><ul><ul><ul><li>Neural Networks </li></ul></ul></ul><ul><ul><ul><li>Support Vector Machine (SVM) </li></ul></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
21. 21. Which Method is Best? <ul><li>In general there is no overall best method. What is best will depend on the details of the problem: </li></ul><ul><ul><li>The data structure </li></ul></ul><ul><ul><li>The characteristics used </li></ul></ul><ul><ul><li>The extent to which it is possible to separate the classes by using those characteristics </li></ul></ul><ul><ul><li>The objective of the classification (overall misclassification rate, cost-weighted misclassification rate, bad risk rate among those accepted, some measure of profitability, etc.) </li></ul></ul><ul><li>In the following slides, we will introduce three models, Logistic, Neural Networks, and SVM in detail, which are used widely today </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
22. 22. Logistic Regression <ul><li>Empirical studies show, logistic regression may perform better than linear models (Hence, better than Discriminant Analysis), when data is nonnormal (particularly for binary data), or when covariance matrices of the two groups are not identical. </li></ul><ul><li>Therefore, logistic regression is the preferred method among the statistical methods </li></ul><ul><li>Probit regression is similar to logistic regression </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
23. 23. Performing Logistic Regression <ul><li>Logistic Regression can be performed using the Maximum Likelihood method </li></ul><ul><li>In the maximum likelihood method, we are seeking parameter values that maximize the likelihood of the observations occurring </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
24. 24. Logistic Regression: Setup <ul><li>Directly models the default probability as a function of the input variables X (a vector) </li></ul><ul><li>Define </li></ul><ul><li>Assume </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
25. 25. Logistic Regression: Setup <ul><li>Assume the observations are independent, the probability (likelihood) of the observed sample is given by </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
26. 26. Logistic Regression and ML <ul><li>ML estimator (of the coefficients a’s) for Logistic Regression can be found by applying non-linear optimization on the above likelihood function. </li></ul><ul><li>The simplified version is given by </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
27. 27. Logistic Regression and ML <ul><li>It is easy to show that the log of the odds (= logit) are a linear function: </li></ul><ul><li>Therefore, the odds per se are a multiplicative function. </li></ul><ul><li>Since probability takes on values between (0,1), the odds take on values between (0,∞), logits take on values between (-∞,∞). So, it looks very much like linear regression, and it does not need to restrict the dependent variable to values of {0, 1}. </li></ul><ul><li>It is not solvable using OLS. </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
28. 28. Logistic Function and Distribution Summer 08, MFIN7011, Tang Consumer Credit Risk
29. 29. Normal Distribution Summer 08, MFIN7011, Tang Consumer Credit Risk The tails are much thinner than Logistic
30. 30. RiskCalc: Moody ’ s Default Model <ul><li>Probit Regression </li></ul><ul><ul><li>Where x is the vector of the ratios </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
31. 31. Neural Networks <ul><li>Non-parametric method </li></ul><ul><li>Non-linear model estimation technique: e.g. </li></ul><ul><ul><li>Saturation effect: i.e. marginal effect of a financial ratio may decline quickly </li></ul></ul><ul><ul><li>Multiplicative factors: highly leveraged firms have a harder time borrowing money </li></ul></ul><ul><li>Neural networks decide how to combine and transform the raw characteristics in the data, as well as yielding estimates of the parameters of the decision surface </li></ul><ul><li>Well suited to situations where we have a poor understanding of the data structure </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
32. 32. Neural Networks <ul><li>Use the logistic function as the activation function in all the nodes </li></ul><ul><li>Works well with classification problems </li></ul><ul><li>Drawbacks </li></ul><ul><ul><li>May take much longer to train </li></ul></ul><ul><ul><li>In credit scoring, there is solid understanding of data </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
33. 33. Multilayer Perceptron (MLP) <ul><li>The input values X are sent along with 1 to the hidden layer neuron </li></ul><ul><li>The hidden layer generates a weight and generates a nonlinear output that is sent to the next layer </li></ul><ul><li>The output neuron takes 1 with input from the hidden layer and generates the output signal </li></ul><ul><li>When learning occurs, the weights are adjusted so that the final OUTs produce the least error (The output of a single neuron is called OUT) </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk X1 X2 1 H1 H2 1 O Input Layer Hidden Layer Output Layer w01 w12 w21 w22 w11 w02 w1 w2 w0
34. 34. Multilayer Perceptron (MLP) <ul><li>Input nodes do not perform processing </li></ul><ul><li>Each hidden and output node processes the signals by an activation function. The most frequently used is given on the right. </li></ul><ul><li>The parameters, w, are obtained by “training” the Neural Net to historical data. </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
35. 35. Support Vector Machine (SVM) <ul><li>A relatively new promising supervised learning method for </li></ul><ul><ul><li>Pattern recognition (Classification) </li></ul></ul><ul><ul><li>Regression estimation </li></ul></ul><ul><li>This originates from the statistical learning theory developed by Vaqnik and Chervonenkis </li></ul><ul><ul><li>1960s, Vapnik V. N., Support Vector </li></ul></ul><ul><ul><li>1995, Statistical Learning Theory </li></ul></ul><ul><ul><ul><li>Vapnik, V. N., “The Nature of Statistical Learning Theory”. New York: Springer-Verlag, 1995 2 </li></ul></ul></ul><ul><ul><ul><li>Cortes C. and Vapnik, V. N., “Support Vector Networks”, Machine Learning, 20:1-25,1995 </li></ul></ul></ul><ul><ul><li>Development, from 1995 to now </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
36. 36. SVM Extension <ul><li>Proximal Support Vector Machine (PSVM) </li></ul><ul><ul><li>Glenn Fung and Olvi L. Mangasariany 2001 </li></ul></ul><ul><li>Incremental and Decremental Support Vector Machine Learning </li></ul><ul><li>Least Squares Support Vector Machine (LS-SVM) </li></ul><ul><li>Also, SVMs can be seen as a new training method for learning machines (such as NNs) </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
37. 37. Linear Classifier <ul><li>There are infinitely many lines that have zero training error. </li></ul><ul><li>Which line should we choose? </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
38. 38. <ul><li>Choose the line with the largest margin . </li></ul><ul><ul><li>The optimal separating hyperplane (OSH) </li></ul></ul><ul><li>The “large margin classifier” </li></ul>Linear Classifier Summer 08, MFIN7011, Tang Consumer Credit Risk margin ” Support Vectors”
39. 39. Performance of SVM <ul><li>S&P CreditModel White Paper </li></ul><ul><li>Fan and Palaniswami (2000): </li></ul><ul><ul><li>SVM 70.35%–70.90% </li></ul></ul><ul><ul><li>NN 66.11%–68.33% </li></ul></ul><ul><ul><li>MDA 59.79%–63.68% </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
40. 40. Credit Scoring and Beyond <ul><li>Data collected at application will become outdated pretty fast </li></ul><ul><li>The way a customer uses its credit account is an indicator for future performance (Behavior Scoring) </li></ul><ul><li>This leads to an update path of PD and credit control tools </li></ul><ul><li>The future is moving into profitability scoring. </li></ul><ul><ul><li>Banks should not only care about getting its money back </li></ul></ul><ul><ul><li>Banks want to extend credit to those it can make a positive NPV, risk-adjusted </li></ul></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
41. 41. Best Practice in Consumer Credit Risk Management Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Credit decision-making </li></ul><ul><ul><li>Adopt to changes in economy or within customer segment </li></ul></ul><ul><li>Credit scoring </li></ul><ul><ul><li>Adaptive algorithms using credit bureau data and firm’s own experience </li></ul></ul><ul><li>Loss forecasting </li></ul><ul><ul><li>Historical delinquency rates and charge-off trend analysis </li></ul></ul><ul><ul><li>Delinquency flow and segmented vintage analysis </li></ul></ul><ul><li>Portfolio management </li></ul><ul><ul><li>Risk adjusted return on capital (RAROC) </li></ul></ul>
42. 42. Analytical Techniques Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Response analysis: avoid adverse selection consequences that result in increased concentrations of high-risk borrowers </li></ul><ul><li>Pricing strategies: avoid “follow the competition”, focus on segment profitability and cash flow </li></ul><ul><li>Loan amount determination: avoid to be judgmental, quantify probabilities of losses </li></ul><ul><li>Credit loss forecasting: decompositional roll rate modeling, trend and seasonal indexing, and vintage curve </li></ul><ul><li>Portfolio management strategies: important for repricing and retention, don’t be judgmental, integrating behavioral element and cash flow profitability analysis (underwriting) </li></ul><ul><li>Collection strategies: behavioral models are useful </li></ul>
43. 43. Credit Scoring and Loss Forecasting Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Two critical components of consumer credit risk analysis </li></ul><ul><ul><li>Corresponds to default probabilities and loss given default </li></ul></ul><ul><li>These two are linked </li></ul><ul><ul><li>Loss given default is higher when default probability is greater </li></ul></ul><ul><li>Market and economic variables matter </li></ul><ul><ul><li>In bad economic states, there will be more default and lower recovery </li></ul></ul><ul><li>Good modeling should achieve stability </li></ul>
44. 44. Do Consumers Choose the Right Credit Contracts? Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Agarwal, Chomsisengphet, Liu, and Souleles (2006): </li></ul><ul><li>Some don’t, especially when the stake is small </li></ul><ul><li>But consumers with high balance do! </li></ul><ul><li>Other issues: </li></ul><ul><li>Personal bankruptcy in the U.S. soared! </li></ul><ul><li>Avoid/fight predatory lending! (e.g., subprime lending) </li></ul><ul><li>China is starting to have a consumer credit market </li></ul>
45. 45. China ’ s Consumer Spending Consumer Credit Risk Summer 08, MFIN7011, Tang %Chg 1997 1998 1999 2000 2001 2002 2003 97-03 Food 2684 2756 2845 3029 3326 3487 3789 41% Medicine&Healthcare 213 255 300 356 401 455 506 138% Clothing 785 750 728 791 866 885 958 22% Household Durables 414 485 569 595 657 727 790 91% Transport&Communication 290 337 385 437 498 554 614 112% Education&Entertainment 550 643 739 837 945 1057 1170 113% Housing 424 507 599 663 752 842 931 120% Services 244 268 296 330 367 400 441 80% TOTAL 5603 6001 6462 7037 7811 8407 9198 64%
46. 46. China ’ s Consumer Credit Market <ul><li>1999-2004: Growth rate 52% </li></ul><ul><ul><li>Automobile loans: 110% </li></ul></ul><ul><ul><ul><li>Only 15% of auto sales, compared to 80% in U.S. </li></ul></ul></ul><ul><ul><li>Bankcard: 36% </li></ul></ul><ul><ul><ul><li>Mostly debit cards </li></ul></ul></ul><ul><ul><li>Mortgage: 1000% </li></ul></ul><ul><ul><ul><li>Still a long way to go! Only 8% of GDP, compared to 45% in developed economies </li></ul></ul></ul><ul><li>Other markets </li></ul><ul><ul><li>Student loan </li></ul></ul><ul><ul><li>Credit cards! </li></ul></ul><ul><li>More opportunities are waiting! </li></ul>Consumer Credit Risk Summer 08, MFIN7011, Tang
47. 47. Summer 08, MFIN7011, Tang Consumer Credit Risk 6
48. 48. Summary Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Introduction to Consumer Credit Risk: </li></ul><ul><ul><li>Credit scoring methods </li></ul></ul><ul><ul><li>Practical issues </li></ul></ul><ul><li>Next: Credit VaR and Basel II </li></ul>
49. 49. SVM Approach Details Summer 08, MFIN7011, Tang Consumer Credit Risk
50. 50. <ul><li>The plane separating and is defined by </li></ul><ul><li>The dashed planes are given by </li></ul>Computing the Margin Summer 08, MFIN7011, Tang Consumer Credit Risk margin w
51. 51. <ul><li>Divide by b </li></ul><ul><li>Define new w = w/ b and α = a/b </li></ul>Computing the Margin Summer 08, MFIN7011, Tang Consumer Credit Risk margin w We have defined a scale for w and a
52. 52. <ul><li>We have </li></ul><ul><li>which gives </li></ul>Computing the Margin Summer 08, MFIN7011, Tang Consumer Credit Risk margin  w) x x +  w)
53. 53. Quadratic Programming Problem Summer 08, MFIN7011, Tang Consumer Credit Risk <ul><li>Maximizing the margin is equivalent to minimizing || w || 2 . </li></ul><ul><li>Minimize || w || 2 subject to the constraints: </li></ul>Where we have defined y(n) = +1 for all y(n) = –1 for all This enables us to write the constraints as
54. 54. Quadratic Programming Problem Summer 08, MFIN7011, Tang Consumer Credit Risk Minimize the cost function (Lagrangian) Here we have introduced non-negative Lagrange multipliers l n  0 that express the constraints
55. 55. Quadratic Programming Problem <ul><li>The first order conditions evaluated at the optimal solution are </li></ul><ul><li>The solution can be derived (together with the constraint) </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
56. 56. Quadratic Programming Problem <ul><li>The original minimizing problem is equivalent to the following maximizing problem (dual) </li></ul><ul><li>For non-support vectors, λ will be zero, as the original constraint is not binding; only a few λ ’s would be nonzero. </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk
57. 57. Quadratic Programming Problem <ul><li>Having solved for the optimal λ ’s (denoted as ), we can derive others </li></ul><ul><li>To classify a new data point x, simply solve </li></ul>Summer 08, MFIN7011, Tang Consumer Credit Risk