Summer 07-mfin7011-tang1922


Published on

Credit Risk Management

Published in: Education
  • Be the first to comment

  • Be the first to like this

No Downloads
Total views
On SlideShare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide
  • will not calculate until another chapter.
  • will not calculate until another chapter.
  • Revolving credit: e.g. credit card
  • Consumer lending exception: CapitalOne , “a high tech firm that happens to be in the credit card industry”, quote of the founder
  • will not calculate until another chapter.
  • Consumer lending extensively uses credit scoring technique.
  • Categorical data. Dataset is usually very big : say 100,000 individuals.
  • Categorical data. Dataset is usually very big : say 100,000 individuals.
  • 1. Distributional features of data
  • The nature of consumer credit lends itself to statistical analysis.
  • DA: e.g. Altman’s Z-score model. Normality: law of large numbers for large samples; so important for small samples
  • Apparently the further the two distributions are separated, the better the credit score model can discriminate good and bad credits. There are several measures that can be used to gauge the difference between the two distribution, e.g. Wilks’ lamda, information value, alpha/beta error etc. see pp92 – 99 of Handbook.
  • How are the coefficients derived ? – see Altman’s paper. Altman selects the 5 variables from a list of 22 variables as doing the best overall job together in predicting corporate bankruptcy. Uses iterative procedures to evaluate different combinations of eligible variables and selects the profile that does the best job among the alternatives. Two sets of sample firms are used: healthy and bankrupt to find the discriminant function .
  • heteroscedasticity: use general least square method U taking 2 values: u = y – beta*x and y = 0, or 1; so 2 values for each i.
  • Mathematical programming: An objective criterion to optimize: e.g. the proportion of applicants correctly classified Subject to certain discriminating conditions (to discriminate good and bad credits) Recursive partitioning algorithm: A computerized nonparametric technique based on pattern recognition Results: a binary classification tree which assigns objects into selected a priori groups Terminal nodes represent final classification of all objects. From two samples of default and non-default firms, e.g., one can calculate the misclassified numbers. Expert systems : evidence shows the predictive performances are quite poor . Also known as artificial intelligence systems Computer-based decision-support systems A consultation module: asking users questions until enough evidence has been collected to support a final recommendation A knowledge base containing static data, algorithms, and rules that tell the system ‘what to do if’ A knowledge acquisition and learning module: rules from the lending officer and rules of its own
  • Classification methods that are easy to understand (such as regression, nearest neighbor and tree-based methods) are much more appealing , than those which are essentially black boxes (such as neural networks) But neural networks have advantages too
  • Two group : one is defaulted, the other non-defaulted. Logistic regression is hence more robust than linear models. Normality : for binary data , such as 0 or 1, it’s hard to justify they have normal distribution. This is severe in significance testing.
  • ML method: we want to obtain estimates of parameters that make the observations most likely to happen . This is usually done by specifying the exact distributions of error terms, i.e. the likelihood function.
  • The inverse of h is called the logit function : g=log[P/(1-P)] Note that function h guarantees P is between 0 and 1 , as required by probability.
  • M is the number of 1’s. that is, the number of defaults in the sample. Order l like this: the first are defaulted ones, then the non-defaulted ones. The probability is conditional here, i.e. conditional on X of the sample Likelihood function serves as the (minus) loss function that is to be optimized. The last function is the cumulative logistic distribution function
  • Simplification is correct: Pi_1_m (P)*Pi_m+1_n (1-P) = Pi_1_m (P/(1-P))*Pi_1_n (1-P) When do MLE, use the log version of the above formula NEED TO correct
  • All the observations are either default or non-default (so P’s are either 1 or 0 for the observations in the sample). Logit = log(odds) Log(odds) translates the odds of default to nondefault to be the opposite of the odds of nondefault to default (e.g. 2 vs. -2) Then the logit function is assumed to be linear . Not solvable by OLS: coz P is wither 1 or 0. coefficients are solved through MLE . If data set is large enough, we can use the sample relative frequency as an estimate of the true probability for each X level, then we have values of the logits, then OLS can be used This website is a logistic calculator:
  • The regression line would be nonlinear None of the observations actually fall on the regression line. They all fall on 0 or 1 .
  • 1. Explain thin tail implications: extreme events
  • Probit model uses probit function that maps the probability to a numerical value between –inf to inf. The probit function is the inverse of the normal cdf. Assume the observations are independent, one solves for beta’s using the ML estimation . See pp100 – 101 Handbook. The other link function often used is the logit function
  • Transfer function : that converts the combination of inputs to an output.
  • Several Neural network models: multilayer perceptron (MLP): best model for credit scoring purpose. mixture of experts (MOE) radial basis function (RBF) learning vector quantization (LVQ) fuzzy adaptive resonance (FAR) The weighted combination of inputs is called NET
  • The overall input g for a neuron is called potential . potential g is a linear combination of weights for the inputs X to the neuron The activation function is also called the transfer function that converts the potential to an output f. W0 is called bias , or the excitation threshold value . It’s like a constant in the regression model. One can set x0=1 and i starts from 0 in the summation sign
  • SVM can perform binary classification (pattern recognition) and real valued function approximation (regression estimation) tasks.
  • 1. Smiley faces: good credits; stars: bad credits
  • Largest margin means the strongest differentiating power/most robust as any additional obligor that falls out of the margin region can be clearly identified; that falls into the middle is hard to label but this would incur no error in labeling. Support vectors: X (the data) associated with the two obligors are called support vectors .
  • NW regression: a nonparametric kernel method Fan’s paper is cited in Atiya’s review paper: predicting bankruptcies
  • Data get outdated : e.g. income will change; so behavior will change Application scores: the scores computed for applications, i.e. whether to extend facility based on this score Behavior scores: after facility has been granted Probability scores : not only want to know binary result, i.e. 0 and 1, but also the expected probability. This is important , e.g. calculating capitals and expected returns
  • will not calculate until another chapter.
  • will not calculate until another chapter.
  • will not calculate until another chapter.
  • will not calculate until another chapter.
  • Revolving credit: e.g. credit card
  • Revolving credit: e.g. credit card
  • will not calculate until another chapter.
  • will not calculate until another chapter.
  • In this example, X has 2 dimensions. W is perpendicular to the line.
  • Use the 2 nd equation minus the 1 st one: w’lamda = 2 => margin = 2/|w|
  • Minimizing |w| is equivalent to maximizing margin n refers to the number of obligors . The constraints: two groups must lie on either side of the margin y identifies default or non-default. Note the label here is different from other methods (say 0 and 1) This is linear SVM , the data are assumed to be linearly separable . (the constraint) The constraint is binding only for support vectors!
  • If the constraint is binding (i.e. =0), then lamda > 0 ( economic meaning : the shadow price of the constraint) If not binding (>0), then lamda = 0
  • Using dual is more convenient. See: Intuition: in primal problem, max the constraint to meet objective; in dual problem, min objective to meet constraint (use a graph to show this). Lamda is price, want to max it so more constraints are binding (less slack).
  • Summer 07-mfin7011-tang1922

    1. 1. MFIN 7011: Credit Risk Management Summer, 2007 Dragon Tang <ul><li>Lecture 18 </li></ul><ul><li>Consumer Credit Risk </li></ul><ul><li>Thursday, August 2, 2007 </li></ul><ul><li>Readings: </li></ul><ul><li>Niu (2004); Agarwal, Chomsisengphet, Liu, and Souleles (2006) </li></ul>
    2. 2. Consumer Credit Risk <ul><li>Objectives : </li></ul><ul><li>Credit scoring approach for consumer credit risk </li></ul><ul><li>Practice, challenge, and opportunity </li></ul>
    3. 3. Consumer Credit Default Risk (low in general) Low High Credit Products Fixed Term Revolving Residential Mortgage Retail Finance Personal Loans Overdrafts Credit Cards
    4. 4. Consumer Lending <ul><li>Examples: </li></ul><ul><ul><li>Automobile loans </li></ul></ul><ul><ul><li>Home equity loans </li></ul></ul><ul><ul><li>Revolving credit </li></ul></ul><ul><li>There is an exponential growth in consumer credit outstanding in the US, from USD 9.8 billion in 1946 to USD 2411 billion in January 2007 </li></ul><ul><ul><li>$878 billion revolving; $1526 billion non-revolving </li></ul></ul><ul><ul><li>Currently interest rate is 13%; interest accessed is 15% </li></ul></ul>
    5. 5. Consumer vs. Corporate Lending <ul><li>Consumer lending is not as glamorous as corporate lending </li></ul><ul><li>Consumer lending is a volume business, where low cost producers who can manage the credit losses are able to enjoy profitable margins </li></ul><ul><li>Corporate lending is often unprofitable as every bank is chasing the same corporate customers, depressing margins </li></ul>
    6. 6. Consumer Credit Risk: Art or Science? <ul><li>Art: consumers care about reputation </li></ul><ul><ul><li>Value of reputation is hard to model </li></ul></ul><ul><ul><li>Reduced form model may be useful </li></ul></ul><ul><li>Science: creditworthiness can be predicted from financial health </li></ul><ul><ul><li>Using structural models of Merton type </li></ul></ul><ul><li>The answer is probably both! </li></ul><ul><ul><li>Hybrid structural-reduced form model should be most promising </li></ul></ul>
    7. 7. Never make predictions, especially about the future. — Casey Stengel
    8. 8. The credit Decision Scoring vs. Judgmental <ul><li>Both methods </li></ul><ul><ul><li>Assume that the future will resemble the past </li></ul></ul><ul><ul><li>Compare applicants to past experience </li></ul></ul><ul><ul><li>Aim to grant credit only to acceptable risks </li></ul></ul><ul><li>Added value of scoring </li></ul><ul><ul><li>Defines degree of credit risk for each applicant </li></ul></ul><ul><ul><li>Ranks risk relative to other applicants </li></ul></ul><ul><ul><li>Allows decisions based on degree of risk </li></ul></ul><ul><ul><li>Enables tracking of performance over time </li></ul></ul><ul><ul><li>Permits known and measurable adjustments </li></ul></ul><ul><ul><li>Permits decision automation </li></ul></ul>
    9. 9. Evaluating the credit applicant Time at present address Time at present job Residential status Debt ratio Bank reference Age Income # of Recent inquiries % of Balance to avail. lines # of Major derogs. Overall Decision Odds of repayment • • • CHARACTERISTICS + + - + + N / A - - + + + Accept ? • • • JUDGMENT 12 20 5 21 28 15 5 -7 10 35 212 Accept 11:1 • • • CREDIT SCORING
    10. 10. Credit Scoring <ul><li>Project </li></ul><ul><ul><li>Input x feature vector </li></ul></ul><ul><ul><li>Label y, default or not </li></ul></ul><ul><ul><li>Data (x i , y i ) </li></ul></ul><ul><ul><li>Target y=f(x) </li></ul></ul><ul><li>Objective </li></ul><ul><ul><li>Given new x, predict y so that probability of error is minimal </li></ul></ul>
    11. 11. Typical Input Data <ul><li>Time at present address 0-1, 1-2, 3-4, 5+ years </li></ul><ul><li>Home status Owner, tenant, other </li></ul><ul><li>Telephone Yes, no </li></ul><ul><li>Applicant's annual income $(0-10000), $(11000-20000), $(21000+) </li></ul><ul><li>Credit card Yes, no </li></ul><ul><li>Type of bank account Cheque and/or savings, none </li></ul><ul><li>Age 18-25, 26-40, 41-55, 55+ years </li></ul><ul><li>Type of occupation Coded </li></ul><ul><li>Purpose of loan Coded </li></ul><ul><li>Marital status Married, divorced, single, widow </li></ul><ul><li>Time with bank Years </li></ul><ul><li>Time with employer Years </li></ul>
    12. 12. Input Data: FICO Score Not in the score: demographic data
    13. 13. Characteristics of Data <ul><li>X: </li></ul><ul><ul><li>Continuous </li></ul></ul><ul><ul><li>Discrete </li></ul></ul><ul><ul><li>Normal distribution? </li></ul></ul><ul><li>Y: </li></ul><ul><ul><li>Binary data: 0 or 1 (=default) </li></ul></ul>
    14. 14. Scoring Models <ul><li>Statistical Methods </li></ul><ul><ul><li>DA (Discriminant Analysis) </li></ul></ul><ul><ul><li>Linear regression </li></ul></ul><ul><ul><li>Logistic regression </li></ul></ul><ul><ul><li>Probit analysis </li></ul></ul><ul><ul><li>Non-parametric models </li></ul></ul><ul><ul><ul><li>Nearest-neighbor approach </li></ul></ul></ul>
    15. 15. Statistical Methods: Discriminant Analysis <ul><li>Multivariate statistical analysis: several predictors (independent variables) and several groups (categorical dependent variable, e.g. 0 and 1) </li></ul><ul><li>Predictive DA: for a new observation, calculate the discriminant score, then classify it according to the score </li></ul><ul><li>The objective is to maximize the between group to within group sum of squares ratio that results in the best discrimination between the groups (within group variance is solely due to randomness; between group variability is due to the difference of the means) </li></ul><ul><li>Normal distribution for the response variables (dependent variables) is assumed (but normality only becomes important if significance tests are to be taken for small samples) </li></ul>
    16. 16. Statistical Credit Scoring Credit Score #Customers Good Credit Bad Credit Cut-off Score
    17. 17. Statistical Credit Scoring <ul><li>Credit scoring systems: </li></ul><ul><li>Altman Z-score model: </li></ul><ul><li>Z = .012 X 1 +.014 X 2 +.033 X 3 +.006 X 4 +1.0 X 5 </li></ul><ul><ul><li>X 1 = working capital/total assets ratio </li></ul></ul><ul><ul><li>X 2 = retained earnings/total assets ratio </li></ul></ul><ul><ul><li>X 3 = earnings before interest and taxes/total assets ratio </li></ul></ul><ul><ul><li>X 4 = market value of equity/book value of total liabilities ratio </li></ul></ul><ul><ul><li>X 5 = sales/total assets ratio </li></ul></ul>
    18. 18. Statistical Methods: Linear Regression <ul><li>The regression model is like: </li></ul><ul><li>For the true model, u can take only two values as Y; thus u can’t be normally distributed. </li></ul><ul><li>u has heteroskedastic variances, which makes the OLS inefficient </li></ul><ul><li>The estimated probability may well lie outside [0,1]. </li></ul>
    19. 19. Statistical Methods: Nearest-Neighbor Approach <ul><li>A historical database has been divided into two groups (good and bad) </li></ul><ul><li>When a consumer comes, calculate the distance between the consumer and everyone in the database </li></ul><ul><li>The consumer will be classified in the category which is the same as the nearest one(s) </li></ul><ul><li>Problems: </li></ul><ul><ul><li>The definition of distance and the number of the nearest ones </li></ul></ul><ul><ul><li>Scoring speed: when a new x comes, we need calculate the distance between the new x and all of the historical data; too much calculation! </li></ul></ul>
    20. 20. Scoring Models <ul><li>Non-statistical Methods </li></ul><ul><ul><li>Mathematical programming </li></ul></ul><ul><ul><li>Recursive partitioning </li></ul></ul><ul><ul><li>Expert systems </li></ul></ul><ul><ul><li>Machine Learning </li></ul></ul><ul><ul><ul><li>Neural Networks </li></ul></ul></ul><ul><ul><ul><li>Support Vector Machine (SVM) </li></ul></ul></ul>
    21. 21. Which Method is Best? <ul><li>In general there is no overall best method. What is best will depend on the details of the problem: </li></ul><ul><ul><li>The data structure </li></ul></ul><ul><ul><li>The characteristics used </li></ul></ul><ul><ul><li>The extent to which it is possible to separate the classes by using those characteristics </li></ul></ul><ul><ul><li>The objective of the classification (overall misclassification rate, cost-weighted misclassification rate, bad risk rate among those accepted, some measure of profitability, etc.) </li></ul></ul><ul><li>In the following slides, we will introduce three models, Logistic, Neural Networks, and SVM in detail, which are used widely today </li></ul>
    22. 22. Logistic Regression <ul><li>Empirical studies show, logistic regression may perform better than linear models (Hence, better than Discriminant Analysis), when data is nonnormal (particularly for binary data), or when covariance matrices of the two groups are not identical. </li></ul><ul><li>Therefore, logistic regression is the preferred method among the statistical methods </li></ul><ul><li>Probit regression is similar to logistic regression </li></ul>
    23. 23. Performing Logistic Regression <ul><li>Logistic Regression can be performed using the Maximum Likelihood method </li></ul><ul><li>In the maximum likelihood method, we are seeking parameter values that maximize the likelihood of the observations occurring </li></ul>
    24. 24. Logistic Regression: Setup <ul><li>Directly models the default probability as a function of the input variables X (a vector) </li></ul><ul><li>Define </li></ul><ul><li>Assume </li></ul>
    25. 25. Logistic Regression: Setup <ul><li>Assume the observations are independent, the probability (likelihood) of the observed sample is given by </li></ul>
    26. 26. Logistic Regression and ML <ul><li>ML estimator (of the coefficients a’s) for Logistic Regression can be found by applying non-linear optimization on the above likelihood function. </li></ul><ul><li>The simplified version is given by </li></ul>
    27. 27. Logistic Regression and ML <ul><li>It is easy to show that the log of the odds (= logit) are a linear function: </li></ul><ul><li>Therefore, the odds per se are a multiplicative function. </li></ul><ul><li>Since probability takes on values between (0,1), the odds take on values between (0,∞), logits take on values between (-∞,∞). So, it looks very much like linear regression, and it does not need to restrict the dependent variable to values of {0, 1}. </li></ul><ul><li>It is not solvable using OLS. </li></ul>
    28. 28. Logistic Function and Distribution
    29. 29. Normal Distribution The tails are much thinner than Logistic
    30. 30. RiskCalc: Moody’s Default Model <ul><li>Probit Regression </li></ul><ul><ul><li>Where x is the vector of the ratios </li></ul></ul>
    31. 31. Neural Networks <ul><li>Non-parametric method </li></ul><ul><li>Non-linear model estimation technique: e.g. </li></ul><ul><ul><li>Saturation effect: i.e. marginal effect of a financial ratio may decline quickly </li></ul></ul><ul><ul><li>Multiplicative factors: highly leveraged firms have a harder time borrowing money </li></ul></ul><ul><li>Neural networks decide how to combine and transform the raw characteristics in the data, as well as yielding estimates of the parameters of the decision surface </li></ul><ul><li>Well suited to situations where we have a poor understanding of the data structure </li></ul>
    32. 32. Neural Networks <ul><li>Use the logistic function as the activation function in all the nodes </li></ul><ul><li>Works well with classification problems </li></ul><ul><li>Drawbacks </li></ul><ul><ul><li>May take much longer to train </li></ul></ul><ul><ul><li>In credit scoring, there is solid understanding of data </li></ul></ul>
    33. 33. Multilayer Perceptron (MLP) <ul><li>The input values X are sent along with 1 to the hidden layer neuron </li></ul><ul><li>The hidden layer generates a weight and generates a nonlinear output that is sent to the next layer </li></ul><ul><li>The output neuron takes 1 with input from the hidden layer and generates the output signal </li></ul><ul><li>When learning occurs, the weights are adjusted so that the final OUTs produce the least error (The output of a single neuron is called OUT) </li></ul>X1 X2 1 H1 H2 1 O Input Layer Hidden Layer Output Layer w01 w12 w21 w22 w11 w02 w1 w2 w0
    34. 34. Multilayer Perceptron (MLP) <ul><li>Input nodes do not perform processing </li></ul><ul><li>Each hidden and output node processes the signals by an activation function. The most frequently used is given on the right. </li></ul><ul><li>The parameters, w, are obtained by “training” the Neural Net to historical data. </li></ul>
    35. 35. Support Vector Machine (SVM) <ul><li>A relatively new promising supervised learning method for </li></ul><ul><ul><li>Pattern recognition (Classification) </li></ul></ul><ul><ul><li>Regression estimation </li></ul></ul><ul><li>This originates from the statistical learning theory developed by Vaqnik and Chervonenkis </li></ul><ul><ul><li>1960s, Vapnik V. N., Support Vector </li></ul></ul><ul><ul><li>1995, Statistical Learning Theory </li></ul></ul><ul><ul><ul><li>Vapnik, V. N., “The Nature of Statistical Learning Theory”. New York: Springer-Verlag, 1995 2 </li></ul></ul></ul><ul><ul><ul><li>Cortes C. and Vapnik, V. N., “Support Vector Networks”, Machine Learning, 20:1-25,1995 </li></ul></ul></ul><ul><ul><li>Development, from 1995 to now </li></ul></ul>
    36. 36. SVM Extension <ul><li>Proximal Support Vector Machine (PSVM) </li></ul><ul><ul><li>Glenn Fung and Olvi L. Mangasariany 2001 </li></ul></ul><ul><li>Incremental and Decremental Support Vector Machine Learning </li></ul><ul><li>Least Squares Support Vector Machine (LS-SVM) </li></ul><ul><li>Also, SVMs can be seen as a new training method for learning machines (such as NNs) </li></ul>
    37. 37. Linear Classifier <ul><li>There are infinitely many lines that have zero training error. </li></ul><ul><li>Which line should we choose? </li></ul>
    38. 38. <ul><li>Choose the line with the largest margin . </li></ul><ul><ul><li>The optimal separating hyperplane (OSH) </li></ul></ul><ul><li>The “large margin classifier” </li></ul>Linear Classifier margin ” Support Vectors”
    39. 39. Performance of SVM <ul><li>S&P CreditModel White Paper </li></ul><ul><li>Fan and Palaniswami (2000): </li></ul><ul><ul><li>SVM 70.35%–70.90% </li></ul></ul><ul><ul><li>NN 66.11%–68.33% </li></ul></ul><ul><ul><li>MDA 59.79%–63.68% </li></ul></ul>
    40. 40. Credit Scoring and Beyond <ul><li>Data collected at application will become outdated pretty fast </li></ul><ul><li>The way a customer uses its credit account is an indicator for future performance (Behavior Scoring) </li></ul><ul><li>This leads to an update path of PD and credit control tools </li></ul><ul><li>The future is moving into profitability scoring. </li></ul><ul><ul><li>Banks should not only care about getting its money back </li></ul></ul><ul><ul><li>Banks want to extend credit to those it can make a positive NPV, risk-adjusted </li></ul></ul>
    41. 41. Best Practice in Consumer Credit Risk Management <ul><li>Credit decision-making </li></ul><ul><ul><li>Adopt to changes in economy or within customer segment </li></ul></ul><ul><li>Credit scoring </li></ul><ul><ul><li>Adaptive algorithms using credit bureau data and firm’s own experience </li></ul></ul><ul><li>Loss forecasting </li></ul><ul><ul><li>Historical delinquency rates and charge-off trend analysis </li></ul></ul><ul><ul><li>Delinquency flow and segmented vintage analysis </li></ul></ul><ul><li>Portfolio management </li></ul><ul><ul><li>Risk adjusted return on capital (RAROC) </li></ul></ul>
    42. 42. Analytical Techniques <ul><li>Response analysis: avoid adverse selection consequences that result in increased concentrations of high-risk borrowers </li></ul><ul><li>Pricing strategies: avoid “follow the competition”, focus on segment profitability and cash flow </li></ul><ul><li>Loan amount determination: avoid to be judgmental, quantify probabilities of losses </li></ul><ul><li>Credit loss forecasting: decompositional roll rate modeling, trend and seasonal indexing, and vintage curve </li></ul><ul><li>Portfolio management strategies: important for repricing and retention, don’t be judgmental, integrating behavioral element and cash flow profitability analysis ( underwriting ) </li></ul><ul><li>Collection strategies: behavioral models are useful </li></ul>
    43. 43. Credit Scoring and Loss Forecasting <ul><li>Two critical components of consumer credit risk analysis </li></ul><ul><ul><li>Corresponds to default probabilities and loss given default </li></ul></ul><ul><li>These two are linked </li></ul><ul><ul><li>Loss given default is higher when default probability is greater </li></ul></ul><ul><li>Market and economic variables matter </li></ul><ul><ul><li>In bad economic states, there will be more default and lower recovery </li></ul></ul><ul><li>Good modeling should achieve stability </li></ul>
    44. 44. Do Consumers Choose the Right Credit Contracts? <ul><li>Agarwal, Chomsisengphet, Liu, and Souleles (2006): </li></ul><ul><li>Some don’t, especially when the stake is small </li></ul><ul><li>But consumers with high balance do! </li></ul><ul><li>Other issues: </li></ul><ul><li>Personal bankruptcy in the U.S. soared! </li></ul><ul><li>Avoid/fight predatory lending! (e.g., subprime lending) </li></ul><ul><li>China is starting to have a consumer credit market </li></ul>
    45. 45. China’s Consumer Spending 64% 9198 8407 7811 7037 6462 6001 5603 TOTAL 80% 441 400 367 330 296 268 244 Services 120% 931 842 752 663 599 507 424 Housing 113% 1170 1057 945 837 739 643 550 Education&Entertainment 112% 614 554 498 437 385 337 290 Transport&Communication 91% 790 727 657 595 569 485 414 Household Durables 22% 958 885 866 791 728 750 785 Clothing 138% 506 455 401 356 300 255 213 Medicine&Healthcare 41% 3789 3487 3326 3029 2845 2756 2684 Food 97-03 2003 2002 2001 2000 1999 1998 1997 %Chg
    46. 46. China’s Consumer Credit Market <ul><li>1999-2004: Growth rate 52% </li></ul><ul><ul><li>Automobile loans: 110% </li></ul></ul><ul><ul><ul><li>Only 15% of auto sales, compared to 80% in U.S. </li></ul></ul></ul><ul><ul><li>Bankcard: 36% </li></ul></ul><ul><ul><ul><li>Mostly debit cards </li></ul></ul></ul><ul><ul><li>Mortgage: 1000% </li></ul></ul><ul><ul><ul><li>Still a long way to go! Only 8% of GDP, compared to 45% in developed economies </li></ul></ul></ul><ul><li>Other markets </li></ul><ul><ul><li>Student loan </li></ul></ul><ul><ul><li>Credit cards! </li></ul></ul><ul><li>More opportunities are waiting! </li></ul>
    47. 47. 6
    48. 48. Summary <ul><li>Introduction to Consumer Credit Risk: </li></ul><ul><ul><li>Credit scoring methods </li></ul></ul><ul><ul><li>Practical issues </li></ul></ul><ul><li>Exam: Saturday, August 4, 2PM </li></ul>
    49. 49. Review for Exam <ul><li>Topics: </li></ul><ul><li>Credit risk modeling: structural/reduced-form/incomplete information </li></ul><ul><li>Recovery rate & default correlation </li></ul><ul><li>Credit derivatives </li></ul><ul><li>Credit VaR/Basel II/consumer credit risk </li></ul><ul><li>Question Types (tentative!): </li></ul><ul><li>True or False (20%) </li></ul><ul><li>Multiple Choice (20%) </li></ul><ul><li>Short Answers (20%) </li></ul><ul><li>Problems (40%) </li></ul><ul><li>60% conceptual; 40% analytical </li></ul><ul><li>Formulas will be provided if needed. </li></ul>
    50. 50. SVM Approach Details
    51. 51. <ul><li>The plane separating and is defined by </li></ul><ul><li>The dashed planes are given by </li></ul>Computing the Margin margin w
    52. 52. <ul><li>Divide by b </li></ul><ul><li>Define new w = w/ b and α = a/b </li></ul>Computing the Margin margin w We have defined a scale for w and a
    53. 53. <ul><li>We have </li></ul><ul><li>which gives </li></ul>Computing the Margin margin  w) x x +  w)
    54. 54. Quadratic Programming Problem <ul><li>Maximizing the margin is equivalent to minimizing || w || 2 . </li></ul><ul><li>Minimize || w || 2 subject to the constraints: </li></ul>Where we have defined y(n) = +1 for all y(n) = –1 for all This enables us to write the constraints as
    55. 55. Quadratic Programming Problem Minimize the cost function (Lagrangian) Here we have introduced non-negative Lagrange multipliers l n  0 that express the constraints
    56. 56. Quadratic Programming Problem <ul><li>The first order conditions evaluated at the optimal solution are </li></ul><ul><li>The solution can be derived (together with the constraint) </li></ul>
    57. 57. Quadratic Programming Problem <ul><li>The original minimizing problem is equivalent to the following maximizing problem (dual) </li></ul><ul><li>For non-support vectors, λ will be zero, as the original constraint is not binding; only a few λ ’s would be nonzero. </li></ul>
    58. 58. Quadratic Programming Problem <ul><li>Having solved for the optimal λ ’s (denoted as ), we can derive others </li></ul><ul><li>To classify a new data point x, simply solve </li></ul>