SlideShare a Scribd company logo
Summary
 Problem Statement :
 Other than “Known Good/Bad Sample” in models such as behavior models or loss forecasting type of models,
application scorecards are developed to predict the behavior of all applicants, and using a model based on only
preciously approved applicants can be inaccurate (“Sample Bias”)
 Previous Accept/ Decline decisions were made systematically and were not random. Accept population is a
biased sample and not representative of the rejects.
 Reject Inference:
 A process that the performance of previously rejected applications is analyzed to estimate their behavior.
Equivalent to saying “if these applicants had been accepted, this is how they would have performed”. This
process gives relevance to the scorecard development process by recreating the population performance for a
100% approval rate.
Reasons for Reject Inference
 Reasons for Reject Inference:
 Ignoring rejects would produce a scorecard that is not applicable to the total applicant population, the issue of
sample bias mentioned in above page.
 Reject inference also incorporates the influence of past decision making into the scorecard development
process, such as Low-side Override. If a scorecard is now developed using the known goods and bads, it will tell
us that those who have serious delinquency are very good credit risk.
 From a decision-making perspective, such as cut-off adjustment, reject inference enables accurate and realistic
expected performance forecasts for all applicant. Also coupled with swap set analysis, it will allow us to approve
the same number of people but obtain better performance through better selection. Or it will allow us to approve
more good customers but keeping same existing bad rate.
 In environments with both low or medium approval rates and low bad rates, reject inference helps in indentifying
opportunities to increase market share with risk-adjusted strategies.
Reject Inference Techniques –Method 1 (Requires High Performance Match Rate)
 Use performance at credit bureaus of those declined by one creditor but approved for a similar product
elsewhere. The performance with other products or companies is taken as proxy for how the declined applicants
would have performed has they been originally accepted.
 This method approximates actual performance, but has a few drawbacks.
 The applicants chosen must also obtain similar credit during similar time frame (i.e., soon after being
declined). Applicant declined at one institution or for one product are also likely to be declined elsewhere,
thus reducing the sample size.
 The “BAD” definition chosen through analysis of known goods and known bads must be used for these
accounts using different data sources. The bad definition from bureau will not be 100% consistent with
the definition from line of business.
 Different portfolios exhibits very different performance match rate.
Reject Inference Techniques- Method 2
2. Approve All Application
 Methods of data collection are:
1) Approving all applicants for a specific period, enough to generate a sample for model development data,
with enough bads.
2) Approving all applications above cutoff, but only randomly selected 1/N below cut-off.
3) Approving all applications up to 10 or 20 points below cut-off, and randomly sampling the rest, in order to
get a better sample applications in the decision making (i.e., where cut-off decision are likely to be made)
 Advantage:
1) The only method to find out the actual performance of rejected accounts. The most scientific and simple
way.
 Disadvantage:.
1) Approve the applicants that are known to be very high-risk can be in potential high cost in losses. A
further strategy for this can be granting the lower loans / credit lines.
2) In certain jurisdictions, there may be legal hurdles to this method. Approving some and declining others
with similar characteristics, or randomly approving applicants may present problems. A further strategy
for this can be avoiding these states.
Reject Inference Techniques- Method 3 (Requires High Performance Match Rate)
3. Supplemental Bureau All Data on Reject Inference
 Key Assumption:
1) Obtain bureau data on accepts and rejects at the end of the observation period. Use the performance
with other creditors over the observation period to infer how the rejects would have performed had they
been accepted.
2) P(bad | X, Z, rejected) = P(bad |X, Z). The bureau data at time of application (X) and the downstream
bureau data (Z) contain all the relevant information about P(bad).
 Method:
I. Step 1, fit a model for P(on-us-bad | Off-us performance) over the booked population.
II. Step 2, apply the model from step one to the reject population and get a predicted probability of bad for
each reject applicant. p=P(bad | Off-us performance).
III. Step 3, replicate each reject account to two with both good and bad. Assign p as the weight to the
account with outcome ’bad’ and 1-p as the weight to the account with outcome ’good’.
IV. Step 4, score the whole data (booked population weights=1 and reject population weights by step 3) to fit
a statistical model.
 Advantage:
1) Weak assumption
2) Incorporate additional information
 Weakness:
1) Requires Quality bureau match.
2) Costly.
Reject Inference Techniques- Method 4 (Low/No Performance Match Rate)
4. Reject Inference when match rate is low/none
 Reject Inference with supplemental Bureau Attributes Data
1) Obtain bureau data on accepts at the end of the observation period. Use the performance with other
creditors over the observation period to infer how the rejects would have performed had they been
accepted.
2) Method:
1) Step 1, fit a model for P(on-us bad | all data at time of application) over the booked population.
2) Step 2, apply the model from step one to the reject population and get a predicted probability of
bad for each reject applicant. p=P(bad | all data at time of application).
3) Step 3, replicate each reject account to two with both good and bad. Assign p as the weight to the
account with outcome ’bad’ and 1-p as the weight to the account with outcome ’good’.
4) Step 4, score the whole data (booked population weights=1 and reject population weights by step
3) to fit a statistical model.
Continue Method 4
This method involves using rejects with weight values that correspond to the probability of a given loan
application being approved or rejected.
1.The first step involves developing a scorecard using the information on approved loan applications:
2. Then, using the resulting scorecard, we evaluate the set of rejects:
3. For each rejected application, there are two records containing the weight values that correspond to the
probabilities:
Continue Method 4
4. The joint dataset that is extended due to the "double" number of rejects is used to adjust the parameters of
the scorecard:
The use of both the probability of rejection and the probability of approval of a reject allows adjusting the
parameters of the score to cover either of the two possible types of behavior ("good" or "bad").
For any comment or question,
please contact @
Kaitlyn.S.Hu@Gmail.Com
Thank You!

More Related Content

What's hot

Ankit Patel - Tableau Developer
Ankit Patel - Tableau DeveloperAnkit Patel - Tableau Developer
Ankit Patel - Tableau Developer
Ankit Patel
 
LLMs_talk_March23.pdf
LLMs_talk_March23.pdfLLMs_talk_March23.pdf
LLMs_talk_March23.pdf
ChaoYang81
 
Introduction to LLMs
Introduction to LLMsIntroduction to LLMs
Introduction to LLMs
Loic Merckel
 
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
Mihai Criveti
 
Data Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
Data Visualisation & Analytics with Tableau (Beginner) - by Maria KoumandrakiData Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
Data Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
Outreach Digital
 
3. Regression.pdf
3. Regression.pdf3. Regression.pdf
3. Regression.pdf
Jyoti Yadav
 
‘Big models’: the success and pitfalls of Transformer models in natural langu...
‘Big models’: the success and pitfalls of Transformer models in natural langu...‘Big models’: the success and pitfalls of Transformer models in natural langu...
‘Big models’: the success and pitfalls of Transformer models in natural langu...
Leiden University
 
Granger Causality
Granger CausalityGranger Causality
Granger Causality
Gaetan Lion
 
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
SOLTUIONSpeople, THINKubators, THINKathons
 
Introduction To Data Science Using R
Introduction To Data Science Using RIntroduction To Data Science Using R
Introduction To Data Science Using R
ANURAG SINGH
 
Knowledge Graphs and Generative AI
Knowledge Graphs and Generative AIKnowledge Graphs and Generative AI
Knowledge Graphs and Generative AI
Neo4j
 
Causal inference in practice
Causal inference in practiceCausal inference in practice
Causal inference in practice
Amit Sharma
 
Machine learning in medicine: calm down
Machine learning in medicine: calm downMachine learning in medicine: calm down
Machine learning in medicine: calm down
BenVanCalster
 
Power BI Overview
Power BI Overview Power BI Overview
Power BI Overview
Gal Vekselman
 
Machine Learning on IBM Watson Studio
Machine Learning on IBM Watson StudioMachine Learning on IBM Watson Studio
Machine Learning on IBM Watson Studio
Upkar Lidder
 
Data science benefits
Data science benefitsData science benefits
Data science benefits
Marian Siwiak, Ph.D.
 
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
Edureka!
 
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHMHEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
amiteshg
 
AI-powered tools for Creatives
AI-powered tools for CreativesAI-powered tools for Creatives
AI-powered tools for Creatives
Remy Rey De Barros
 
PhD Thesis Proposal
PhD Thesis Proposal PhD Thesis Proposal
PhD Thesis Proposal
Ziqiang Feng
 

What's hot (20)

Ankit Patel - Tableau Developer
Ankit Patel - Tableau DeveloperAnkit Patel - Tableau Developer
Ankit Patel - Tableau Developer
 
LLMs_talk_March23.pdf
LLMs_talk_March23.pdfLLMs_talk_March23.pdf
LLMs_talk_March23.pdf
 
Introduction to LLMs
Introduction to LLMsIntroduction to LLMs
Introduction to LLMs
 
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
Retrieval Augmented Generation in Practice: Scalable GenAI platforms with k8s...
 
Data Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
Data Visualisation & Analytics with Tableau (Beginner) - by Maria KoumandrakiData Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
Data Visualisation & Analytics with Tableau (Beginner) - by Maria Koumandraki
 
3. Regression.pdf
3. Regression.pdf3. Regression.pdf
3. Regression.pdf
 
‘Big models’: the success and pitfalls of Transformer models in natural langu...
‘Big models’: the success and pitfalls of Transformer models in natural langu...‘Big models’: the success and pitfalls of Transformer models in natural langu...
‘Big models’: the success and pitfalls of Transformer models in natural langu...
 
Granger Causality
Granger CausalityGranger Causality
Granger Causality
 
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
arbar https://www.slideshare.net/Solutionman/charles-caldwell-improve-your-li...
 
Introduction To Data Science Using R
Introduction To Data Science Using RIntroduction To Data Science Using R
Introduction To Data Science Using R
 
Knowledge Graphs and Generative AI
Knowledge Graphs and Generative AIKnowledge Graphs and Generative AI
Knowledge Graphs and Generative AI
 
Causal inference in practice
Causal inference in practiceCausal inference in practice
Causal inference in practice
 
Machine learning in medicine: calm down
Machine learning in medicine: calm downMachine learning in medicine: calm down
Machine learning in medicine: calm down
 
Power BI Overview
Power BI Overview Power BI Overview
Power BI Overview
 
Machine Learning on IBM Watson Studio
Machine Learning on IBM Watson StudioMachine Learning on IBM Watson Studio
Machine Learning on IBM Watson Studio
 
Data science benefits
Data science benefitsData science benefits
Data science benefits
 
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
What Is Power BI? | Introduction To Microsoft Power BI | Power BI Training | ...
 
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHMHEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
HEART DISEASE PREDICTION USING NAIVE BAYES ALGORITHM
 
AI-powered tools for Creatives
AI-powered tools for CreativesAI-powered tools for Creatives
AI-powered tools for Creatives
 
PhD Thesis Proposal
PhD Thesis Proposal PhD Thesis Proposal
PhD Thesis Proposal
 

Similar to Reject Inference Methodologies on Underwriting Model

ICAR - IFPRI- Power Calculation
ICAR - IFPRI- Power CalculationICAR - IFPRI- Power Calculation
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
55296
 
Credit Risk Evaluation Model
Credit Risk Evaluation ModelCredit Risk Evaluation Model
Credit Risk Evaluation Model
Mihai Enescu
 
Credit Card Marketing Classification Trees Fr.docx
 Credit Card Marketing Classification Trees Fr.docx Credit Card Marketing Classification Trees Fr.docx
Credit Card Marketing Classification Trees Fr.docx
ShiraPrater50
 
Audit sampling
Audit samplingAudit sampling
Audit sampling
Heickal Pradinanta
 
Predictive_Analytics_A_WC_Game_Changer
Predictive_Analytics_A_WC_Game_ChangerPredictive_Analytics_A_WC_Game_Changer
Predictive_Analytics_A_WC_Game_Changer
Jeff Viene
 
Certified Specialist Business Intelligence (.docx
Certified     Specialist     Business  Intelligence     (.docxCertified     Specialist     Business  Intelligence     (.docx
Certified Specialist Business Intelligence (.docx
durantheseldine
 
Project Report for Mostan Superstore.pptx
Project Report for Mostan Superstore.pptxProject Report for Mostan Superstore.pptx
Project Report for Mostan Superstore.pptx
ChristianahEfunniyi
 
Reduction in customer complaints - Mortgage Industry
Reduction in customer complaints - Mortgage IndustryReduction in customer complaints - Mortgage Industry
Reduction in customer complaints - Mortgage Industry
Pranov Mishra
 
Credit scoring i financial sector
Credit scoring i financial  sector Credit scoring i financial  sector
Credit scoring i financial sector
Chandrasekhar Subramanyam
 
churn_detection.pptx
churn_detection.pptxchurn_detection.pptx
churn_detection.pptx
DhanuDhanu49
 
Chapter16For all types of project and in their different sizes, .docx
Chapter16For all types of project and in their different sizes, .docxChapter16For all types of project and in their different sizes, .docx
Chapter16For all types of project and in their different sizes, .docx
christinemaritza
 
How Big Data Analysis Can Protect You From Fraud
How Big Data Analysis Can Protect You From FraudHow Big Data Analysis Can Protect You From Fraud
How Big Data Analysis Can Protect You From Fraud
Trendwise Analytics
 
Creditscore
CreditscoreCreditscore
Creditscore
kevinlan
 
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
Pratik Meshram
 
Modeling results from Health Sciences data
Modeling results from Health Sciences dataModeling results from Health Sciences data
Modeling results from Health Sciences data
Judson Chase
 
Assessment Group Managers Training-SMCC1.pptx
Assessment Group Managers Training-SMCC1.pptxAssessment Group Managers Training-SMCC1.pptx
Assessment Group Managers Training-SMCC1.pptx
HctorMurciaForero
 
How Long Is The Tail?
How Long Is The Tail?How Long Is The Tail?
How Long Is The Tail?
Marius88
 
03_AJMS_298_21.pdf
03_AJMS_298_21.pdf03_AJMS_298_21.pdf
03_AJMS_298_21.pdf
BRNSS Publication Hub
 
Data Analytics for Internal Auditors - Understanding Sampling
Data Analytics for Internal Auditors - Understanding SamplingData Analytics for Internal Auditors - Understanding Sampling
Data Analytics for Internal Auditors - Understanding Sampling
Jim Kaplan CIA CFE
 

Similar to Reject Inference Methodologies on Underwriting Model (20)

ICAR - IFPRI- Power Calculation
ICAR - IFPRI- Power CalculationICAR - IFPRI- Power Calculation
ICAR - IFPRI- Power Calculation
 
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
7. Plan, perform, and evaluate samples for substantive procedures IPPTChap009...
 
Credit Risk Evaluation Model
Credit Risk Evaluation ModelCredit Risk Evaluation Model
Credit Risk Evaluation Model
 
Credit Card Marketing Classification Trees Fr.docx
 Credit Card Marketing Classification Trees Fr.docx Credit Card Marketing Classification Trees Fr.docx
Credit Card Marketing Classification Trees Fr.docx
 
Audit sampling
Audit samplingAudit sampling
Audit sampling
 
Predictive_Analytics_A_WC_Game_Changer
Predictive_Analytics_A_WC_Game_ChangerPredictive_Analytics_A_WC_Game_Changer
Predictive_Analytics_A_WC_Game_Changer
 
Certified Specialist Business Intelligence (.docx
Certified     Specialist     Business  Intelligence     (.docxCertified     Specialist     Business  Intelligence     (.docx
Certified Specialist Business Intelligence (.docx
 
Project Report for Mostan Superstore.pptx
Project Report for Mostan Superstore.pptxProject Report for Mostan Superstore.pptx
Project Report for Mostan Superstore.pptx
 
Reduction in customer complaints - Mortgage Industry
Reduction in customer complaints - Mortgage IndustryReduction in customer complaints - Mortgage Industry
Reduction in customer complaints - Mortgage Industry
 
Credit scoring i financial sector
Credit scoring i financial  sector Credit scoring i financial  sector
Credit scoring i financial sector
 
churn_detection.pptx
churn_detection.pptxchurn_detection.pptx
churn_detection.pptx
 
Chapter16For all types of project and in their different sizes, .docx
Chapter16For all types of project and in their different sizes, .docxChapter16For all types of project and in their different sizes, .docx
Chapter16For all types of project and in their different sizes, .docx
 
How Big Data Analysis Can Protect You From Fraud
How Big Data Analysis Can Protect You From FraudHow Big Data Analysis Can Protect You From Fraud
How Big Data Analysis Can Protect You From Fraud
 
Creditscore
CreditscoreCreditscore
Creditscore
 
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
pratik meshram-Unit 4 contemporary marketing research full notes pune univers...
 
Modeling results from Health Sciences data
Modeling results from Health Sciences dataModeling results from Health Sciences data
Modeling results from Health Sciences data
 
Assessment Group Managers Training-SMCC1.pptx
Assessment Group Managers Training-SMCC1.pptxAssessment Group Managers Training-SMCC1.pptx
Assessment Group Managers Training-SMCC1.pptx
 
How Long Is The Tail?
How Long Is The Tail?How Long Is The Tail?
How Long Is The Tail?
 
03_AJMS_298_21.pdf
03_AJMS_298_21.pdf03_AJMS_298_21.pdf
03_AJMS_298_21.pdf
 
Data Analytics for Internal Auditors - Understanding Sampling
Data Analytics for Internal Auditors - Understanding SamplingData Analytics for Internal Auditors - Understanding Sampling
Data Analytics for Internal Auditors - Understanding Sampling
 

Reject Inference Methodologies on Underwriting Model

  • 1.
  • 2. Summary  Problem Statement :  Other than “Known Good/Bad Sample” in models such as behavior models or loss forecasting type of models, application scorecards are developed to predict the behavior of all applicants, and using a model based on only preciously approved applicants can be inaccurate (“Sample Bias”)  Previous Accept/ Decline decisions were made systematically and were not random. Accept population is a biased sample and not representative of the rejects.  Reject Inference:  A process that the performance of previously rejected applications is analyzed to estimate their behavior. Equivalent to saying “if these applicants had been accepted, this is how they would have performed”. This process gives relevance to the scorecard development process by recreating the population performance for a 100% approval rate.
  • 3. Reasons for Reject Inference  Reasons for Reject Inference:  Ignoring rejects would produce a scorecard that is not applicable to the total applicant population, the issue of sample bias mentioned in above page.  Reject inference also incorporates the influence of past decision making into the scorecard development process, such as Low-side Override. If a scorecard is now developed using the known goods and bads, it will tell us that those who have serious delinquency are very good credit risk.  From a decision-making perspective, such as cut-off adjustment, reject inference enables accurate and realistic expected performance forecasts for all applicant. Also coupled with swap set analysis, it will allow us to approve the same number of people but obtain better performance through better selection. Or it will allow us to approve more good customers but keeping same existing bad rate.  In environments with both low or medium approval rates and low bad rates, reject inference helps in indentifying opportunities to increase market share with risk-adjusted strategies.
  • 4. Reject Inference Techniques –Method 1 (Requires High Performance Match Rate)  Use performance at credit bureaus of those declined by one creditor but approved for a similar product elsewhere. The performance with other products or companies is taken as proxy for how the declined applicants would have performed has they been originally accepted.  This method approximates actual performance, but has a few drawbacks.  The applicants chosen must also obtain similar credit during similar time frame (i.e., soon after being declined). Applicant declined at one institution or for one product are also likely to be declined elsewhere, thus reducing the sample size.  The “BAD” definition chosen through analysis of known goods and known bads must be used for these accounts using different data sources. The bad definition from bureau will not be 100% consistent with the definition from line of business.  Different portfolios exhibits very different performance match rate.
  • 5. Reject Inference Techniques- Method 2 2. Approve All Application  Methods of data collection are: 1) Approving all applicants for a specific period, enough to generate a sample for model development data, with enough bads. 2) Approving all applications above cutoff, but only randomly selected 1/N below cut-off. 3) Approving all applications up to 10 or 20 points below cut-off, and randomly sampling the rest, in order to get a better sample applications in the decision making (i.e., where cut-off decision are likely to be made)  Advantage: 1) The only method to find out the actual performance of rejected accounts. The most scientific and simple way.  Disadvantage:. 1) Approve the applicants that are known to be very high-risk can be in potential high cost in losses. A further strategy for this can be granting the lower loans / credit lines. 2) In certain jurisdictions, there may be legal hurdles to this method. Approving some and declining others with similar characteristics, or randomly approving applicants may present problems. A further strategy for this can be avoiding these states.
  • 6. Reject Inference Techniques- Method 3 (Requires High Performance Match Rate) 3. Supplemental Bureau All Data on Reject Inference  Key Assumption: 1) Obtain bureau data on accepts and rejects at the end of the observation period. Use the performance with other creditors over the observation period to infer how the rejects would have performed had they been accepted. 2) P(bad | X, Z, rejected) = P(bad |X, Z). The bureau data at time of application (X) and the downstream bureau data (Z) contain all the relevant information about P(bad).  Method: I. Step 1, fit a model for P(on-us-bad | Off-us performance) over the booked population. II. Step 2, apply the model from step one to the reject population and get a predicted probability of bad for each reject applicant. p=P(bad | Off-us performance). III. Step 3, replicate each reject account to two with both good and bad. Assign p as the weight to the account with outcome ’bad’ and 1-p as the weight to the account with outcome ’good’. IV. Step 4, score the whole data (booked population weights=1 and reject population weights by step 3) to fit a statistical model.  Advantage: 1) Weak assumption 2) Incorporate additional information  Weakness: 1) Requires Quality bureau match. 2) Costly.
  • 7. Reject Inference Techniques- Method 4 (Low/No Performance Match Rate) 4. Reject Inference when match rate is low/none  Reject Inference with supplemental Bureau Attributes Data 1) Obtain bureau data on accepts at the end of the observation period. Use the performance with other creditors over the observation period to infer how the rejects would have performed had they been accepted. 2) Method: 1) Step 1, fit a model for P(on-us bad | all data at time of application) over the booked population. 2) Step 2, apply the model from step one to the reject population and get a predicted probability of bad for each reject applicant. p=P(bad | all data at time of application). 3) Step 3, replicate each reject account to two with both good and bad. Assign p as the weight to the account with outcome ’bad’ and 1-p as the weight to the account with outcome ’good’. 4) Step 4, score the whole data (booked population weights=1 and reject population weights by step 3) to fit a statistical model.
  • 8. Continue Method 4 This method involves using rejects with weight values that correspond to the probability of a given loan application being approved or rejected. 1.The first step involves developing a scorecard using the information on approved loan applications: 2. Then, using the resulting scorecard, we evaluate the set of rejects: 3. For each rejected application, there are two records containing the weight values that correspond to the probabilities:
  • 9. Continue Method 4 4. The joint dataset that is extended due to the "double" number of rejects is used to adjust the parameters of the scorecard: The use of both the probability of rejection and the probability of approval of a reject allows adjusting the parameters of the score to cover either of the two possible types of behavior ("good" or "bad").
  • 10. For any comment or question, please contact @ Kaitlyn.S.Hu@Gmail.Com Thank You!