Combining Linear and Non Linear Modeling Techniques

Like this? Share it with your network

Share

Combining Linear and Non Linear Modeling Techniques

  • 1,244 views
Uploaded on

Salford Systems demonstrates how to combine linear and non-linear predictive modeling. ...

Salford Systems demonstrates how to combine linear and non-linear predictive modeling.

More in: Technology
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
    Be the first to like this
No Downloads

Views

Total Views
1,244
On Slideshare
1,164
From Embeds
80
Number of Embeds
1

Actions

Shares
Downloads
25
Comments
0
Likes
0

Embeds 80

http://www.salford-systems.com 80

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
    No notes for slide

Transcript

  • 1. EMB America + Salford SystemsGetting the best of Two Worlds
  • 2.  Who is EMB? Insurance industry predictive modeling applications EMBLEM- our GLM tool How we have used CART with EMBLEM Case studies Other areas of expected synergies
  • 3.  Global network of p&c insurance consultants servicing clients throughout the world (insert globe)
  • 4.  Predictive Modeling Ratemaking & Profitability Analysis Underwriting & Credit Scoring Enterprise Risk Management, Pro Forma, Business Planning Retention & Conversion Modeling New Program Development Competitive Analysis Reinsurance Program Analysis Reserve Analysis & Opinion Letters Software Development & Software Support Expert Witness Testimony Regulatory Support & Law Analysis
  • 5.  EMB’s suite of software products cover all aspects of personal and commercial lines of insurance ◦ EMBLEM ◦ Rate Assessor ◦ Classifier ◦ Igloo Professional ◦ ExtrEMB ◦ ResQ Professional ◦ PrisEMB ◦ RePro
  • 6.  We use EMBLEM, a GLM tool, for our predictive modeling needs Why?
  • 7.  Primary application: ◦ Estimating the cost of the product they sell (insurance) two steps:  Reserving= estimating the cost of outstanding insurance claims  Pricing= estimating the cost of future insurance coverage Secondary applications ◦ Retention Modeling= probability that a policyholder will renew ◦ Conversion Modeling= probability that a prospective policyholder will purchase a policy ◦ Price Optimization ◦ Claim fraud detection ◦ Marketing
  • 8.  Goal is to develop a unique rate for every risk ◦ Don’t think in terms of good/bad risks ◦ State Farm/Allstate vs GEICO/Progressive ◦ Quickly exhausts the data  Credibility/ variability/ stability Risks are described by the predictor variables, not the target. ◦ Need to have a mapping of the predictor variable levels to a target value- not the other way around  Other way around makes it difficult to derive impact of individual predictor variables  Important because actual data often does not describe all possible combinations of potential customers
  • 9.  Highly regulated marketplace ◦ Restrictions  Predictors can and cannot use  Credit scores  Rules on values for the predictors  Ages 65+ relativities cannot be >110% of ages 40-60  Maximum rate change between adjacent territories  Rules on predictor order and magnitude of importance  CA Sequential Analysis (driving record>annual mileage>years held license) ◦ Regulatory Approval  Rates need to be supported  Black box methodologies will not be accepted
  • 10.  Response variable is continuous/discrete function (insert graph) ◦ Gamma consistent with severity modeling, or even Inverse Gaussian (insert graph) ◦ Poisson consistent with frequency modeling No single trial/outcome ◦ Trial is measured in terms of time ◦ Actual policy length varies tremendously because of changes  Marital status  New car  moved
  • 11.  In 1996, EMB designed EMBLEM to provide access to GLM for statisticians and non-statisticians pricing personal and commercial insurance EMBLEM revolutionized the use of GLM’s, enabling analysis that was previously either impossible or too time-consuming to be worth attempting EMBLEM is now used by over 100 insurance companies globally: ◦ 18 of the top 20 personal auto writers in the UK ◦ 50 companies in the US including 8 of the top 10 personal auto writers Fastest GLM tool with the capability to model millions of observations in seconds with a host of diagnostic tools: ◦ Graphical, practical, statistical, automated. ◦ Stand-alone software package that can be integrated with a variety of external software including SAS® ◦ Microsoft® Visual Basic® for Applications provides ultimate flexibility
  • 12.  GLM characteristics work to our advantage ◦ Exponential family does an excellent job of describing the underlying components of insurance losses ◦ Output of the model is in the form of Beta parameters which can easily be converted to rate relativities ◦ EMBLEM is not automated  User has complete control over the model structure  Complete diagnostic tools to assist the modeler with decisions
  • 13.  In terms of estimating the cost of insurance: ◦ UK has embraced predictive modeling  Experienced with its techniques  Knowledgeable with the factors that tend to be predictive ◦ US is learning about predictive modeling  Saturation with big players in personal lines marketplace  Companies not using predictive modeling techniques are being adversely selected against  Now expanding dimensionality of databases  Still fairly new concept in commercial lines marketplace  Big players are using techniques but historical rating structures are hindering the rapid expansion
  • 14.  Result? ◦ UK is expanding into secondary applications  Retention modeling  Conversion modeling  Price optimization  Claim fraud detection ◦ Because Predictive Modeling has been around for some time in the UK, the datasets are getting larger in terms of the number of predictors to evaluate ◦ Experienced US companies are beginning to evaluate the secondary applications ◦ Marketing is used in a manner similar to other industries
  • 15.  How does CART fit into this? ◦ As we transition into the secondary applications we move from modeling a continuous function to a binary function  Tree-based techniques can add value to the analysis Retention and Conversion modeling ◦ Accept/ Reject target variable ◦ Desirable smooth surface ◦ Price optimization integrates these with premium models Marketing and Fraud detection ◦ Classic tree applications
  • 16.  Using CART and EMBLEM ◦ Goal is to play off of the strengths of each tool CART strengths ◦ Automatic separation of relevant from irrelevant predictors ◦ Easily rank-orders variable importance ◦ Automatic interaction detection (requires additional work) ◦ Captures multiple structures within a dataset rather than a single dominant structure ◦ Can handle missing values and is impervious to outliers
  • 17.  EMBLEM Strengths ◦ User has control over the model structure ◦ Ease of communication/conceptualization- effects of each explanatory variable is transparent ◦ Provides predicted response values for new data points
  • 18.  CART ◦ Factor selection ◦ Interaction detection ◦ Model validation EMBLEM ◦ Model structure ◦ Incorporating time/seasonality trend effects ◦ Implementation of results
  • 19.  Both CART and EMBLEM are excellent tools both of which produce consistent results in similar situations ◦ This is not an exercise of seeing which is better The purpose of this discussion is to show how efficiencies can be gained in the modeling process ◦ As datasets get larger in terms of the number of predictors time becomes a crucial element
  • 20.  Retention modeling assignment ◦ 97,227 observations  Each observation represents one trial/outcome  Split 50/50 between training/test datasets ◦ 11 predictors  Grand total number of levels:147
  • 21.  Modeling Process ◦ Started with Forward Entry Regression  Automated process  Used Chi-Squared statistic for testing significance  Took about 30 minutes to run ◦ Significant factors (8)  Rating Area  Vehicle Category  Age  NCD  Driver Restriction  Vehicle Age  Change Over Last Year’s Premium  Market Competitiveness
  • 22.  Build a model with no factors and add based on prespecified criteria regarding improvement in model fit: (insert table) Add the factor that performed the best on the Chi Square test. (Policyholder Age) Iterate process with the new base model until no further factors indicated removal
  • 23.  Compared results with CART/ TreeNet ◦ Significant factors were essentially the same ◦ Model predictiveness was the same (ROC=0.7) Interactions ◦ No significant interactions were found by EMBLEM or CART Test Dataset ◦ ROC=0.7
  • 24.  Retention modeling assignment ◦ 198,386 observations  Each observation represented one trial/outcome  Split 50/50 between training/test datasets ◦ 135 predictors  Grand total number of levels: approx 3,752
  • 25.  Forward Entry Regression ◦ Found 57 predictors to be significant ◦ Took a weekend to run Comparison to CART/ TreeNet ◦ Found 24 significant predictors ◦ Top 15 based on variable importance were also found by EMBLEM ◦ Correlations with the rest of the predictors Through the modeling process we reduced the number of predictors to 26
  • 26.  Interactions ◦ We relied on indications from CART/ TreeNet ◦ 6 interactions were identified and included in the model EMBLEM Results ◦ Training ROC= .862 ◦ Test ROC= .85
  • 27.  Variable importance Segmentation Super-Profiling
  • 28.  CART excels at identifying different segments in data CART may also help determine where to segment data Segmentation is a useful alternative to fitting many interactions ◦ Example: In a automobile insurance renewal problem, a CART analysis showed several occurrences of a split between those policyholders with just one years duration and those with a greater duration This suggests segmenting the data into two parts: ◦ Policies renewing with one year duration ◦ Policies renewing with more than one year
  • 29.  After a GLM model is constructed use CART to model the residuals to see if any patterns exists ◦ If a pattern is discovered, go back to the model structure and incorporate the findings ◦ Test to see if model structure was inadvertently over-simplified