1. Financial institutions need to construct proxy CDS rates for counterparties lacking liquid CDS quotes, which are required for CVA pricing, CVA risk charge calculation, etc;
2. Existing CDS Proxy Methods do not meet regulatory requirements and are vulnerable to arbitrage;
3. After investigating 8 most popular Machine Learning algorithms, we show that Machine Learning techniques can be used to construct reliable CDS proxies that meet regulatory regulations while free from the above problem
4. Feature variable selection can be critical for performance of CDS-proxy construction methods
5. Effects of feature variable correlations on classification performances have to be investigated in the case of financial data
Credit Default Swap (CDS) Rate Construction by Machine Learning Techniques
1. Electronic copy available at: https://ssrn.com/abstract=3142258
CDS Rate Construction Methods
by Machine Learning Techniques
Zhongmin Luo
7-Mar-2018 * **
Independent Consultant and Researcher
Department of Economics, Mathematics and Statistics,
Birkbeck, University of London
** A Presentation at Riskβs Quant Summit Europe Conference on 7 Mar 2018 in London, UK. Disclaimer: thanks for feedbacks from participants and the views and
opinions expressed in the presentation are those of the authors and do not necessarily reflect those of above affiliated institutions.
* 2018 Call for Paper Winner: Based on a joint work with Raymond Brummelhuis, University of Reims on the paper titled
CDS Rate Construction Methods by Machine Learning Techniques. Available at SSRN: https://ssrn.com/abstract=2967184
2. Electronic copy available at: https://ssrn.com/abstract=3142258
Agenda
1. Motivations
2. Machine Learning Techniques based CDS Proxy Method
1) The Method
2) Top 3 classifiers: Neural Network, Support Vector Machine,
Ensemble/Bagged Trees
3) Empirical Results (Summary, Parameter Tuning, Regularization and
Correlation impacts)
3. Conclusions
4. Reference and Q&A
2
3. Electronic copy available at: https://ssrn.com/abstract=3142258
Question?
3
Besides they all belong to European Region/Banking Sector, what else do they have
in common on 15Sep08?
Lehman EU,
Fortis, AIB,
Northern Rock
Commerzbank, Credit Suisse,
Macqaurie UK, Wachovia EU,
Standard Chartered,
UniCredit
A
4. Motivations
β’ After the 2007-09 Financial Crisis, financial institutions have to answer two questions:
1. How much Value Adjustments (CVA/XVA) are needed for Derivative Bookβs MtM to
reflect the counterparty default risk?
2. How much capital do banks need to hold against the volatility of CVA?
4
β’ One core input is risk-neutral Counterparty Default Probability π·π« π, π to calculate
CVA or CVA Risk Capital[Basel 4 (2017) [Pykhtin and Zhu (2007) ][Brigo et al (2013)].
πΆππ΄ π‘ = 1 β π
π‘
π
πΈ π
π΅0
π΅π‘
πΈ π‘ π π·π«(π, π)
β’ Financial regulators and accounting standard bodies require us to derive the risk-neutral
π·π«(π, π) from counterpartyβs liquid CDS quotes if available; otherwise, a so-called CDS
Proxy Method has to be applied.
5. 1.1 A Shortage of Liquidity Problem
β’ Shortage of Liquidity problem: in reality, the vast majority of FIsβ counterparties donβt have
liquid CDS quotes; thus, a CDS Proxy Method has to be used to construct Proxy CDS Rates.
5
87%
70%
84%
90%
96%
82%
99%
94%
89%
83%
63%
91%
95%
0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
%ofCounterpartiesinRegions/Sectors
A Typical European Bank Counterparty Distribution by Regions/Sectors
(Overall: 84.4%; EBA Survey: >75%)
Observables Nonobservables
6. 1.2. Regulatory Criteria and Two Existing CDS Proxy Methods
Two Existing CDS Proxy Methods (both violate #3)
1. Credit Curve Mapping: proxies CDS rates by the mean/median of CDS rates within a
Region/Sector/Credit Quality (Rating) bucket.
2. Cross-sectional Regression: explains a term-specific CDS rate for counterparty π (denoted
by ππ) by its response (π½) to the event whether the counterparty belongs to a region (π ),
sector (π), rating (π) or seniority (s), indicated by respective indicator function πΌ, estimated
by running a Cross-sectional regression for each CDS term as shown below:
6
Regulatory Criteria [Basel 2015, 2017 and EBA 2015]
1. The CDS Proxy Method has to be based on an algorithm that discriminates at least 3 types
of variables: Region, Sector and Credit Quality (e.g., rating or PDs).
2. Both the observable and the nonobservable counterparties come from the same peer
group defined by the above 3 variables.
3. The appropriateness of a Proxy CDS Spread should be determined by its CDS spread
volatility across the constituents within the bucket and not by its level; i.e., any CDS Proxy
Method should reflect the idiosyncratic component of a counterpartyβs default risk.
7. 1.3. Research Gaps and Research Objectives
Two Types of Research Gaps
1. As CDS Curve Proxy Method
β’ Credit Curve Mapping: failing #3
β’ Cross-sectional Regression: failing #3. it can introduce Arbitrage for CDS
curves. [Brummelhuis and Luo (2018)]
β’ Bond spreads include significant liquidity premiums (Longstaff et al,2005), thus, are
not good choice for CDS Proxy.
β’ Rating-implied PDs: real-world PDs implied from (Credit/IRB) ratings for CVA.
2. As a Classifier Performance Comparison study based on financial market
data:
β’ Existing Classifier Performance Comparison studies[Delgado et al (2014), King et al (1995)]
are based on non-financial market data;
β’ our study is cross-classifier performance comparison for financial market
data.
Research Objectives: fill in the gaps identified above based on Machine Learning
or ML-based Techniques.
7
8. 1.4. Criteria for a Sound CDS Proxy Method
8
β’ Meet regulatory requirements β Region/Sector/Credit Quality while
accounting for idiosyncratic part of counterparty default risk based on
liquid CDS quotes; i.e., avoid β¦
β’ No model-induced CDS Curve Arbitrage [Brummelhuis and Luo 2018].
β’ Training and Cross-validation based on established statistical principles.
Lehman EU,
Fortis, AIB,
Northern Rock
Commerzbank, Credit
Suisse, Standard
Chartered, Macqaurie UK,
Wachovia EU, UniCredit
A
9. 2.1 First attempt
9
Extended Cross-sectional Regression by including ππ· 0, π‘ .
β’ LEFT: But the regression line doesnβt fit the data well due to the nonlinearity/outliers. (1.83% estimated as 4.98%).
β’ RIGHT: we cannot use real-world PDs directly for pricing; Risk-neutral PDs > Real-world ones (explained next)!
10. 2.2 What is the Real-world PD?
10
But we have other information about illiquid counterparties, e.g.,
1. Public firms: we have Equity prices and firmsβ Balance Sheet information.
2. Firms with Equity option prices, we have implied vols that are
explanatory for ππ·(0, π‘) [Berndt et al, 2005].
From #1, based on 1st-passage-time Structural Model [Black and Cox 1976], we can get
Real-world ππ·(0, π‘) after an empirical transformation from risk-neutral
Distance to Default π·π· , which is a common practice followed by
Bloombergβ’ [BBG, 2014]and MoodysKMVβ’ [MoodysKMV].
Equity +
Balance
Sheet Data
Risk-
neutral
DD
Real-
world
PDs
Structural
Model
Empirical
Transformation
11. 2.3 The ML-Technique based CDS Proxy Method
11
Given a Training Set π· π
with π¦π for Class Label and ππ for Feature Vector,
as shown below:
ML-Techniques construct a mapping called πΉ π(π₯) below; π is learned
from π· π
based on a algorithm called a Classifier Family.
A Classifier Family with a parameterization choice is called a Classifier.
In this paper, we studied 8 Classifier Families and presented 156
Classifiers. ML also has a large number of Regression Techniques; one
application in finance is: [Brummelhuis and Luo, 2018].
12. 2.4 List of Eight Classifier Families and 156 Classifiers
12
1. Neural Network (NN): e.g., Activation Functions, # of hidden units
2. Support Vector Machine (SVM): e.g., kernel functions
3. Ensemble Bagged Tree (BT): e.g., # of learning cycles.
4. Discriminant Analysis (DA): e.g., Linear/Quadratic; regularization.
5. NaΓ―ve Bayes (NB): e.g., Kernel choices; bandwidth selections.
6. π Nearest Neighbours (πNN): e.g., Distance metrics; π in πNN.
7. Logistic Regression (LR).
8. Decision Tree (DT): e.g., Impurity measure choices; Tree sizes.
18. 2.10 SVM for Linearly Separable Data
18
Maximizing the margin
19. 2.11 SVM for Nonlinearly Separable Data
19
β’ For non-linearly separable data, transformed it into a linearly separable one
first; by limiting ourselves to π½ = π πΌπ π₯π, the previous optimization problem
becomes one on the Left.
β’ Then, one can replace the πΏ π
πΏ with a kernel function denoted by π(ππ, ππ),
which is also called βkernel trickβ as indicated on the Right.
β’ Linear kernel; Polynomial kernel; Gaussian kernel
21. 2.13 Bagged Tree / Ensemble
21
β’ Ensemble is based on a committee of learning algorithms; e.g., Bagged Trees
is based on Bootstrapping;
β’ The learning outcome is determined by Majority Vote Rule from a sequence
of Decision Tree classification results.
22. 2.14 Model Assessments: K-fold Cross Validation
22
β’ First, we split observable data π· πinto πΎ folds typically of equal sizes
π« πΆ =
π=π
π²
π« π(π²)
β’ Second, for π = 1,2, β¦ , πΎ, define holdout sample π· π» π = π· π(πΎ) and define
the π-th Training Set by
π« π» π = π« πΆ β π« π― π
β’ Third, for π = 1,2, β¦ , πΎ, we apply the Classifier trained from Training Set
π· π π to estimate π¦π for data (π₯, π¦) of π· π» π and calculate the expected
Misclassification Rate as:
π π
π― =
π
#π« π― π
π,π βπ« π― π
π°(π, π π )
23. 3.1 Summary of Cross-classifier Performances
23
β’ Other 5 classifier families: Discriminant Analysis (DA), NaΓ―ve Bayes (NB), π Nearest
Neighbours (πNN), Logistic Regression (LR) and Decision Tree (DT).
β’ The ranking of top performing Classifier Families is in line with those reported in
classifier performance comparison literature based on non-financial (Delgado et al 2014, King
et al 1995).
24. 3.2. Conclusions
1. ML-Technique based CDS Proxy Method satisfies regulatory requirements,
account for counterparty-specific default risk, appropriate for CVA pricing,
Counterparty Credit risk management (Success Criteria #1).
2. No Arbitrage introduced by the model (Success Criteria #2) [Brummelhuis and Luo 2018].
3. Model assessment is based on Statistical/Machine Learning theories,
produces satisfactory results based on Cross-validation procedure (Success
Criteria #3).
β’ Based on studies of 156 classifiers across 8 algorithms, Neural Network [99.3% (0.6%)], SVM [96.8% (1.6%)]
and Ensemble/Bagged Tree [96.0% (2.2%)]are top 3 performers; the ranking is in line with Classifier
Comparison literature.
4. To the best of our knowledge, the study is:
β’ The 1st Machine Learning Technique based CDS Proxy Method.
β’ The 1st Classifier Performance Comparison research based on Financial Market Data.
β’ The 1st Research effort to look at Correlation impacts on cross-classifier performance.
24
25. References
25
1. Berndt, A., Duffie, D., Douglas, R., Ferguson M., Schranz, D., 2005, Measure Default Risk Premia from Default
Swap Rates and EDFs.
2. BCBS, July 2015, Review of the Credit Valuation Adjustment Risk Framework, Consultative Document, Bank
for International Settlements.
3. BCBS, Basel III, Finalizing post-crisis reforms, December 2017.
4. Bloomberg, Bloomberg Credit Risk, Framework, Usage and Methodology, 2014
5. Brigo, D., Morini M. and Pallavicini A., 2013, Counterparty Credit Risk, Collateral and Funding: With Pricing
Cases for All Asset Classes, John Wiley and Sons Ltd.
6. Brummelhuis, Raymond and Luo, Zhongmin, CDS Rate Construction Methods by Machine Learning
Techniques (May 12, 2017). SSRN: https://ssrn.com/abstract=2967184
7. Brummelhuis, Raymond and Luo, Zhongmin, A Note on No-arbitrage Restrictions on CDS Curve, 2018.
8. Brummelhuis, Raymond and Luo, Zhongmin, Bank Capital, Net Interest Margin and Stress Testing by
Machine Learning Techniques, 2018
9. EBA Report, 22 February 2015, On Credit Valuation Adjustment (CVA) under Article 456(2) of Regulation (EU)
No 575/2013 (Capital Requirements Regulation).
10. King, R., Feng, C., and Shutherland, A., Statlog: comparison of classification algorithms on large real-world
problems, Applied Artificial Intelligence, 1995, 9(3), 289-333.
11. Pykhtin M. and Zhu S., 2007, A Guide to Modelling Counterparty Credit Risk.
12. Wu, X., Kumar, V., Quinlan, J., Ghosh, J., Yang, Q., Motoda, H., McLachlan, G., Ng, A., Liu, B., Yu, P., Zhou, Z.,
Steinbach, M., Hand, D., Steinberg, D., 2008, Top 10 algorithms in data mining, Knowl Inf Syst (2008).