Risk Classification with an Adaptive Naive Bayes Kernel Machine Model

Risk Classification with an
Adaptive Naive Bayes Kernel Machine Model
Jessica Minnier1,
Ming Yuan3, Jun Liu4, and Tianxi Cai2
1Department of Public Health & Preventive Medicine, Oregon Health & Science University
2Department of Biostatistics, Harvard School of Public Health
3Department of Statistics, University of Wisconsin-Madison
4Department of Statistics, Harvard University
June 30, 2015
ASA Oregon Chapter Meeting
Outline
1 Background and Motivation
2 Model and Methods
Kernels
Blockwise Kernel PCA Estimation
Regularized Selection of Informative Regions
Theoretical Results
3 Simulation Studies
4 Genetic Risk of Type I Diabetes
5 Conclusions
Adaptive Naive Bayes Kernel Machine Model 2
Background and Motivation
Adaptive Naive Bayes (Blockwise) Kernel Machine Classification
• Goal: genetic data → quantify disease risk, predict therapeutic
efficacy, determine disease subtypes
• Goal: build an accurate parsimonious prediction model
– reduce the cost of unnecessary marker measurements
– improve the prediction precision for future patients
– improve over modest prediction precision obtained with clinical
predictors and/or known risk alleles
Adaptive Naive Bayes Kernel Machine Model Background and Motivation 3
Background and Motivation
Adaptive Naive Bayes (Blockwise) Kernel Machine Classification
• Goal: genetic data → quantify disease risk, predict therapeutic
efficacy, determine disease subtypes
• Goal: build an accurate parsimonious prediction model
– reduce the cost of unnecessary marker measurements
– improve the prediction precision for future patients
– improve over modest prediction precision obtained with clinical
predictors and/or known risk alleles
• Complex diseases
– many alleles contribute to risk
– many distinct combinations of risk factors lead to disease
Adaptive Naive Bayes Kernel Machine Model Background and Motivation 3
Background and Motivation
• Genome wide association studies (GWAS)
– identifying SNPs associated with disease risk
– primary goal of testing
– accurate risk prediction remains difficult
• Common approach:
– select top ranked SNPs based on large scale testing
– construct a composite genetic score w/ selected SNPs
Adaptive Naive Bayes Kernel Machine Model Background and Motivation 4
Background and Motivation
• Genome wide association studies (GWAS)
– identifying SNPs associated with disease risk
– primary goal of testing
– accurate risk prediction remains difficult
• Common approach:
– select top ranked SNPs based on large scale testing
– construct a composite genetic score w/ selected SNPs
– may not work well due to
false +/− errors in identifying predictive SNPs
over-fitting
using only subset of SNPs available
additive effects only
Adaptive Naive Bayes Kernel Machine Model Background and Motivation 4
Background and Motivation
Recent progress in prediction with high dimensional data
• Regularized estimation: LASSO (Tibshirani, 1996); SCAD (Fan and Li,
2001); Adaptive LASSO (Zou, 2006)
• Machine learning: Support vector machine (Cristianini, Shawe-Taylor,
2000); Least square Kernel Machine Regression (Liu, Lin, Ghosh, 2007);
Kernel logistic regression (Zhu and Hastie, 2005; Liu, Ghosh and Lin,
2008)
• Screening + Regularized estimation: Sure independence screening
(Fan and Lv, 2008; Fan and Song, 2009)
Global methods: may be unstable for large p, high correlation
Adaptive Naive Bayes Kernel Machine Model Background and Motivation 5
Approach
Challenge:
• Prediction models based on univariate testing, additive models, global
methods → low prediction accuracy, low AUC, missing heritability
• Non-linear effects? testing for interactions → low power
Adaptive Naive Bayes Kernel Machine Model Model and Methods 6
Approach
Challenge:
• Prediction models based on univariate testing, additive models, global
methods → low prediction accuracy, low AUC, missing heritability
• Non-linear effects? testing for interactions → low power
Our approach [Minnier et al., 2015]:
• Blockwise method:
leverage biological knowledge to build models at the gene-set level
genes, gene-pathways, linkage disequilibrium blocks
• Kernel machine regression:
allow for complex and nonlinear effects
implicitly specify underlying complex functional form of covariate
effects via similarity measures (kernels) that define the distance
between two sets of covariates
Adaptive Naive Bayes Kernel Machine Model Model and Methods 6
Kernel Methods: similar inputs to similar outputs
• transform data to feature space H with non-linear map φ
• “kernel trick” lets us use K(, ) similarity function instead of φ
• K induces the feature space
N. Takahashi’s webpage
Adaptive Naive Bayes Kernel Machine Model Model and Methods 7
Previous Methods
Blockwise methods
• Inference: Gene-set testing
Gene burden tests
Gene Set Enrichment Analysis (GSEA)
SNP-set Sequence Kernel Association Test (SKAT, SKAT-O; Wu et al.
2010; Wu, Lee, et al. 2011)
Adaptive Naive Bayes Kernel Machine Model Model and Methods 8
Previous Methods
Blockwise methods
• Inference: Gene-set testing
Gene burden tests
Gene Set Enrichment Analysis (GSEA)
SNP-set Sequence Kernel Association Test (SKAT, SKAT-O; Wu et al.
2010; Wu, Lee, et al. 2011)
Kernel machine methods
• Support Vector Machine (SVM) classification methods
• Inference
KM SNP-set Testing (Liu et al. 2007, 2008; SKAT methods)
Gene expression test with kernel Cox model (Li and Luan 2003)
Adaptive Naive Bayes Kernel Machine Model Model and Methods 8
Notations and Model Assumptions
• Data
– Response: Y = (Y1, ..., Yn)T
– Predictors: M blocks of genomic regions, for b = 1, ..., M,
X(b)
= (X
(b)
1 , ..., X(b)
n )T
n×pb
,
Adaptive Naive Bayes Kernel Machine Model Model and Methods 9
Notations and Model Assumptions
• Data
– Response: Y = (Y1, ..., Yn)T
– Predictors: M blocks of genomic regions, for b = 1, ..., M,
X(b)
= (X
(b)
1 , ..., X(b)
n )T
n×pb
,
• Blockwise: Partition genome into gene-sets
– Recombination hotspots, gene-pathways
Adaptive Naive Bayes Kernel Machine Model Model and Methods 9
Notations and Model Assumptions
• Data
– Response: Y = (Y1, ..., Yn)T
– Predictors: M blocks of genomic regions, for b = 1, ..., M,
X(b)
= (X
(b)
1 , ..., X(b)
n )T
n×pb
,
• Model under blockwise Naive Bayes (NB) assumption:
X(1)
, ..., X(M)
| Y independent
Adaptive Naive Bayes Kernel Machine Model Model and Methods 10
Notations and Model Assumptions
• Data
– Response: Y = (Y1, ..., Yn)T
– Predictors: M blocks of genomic regions, for b = 1, ..., M,
X(b)
= (X
(b)
1 , ..., X(b)
n )T
n×pb
,
• Model under blockwise Naive Bayes (NB) assumption:
X(1)
, ..., X(M)
| Y independent ⇒
logit{pr(Y = 1 | X(1)
, ..., X(M)
)} = c +
M
b=1
logit{pr(Y = 1 | X(b)
)}
– NB assumption allows separate estimation by block and reduces overfitting
– Performs well for zero-one loss L(X) = I( ˆY (X) = Y ) [Domingos and
Pazzani, 1997]
Adaptive Naive Bayes Kernel Machine Model Model and Methods 10
Notations and Model Assumptions
• Within each region, the effect may be complex and interactive due to
– multiple causal variants
– un-typed causal variants in the presence of high LD
Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
Notations and Model Assumptions
• Within each region, the effect may be complex and interactive due to
– multiple causal variants
– un-typed causal variants in the presence of high LD
• Blockwise Kernel Machine Regression
logit{pr(Y = 1 | X(b)
)} = a(b)
+h(b)
(X(b)
)
h(b)
(X(b)
) =
l
β(b)
l ψ(b)
l (X(b)
) ∈ HK(b)
{ψ(b)
l } = { λ(b)
l φ(b)
l } implicitly specified via a symmetric positive
definite kernel K(b)
(·, ·).
Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
Notations and Model Assumptions
• Within each region, the effect may be complex and interactive due to
– multiple causal variants
– un-typed causal variants in the presence of high LD
• Blockwise Kernel Machine Regression
logit{pr(Y = 1 | X(b)
)} = a(b)
+h(b)
(X(b)
)
h(b)
(X(b)
) =
l
β(b)
l ψ(b)
l (X(b)
) ∈ HK(b)
{ψ(b)
l } = { λ(b)
l φ(b)
l } implicitly specified via a symmetric positive
definite kernel K(b)
(·, ·).
K(b)
(X(b)
i , X(b)
j ) defines the similarity between X(b)
i and X(b)
j .
Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
Notations and Model Assumptions
• Within each region, the effect may be complex and interactive due to
– multiple causal variants
– un-typed causal variants in the presence of high LD
• Blockwise Kernel Machine Regression
logit{pr(Y = 1 | X(b)
)} = a(b)
+h(b)
(X(b)
)
h(b)
(X(b)
) =
l
β(b)
l ψ(b)
l (X(b)
) ∈ HK(b)
{ψ(b)
l } = { λ(b)
l φ(b)
l } implicitly specified via a symmetric positive
definite kernel K(b)
(·, ·).
K(b)
(X(b)
i , X(b)
j ) defines the similarity between X(b)
i and X(b)
j .
HK(b) , the functional space spanned by K(b)
(·, ·), is a reproducible
kernel hilbert space (RKHS)
Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
Choices of Kernel Functions
Linear kernel: K(Xi , Xj ) = ρ + XT
i Xj ,
h(X) =
p
k=1
βkXk
Fitting logistic regression with linear kernel ⇔ logistic ridge regression.
Adaptive Naive Bayes Kernel Machine Model Model and Methods 12
Choices of Kernel Functions
Linear kernel: K(Xi , Xj ) = ρ + XT
i Xj ,
h(X) =
p
k=1
βkXk
Fitting logistic regression with linear kernel ⇔ logistic ridge regression.
IBS kernel: K(Xi , Xj ) = p
k=1(2 − |Xik − Xjk|),
powerful in detecting non-linear effects with SNP data [Wu et al, 2010]
Adaptive Naive Bayes Kernel Machine Model Model and Methods 12
Estimation of h: Kernel PCA
• primal form: h = l βl ψl = l βl
√
λl φl
• Kernel PCA approximation:
K = [K(Xi , Xj )]1≤i,j≤n =
n
l=1
λl φl φ
T
l
K =
0
l=1
λl φl φ
T
l = ΨΨ
T
; Ψ = [λ
1
2
1 φ1, ..., λ
1
2
0
φ 0
]n× 0
Scholkopf et al. [1999]; Williams and Seeger [2000]; Braun et al. [2008]; Zhang et al. [2010]
• h(b)
(X(b)
) = Ψβ
• obtain (a, β) as the minimizer of ridge logistic objective function
L(Y , a, Ψβ) + τ β 2
Adaptive Naive Bayes Kernel Machine Model Model and Methods 13
Regularized Selection of Informative Regions
• For b = 1, ..., M, perform kernel PCA regression and obtain h(b)
logit{pr(Y = 1 | X(b)
)} = a(b)
+ h(b)
(X(b)
)
• Classify a future subject with X = {X(b)
, b = 1, ..., M} based on
M
b=1
h(b)
(X(b)
) ≥ c
• Final prediction rule with weighted block effects
– Some regions may not be predictive of the outcome due to false
discovery
– Inclusion of all regions for prediction may lead to reduced accuracy
– Regularized estimation of block effects using LASSO:
M
b=1
γbh(b)
(X(b)
) ≥ c
Adaptive Naive Bayes Kernel Machine Model Model and Methods 14
Regularized Selection of Informative Regions
• For b = 1, ..., M, perform kernel PCA regression and obtain h(b)
logit{pr(Y = 1 | X(b)
)} = a(b)
+ h(b)
(X(b)
)
• Classify a future subject with X = {X(b)
, b = 1, ..., M} based on
M
b=1
h(b)
(X(b)
) ≥ c
• Final prediction rule with weighted block effects
– Regularized estimation of block effects using LASSO, pseudo-data H
estimated with cross-validation:
K
k=1
YT
log g(b + Hγ) + (1 − Y)T
log{1 − g(b + Hγ)} − τ2 γ 1,
M
b=1
γbh(b)
(X(b)
) ≥ c
Adaptive Naive Bayes Kernel Machine Model Model and Methods 15
Theoretical Results
• Consistency of h(b)(x):
– h(b)
(x) → h(b)
(x) at
√
n rate for finite dimensional HK
– Relies on convergence of sample eigen-values and -vectors from kernel
PCA to the true eigensystem of HK
Ψ → Ψ = {ψ(b)
1 , . . . , ψ(b)
0
}
• Oracle property of γ:
– Gene-set selection consistency
P(A = A) → 1
where A = {b|h(b)
(x) = 0}, A = {b|h(b)
(x) = 0}
Adaptive Naive Bayes Kernel Machine Model Model and Methods 16
Simulation Studies for NBKM
• SNP data sampled from gene-sets in a GWAS dataset (from type I
diabetes study, Affy 500k)
• 350 regions, 9256 SNPs
• Only the first 4 regions are associated with the outcome
• the joint effects of the SNPs in each of these regions set as
– linear for the first two regions and non-linear for the other 2 regions
– linear for all 4 regions
– nonlinear for all 4 regions
Adaptive Naive Bayes Kernel Machine Model Simulation Studies 17
Prediction Accuracy
Simulations: nt = 1000, nv = 500, # of genes = 350 total # of SNPs = 9256
Adaptive Naive Bayes Kernel Machine Model Simulation Studies 18
Gene-set selection
Simulations: nt = 1000, nv = 500, # of genes = 350 total # of SNPs = 9256
Adaptive Naive Bayes Kernel Machine Model Simulation Studies 19
Genetic Risk of Type I Diabetes
• Autoimmune disease, usually diagnosed in childhood
• T1D
– 75 SNPs have been identified as T1D risk alleles (National Human
Genome Research Institute, Hindorff et al. [2009])
– 91 genes that either contain these SNPs or flank the SNP on either
side on the chromosome
Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 20
Genetic Risk of Type I Diabetes
• Autoimmune disease, usually diagnosed in childhood
• T1D
– 75 SNPs have been identified as T1D risk alleles (National Human
Genome Research Institute, Hindorff et al. [2009])
– 91 genes that either contain these SNPs or flank the SNP on either
side on the chromosome
• T1D + Other autoimmune diseases (Rheumatoid arthritis, Celiac
disease, Crohns disease, Lupus, Inflammatory bowel disease)
– 365 SNPs have been identified as T1D+other autoimmune disease risk
alleles (NHGRI)
– 375 genes that either contain these SNPs or flank the SNP on either
side on the chromosome
Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 20
Genetic Risk of Type I Diabetes
GWAS data collected by Welcome Trust Case Control Consortium
(WTCCC)
• 2000 cases, 3000 controls of European descent from Great Britain
• segment the genome into gene-sets: gene and a flanking region of
20KB on either side of the gene
• The WTCCC data includes
– 350 of the gene-sets listed in the NHGRI catalog
– covering 9,256 SNPs in the WTCCC data
Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 21
T1D Prediction Results
Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 22
Conclusions
• Kernel Machine Regression provides a useful tool for incorporating
non-linear complex effects
• Blockwise KM regression achieves a nice balance between capturing
complex effects and overfitting
• IBS kernel performs well under both linear and non-linear settings
Remarks
• May use SKAT to screen blocks for initial stage
• Can be extended to data with other covariates such as clinical
variables
• Possible extensions might incorporate more complex block structure,
different types of outcomes, interactions, and beyond!
Adaptive Naive Bayes Kernel Machine Model Conclusions 23
Thank you!
Adaptive Naive Bayes Kernel Machine Model Conclusions 24
References I
M. Braun, J. Buhmann, and K. M¨uller. On relevant dimensions in kernel feature spaces. The Journal of Machine Learning
Research, 9:1875–1908, 2008.
P. Domingos and M. Pazzani. On the optimality of the simple bayesian classifier under zero-one loss. Machine learning, 29(2):
103–130, 1997.
L. Hindorff, P. Sethupathy, H. Junkins, E. Ramos, J. Mehta, F. Collins, and T. Manolio. Potential etiologic and functional
implications of genome-wide association loci for human diseases and traits. Proceedings of the National Academy of
Sciences, 106(23):9362, 2009.
J. Minnier, M. Yuan, J. S. Liu, and T. Cai. Risk classification with an adaptive naive bayes kernel machine model. Journal of the
American Statistical Association, 110(509):393–404, 2015.
B. Scholkopf, S. Mika, C. Burges, P. Knirsch, K. Muller, G. Ratsch, and A. Smola. Input space versus feature space in
kernel-based methods. Neural Networks, IEEE Transactions on, 10(5):1000–1017, 1999.
C. Williams and M. Seeger. The effect of the input density distribution on kernel-based classifiers. In Proceedings of the 17th
International Conference on Machine Learning. Citeseer, 2000.
R. Zhang, W. Wang, and Y. Ma. Approximations of the standard principal components analysis and kernel pca. Expert Systems
with Applications, 37(9):6531–6537, 2010.
Adaptive Naive Bayes Kernel Machine Model 25
1 of 36

Recommended

Math, Stats and CS in Public Health and Medical Research by
Math, Stats and CS in Public Health and Medical ResearchMath, Stats and CS in Public Health and Medical Research
Math, Stats and CS in Public Health and Medical ResearchJessica Minnier
865 views39 slides
Deep learning application to medical imaging: Perspectives as a physician by
Deep learning application to medical imaging: Perspectives as a physicianDeep learning application to medical imaging: Perspectives as a physician
Deep learning application to medical imaging: Perspectives as a physicianHongyoon Choi
1.5K views53 slides
Amia tb-review-08 by
Amia tb-review-08Amia tb-review-08
Amia tb-review-08Russ Altman
291 views58 slides
Rodriguez_UROC_Final_Poster by
Rodriguez_UROC_Final_PosterRodriguez_UROC_Final_Poster
Rodriguez_UROC_Final_Poster​Iván Rodríguez
34 views1 slide
Ai in drug design webinar 26 feb 2019 by
Ai in drug design webinar 26 feb 2019Ai in drug design webinar 26 feb 2019
Ai in drug design webinar 26 feb 2019Pistoia Alliance
791 views72 slides
Ml in genomics by
Ml in genomicsMl in genomics
Ml in genomicsBrianSchilder
2.5K views21 slides

More Related Content

What's hot

CariesICTAI2008 by
CariesICTAI2008CariesICTAI2008
CariesICTAI2008Robson D. Montenegro
349 views5 slides
GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION... by
GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...
GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...ijaia
31 views17 slides
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASE by
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASEMISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASE
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASEIJDKP
23 views7 slides
AI for drug discovery by
AI for drug discoveryAI for drug discovery
AI for drug discoveryTruyen Tran
2.1K views31 slides
American Statistical Association October 23 2009 Presentation Part 1 by
American Statistical Association October 23 2009 Presentation Part 1American Statistical Association October 23 2009 Presentation Part 1
American Statistical Association October 23 2009 Presentation Part 1Double Check ĆŐNSULTING
308 views24 slides
NLP tutorial at AIME 2020 by
NLP tutorial at AIME 2020NLP tutorial at AIME 2020
NLP tutorial at AIME 2020Rui Zhang
683 views264 slides

What's hot(19)

GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION... by ijaia
GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...
GRAPHICAL MODEL AND CLUSTERINGREGRESSION BASED METHODS FOR CAUSAL INTERACTION...
ijaia31 views
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASE by IJDKP
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASEMISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASE
MISSING DATA CLASSIFICATION OF CHRONIC KIDNEY DISEASE
IJDKP23 views
AI for drug discovery by Truyen Tran
AI for drug discoveryAI for drug discovery
AI for drug discovery
Truyen Tran2.1K views
NLP tutorial at AIME 2020 by Rui Zhang
NLP tutorial at AIME 2020NLP tutorial at AIME 2020
NLP tutorial at AIME 2020
Rui Zhang683 views
The Role of Statistician in Personalized Medicine: An Overview of Statistical... by Setia Pramana
The Role of Statistician in Personalized Medicine: An Overview of Statistical...The Role of Statistician in Personalized Medicine: An Overview of Statistical...
The Role of Statistician in Personalized Medicine: An Overview of Statistical...
Setia Pramana1.9K views
Analysis of Imbalanced Classification Algorithms A Perspective View by ijtsrd
Analysis of Imbalanced Classification Algorithms A Perspective ViewAnalysis of Imbalanced Classification Algorithms A Perspective View
Analysis of Imbalanced Classification Algorithms A Perspective View
ijtsrd39 views
Very brief overview of AI in drug discovery by Dr. Gerry Higgins
Very brief overview of AI in drug discoveryVery brief overview of AI in drug discovery
Very brief overview of AI in drug discovery
Dr. Gerry Higgins3.2K views
September Journal Club -Aishwarya by RSG Luxembourg
September Journal Club -AishwaryaSeptember Journal Club -Aishwarya
September Journal Club -Aishwarya
RSG Luxembourg248 views
Machine Learning in Biology and Why It Doesn't Make Sense - Theo Knijnenburg,... by Seattle DAML meetup
Machine Learning in Biology and Why It Doesn't Make Sense - Theo Knijnenburg,...Machine Learning in Biology and Why It Doesn't Make Sense - Theo Knijnenburg,...
Machine Learning in Biology and Why It Doesn't Make Sense - Theo Knijnenburg,...
Simplified Knowledge Prediction: Application of Machine Learning in Real Life by Peea Bal Chakraborty
Simplified Knowledge Prediction: Application of Machine Learning in Real LifeSimplified Knowledge Prediction: Application of Machine Learning in Real Life
Simplified Knowledge Prediction: Application of Machine Learning in Real Life
SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu... by IJECEIAES
SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu...SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu...
SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu...
IJECEIAES9 views
Evaluation of Logistic Regression and Neural Network Model With Sensitivity A... by CSCJournals
Evaluation of Logistic Regression and Neural Network Model With Sensitivity A...Evaluation of Logistic Regression and Neural Network Model With Sensitivity A...
Evaluation of Logistic Regression and Neural Network Model With Sensitivity A...
CSCJournals529 views
Systems biology in polypharmacology: explaining and predicting drug secondary... by Andrei KUCHARAVY
Systems biology in polypharmacology: explaining and predicting drug secondary...Systems biology in polypharmacology: explaining and predicting drug secondary...
Systems biology in polypharmacology: explaining and predicting drug secondary...
Andrei KUCHARAVY1.4K views
Machine Learning Based Approaches for Prediction of Parkinson's Disease by mlaij
Machine Learning Based Approaches for Prediction of Parkinson's Disease  Machine Learning Based Approaches for Prediction of Parkinson's Disease
Machine Learning Based Approaches for Prediction of Parkinson's Disease
mlaij181 views
A Self-Adaptive Evolutionary Negative Selection Approach for Anom by Luis J. Gonzalez, PhD
A Self-Adaptive Evolutionary Negative Selection Approach for AnomA Self-Adaptive Evolutionary Negative Selection Approach for Anom
A Self-Adaptive Evolutionary Negative Selection Approach for Anom
FunGen JC Presentation - Mostafavi et al. (2019) by BrianSchilder
FunGen JC Presentation - Mostafavi et al. (2019)FunGen JC Presentation - Mostafavi et al. (2019)
FunGen JC Presentation - Mostafavi et al. (2019)
BrianSchilder3.1K views

Viewers also liked

Naive bayes classifier by
Naive bayes classifierNaive bayes classifier
Naive bayes classifierAbhijit Sengupta
660 views7 slides
Datamining with nb by
Datamining with nbDatamining with nb
Datamining with nbJames Wong
268 views53 slides
Dwdm naive bayes_ankit_gadgil_027 by
Dwdm naive bayes_ankit_gadgil_027Dwdm naive bayes_ankit_gadgil_027
Dwdm naive bayes_ankit_gadgil_027ankitgadgil
1.8K views11 slides
Ing transito (1) by
Ing transito (1)Ing transito (1)
Ing transito (1)erika acuña noriega
250 views3 slides
REALIDAD AUMENTADA EVARISTO GLEZ PORTAS by
REALIDAD AUMENTADA EVARISTO GLEZ PORTASREALIDAD AUMENTADA EVARISTO GLEZ PORTAS
REALIDAD AUMENTADA EVARISTO GLEZ PORTASVani González Portas
115 views11 slides
About-Callbox by
About-CallboxAbout-Callbox
About-CallboxDave Jensen
133 views18 slides

Viewers also liked(18)

Datamining with nb by James Wong
Datamining with nbDatamining with nb
Datamining with nb
James Wong268 views
Dwdm naive bayes_ankit_gadgil_027 by ankitgadgil
Dwdm naive bayes_ankit_gadgil_027Dwdm naive bayes_ankit_gadgil_027
Dwdm naive bayes_ankit_gadgil_027
ankitgadgil1.8K views
explain what values and attitudes are and describe their impact on managerial... by evangeline jumalon
explain what values and attitudes are and describe their impact on managerial...explain what values and attitudes are and describe their impact on managerial...
explain what values and attitudes are and describe their impact on managerial...
evangeline jumalon1.7K views
Ensayo cientifico accesibilidad y circulacion peatonal. electiva iv. by erika acuña noriega
Ensayo cientifico accesibilidad y circulacion peatonal. electiva iv.Ensayo cientifico accesibilidad y circulacion peatonal. electiva iv.
Ensayo cientifico accesibilidad y circulacion peatonal. electiva iv.
Text classification by James Wong
Text classificationText classification
Text classification
James Wong2.7K views
A short tutorial on r by Ashraf Uddin
A short tutorial on rA short tutorial on r
A short tutorial on r
Ashraf Uddin1.9K views
Introduction to text classification using naive bayes by Dhwaj Raj
Introduction to text classification using naive bayesIntroduction to text classification using naive bayes
Introduction to text classification using naive bayes
Dhwaj Raj6.5K views
Sentiment analysis using naive bayes classifier by Dev Sahu
Sentiment analysis using naive bayes classifier Sentiment analysis using naive bayes classifier
Sentiment analysis using naive bayes classifier
Dev Sahu27.3K views
Minor disorder of pregnancy ppt by pinal darji
Minor disorder of pregnancy pptMinor disorder of pregnancy ppt
Minor disorder of pregnancy ppt
pinal darji107K views

Similar to Risk Classification with an Adaptive Naive Bayes Kernel Machine Model

32_Nov07_MachineLear.. by
32_Nov07_MachineLear..32_Nov07_MachineLear..
32_Nov07_MachineLear..butest
464 views52 slides
04 1 evolution by
04 1 evolution04 1 evolution
04 1 evolutionTianlu Wang
345 views33 slides
Data driven model optimization [autosaved] by
Data driven model optimization [autosaved]Data driven model optimization [autosaved]
Data driven model optimization [autosaved]Russell Jarvis
109 views42 slides
Integrative Networks Centric Bioinformatics by
Integrative Networks Centric BioinformaticsIntegrative Networks Centric Bioinformatics
Integrative Networks Centric BioinformaticsNatalio Krasnogor
9.9K views73 slides
ABC short course: final chapters by
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chaptersChristian Robert
3.4K views100 slides
5 5 10 by
5 5 105 5 10
5 5 10kohannim
293 views27 slides

Similar to Risk Classification with an Adaptive Naive Bayes Kernel Machine Model(20)

32_Nov07_MachineLear.. by butest
32_Nov07_MachineLear..32_Nov07_MachineLear..
32_Nov07_MachineLear..
butest464 views
Data driven model optimization [autosaved] by Russell Jarvis
Data driven model optimization [autosaved]Data driven model optimization [autosaved]
Data driven model optimization [autosaved]
Russell Jarvis109 views
Integrative Networks Centric Bioinformatics by Natalio Krasnogor
Integrative Networks Centric BioinformaticsIntegrative Networks Centric Bioinformatics
Integrative Networks Centric Bioinformatics
Natalio Krasnogor9.9K views
5 5 10 by kohannim
5 5 105 5 10
5 5 10
kohannim293 views
Diversity mechanisms for evolutionary populations in Search-Based Software En... by Annibale Panichella
Diversity mechanisms for evolutionary populations in Search-Based Software En...Diversity mechanisms for evolutionary populations in Search-Based Software En...
Diversity mechanisms for evolutionary populations in Search-Based Software En...
. An introduction to machine learning and probabilistic ... by butest
. An introduction to machine learning and probabilistic .... An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic ...
butest1.2K views
Large Scale Data Clustering: an overview by Vahid Mirjalili
Large Scale Data Clustering: an overviewLarge Scale Data Clustering: an overview
Large Scale Data Clustering: an overview
Vahid Mirjalili2.9K views
Oral presentation at Protein Folding Consortium Workshop in Berkeley (2017) by Temple University
Oral presentation at Protein Folding Consortium Workshop in Berkeley (2017)Oral presentation at Protein Folding Consortium Workshop in Berkeley (2017)
Oral presentation at Protein Folding Consortium Workshop in Berkeley (2017)
randomization approach in case-based reasoning: case of study of mammography ... by Miled Basma Bentaiba
randomization approach in case-based reasoning: case of study of mammography ...randomization approach in case-based reasoning: case of study of mammography ...
randomization approach in case-based reasoning: case of study of mammography ...
Multi-omics infrastructure and data for R/Bioconductor by Levi Waldron
Multi-omics infrastructure and data for R/BioconductorMulti-omics infrastructure and data for R/Bioconductor
Multi-omics infrastructure and data for R/Bioconductor
Levi Waldron895 views
Introduction to Supervised ML Concepts and Algorithms by NBER
Introduction to Supervised ML Concepts and AlgorithmsIntroduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and Algorithms
NBER31K views
2008: Applied AIS - A Roadmap of AIS Research in Brazil and Sample Applications by Leandro de Castro
2008: Applied AIS - A Roadmap of AIS Research in Brazil and Sample Applications2008: Applied AIS - A Roadmap of AIS Research in Brazil and Sample Applications
2008: Applied AIS - A Roadmap of AIS Research in Brazil and Sample Applications
Leandro de Castro1.2K views
Affymetrix OncoScan®* data analysis with Nexus Copy Number™ by Affymetrix
Affymetrix OncoScan®* data analysis with Nexus Copy Number™Affymetrix OncoScan®* data analysis with Nexus Copy Number™
Affymetrix OncoScan®* data analysis with Nexus Copy Number™
Affymetrix652 views
R Analytics in the Cloud by DataMine Lab
R Analytics in the CloudR Analytics in the Cloud
R Analytics in the Cloud
DataMine Lab4.3K views
The application of artificial intelligence by Pallavi Vashistha
The application of artificial intelligenceThe application of artificial intelligence
The application of artificial intelligence
Pallavi Vashistha1.2K views

Recently uploaded

SUPER STORE SQL PROJECT.pptx by
SUPER STORE SQL PROJECT.pptxSUPER STORE SQL PROJECT.pptx
SUPER STORE SQL PROJECT.pptxkhan888620
12 views16 slides
Short Story Assignment by Kelly Nguyen by
Short Story Assignment by Kelly NguyenShort Story Assignment by Kelly Nguyen
Short Story Assignment by Kelly Nguyenkellynguyen01
19 views17 slides
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M... by
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...DataScienceConferenc1
6 views11 slides
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an... by
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...StatsCommunications
5 views26 slides
[DSC Europe 23] Ivana Sesic - Use of AI in Public Health.pptx by
[DSC Europe 23] Ivana Sesic - Use of AI in Public Health.pptx[DSC Europe 23] Ivana Sesic - Use of AI in Public Health.pptx
[DSC Europe 23] Ivana Sesic - Use of AI in Public Health.pptxDataScienceConferenc1
5 views15 slides
[DSC Europe 23][AI:CSI] Dragan Pleskonjic - AI Impact on Cybersecurity and P... by
[DSC Europe 23][AI:CSI]  Dragan Pleskonjic - AI Impact on Cybersecurity and P...[DSC Europe 23][AI:CSI]  Dragan Pleskonjic - AI Impact on Cybersecurity and P...
[DSC Europe 23][AI:CSI] Dragan Pleskonjic - AI Impact on Cybersecurity and P...DataScienceConferenc1
5 views36 slides

Recently uploaded(20)

SUPER STORE SQL PROJECT.pptx by khan888620
SUPER STORE SQL PROJECT.pptxSUPER STORE SQL PROJECT.pptx
SUPER STORE SQL PROJECT.pptx
khan88862012 views
Short Story Assignment by Kelly Nguyen by kellynguyen01
Short Story Assignment by Kelly NguyenShort Story Assignment by Kelly Nguyen
Short Story Assignment by Kelly Nguyen
kellynguyen0119 views
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M... by DataScienceConferenc1
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...
[DSC Europe 23] Milos Grubjesic Empowering Business with Pepsico s Advanced M...
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an... by StatsCommunications
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...
OECD-Persol Holdings Workshop on Advancing Employee Well-being in Business an...
[DSC Europe 23][AI:CSI] Dragan Pleskonjic - AI Impact on Cybersecurity and P... by DataScienceConferenc1
[DSC Europe 23][AI:CSI]  Dragan Pleskonjic - AI Impact on Cybersecurity and P...[DSC Europe 23][AI:CSI]  Dragan Pleskonjic - AI Impact on Cybersecurity and P...
[DSC Europe 23][AI:CSI] Dragan Pleskonjic - AI Impact on Cybersecurity and P...
[DSC Europe 23][Cryptica] Martin_Summer_Digital_central_bank_money_Ideas_init... by DataScienceConferenc1
[DSC Europe 23][Cryptica] Martin_Summer_Digital_central_bank_money_Ideas_init...[DSC Europe 23][Cryptica] Martin_Summer_Digital_central_bank_money_Ideas_init...
[DSC Europe 23][Cryptica] Martin_Summer_Digital_central_bank_money_Ideas_init...
Organic Shopping in Google Analytics 4.pdf by GA4 Tutorials
Organic Shopping in Google Analytics 4.pdfOrganic Shopping in Google Analytics 4.pdf
Organic Shopping in Google Analytics 4.pdf
GA4 Tutorials14 views
Chapter 3b- Process Communication (1) (1)(1) (1).pptx by ayeshabaig2004
Chapter 3b- Process Communication (1) (1)(1) (1).pptxChapter 3b- Process Communication (1) (1)(1) (1).pptx
Chapter 3b- Process Communication (1) (1)(1) (1).pptx
ayeshabaig20046 views
CRIJ4385_Death Penalty_F23.pptx by yvettemm100
CRIJ4385_Death Penalty_F23.pptxCRIJ4385_Death Penalty_F23.pptx
CRIJ4385_Death Penalty_F23.pptx
yvettemm1006 views
Survey on Factuality in LLM's.pptx by NeethaSherra1
Survey on Factuality in LLM's.pptxSurvey on Factuality in LLM's.pptx
Survey on Factuality in LLM's.pptx
NeethaSherra16 views
[DSC Europe 23] Zsolt Feleki - Machine Translation should we trust it.pptx by DataScienceConferenc1
[DSC Europe 23] Zsolt Feleki - Machine Translation should we trust it.pptx[DSC Europe 23] Zsolt Feleki - Machine Translation should we trust it.pptx
[DSC Europe 23] Zsolt Feleki - Machine Translation should we trust it.pptx
[DSC Europe 23] Stefan Mrsic_Goran Savic - Evolving Technology Excellence.pptx by DataScienceConferenc1
[DSC Europe 23] Stefan Mrsic_Goran Savic - Evolving Technology Excellence.pptx[DSC Europe 23] Stefan Mrsic_Goran Savic - Evolving Technology Excellence.pptx
[DSC Europe 23] Stefan Mrsic_Goran Savic - Evolving Technology Excellence.pptx
Ukraine Infographic_22NOV2023_v2.pdf by AnastosiyaGurin
Ukraine Infographic_22NOV2023_v2.pdfUkraine Infographic_22NOV2023_v2.pdf
Ukraine Infographic_22NOV2023_v2.pdf
AnastosiyaGurin1.4K views
Data Journeys Hard Talk workshop final.pptx by info828217
Data Journeys Hard Talk workshop final.pptxData Journeys Hard Talk workshop final.pptx
Data Journeys Hard Talk workshop final.pptx
info82821710 views
CRM stick or twist.pptx by info828217
CRM stick or twist.pptxCRM stick or twist.pptx
CRM stick or twist.pptx
info82821710 views
CRM stick or twist workshop by info828217
CRM stick or twist workshopCRM stick or twist workshop
CRM stick or twist workshop
info8282179 views

Risk Classification with an Adaptive Naive Bayes Kernel Machine Model

  • 1. Risk Classification with an Adaptive Naive Bayes Kernel Machine Model Jessica Minnier1, Ming Yuan3, Jun Liu4, and Tianxi Cai2 1Department of Public Health & Preventive Medicine, Oregon Health & Science University 2Department of Biostatistics, Harvard School of Public Health 3Department of Statistics, University of Wisconsin-Madison 4Department of Statistics, Harvard University June 30, 2015 ASA Oregon Chapter Meeting
  • 2. Outline 1 Background and Motivation 2 Model and Methods Kernels Blockwise Kernel PCA Estimation Regularized Selection of Informative Regions Theoretical Results 3 Simulation Studies 4 Genetic Risk of Type I Diabetes 5 Conclusions Adaptive Naive Bayes Kernel Machine Model 2
  • 3. Background and Motivation Adaptive Naive Bayes (Blockwise) Kernel Machine Classification • Goal: genetic data → quantify disease risk, predict therapeutic efficacy, determine disease subtypes • Goal: build an accurate parsimonious prediction model – reduce the cost of unnecessary marker measurements – improve the prediction precision for future patients – improve over modest prediction precision obtained with clinical predictors and/or known risk alleles Adaptive Naive Bayes Kernel Machine Model Background and Motivation 3
  • 4. Background and Motivation Adaptive Naive Bayes (Blockwise) Kernel Machine Classification • Goal: genetic data → quantify disease risk, predict therapeutic efficacy, determine disease subtypes • Goal: build an accurate parsimonious prediction model – reduce the cost of unnecessary marker measurements – improve the prediction precision for future patients – improve over modest prediction precision obtained with clinical predictors and/or known risk alleles • Complex diseases – many alleles contribute to risk – many distinct combinations of risk factors lead to disease Adaptive Naive Bayes Kernel Machine Model Background and Motivation 3
  • 5. Background and Motivation • Genome wide association studies (GWAS) – identifying SNPs associated with disease risk – primary goal of testing – accurate risk prediction remains difficult • Common approach: – select top ranked SNPs based on large scale testing – construct a composite genetic score w/ selected SNPs Adaptive Naive Bayes Kernel Machine Model Background and Motivation 4
  • 6. Background and Motivation • Genome wide association studies (GWAS) – identifying SNPs associated with disease risk – primary goal of testing – accurate risk prediction remains difficult • Common approach: – select top ranked SNPs based on large scale testing – construct a composite genetic score w/ selected SNPs – may not work well due to false +/− errors in identifying predictive SNPs over-fitting using only subset of SNPs available additive effects only Adaptive Naive Bayes Kernel Machine Model Background and Motivation 4
  • 7. Background and Motivation Recent progress in prediction with high dimensional data • Regularized estimation: LASSO (Tibshirani, 1996); SCAD (Fan and Li, 2001); Adaptive LASSO (Zou, 2006) • Machine learning: Support vector machine (Cristianini, Shawe-Taylor, 2000); Least square Kernel Machine Regression (Liu, Lin, Ghosh, 2007); Kernel logistic regression (Zhu and Hastie, 2005; Liu, Ghosh and Lin, 2008) • Screening + Regularized estimation: Sure independence screening (Fan and Lv, 2008; Fan and Song, 2009) Global methods: may be unstable for large p, high correlation Adaptive Naive Bayes Kernel Machine Model Background and Motivation 5
  • 8. Approach Challenge: • Prediction models based on univariate testing, additive models, global methods → low prediction accuracy, low AUC, missing heritability • Non-linear effects? testing for interactions → low power Adaptive Naive Bayes Kernel Machine Model Model and Methods 6
  • 9. Approach Challenge: • Prediction models based on univariate testing, additive models, global methods → low prediction accuracy, low AUC, missing heritability • Non-linear effects? testing for interactions → low power Our approach [Minnier et al., 2015]: • Blockwise method: leverage biological knowledge to build models at the gene-set level genes, gene-pathways, linkage disequilibrium blocks • Kernel machine regression: allow for complex and nonlinear effects implicitly specify underlying complex functional form of covariate effects via similarity measures (kernels) that define the distance between two sets of covariates Adaptive Naive Bayes Kernel Machine Model Model and Methods 6
  • 10. Kernel Methods: similar inputs to similar outputs • transform data to feature space H with non-linear map φ • “kernel trick” lets us use K(, ) similarity function instead of φ • K induces the feature space N. Takahashi’s webpage Adaptive Naive Bayes Kernel Machine Model Model and Methods 7
  • 11. Previous Methods Blockwise methods • Inference: Gene-set testing Gene burden tests Gene Set Enrichment Analysis (GSEA) SNP-set Sequence Kernel Association Test (SKAT, SKAT-O; Wu et al. 2010; Wu, Lee, et al. 2011) Adaptive Naive Bayes Kernel Machine Model Model and Methods 8
  • 12. Previous Methods Blockwise methods • Inference: Gene-set testing Gene burden tests Gene Set Enrichment Analysis (GSEA) SNP-set Sequence Kernel Association Test (SKAT, SKAT-O; Wu et al. 2010; Wu, Lee, et al. 2011) Kernel machine methods • Support Vector Machine (SVM) classification methods • Inference KM SNP-set Testing (Liu et al. 2007, 2008; SKAT methods) Gene expression test with kernel Cox model (Li and Luan 2003) Adaptive Naive Bayes Kernel Machine Model Model and Methods 8
  • 13. Notations and Model Assumptions • Data – Response: Y = (Y1, ..., Yn)T – Predictors: M blocks of genomic regions, for b = 1, ..., M, X(b) = (X (b) 1 , ..., X(b) n )T n×pb , Adaptive Naive Bayes Kernel Machine Model Model and Methods 9
  • 14. Notations and Model Assumptions • Data – Response: Y = (Y1, ..., Yn)T – Predictors: M blocks of genomic regions, for b = 1, ..., M, X(b) = (X (b) 1 , ..., X(b) n )T n×pb , • Blockwise: Partition genome into gene-sets – Recombination hotspots, gene-pathways Adaptive Naive Bayes Kernel Machine Model Model and Methods 9
  • 15. Notations and Model Assumptions • Data – Response: Y = (Y1, ..., Yn)T – Predictors: M blocks of genomic regions, for b = 1, ..., M, X(b) = (X (b) 1 , ..., X(b) n )T n×pb , • Model under blockwise Naive Bayes (NB) assumption: X(1) , ..., X(M) | Y independent Adaptive Naive Bayes Kernel Machine Model Model and Methods 10
  • 16. Notations and Model Assumptions • Data – Response: Y = (Y1, ..., Yn)T – Predictors: M blocks of genomic regions, for b = 1, ..., M, X(b) = (X (b) 1 , ..., X(b) n )T n×pb , • Model under blockwise Naive Bayes (NB) assumption: X(1) , ..., X(M) | Y independent ⇒ logit{pr(Y = 1 | X(1) , ..., X(M) )} = c + M b=1 logit{pr(Y = 1 | X(b) )} – NB assumption allows separate estimation by block and reduces overfitting – Performs well for zero-one loss L(X) = I( ˆY (X) = Y ) [Domingos and Pazzani, 1997] Adaptive Naive Bayes Kernel Machine Model Model and Methods 10
  • 17. Notations and Model Assumptions • Within each region, the effect may be complex and interactive due to – multiple causal variants – un-typed causal variants in the presence of high LD Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
  • 18. Notations and Model Assumptions • Within each region, the effect may be complex and interactive due to – multiple causal variants – un-typed causal variants in the presence of high LD • Blockwise Kernel Machine Regression logit{pr(Y = 1 | X(b) )} = a(b) +h(b) (X(b) ) h(b) (X(b) ) = l β(b) l ψ(b) l (X(b) ) ∈ HK(b) {ψ(b) l } = { λ(b) l φ(b) l } implicitly specified via a symmetric positive definite kernel K(b) (·, ·). Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
  • 19. Notations and Model Assumptions • Within each region, the effect may be complex and interactive due to – multiple causal variants – un-typed causal variants in the presence of high LD • Blockwise Kernel Machine Regression logit{pr(Y = 1 | X(b) )} = a(b) +h(b) (X(b) ) h(b) (X(b) ) = l β(b) l ψ(b) l (X(b) ) ∈ HK(b) {ψ(b) l } = { λ(b) l φ(b) l } implicitly specified via a symmetric positive definite kernel K(b) (·, ·). K(b) (X(b) i , X(b) j ) defines the similarity between X(b) i and X(b) j . Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
  • 20. Notations and Model Assumptions • Within each region, the effect may be complex and interactive due to – multiple causal variants – un-typed causal variants in the presence of high LD • Blockwise Kernel Machine Regression logit{pr(Y = 1 | X(b) )} = a(b) +h(b) (X(b) ) h(b) (X(b) ) = l β(b) l ψ(b) l (X(b) ) ∈ HK(b) {ψ(b) l } = { λ(b) l φ(b) l } implicitly specified via a symmetric positive definite kernel K(b) (·, ·). K(b) (X(b) i , X(b) j ) defines the similarity between X(b) i and X(b) j . HK(b) , the functional space spanned by K(b) (·, ·), is a reproducible kernel hilbert space (RKHS) Adaptive Naive Bayes Kernel Machine Model Model and Methods 11
  • 21. Choices of Kernel Functions Linear kernel: K(Xi , Xj ) = ρ + XT i Xj , h(X) = p k=1 βkXk Fitting logistic regression with linear kernel ⇔ logistic ridge regression. Adaptive Naive Bayes Kernel Machine Model Model and Methods 12
  • 22. Choices of Kernel Functions Linear kernel: K(Xi , Xj ) = ρ + XT i Xj , h(X) = p k=1 βkXk Fitting logistic regression with linear kernel ⇔ logistic ridge regression. IBS kernel: K(Xi , Xj ) = p k=1(2 − |Xik − Xjk|), powerful in detecting non-linear effects with SNP data [Wu et al, 2010] Adaptive Naive Bayes Kernel Machine Model Model and Methods 12
  • 23. Estimation of h: Kernel PCA • primal form: h = l βl ψl = l βl √ λl φl • Kernel PCA approximation: K = [K(Xi , Xj )]1≤i,j≤n = n l=1 λl φl φ T l K = 0 l=1 λl φl φ T l = ΨΨ T ; Ψ = [λ 1 2 1 φ1, ..., λ 1 2 0 φ 0 ]n× 0 Scholkopf et al. [1999]; Williams and Seeger [2000]; Braun et al. [2008]; Zhang et al. [2010] • h(b) (X(b) ) = Ψβ • obtain (a, β) as the minimizer of ridge logistic objective function L(Y , a, Ψβ) + τ β 2 Adaptive Naive Bayes Kernel Machine Model Model and Methods 13
  • 24. Regularized Selection of Informative Regions • For b = 1, ..., M, perform kernel PCA regression and obtain h(b) logit{pr(Y = 1 | X(b) )} = a(b) + h(b) (X(b) ) • Classify a future subject with X = {X(b) , b = 1, ..., M} based on M b=1 h(b) (X(b) ) ≥ c • Final prediction rule with weighted block effects – Some regions may not be predictive of the outcome due to false discovery – Inclusion of all regions for prediction may lead to reduced accuracy – Regularized estimation of block effects using LASSO: M b=1 γbh(b) (X(b) ) ≥ c Adaptive Naive Bayes Kernel Machine Model Model and Methods 14
  • 25. Regularized Selection of Informative Regions • For b = 1, ..., M, perform kernel PCA regression and obtain h(b) logit{pr(Y = 1 | X(b) )} = a(b) + h(b) (X(b) ) • Classify a future subject with X = {X(b) , b = 1, ..., M} based on M b=1 h(b) (X(b) ) ≥ c • Final prediction rule with weighted block effects – Regularized estimation of block effects using LASSO, pseudo-data H estimated with cross-validation: K k=1 YT log g(b + Hγ) + (1 − Y)T log{1 − g(b + Hγ)} − τ2 γ 1, M b=1 γbh(b) (X(b) ) ≥ c Adaptive Naive Bayes Kernel Machine Model Model and Methods 15
  • 26. Theoretical Results • Consistency of h(b)(x): – h(b) (x) → h(b) (x) at √ n rate for finite dimensional HK – Relies on convergence of sample eigen-values and -vectors from kernel PCA to the true eigensystem of HK Ψ → Ψ = {ψ(b) 1 , . . . , ψ(b) 0 } • Oracle property of γ: – Gene-set selection consistency P(A = A) → 1 where A = {b|h(b) (x) = 0}, A = {b|h(b) (x) = 0} Adaptive Naive Bayes Kernel Machine Model Model and Methods 16
  • 27. Simulation Studies for NBKM • SNP data sampled from gene-sets in a GWAS dataset (from type I diabetes study, Affy 500k) • 350 regions, 9256 SNPs • Only the first 4 regions are associated with the outcome • the joint effects of the SNPs in each of these regions set as – linear for the first two regions and non-linear for the other 2 regions – linear for all 4 regions – nonlinear for all 4 regions Adaptive Naive Bayes Kernel Machine Model Simulation Studies 17
  • 28. Prediction Accuracy Simulations: nt = 1000, nv = 500, # of genes = 350 total # of SNPs = 9256 Adaptive Naive Bayes Kernel Machine Model Simulation Studies 18
  • 29. Gene-set selection Simulations: nt = 1000, nv = 500, # of genes = 350 total # of SNPs = 9256 Adaptive Naive Bayes Kernel Machine Model Simulation Studies 19
  • 30. Genetic Risk of Type I Diabetes • Autoimmune disease, usually diagnosed in childhood • T1D – 75 SNPs have been identified as T1D risk alleles (National Human Genome Research Institute, Hindorff et al. [2009]) – 91 genes that either contain these SNPs or flank the SNP on either side on the chromosome Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 20
  • 31. Genetic Risk of Type I Diabetes • Autoimmune disease, usually diagnosed in childhood • T1D – 75 SNPs have been identified as T1D risk alleles (National Human Genome Research Institute, Hindorff et al. [2009]) – 91 genes that either contain these SNPs or flank the SNP on either side on the chromosome • T1D + Other autoimmune diseases (Rheumatoid arthritis, Celiac disease, Crohns disease, Lupus, Inflammatory bowel disease) – 365 SNPs have been identified as T1D+other autoimmune disease risk alleles (NHGRI) – 375 genes that either contain these SNPs or flank the SNP on either side on the chromosome Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 20
  • 32. Genetic Risk of Type I Diabetes GWAS data collected by Welcome Trust Case Control Consortium (WTCCC) • 2000 cases, 3000 controls of European descent from Great Britain • segment the genome into gene-sets: gene and a flanking region of 20KB on either side of the gene • The WTCCC data includes – 350 of the gene-sets listed in the NHGRI catalog – covering 9,256 SNPs in the WTCCC data Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 21
  • 33. T1D Prediction Results Adaptive Naive Bayes Kernel Machine Model Genetic Risk of Type I Diabetes 22
  • 34. Conclusions • Kernel Machine Regression provides a useful tool for incorporating non-linear complex effects • Blockwise KM regression achieves a nice balance between capturing complex effects and overfitting • IBS kernel performs well under both linear and non-linear settings Remarks • May use SKAT to screen blocks for initial stage • Can be extended to data with other covariates such as clinical variables • Possible extensions might incorporate more complex block structure, different types of outcomes, interactions, and beyond! Adaptive Naive Bayes Kernel Machine Model Conclusions 23
  • 35. Thank you! Adaptive Naive Bayes Kernel Machine Model Conclusions 24
  • 36. References I M. Braun, J. Buhmann, and K. M¨uller. On relevant dimensions in kernel feature spaces. The Journal of Machine Learning Research, 9:1875–1908, 2008. P. Domingos and M. Pazzani. On the optimality of the simple bayesian classifier under zero-one loss. Machine learning, 29(2): 103–130, 1997. L. Hindorff, P. Sethupathy, H. Junkins, E. Ramos, J. Mehta, F. Collins, and T. Manolio. Potential etiologic and functional implications of genome-wide association loci for human diseases and traits. Proceedings of the National Academy of Sciences, 106(23):9362, 2009. J. Minnier, M. Yuan, J. S. Liu, and T. Cai. Risk classification with an adaptive naive bayes kernel machine model. Journal of the American Statistical Association, 110(509):393–404, 2015. B. Scholkopf, S. Mika, C. Burges, P. Knirsch, K. Muller, G. Ratsch, and A. Smola. Input space versus feature space in kernel-based methods. Neural Networks, IEEE Transactions on, 10(5):1000–1017, 1999. C. Williams and M. Seeger. The effect of the input density distribution on kernel-based classifiers. In Proceedings of the 17th International Conference on Machine Learning. Citeseer, 2000. R. Zhang, W. Wang, and Y. Ma. Approximations of the standard principal components analysis and kernel pca. Expert Systems with Applications, 37(9):6531–6537, 2010. Adaptive Naive Bayes Kernel Machine Model 25