This document discusses quantifying uncertainty in decision curve analysis when evaluating clinical prediction models. It presents several methods for quantifying uncertainty in net benefit, including confidence intervals, P(useful), and expected value of perfect information (EVPI). Limited sample sizes and heterogeneity between populations can introduce uncertainty about the optimal clinical strategy. Quantifying this uncertainty is important for understanding the value of further external validation of prediction models but traditional hypothesis testing is not always appropriate.
ISCB 2023 Sources of uncertainty b.pptxBenVanCalster
This talk gives an overview and illustration of various sources of uncertainty when developing clinical risk prediction models: aleatory, approximation, model, modeler, data, and population uncertainty. Presented at the International Conference of Clinical Biostatistics, 29th of August 2023, Milan, Italy.
These slides were presented on November 22 2016 during the Annual Julius Symposium, organised by the Julius Center for Health Sciences and Primary Care, University Medical Hospital Utrecht.
Only a few months ago, the American Statistical Association authoritatively issued an official statement on significance and p-values (American Statistician, 2016, 70:2, 129-133), claiming that the p-value is: “commonly misused and misinterpreted.”
In this presentation I focus on the principles of the ASA statement.
Dans le contexte judiciaire, le raisonnement
bayésien offre une approche logique et rigoureuse pour évaluer la nature d’une trace, la source de cette trace et l’activité ayant mené au dépôt de cette trace, qu’il s’agisse de sperme, de sang, de morsures, de fibres textiles ou de résidus de tir, etc. À travers l’exposition de divers scénarios fictifs, cet article qui se veut pédagogique met en lumière les bases fondamentales de l’approche bayésienne permettant l’évaluation probabiliste des traces et indices.
Algorithmic Bias: Challenges and Opportunities for AI in HealthcareGregory Nelson
Gregory S. Nelson, VP, Analytics and Strategy – Vidant Health | Adjunct Faculty Duke University
The promise of AI is quickly becoming a reality for a number of industries including healthcare. For example, we have seen early successes in the augmenting clinical intelligence for diagnostic imaging and in early detection of pneumonia and sepsis. But what happens when the algorithms are biased? In this presentation, we will outline a framework for AI governance and discuss ways in which we can address algorithmic bias in machine learning.
Objective 1: Illustrate the issues of bias in AI through examples specific to healthcare.
Objective 2: Summarize the growing body of work in the legal, regulatory, and ethical oversight of AI models and the implications for healthcare.
Objective 3: Outline steps that we can take to establish an AI governance strategy for our organizations.
This presentation explores the strengths and weaknesses of ordinary least squares and propensity score matching. Matching alone cannot solve endogeneity problems faced by OLS. The presentation shows how PSM and OLS can be combined to yield less-biased estimators than either method alone.
ISCB 2023 Sources of uncertainty b.pptxBenVanCalster
This talk gives an overview and illustration of various sources of uncertainty when developing clinical risk prediction models: aleatory, approximation, model, modeler, data, and population uncertainty. Presented at the International Conference of Clinical Biostatistics, 29th of August 2023, Milan, Italy.
These slides were presented on November 22 2016 during the Annual Julius Symposium, organised by the Julius Center for Health Sciences and Primary Care, University Medical Hospital Utrecht.
Only a few months ago, the American Statistical Association authoritatively issued an official statement on significance and p-values (American Statistician, 2016, 70:2, 129-133), claiming that the p-value is: “commonly misused and misinterpreted.”
In this presentation I focus on the principles of the ASA statement.
Dans le contexte judiciaire, le raisonnement
bayésien offre une approche logique et rigoureuse pour évaluer la nature d’une trace, la source de cette trace et l’activité ayant mené au dépôt de cette trace, qu’il s’agisse de sperme, de sang, de morsures, de fibres textiles ou de résidus de tir, etc. À travers l’exposition de divers scénarios fictifs, cet article qui se veut pédagogique met en lumière les bases fondamentales de l’approche bayésienne permettant l’évaluation probabiliste des traces et indices.
Algorithmic Bias: Challenges and Opportunities for AI in HealthcareGregory Nelson
Gregory S. Nelson, VP, Analytics and Strategy – Vidant Health | Adjunct Faculty Duke University
The promise of AI is quickly becoming a reality for a number of industries including healthcare. For example, we have seen early successes in the augmenting clinical intelligence for diagnostic imaging and in early detection of pneumonia and sepsis. But what happens when the algorithms are biased? In this presentation, we will outline a framework for AI governance and discuss ways in which we can address algorithmic bias in machine learning.
Objective 1: Illustrate the issues of bias in AI through examples specific to healthcare.
Objective 2: Summarize the growing body of work in the legal, regulatory, and ethical oversight of AI models and the implications for healthcare.
Objective 3: Outline steps that we can take to establish an AI governance strategy for our organizations.
This presentation explores the strengths and weaknesses of ordinary least squares and propensity score matching. Matching alone cannot solve endogeneity problems faced by OLS. The presentation shows how PSM and OLS can be combined to yield less-biased estimators than either method alone.
5 essential steps for sample size determination in clinical trials slidesharenQuery
In this free webinar hosted by nQuery Researcher & Statistician Eimear Keyes, we map out the 5 essential steps for sample size determination in clinical trials. At each step, Eimear will highlight the important function it plays and how to avoid the errors that will negatively impact your sample size determination and therefore your study.
Watch the Video: https://www.statsols.com/webinar/the-5-essential-steps-for-sample-size-determination
Presenting a published paper:
"Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review
approach"
Diagnostic accuracy of MALDI-TOF mass spectrometry for the direct identification of clinical pathogens from urine
Journal Club (Systematic Review & Meta Analysis)
Clinical Microbiology Fellowship
Prediction, Big Data, and AI: Steyerberg, Basel Nov 1, 2019Ewout Steyerberg
Title"Clinical prediction models in the age of artificial intelligence and big data", presented at the Basel Biometrics Society seminar Nov 1, 2019, Basel, by Ewout Steyerberg, with substantial inout from Maarten van Smeden and Ben van Calster
How to combine results from randomised clinical trials on the additive scale with real world data to provide predictions on the clinically relevant scale for individual patients
5 essential steps for sample size determination in clinical trials slidesharenQuery
In this free webinar hosted by nQuery Researcher & Statistician Eimear Keyes, we map out the 5 essential steps for sample size determination in clinical trials. At each step, Eimear will highlight the important function it plays and how to avoid the errors that will negatively impact your sample size determination and therefore your study.
Watch the Video: https://www.statsols.com/webinar/the-5-essential-steps-for-sample-size-determination
Presenting a published paper:
"Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review
approach"
Diagnostic accuracy of MALDI-TOF mass spectrometry for the direct identification of clinical pathogens from urine
Journal Club (Systematic Review & Meta Analysis)
Clinical Microbiology Fellowship
Prediction, Big Data, and AI: Steyerberg, Basel Nov 1, 2019Ewout Steyerberg
Title"Clinical prediction models in the age of artificial intelligence and big data", presented at the Basel Biometrics Society seminar Nov 1, 2019, Basel, by Ewout Steyerberg, with substantial inout from Maarten van Smeden and Ben van Calster
How to combine results from randomised clinical trials on the additive scale with real world data to provide predictions on the clinically relevant scale for individual patients
Developing and validating statistical models for clinical prediction and prog...Evangelos Kritsotakis
Talk on clinical prediction models presented at the Joint Seminar Series in Translational and Clinical Medicine organised by the University of Crete Medical School, the Institute of Molecular Biology and Biotechnology of the Foundation for Research and Technology Hellas (IMBB-FORTH), and the University of Crete Research Center (UCRC), Heraklion [online], Greece, April 7, 2021.
Practical Methods To Overcome Sample Size ChallengesnQuery
Watch the video at: https://www.statsols.com/webinars/practical-methods-to-overcome-sample-size-challenges
In this webinar hosted by Ronan Fitzpatrick - Head of Statistics and nQuery Lead Researcher at Statsols - we will examine some of the most common practical challenges you will experience while calculating sample size for your study. These challenges will be split into two categories:
1. Overcoming Sample Size Calculation Challenges
(Survival Analysis Example)
We will examine practical methods to overcome common sample size calculation issues by focusing in on one of the more complex areas for sample size determination; Survival analysis. We will cover difficulties and potential issues surrounding challenges such as:
Drop Out: How to deal with expected dropouts or censoring. We compare the simple loss-to-follow-up method and integrating a dropout process into the sample size model?
Planning Uncertainty: How best to deal with the inevitable uncertainty at the planning stage? We examine how best to apply a sensitivity analysis and Bayesian approaches to explore the uncertainty in your sample size calculations.
Choosing the Effect Size: Various approaches and interpretations exist for how to find the effect size value. We examine those contrasting interpretations and determine the best method and also how to deal with parameterization options.
2. Overcoming Study Design Challenges
(Vaccine Efficacy Example)
The Randomised Controlled Trial (RCT) is considered the gold standard in trial design in drug development. However, there are often practical impediments which mean that adjustments or pragmatic approaches are needed for some trials and studies.
We will examine practical methods how to overcome common study design challenges and how these affect your sample size calculations. In this webinar, we will use common issues in vaccine study design to examine difficulties surrounding issues such as:
Case-Control Analysis: We will examine how to deal with study constraints and how to deal with analyses done during an observational study.
Alternative Randomization Methods: How best to address randomization in your vaccine trial design when full randomization is difficult, expensive or impractical. We examine how sample size calculations are affected with cluster or Mendelian randomization.
Rare Events: How does an outcome being rare affect the types of study design and statistical methods chosen in your study.
ENRICH Trial - Clinial Outcomes for surgical treatment of ICHPSek
The ENRICH (Early MiNimally-invasive Removal of ICH) trial was designed to evaluate minimally invasive parafascicular surgery (MIPS) and ICH removal using the BrainPath® and Myriad® devices versus medical management alone, as defined by the American Heart Association/American Stroke Association guidelines.
Large amounts of heterogeneous medical data have become available in various healthcare organizations (payers, providers, pharmaceuticals). Those data could be an enabling resource for deriving insights for improving care delivery and reducing waste. The enormity and complexity of these datasets present great challenges in analyses and subsequent applications to a practical clinical environment. More details are available here http://dmkd.cs.wayne.edu/TUTORIAL/Healthcare/
Avoid overfitting in precision medicine: How to use cross-validation to relia...Nicole Krämer
The identification of patient subgroups who may derive benefit from a treatment is of crucial importance in precision medicine. Many different algorithms have been proposed and studied in the literature.
We illustrate that many of these algorithms overfit in the sense that the treatment benefit for the identified patients is substantially overestimated. Further, we show that with cross-validation, it is possible to obtain more realistic estimates.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
2. impact of
prediction models
in clinical practice
Example: The IOTA ADNEX model
Multinomial logistic regression model based on age, ca-125, 6 ultrasound characteristics and
center type
Among patients with ovarian masses, refer patients with highest risk of malignancy to
oncological care/surgery
Wynants, L., van Smeden, M., McLernon, D.J. et al. Three myths about risk thresholds for prediction models. BMC Med 17, 192 (2019). 2
3. A link between costs,
benefits and risk
threshold to intervene
The risk threshold can be chosen to minimize the expected total costs. For a (calibrated) risk model:
𝑡 =
𝐶𝐹𝑃 − 𝐶𝑇𝑁
𝐶𝐹𝑃 + 𝐶𝐹𝑁 − 𝐶𝑇𝑃 − 𝐶𝑇𝑁
Assuming 𝐶𝑇𝑁 = 0 and recognizing the benefit of intervening when needed 𝐵𝑇𝑃 = 𝐶𝐹𝑁 − 𝐶𝑇𝑃:
𝑡 =
𝐶𝐹𝑃
𝐶𝐹𝑃 + 𝐵𝑇𝑃
And
𝑡
1−𝑡
=
𝐶𝐹𝑃
𝐵𝑇𝑃
Wynants, L., van Smeden, M., McLernon, D.J. et al. Three myths about risk thresholds for prediction models. BMC Med 17, 192 (2019). 3
4. Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Making. 2006;26:565–74.
Van Calster B, Valentin L, Froyman W, Landolfo C, Ceusters J, Testa A C et al. Validation of models to diagnose ovarian cancer in patients managed surgically or
conservatively: multicentre cohort study BMJ 2020; 370 :m2614 doi:10.1136/bmj.m2614
5
Net Benefit
5. How certain are you?
Kerr KF, Marsh TL, Janes H. The Importance of Uncertainty and Opt-In v. Opt-Out: Best Practices for Decision Curve Analysis. Med Decis Making. 2019 Jul;39(5):491-
492.
6
“ In contemplating a change from a default policy
(treat-all or treat-none) to a policy based on
estimated risk, policymakers should know the
strength of the evidence in favor of the policy change
– they need some quantification of uncertainty such
as confidence intervals”
Kerr, Marsh, Janes 2019
6. 7
N = 4905, 1039 events
N = 250, 53 events
(average external validation sample size)
7. CI around NBmodel
Pro
• Bootstrap, closed form asymptotic methods, and Bayesian methods available
for pointwise CI or confidence bands for NBmodel (Zhang 2018, Marsh 2020, Pfeiffer 2020,
Sande 2020, Cruz 2023)
• Reflects uncertainty when treat none is the only competing strategy
Con
• Does not reflect uncertainty if treat all or other models/tests are (among the)
competing strategies
Zhang Z et al Transl Med. 2018;6:308; Marsh et al Biometrics. 2020;76:843–52; Pfeiffer & Gail Biom J. 2020;62:764–76; Sande SZ et al Stat Med. 2020;39:2980–3002;
Cruz & Korthauer arxiv 2023.
8
9. 95% CI difference NB with all strategies
Van Calster B, Valentin L, Froyman W, Landolfo C, Ceusters J, Testa A C et al. Validation of models to diagnose ovarian cancer in patients managed surgically or
conservatively: multicentre cohort study BMJ 2020; 370 :m2614 doi:10.1136/bmj.m2614
https://github.com/mdbrown/rmda
12
t=1%, n=250
0.206 (95%CI 0.153 to 0.254)
-0.003 (95% CI -0.011 to 0.002)
t=10%, n=250
Diff treat none 0.192 (95% CI 0.140 to 0.245)
Diff treat all 0.066 (95% CI 0.052 to 0.078)
10. Statistical test for ΔNB=0?
• Would delay the introduction of reliable models, and lead to patient harm
• Traditional decision theory dictates that we choose the model with the highest NB,
irrespective of statistical significance
• Would make research unfeasible
• Sample sizes in the millions sometimes needed demonstrate a statistically significant
difference between two models
• Only valid for the population the data was gathered in
• Too much uncertainty
• If risk-neutral: use model but continue validation, revise strategy later if needed
• If not risk-neutral: depending on (de-)implementation costs, wait to replace well-
established practice until uncertainty is “sufficiently” reduced -> context-specific α
Vickers AJ, Van Calster B, Wynants L, Steyerberg EW. Decision curve analysis: confidence intervals and hypothesis testing for net benefit. Diagn Progn Res. 2023 Jun
6;7(1):11.; Claxton K. The irrelevance of inference: a decision-making approach to the stochastic evaluation of health care technologies. J Health Econ. 1999;18:341–
64.
14
11. 95% CI for difference in NB
Pro
• Shows uncertainty due to limited sample size
• Bootstrap, closed form asymptotic and Bayesian methods available (Vickers 2008, Zhang
2018, Marsh 2020, Pfeiffer 2020, Sande 2020, Cruz 2023)
Con
• No yes/no decision (justify your alpha)
• Only valid in the population the data came from
Vickers AJ et al BMC Med Inform Decis Mak. 2008;8:53.; Zhang Z et al Transl Med. 2018;6:308; Marsh et al Biometrics. 2020;76:843–52; Pfeiffer & Gail Biom J.
2020;62:764–76; Sande SZ et al Stat Med. 2020;39:2980–3002; Cruz & Korhauer arxiv 2023.
15
13. P(Useful)
• Trivariate meta-analysis model
Wynants L, Riley RD, Timmerman D, Van Calster B. Random-effects meta-analysis of the clinical utility of tests and prediction models. Stat Med. 2018;37:2034–52.
ERRATUM https://doi.org/10.1002/sim.9597
17
Sample logit prevalence, logit
sensitivity, logit specificity for a new
center
Calculate NB of the model the
competing strategies in the new center
14. P(Useful)
Systematic review ADNEX
37 studies reporting sens/spec/prev
P(useful)=0.95
BARREÑADA L, LEDGER A, DHIMAN P, et al. The ADNEX risk prediction model for ovarian cancer diagnosis: A systematic review
and meta-analysis of external validation studies. medRxiv. 2023:2023.07.12.23291935. doi:10.1101/2023.07.12.23291935 19
15. P(Useful)
Wynants L, Riley RD, Timmerman D, Van Calster B. Random-effects meta-analysis of the clinical utility of tests and prediction models. Stat Med. 2018;37:2034–52. 20
2403 patients from 18 centers 2320 patients from 11 studies
16. P(Useful)
Pro
• Bayesian and bootstrap methods available for meta-analysis and single validation (Wynants 2018,
Sadatsafavi 2023, Cruz 2023)
• Meta-analysis recognizes performance (and NB) varies between populations
Con
• No yes/no decision
• Disadvantage: harm induced in specific centers may be very small or even negligible
Wynants et al Stat Med. 2018;37:2034–52.; ERRATUM https://doi.org/10.1002/sim.9597; Sadatsafavi et al Med. Decis. Making. 2023;43(5):564-75; Cruz & Korthauer
2023 doi:10.48550/arXiv.2308.02067.
21
17. EVPI
• Do not just look at the probability that using the model would lead to harm
• Also look at how much harm it would do if the model is not the optimal strategy
• value of information metric to quantify the expected return on investment in research
Wilson, E.C.F. A Practical Guide to Value of Information Analysis. PharmacoEconomics 33, 105–121 (2015). https://doi.org/10.1007/s40273-014-0219-x 22
18. Pick a box
Thanks to Mohsen Sadatsafavi for this example 23
EVPI = expected reward under perfect info - expected reward under current info
EVPI = 62,5 € - 50 €
= 12,5€
Box A: 0€ or 50€ Box B: 0€ or
100€
P
0 0 0,25
0 100 0,25
50 0 0,25
50 100 0,25
E(reward)=25 E(reward)=50
19. What is the maximum bribe you should pay?
24
EVPI = expected reward under perfect info - expected reward under current info
EVPI = 62,5 € - 50 €
= 12,5€
This is the expected gain of having perfect information
Box A: 0€ or 50€ Box B: 0€ or
100€
P Decision if you had
perfect info
Reward if you had
perfect info
0 0 0,25 A or B 0
0 100 0,25 B 100
50 0 0,25 A 50
50 100 0,25 B 100
E(reward)=25 E(reward)=50 E(reward)=62,5
20. Opportunity loss
25
EVPI = expected reward under perfect info - expected reward under current info
EVPI = 62,5 € - 50 €
= 12,5€
Box A: 0€ or 50€ Box B: 0€ or
100€
P Decision if you had
perfect info
Reward if you had
perfect info
B-A Loss under current
info
0 0 0,25 A or B 0 0 0
0 100 0,25 B 100 100 0
50 0 0,25 A 50 -50 50
50 100 0,25 B 100 50 0
E(reward)=25 E(reward)=50 E(reward)=62,5 E(difference)=25 E(loss)= 12,5
21. EVPI
Wilson, E.C.F. A Practical Guide to Value of Information Analysis. PharmacoEconomics 33, 105–121 (2015). https://doi.org/10.1007/s40273-014-0219-x 26
ΔNB
Loss
(net
true
positives/100)
-0,15 -0,10 -0,05 0 0,05 0,10 0,15
-
0,05
0,10
0,15
22. EVPI external validation
Sadatsafavi M, Lee TY, Wynants L, Vickers AJ, Gustafson P. Value-of-Information Analysis for External Validation of Risk Prediction Models. Med. Decis. Making.
2023;43(5):564-75. doi:10.1177/0272989X231178317
28
23. EVPI
29
t=1%, n=250
0.206 (95%CI 0.153 to 0.254)
-0.003 (95% CI -0.011 to 0.002)
0.39
0.0006
Scaled to EU population:
- a gain of 210 correctly detected cancers/year
- equivalent to a gain of 1890 avoided false referrals/year
t=10%, n=250
Diff treat none 0.192 (95% CI 0.140 to 0.245)
Diff treat all 0.066 (95% CI 0.052 to 0.078)
P(useful) 1
EVPI 0
24. EVPI
30
Cruz 2023
Cruz GNF, Korthauer K. Bayesian Decision Curve Analysis with bayesDCA. 2023 doi:10.48550/arXiv.2308.02067.
25. EVPI – extension to heterogeneous data
31
N centers 37
Total N 9989
NB model (t=0.1) 0.28
(Cr I 0.26 to 0.33, Pr I 0.09 to 0.67)
NB TA 0.25
(Cr I 0.18 to 0.29, Pr -0.01 to 0.63)
Prob Useful 0.96
Voi perfect info population level 0
Voi perfect center level info * 0.00038
Voi partial perfect center level info 0.00003
Voi perfect info in the first center 0.0821
*NB perfect center-level info – NB current info
E𝜓E𝜃|𝜓max 𝑁𝐵𝑚𝑜𝑑𝑒𝑙(𝜃), 𝑁𝐵𝑎𝑙𝑙(𝜃) , 0 − max(E𝜓E𝜃|𝜓𝑁𝐵𝑚𝑜𝑑𝑒𝑙 𝜃 , E𝜓E𝜃|𝜓𝑁𝐵𝑎𝑙𝑙(𝜃), 0)
26. EVPI
Pro
Bayesian, bootstrap and closed-form asymptotic methods available (Sadatsafavi 2023, Cruz
2023)
Yes/no answer (need to decide how many € you are willing to pay for 1 net TP)
Con
Interpretation (net number of TPs, expected loss instead of expected difference)
Perfect information may be utopic (EVSI/ expected net gain of sampling)
Sadatsafavi M, Lee TY, Wynants L, Vickers AJ, Gustafson P. Value-of-Information Analysis for External Validation of Risk Prediction Models. Med. Decis. Making.
2023;43(5):564-75. doi:10.1177/0272989X231178317; Cruz GNF, Korthauer K. Bayesian Decision Curve Analysis with bayesDCA. 2023 doi:10.48550/arXiv.2308.02067. 32
27. Conclusion
Limited sample sizes and between-center heterogeneity are driving
uncertainty re. the optimal strategy to use in a DCA
Quantifying the degree of uncertainty is relevant for understanding the value
of additional external validation, but beware of mindless NHST
CIs around ΔNB, P(Useful) and EVPI quantify the degree of uncertainty
In practice, many considerations may play a role in the decision (not) to
introduce a model, e.g. lack of face validity, no practical implementation in
workflow, …
33
28. References
• Wynants L, Riley RD, Timmerman D, Van Calster B. Random-effects meta-analysis of the clinical
utility of tests and prediction models. Stat Med. 2018;37:2034–52.
• Vickers AJ, Van Calster B, Wynants L, Steyerberg EW. Decision curve analysis: confidence intervals
and hypothesis testing for net benefit. Diagn Progn Res. 2023 Jun 6;7(1):11.
• Sadatsafavi M, Lee TY, Wynants L, Vickers AJ, Gustafson P. Value-of-Information Analysis for External
Validation of Risk Prediction Models. Med. Decis. Making. 2023;43(5):564-75.
Laure.wynants@maastrichtuniversity.nl
34