SlideShare a Scribd company logo
1 of 41
(c) Stephen Senn 2016 1
On being Bayesian
Namur 13 October 2016
Stephen Senn
Acknowledgements
Thank you for the kind invitation
This work is partly supported by the European Union’s 7th
Framework Programme for research, technological
development and demonstration under grant agreement
no. 602552. “IDEAL”
The work on historical placebos is joint with Olivier
Collignon and Anna Schritz in my group
(c) Stephen Senn 2
2(c) Stephen Senn 2016
Basic Thesis
• The Bayesian approach holds out the promise of
providing a principled way of synthesizing
difference sources of information
• This is, however, more difficult than many
suppose
• Key tasks are
– Appropriate formulation of prior distributions
– Establishing exactly what the objective content of such prior
distributions is
– Understanding what a prior distribution commits you to believe
– Developing insight (mathematics is not enough)
(c) Stephen Senn 2016 3
Outline
• Brief, basic reminder as to how it works
– Illustrated using a simple example
• What prior distributions have to reflect
• Some examples to check understanding
– A simple binary outcome
– Dawid’s selection paradox
– Historical placebos
• Some advice
(c) Stephen Senn 2016 4
Key features of Bayesian inference
1. Probability is given a personal and subjective
interpretation
2. Probabilities do not have to be defined in terms of
(theoretical) infinite repetitions
3. Probability statements about parameters and
predictions become the goal of inference
4. There is nothing inherently special about a defined set-
up for collecting data
5. To be fully Bayesian utilities should be considered also
(c) Stephen Senn 2016 5
(c) Stephen Senn 2016 6
An Example
My compact disc (CD) player* allowed me to
press tracks in sequential order by pressing
play or in random order by playing shuffle.
One day I was playing the CD Hysteria by Def Leppard. This CD has 12
tracks.
I thought that I had pressed the shuffle button but the first track played
was ‘women’, which is the first track on the CD.
Q. What is the probability that I did, in fact, press the shuffle button as
intended?
*I now have an Ipod nano
(c) Stephen Senn 2016 7
A Bayesian Solution
We have two basic hypotheses:
1) I pressed shuffle.
2) I pressed play.
First we have to establish a so-called prior probability for these
hypotheses: a probability before seeing the evidence.
Suppose that the probability that I press the shuffle button when I
mean to press the shuffle button is 9/10. The probability of
making a mistake and pressing the play button is then 1/10.
(c) Stephen Senn 2016 8
Next we establish probabilities of events given theories. These
particular sorts of probabilities are referred to as likelihoods ,a
term due to RA Fisher(1890-1962).
If I pressed shuffle, then the probability that the first track will be
‘women’ (W) is 1/12. If I pressed play, then the probability that
the first track is W is 1.
For completeness (although it is not necessary for the solution)
we consider the likelihoods had any other track apart from
‘women’ (say X) been played.
If I pressed shuffle then the probability of X is 11/12. If I pressed
play then this probability is 0.
We can put this together as follows
Hypothesis Prior
Probability
P
Evidence Likelihood P x L
Shuffle 9/10 W 1/12 9/120
Shuffle 9/10 X 11/12 99/120
Play 1/10 W 1 12/120
Play 1/10 X 0 0
TOTAL 120/120 = 1
(c) Stephen Senn 2016 9
After seeing (hearing) the evidence, however,
only two rows remain
Hypothesis Prior Probability
P
Evidence Likelihood P x L
Shuffle 9/10 W 1/12 9/120
Shuffle 9/10 X 11/12 99/120
Play 1/10 W 1 12/120
Play 1/10 X 0 0
TOTAL 21/120
(c) Stephen Senn 2016 10
(c) Stephen Senn 2016 11
The probabilities of the two cases which remain do not add up
to 1.
However, since these two cases cover all the possibilities which
remain, their combined probability must be 1.
Therefore we rescale the individual probabilities to make them
add to 1.
We can do this without changing their relative value by dividing
by their total, 21/120.
This has been done in the table below.
So we rescale by dividing by the total
probability
Hypothesis Prior
Probability
P
Evidence Likelihood P x L Posterior
Probability
Shuffle 9/10 W 1/12 9/120 (9/120)/(21/120)
=9/21
Shuffle 9/10 X 11/12 99/120
Play 1/10 W 1 12/120 (12/120)/(21/120)
=12/21
Play 1/10 X 0 0
TOTAL 21/120 21/21=1
(c) Stephen Senn 2016 12
The probability I pressed shuffle is 9/21
This completes the Bayesian solution
Characteristics of prior
distributions
• They must be what you would use to bet on in
advance of getting any further data
• No amount of further data in any form should be
capable of causing you to revise your prior
distribution qua prior
– Updating your prior distribution to become a posterior is quite
another matter
– Remember that to the extent defined by the model the prior
distribution and the data are exchangeable
– Wanting to change your prior is like wanting to change some
data
(c) Stephen Senn 2016 13
An example to get you started
• You are proposing to estimate the probability 
of a binary event
– E.g. cure/no cure
• You use a uniform prior on 
• You now proceed to study 10,000 occurrences
• Which does your prior distribution say is more
likely?
– 10,000 successes
– 5,000 successes 5,000 failures, in any order
(c) Stephen Senn 2016 14
15
Case Ascertainment
• One of the things we learn in statistics is
that it matters how we ascertain cases
• The selection procedure affects our
inferences
• We react differently if we learn that the
results we are being shown are from a
treatment that was one of many if it was
chosen randomly or it was chosen as the
best observed
(c) Stephen Senn 2016
16
A Selection Paradox of Dawid’s
• Suppose that we estimate treatment means from
a number of treatments in clinical research
• We use a standard conjugate prior
• Since Bayesian analysis is full conditioned on
the data, then for any treatment the posterior
mean will not depend on why we have chosen
the treatment
– At random
– Because it gave the largest response
See DAWID, A. P. (1994), in Multivariate Analysis and its Applications, eds. T. W.
Anderson, K. a.-t. a. Fang, & I. Olkin
(c) Stephen Senn 2016
A Simulation to Illustrate This
• Simulate 10 true means
• For each true mean
simulate observed value
• Now select in one of two
ways
1. Randomly choose one
member from each group
of 10
2. Choose the member with
the highest observed
mean
(c) Stephen Senn 2016 17
2
2
prior mean, 0
prior variance, 1.0
data variance, 4.0
cluster size, 10
number of simulations = 500
= true mean
= data mean
ˆ posterior meanBayes
m
m










What the Bayesian theory says
(c) Stephen Senn 2016 18
   
2 2
2 2
2 2
2 2
1 1
ˆ
1 1
ˆ
1
ˆ 0 0 0.2
1 4
Bayes
Bayes
Bayes
m
m
m m

 

 
  

 

   
     
   

  


   

What the simulation shows
(c) Stephen Senn 2016 19
(c) Stephen Senn 2016 20
Theory says 0.2
Does this mean the
frequentist intuition is wrong?
• Not necessarily
• One needs to think carefully about what
the prior distribution implies
• Actually, even if the prior variance were
large the prior distribution would be very
informative about two things
– Normality
– Conditional independence
(c) Stephen Senn 2016 21
(c) Stephen Senn 2016 22
The explanation of the
paradox
• Having a Normal prior is equivalent to having
seen thousands of true means
• Furthermore, a priori, the true mean of any value
in your sample of ten is exchangeable with any
one of these thousands of means
• Why should the fact that it is locally the highest
have any effect on your Bayesian calibration?
• Now let us see what happened when we no
longer make the means exchangeable
(c) Stephen Senn 2016 23
A hierarchical simulation
• Simulate cluster mean
• Then simulate for cluster
members
• Simulation run two
ways
1. Randomly choose
one member from
each group of 10
2. Choose the member
with the highest
observed mean
(c) Stephen Senn 2016 24
2
2
2
prior mean of all clusters, 0
prior variance of cluster means, 0.5
within cluster variance, 0.5
data variance, 4.0
cluster size, 10
number of simulations = 500
= true mean
= data mean
ˆ posterior meanBayes
m
m












What the simulation shows
(c) Stephen Senn 2016 25
(c) Stephen Senn 2016 26
The regression equations are
now quite different depending on
how the means were chosen
Lessons
• As soon as you replace the conjugate prior
with a hierarchical one you get very
different results according to selection
• Be very careful to establish what your prior
implies
• True Bayesian inference does not
necessarily give you the license to ignore
frequentist lessons you might think
(c) Stephen Senn 2016 27
Historical control
• In many indications, the same treatment is often
used as a control
– Either a placebo
– Or a standard treatment
• This means that when a new treatment is trialled
there will be a lot of information from previous
trials on the control being used
• Since Bayes is supposed to be a way of
synthesizing all information, how would we do
this?
(c) Stephen Senn 2016 28
Problem
• Obviously a historical control is not worth
the same as a concurrent control
• How should we deal with this?
• Ask the following question
• Given a choice between an infinite number
of historical controls and n concurrent
controls how large does n have to be
before I prefer the latter?
(c) Stephen Senn 2016 29
Model (frequentist formulation)
(c) Stephen Senn 2016 30
 
 
 
   
2
2
2
2 2 2
2
2 2 2
2
2 *
ic*
,
,
,
1 1 1 1
, 2 ,
1 1
2 2
Hence, 2 , where n is number of concurrent controls
you would p
jc
i
ic i
it i
it ic it jc
it ic it jc
it jc it
n
ic
N
Y N
Y N
Var Y Y Var Y Y i j
n n n n
n n n
n
 
 
  
  

  





  
           
   
 
     
 

:
:
:
refer to infinitely many historical ones
But….
• When you start thinking like this you begin
to wonder
• Is it really the number of historical control
patients that I have that is important?
• Or should I really be thinking about the
data in some other way?
• What do the data really represent?
(c) Stephen Senn 2016 31
Sacred cow
The TARGET study
• One of the largest studies ever run in osteoarthritis
• 18,000 patients
• Randomisation took place in two sub-studies of equal size
o Lumiracoxib versus ibuprofen
o Lumiracoxib versus naproxen
• Purpose to investigate cardiovascular and gastric tolerability of
lumiracoxib
o That is to say side-effects on the heart and the stomach
(c) Stephen Senn 2012 32
(c) Stephen Senn 2016
Baseline Demographics
Sub-Study 1 Sub Study 2
Demographic
Characteristic
Lumiracoxib
n = 4376
Ibuprofen
n = 4397
Lumiracoxib
n = 4741
Naproxen
n = 4730
Use of low-dose
aspirin
975 (22.3) 966 (22.0) 1195 (25.1) 1193 (25.2)
History of vascular
disease
393 (9.0) 340 (7.7) 588 (12.4) 559 (11.8)
Cerebro-vascular
disease
69 (1.6) 65 (1.5) 108 (2.3) 107 (2.3)
Dyslipidaemias 1030 (23.5) 1025 (23.3) 799 (16.9) 809 (17.1)
Nitrate use 105 (2.4) 79 (1.8) 181 (3.8) 165 (3.5)
33(c) Stephen Senn 2016
Baseline Chi-square P-values
Model Term
Demographic
Characteristic
Sub-study
(DF=1)
Treatment given Sub-
study
(DF=2)
Treatment
(DF=2)
Use of low-dose
aspirin
< 0.0001 0.94 0.0012
History of vascular
disease
< 0.0001 0.07 <0.0001
Cerebro-vascular
disease
0.0002 0.93 0.0208
Dyslipidaemias <0.0001 0.92 <0.0001
Nitrate use < 0.0001 0.10 <0.0001
34(c) Stephen Senn 2016
Outcome Variables
All four groups
Sub-Study 1 Sub Study 2
Outcome
Variables
Lumiracoxib
n = 4376
Ibuprofen
n = 4397
Lumiracoxib
n = 4741
Naproxen
n = 4730
Total of
discontinuations
1751
(40.01)
1941
(44.14)
1719
(36.26)
1790
(37.84)
CV events 33
(0.75)
32
(0.73)
52
(1.10)
43
(0.91)
At least one AE 699
(15.97)
789
(17.94)
710
(14.98)
846
(17.89)
Any GI 1855
(42.39)
1851
( 42.10)
1785
(37.65)
1988
(21.87)
Dyspepsia 1230
(28.11)
1205
(27.41)
1037
(21.87)
1119
(23.66)
35(c) Stephen Senn 2016
Deviances and P-Values
Lumiracoxib only fitting Sub-study
Statistic
Outcome
Variables
Deviance P-Value
Total of
discontinuations
13.61 0.0002
CV events 2.92 0.09
At least one AE 1.73 0.19
Any GI 21.31 <0.0001
Dyspepsia 47.34 < 0.0001
36(c) Stephen Senn 2016
A Simple Model
An unrealistic balanced trial
𝑛 patients per arm, 𝑐 centres in total with 𝑝 patients per centre
2𝑛 = 𝑝𝑐, 𝑛 =
𝑝𝑐
2
Between-centres variance is 2 within-centre variance is 2.
Design Variance of Treatment Contrast
Completely randomised
4
𝛾2+𝜎2
𝑐𝑝
Randomised blocks (centre blocks) 4
𝜎2
𝑐𝑝
Cluster randomised
4
𝛾2+
𝜎2
𝑝
𝑐
When using external controls we have at least the variability of a cluster randomised trial
37
(c) Stephen Senn 2016
Lessons from TARGET
• If you want to use historical controls you will have to work very hard
• You need at least two components of variation in your model
o Between centre
o Between trial
• And possibly a third
o Between eras
• What seems like a lot of information may not be much
• Concurrent control and randomisation seems to work well
• Moral for any Bayesian: find out as much as possible about any data
you intend to use
38
(c) Stephen Senn 2016
That example revisited
The question
• You are proposing to estimate
the probability  of a binary
event
oE.g. cure/no cure
• You use a uniform prior on 
• You now proceed to study
10,000 occurrences
• Which does your prior
distribution say is more likely?
o10,000 successes
o5,000 successes 5,000 failures,
in any order
The solution
• You started with an
‘uninformative prior
• After 10,000 trials the observed
proportion must be pretty much
what you believe is the true
probability
• But you said every true
probability is equally likely
• Therefore 5,000 success in any
order is just as likely as 10,000
success
(c) Stephen Senn 2016 39
Advice
• Think hard about any prior distribution
• Try to establish the objective content of any prior distribution
• Uninformative prior distributions are not appropriate for nuisance
parameters
• Be prepared to think hierarchically
• Check that
o The prior distribution states your current belief
o No data in any shape of form would cause you to abandon it
• If the result seems to contradict frequentist wisdom think carefully
why
• Develop statistical insight – understand what being Bayesian means
• As in any statistically system, ask yourself the question
o How did I get to see what I see?
40(c) Stephen Senn 2016
(c) Stephen Senn 2016 41
References
1 Senn, S. J. Bayesian, likelihood and frequentist approaches to
statistics. Applied Clinical Trials 12, 35-38 (2003).
2 Senn, S. J. Trying to be precise about vagueness. Statistics in
Medicine 26, 1417-1430 (2007).
3 Senn, S. A note concerning a selection "Paradox" of Dawid's.
American Statistician 62, 206-210, doi:10.1198/000313008x331530 (2008).
4 Senn, S. J. Comment on article by Gelman. Bayesian analysis 3,
459-462 (2008).
5 Senn, S. J. Comment on "Harold Jeffreys's Theory of Probability
Revisited". Statistical Science 24, 185-186 (2009).
6 Senn, S. J. You may believe you are a Bayesian but you are probably
wrong. Rationality, Markets and Morals 2, 48-66 (2011).

More Related Content

What's hot

De Finetti meets Popper
De Finetti meets PopperDe Finetti meets Popper
De Finetti meets PopperStephen Senn
 
P values and the art of herding cats
P values  and the art of herding catsP values  and the art of herding cats
P values and the art of herding catsStephen Senn
 
Personalised medicine a sceptical view
Personalised medicine a sceptical viewPersonalised medicine a sceptical view
Personalised medicine a sceptical viewStephen Senn
 
The Seven Habits of Highly Effective Statisticians
The Seven Habits of Highly Effective StatisticiansThe Seven Habits of Highly Effective Statisticians
The Seven Habits of Highly Effective StatisticiansStephen Senn
 
Seven myths of randomisation
Seven myths of randomisation Seven myths of randomisation
Seven myths of randomisation Stephen Senn
 
In search of the lost loss function
In search of the lost loss function In search of the lost loss function
In search of the lost loss function Stephen Senn
 
The Rothamsted school meets Lord's paradox
The Rothamsted school meets Lord's paradoxThe Rothamsted school meets Lord's paradox
The Rothamsted school meets Lord's paradoxStephen Senn
 
Seventy years of RCTs
Seventy years of RCTsSeventy years of RCTs
Seventy years of RCTsStephen Senn
 
Why I hate minimisation
Why I hate minimisationWhy I hate minimisation
Why I hate minimisationStephen Senn
 
The revenge of RA Fisher
The revenge of RA FisherThe revenge of RA Fisher
The revenge of RA FisherStephen Senn
 
The challenge of small data
The challenge of small dataThe challenge of small data
The challenge of small dataStephen Senn
 
Understanding randomisation
Understanding randomisationUnderstanding randomisation
Understanding randomisationStephen Senn
 
To infinity and beyond
To infinity and beyond To infinity and beyond
To infinity and beyond Stephen Senn
 
What is your question
What is your questionWhat is your question
What is your questionStephenSenn2
 
To infinity and beyond v2
To infinity and beyond v2To infinity and beyond v2
To infinity and beyond v2Stephen Senn
 
Is ignorance bliss
Is ignorance blissIs ignorance bliss
Is ignorance blissStephen Senn
 
In Search of Lost Infinities: What is the “n” in big data?
In Search of Lost Infinities: What is the “n” in big data?In Search of Lost Infinities: What is the “n” in big data?
In Search of Lost Infinities: What is the “n” in big data?Stephen Senn
 
The revenge of RA Fisher
The revenge of RA Fisher The revenge of RA Fisher
The revenge of RA Fisher Stephen Senn
 
Take it to the Limit: quantitation, likelihood, modelling and other matters
Take it to the Limit: quantitation, likelihood, modelling and other mattersTake it to the Limit: quantitation, likelihood, modelling and other matters
Take it to the Limit: quantitation, likelihood, modelling and other mattersStephen Senn
 

What's hot (20)

De Finetti meets Popper
De Finetti meets PopperDe Finetti meets Popper
De Finetti meets Popper
 
P values and the art of herding cats
P values  and the art of herding catsP values  and the art of herding cats
P values and the art of herding cats
 
Yates and cochran
Yates and cochranYates and cochran
Yates and cochran
 
Personalised medicine a sceptical view
Personalised medicine a sceptical viewPersonalised medicine a sceptical view
Personalised medicine a sceptical view
 
The Seven Habits of Highly Effective Statisticians
The Seven Habits of Highly Effective StatisticiansThe Seven Habits of Highly Effective Statisticians
The Seven Habits of Highly Effective Statisticians
 
Seven myths of randomisation
Seven myths of randomisation Seven myths of randomisation
Seven myths of randomisation
 
In search of the lost loss function
In search of the lost loss function In search of the lost loss function
In search of the lost loss function
 
The Rothamsted school meets Lord's paradox
The Rothamsted school meets Lord's paradoxThe Rothamsted school meets Lord's paradox
The Rothamsted school meets Lord's paradox
 
Seventy years of RCTs
Seventy years of RCTsSeventy years of RCTs
Seventy years of RCTs
 
Why I hate minimisation
Why I hate minimisationWhy I hate minimisation
Why I hate minimisation
 
The revenge of RA Fisher
The revenge of RA FisherThe revenge of RA Fisher
The revenge of RA Fisher
 
The challenge of small data
The challenge of small dataThe challenge of small data
The challenge of small data
 
Understanding randomisation
Understanding randomisationUnderstanding randomisation
Understanding randomisation
 
To infinity and beyond
To infinity and beyond To infinity and beyond
To infinity and beyond
 
What is your question
What is your questionWhat is your question
What is your question
 
To infinity and beyond v2
To infinity and beyond v2To infinity and beyond v2
To infinity and beyond v2
 
Is ignorance bliss
Is ignorance blissIs ignorance bliss
Is ignorance bliss
 
In Search of Lost Infinities: What is the “n” in big data?
In Search of Lost Infinities: What is the “n” in big data?In Search of Lost Infinities: What is the “n” in big data?
In Search of Lost Infinities: What is the “n” in big data?
 
The revenge of RA Fisher
The revenge of RA Fisher The revenge of RA Fisher
The revenge of RA Fisher
 
Take it to the Limit: quantitation, likelihood, modelling and other matters
Take it to the Limit: quantitation, likelihood, modelling and other mattersTake it to the Limit: quantitation, likelihood, modelling and other matters
Take it to the Limit: quantitation, likelihood, modelling and other matters
 

Viewers also liked

Modeling and Mining Sequential Data
Modeling and Mining Sequential DataModeling and Mining Sequential Data
Modeling and Mining Sequential DataPhilipp Singer
 
Introduction to Bayesian Statistics
Introduction to Bayesian StatisticsIntroduction to Bayesian Statistics
Introduction to Bayesian StatisticsPhilipp Singer
 
Bayesian statistics using r intro
Bayesian statistics using r   introBayesian statistics using r   intro
Bayesian statistics using r introBayesLaplace1
 
Placebos in medical research
Placebos in medical researchPlacebos in medical research
Placebos in medical researchStephen Senn
 
An introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using PythonAn introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using Pythonfreshdatabos
 
An introduction to bayesian statistics
An introduction to bayesian statisticsAn introduction to bayesian statistics
An introduction to bayesian statisticsJohn Tyndall
 
Introduction to Bayesian Methods
Introduction to Bayesian MethodsIntroduction to Bayesian Methods
Introduction to Bayesian MethodsCorey Chivers
 

Viewers also liked (9)

Modeling and Mining Sequential Data
Modeling and Mining Sequential DataModeling and Mining Sequential Data
Modeling and Mining Sequential Data
 
Lecture 2
Lecture 2Lecture 2
Lecture 2
 
Bayesian intro
Bayesian introBayesian intro
Bayesian intro
 
Introduction to Bayesian Statistics
Introduction to Bayesian StatisticsIntroduction to Bayesian Statistics
Introduction to Bayesian Statistics
 
Bayesian statistics using r intro
Bayesian statistics using r   introBayesian statistics using r   intro
Bayesian statistics using r intro
 
Placebos in medical research
Placebos in medical researchPlacebos in medical research
Placebos in medical research
 
An introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using PythonAn introduction to Bayesian Statistics using Python
An introduction to Bayesian Statistics using Python
 
An introduction to bayesian statistics
An introduction to bayesian statisticsAn introduction to bayesian statistics
An introduction to bayesian statistics
 
Introduction to Bayesian Methods
Introduction to Bayesian MethodsIntroduction to Bayesian Methods
Introduction to Bayesian Methods
 

Similar to On being Bayesian

Senn repligate
Senn repligateSenn repligate
Senn repligatejemille6
 
Approximate ANCOVA
Approximate ANCOVAApproximate ANCOVA
Approximate ANCOVAStephen Senn
 
P values and replication
P values and replicationP values and replication
P values and replicationStephen Senn
 
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...nszakir
 
Network meta-analysis & models for inconsistency
Network meta-analysis & models for inconsistencyNetwork meta-analysis & models for inconsistency
Network meta-analysis & models for inconsistencycheweb1
 
What is your question
What is your questionWhat is your question
What is your questionStephen Senn
 
The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...StephenSenn2
 
The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...jemille6
 
Clinical trials: three statistical traps for the unwary
Clinical trials: three statistical traps for the unwaryClinical trials: three statistical traps for the unwary
Clinical trials: three statistical traps for the unwaryStephen Senn
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data sciencepujashri1975
 
Probability Distributions.pdf
Probability Distributions.pdfProbability Distributions.pdf
Probability Distributions.pdfShivakumar B N
 
Statistical tests
Statistical testsStatistical tests
Statistical testsmartyynyyte
 
Introduction to Statistics - Part 2
Introduction to Statistics - Part 2Introduction to Statistics - Part 2
Introduction to Statistics - Part 2Damian T. Gordon
 
Introduction to Bayesian Statistics.ppt
Introduction to Bayesian Statistics.pptIntroduction to Bayesian Statistics.ppt
Introduction to Bayesian Statistics.pptLong Dang
 
QNT 561 Week 4 Weekly Learning Assessments
QNT 561 Week 4 Weekly Learning Assessments  QNT 561 Week 4 Weekly Learning Assessments
QNT 561 Week 4 Weekly Learning Assessments student ehelp
 

Similar to On being Bayesian (20)

P value wars
P value warsP value wars
P value wars
 
Senn repligate
Senn repligateSenn repligate
Senn repligate
 
Approximate ANCOVA
Approximate ANCOVAApproximate ANCOVA
Approximate ANCOVA
 
P values and replication
P values and replicationP values and replication
P values and replication
 
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...
Chapter 5 part2- Sampling Distributions for Counts and Proportions (Binomial ...
 
Network meta-analysis & models for inconsistency
Network meta-analysis & models for inconsistencyNetwork meta-analysis & models for inconsistency
Network meta-analysis & models for inconsistency
 
What is your question
What is your questionWhat is your question
What is your question
 
bayesjaw.ppt
bayesjaw.pptbayesjaw.ppt
bayesjaw.ppt
 
The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...
 
The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...The replication crisis: are P-values the problem and are Bayes factors the so...
The replication crisis: are P-values the problem and are Bayes factors the so...
 
Clinical trials: three statistical traps for the unwary
Clinical trials: three statistical traps for the unwaryClinical trials: three statistical traps for the unwary
Clinical trials: three statistical traps for the unwary
 
non para.doc
non para.docnon para.doc
non para.doc
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data science
 
Probability distribution
Probability distributionProbability distribution
Probability distribution
 
Probability Distributions.pdf
Probability Distributions.pdfProbability Distributions.pdf
Probability Distributions.pdf
 
Statistical tests
Statistical testsStatistical tests
Statistical tests
 
Introduction to Statistics - Part 2
Introduction to Statistics - Part 2Introduction to Statistics - Part 2
Introduction to Statistics - Part 2
 
Probability unit2.pptx
Probability unit2.pptxProbability unit2.pptx
Probability unit2.pptx
 
Introduction to Bayesian Statistics.ppt
Introduction to Bayesian Statistics.pptIntroduction to Bayesian Statistics.ppt
Introduction to Bayesian Statistics.ppt
 
QNT 561 Week 4 Weekly Learning Assessments
QNT 561 Week 4 Weekly Learning Assessments  QNT 561 Week 4 Weekly Learning Assessments
QNT 561 Week 4 Weekly Learning Assessments
 

More from Stephen Senn

Has modelling killed randomisation inference frankfurt
Has modelling killed randomisation inference frankfurtHas modelling killed randomisation inference frankfurt
Has modelling killed randomisation inference frankfurtStephen Senn
 
Vaccine trials in the age of COVID-19
Vaccine trials in the age of COVID-19Vaccine trials in the age of COVID-19
Vaccine trials in the age of COVID-19Stephen Senn
 
A century of t tests
A century of t testsA century of t tests
A century of t testsStephen Senn
 
The story of MTA/02
The story of MTA/02The story of MTA/02
The story of MTA/02Stephen Senn
 
Confounding, politics, frustration and knavish tricks
Confounding, politics, frustration and knavish tricksConfounding, politics, frustration and knavish tricks
Confounding, politics, frustration and knavish tricksStephen Senn
 
And thereby hangs a tail
And thereby hangs a tailAnd thereby hangs a tail
And thereby hangs a tailStephen Senn
 
Real world modified
Real world modifiedReal world modified
Real world modifiedStephen Senn
 

More from Stephen Senn (7)

Has modelling killed randomisation inference frankfurt
Has modelling killed randomisation inference frankfurtHas modelling killed randomisation inference frankfurt
Has modelling killed randomisation inference frankfurt
 
Vaccine trials in the age of COVID-19
Vaccine trials in the age of COVID-19Vaccine trials in the age of COVID-19
Vaccine trials in the age of COVID-19
 
A century of t tests
A century of t testsA century of t tests
A century of t tests
 
The story of MTA/02
The story of MTA/02The story of MTA/02
The story of MTA/02
 
Confounding, politics, frustration and knavish tricks
Confounding, politics, frustration and knavish tricksConfounding, politics, frustration and knavish tricks
Confounding, politics, frustration and knavish tricks
 
And thereby hangs a tail
And thereby hangs a tailAnd thereby hangs a tail
And thereby hangs a tail
 
Real world modified
Real world modifiedReal world modified
Real world modified
 

Recently uploaded

Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Nistarini College, Purulia (W.B) India
 
G9 Science Q4- Week 1-2 Projectile Motion.ppt
G9 Science Q4- Week 1-2 Projectile Motion.pptG9 Science Q4- Week 1-2 Projectile Motion.ppt
G9 Science Q4- Week 1-2 Projectile Motion.pptMAESTRELLAMesa2
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Patrick Diehl
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...RohitNehra6
 
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdf
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdfNAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdf
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdfWadeK3
 
Biological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfBiological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfmuntazimhurra
 
Work, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE PhysicsWork, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE Physicsvishikhakeshava1
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxkessiyaTpeter
 
Artificial Intelligence In Microbiology by Dr. Prince C P
Artificial Intelligence In Microbiology by Dr. Prince C PArtificial Intelligence In Microbiology by Dr. Prince C P
Artificial Intelligence In Microbiology by Dr. Prince C PPRINCE C P
 
Orientation, design and principles of polyhouse
Orientation, design and principles of polyhouseOrientation, design and principles of polyhouse
Orientation, design and principles of polyhousejana861314
 
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCESTERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCEPRINCE C P
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxUmerFayaz5
 
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |aasikanpl
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSarthak Sekhar Mondal
 
Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxpradhanghanshyam7136
 
A relative description on Sonoporation.pdf
A relative description on Sonoporation.pdfA relative description on Sonoporation.pdf
A relative description on Sonoporation.pdfnehabiju2046
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)PraveenaKalaiselvan1
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxyaramohamed343013
 

Recently uploaded (20)

Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...Bentham & Hooker's Classification. along with the merits and demerits of the ...
Bentham & Hooker's Classification. along with the merits and demerits of the ...
 
G9 Science Q4- Week 1-2 Projectile Motion.ppt
G9 Science Q4- Week 1-2 Projectile Motion.pptG9 Science Q4- Week 1-2 Projectile Motion.ppt
G9 Science Q4- Week 1-2 Projectile Motion.ppt
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?
 
Biopesticide (2).pptx .This slides helps to know the different types of biop...
Biopesticide (2).pptx  .This slides helps to know the different types of biop...Biopesticide (2).pptx  .This slides helps to know the different types of biop...
Biopesticide (2).pptx .This slides helps to know the different types of biop...
 
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdf
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdfNAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdf
NAVSEA PEO USC - Unmanned & Small Combatants 26Oct23.pdf
 
The Philosophy of Science
The Philosophy of ScienceThe Philosophy of Science
The Philosophy of Science
 
Biological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfBiological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdf
 
Work, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE PhysicsWork, Energy and Power for class 10 ICSE Physics
Work, Energy and Power for class 10 ICSE Physics
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
 
Artificial Intelligence In Microbiology by Dr. Prince C P
Artificial Intelligence In Microbiology by Dr. Prince C PArtificial Intelligence In Microbiology by Dr. Prince C P
Artificial Intelligence In Microbiology by Dr. Prince C P
 
Orientation, design and principles of polyhouse
Orientation, design and principles of polyhouseOrientation, design and principles of polyhouse
Orientation, design and principles of polyhouse
 
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCESTERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
STERILITY TESTING OF PHARMACEUTICALS ppt by DR.C.P.PRINCE
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 
Animal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptxAnimal Communication- Auditory and Visual.pptx
Animal Communication- Auditory and Visual.pptx
 
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
Call Us ≽ 9953322196 ≼ Call Girls In Mukherjee Nagar(Delhi) |
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
 
Cultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptxCultivation of KODO MILLET . made by Ghanshyam pptx
Cultivation of KODO MILLET . made by Ghanshyam pptx
 
A relative description on Sonoporation.pdf
A relative description on Sonoporation.pdfA relative description on Sonoporation.pdf
A relative description on Sonoporation.pdf
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)
 
Scheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docxScheme-of-Work-Science-Stage-4 cambridge science.docx
Scheme-of-Work-Science-Stage-4 cambridge science.docx
 

On being Bayesian

  • 1. (c) Stephen Senn 2016 1 On being Bayesian Namur 13 October 2016 Stephen Senn
  • 2. Acknowledgements Thank you for the kind invitation This work is partly supported by the European Union’s 7th Framework Programme for research, technological development and demonstration under grant agreement no. 602552. “IDEAL” The work on historical placebos is joint with Olivier Collignon and Anna Schritz in my group (c) Stephen Senn 2 2(c) Stephen Senn 2016
  • 3. Basic Thesis • The Bayesian approach holds out the promise of providing a principled way of synthesizing difference sources of information • This is, however, more difficult than many suppose • Key tasks are – Appropriate formulation of prior distributions – Establishing exactly what the objective content of such prior distributions is – Understanding what a prior distribution commits you to believe – Developing insight (mathematics is not enough) (c) Stephen Senn 2016 3
  • 4. Outline • Brief, basic reminder as to how it works – Illustrated using a simple example • What prior distributions have to reflect • Some examples to check understanding – A simple binary outcome – Dawid’s selection paradox – Historical placebos • Some advice (c) Stephen Senn 2016 4
  • 5. Key features of Bayesian inference 1. Probability is given a personal and subjective interpretation 2. Probabilities do not have to be defined in terms of (theoretical) infinite repetitions 3. Probability statements about parameters and predictions become the goal of inference 4. There is nothing inherently special about a defined set- up for collecting data 5. To be fully Bayesian utilities should be considered also (c) Stephen Senn 2016 5
  • 6. (c) Stephen Senn 2016 6 An Example My compact disc (CD) player* allowed me to press tracks in sequential order by pressing play or in random order by playing shuffle. One day I was playing the CD Hysteria by Def Leppard. This CD has 12 tracks. I thought that I had pressed the shuffle button but the first track played was ‘women’, which is the first track on the CD. Q. What is the probability that I did, in fact, press the shuffle button as intended? *I now have an Ipod nano
  • 7. (c) Stephen Senn 2016 7 A Bayesian Solution We have two basic hypotheses: 1) I pressed shuffle. 2) I pressed play. First we have to establish a so-called prior probability for these hypotheses: a probability before seeing the evidence. Suppose that the probability that I press the shuffle button when I mean to press the shuffle button is 9/10. The probability of making a mistake and pressing the play button is then 1/10.
  • 8. (c) Stephen Senn 2016 8 Next we establish probabilities of events given theories. These particular sorts of probabilities are referred to as likelihoods ,a term due to RA Fisher(1890-1962). If I pressed shuffle, then the probability that the first track will be ‘women’ (W) is 1/12. If I pressed play, then the probability that the first track is W is 1. For completeness (although it is not necessary for the solution) we consider the likelihoods had any other track apart from ‘women’ (say X) been played. If I pressed shuffle then the probability of X is 11/12. If I pressed play then this probability is 0.
  • 9. We can put this together as follows Hypothesis Prior Probability P Evidence Likelihood P x L Shuffle 9/10 W 1/12 9/120 Shuffle 9/10 X 11/12 99/120 Play 1/10 W 1 12/120 Play 1/10 X 0 0 TOTAL 120/120 = 1 (c) Stephen Senn 2016 9
  • 10. After seeing (hearing) the evidence, however, only two rows remain Hypothesis Prior Probability P Evidence Likelihood P x L Shuffle 9/10 W 1/12 9/120 Shuffle 9/10 X 11/12 99/120 Play 1/10 W 1 12/120 Play 1/10 X 0 0 TOTAL 21/120 (c) Stephen Senn 2016 10
  • 11. (c) Stephen Senn 2016 11 The probabilities of the two cases which remain do not add up to 1. However, since these two cases cover all the possibilities which remain, their combined probability must be 1. Therefore we rescale the individual probabilities to make them add to 1. We can do this without changing their relative value by dividing by their total, 21/120. This has been done in the table below.
  • 12. So we rescale by dividing by the total probability Hypothesis Prior Probability P Evidence Likelihood P x L Posterior Probability Shuffle 9/10 W 1/12 9/120 (9/120)/(21/120) =9/21 Shuffle 9/10 X 11/12 99/120 Play 1/10 W 1 12/120 (12/120)/(21/120) =12/21 Play 1/10 X 0 0 TOTAL 21/120 21/21=1 (c) Stephen Senn 2016 12 The probability I pressed shuffle is 9/21 This completes the Bayesian solution
  • 13. Characteristics of prior distributions • They must be what you would use to bet on in advance of getting any further data • No amount of further data in any form should be capable of causing you to revise your prior distribution qua prior – Updating your prior distribution to become a posterior is quite another matter – Remember that to the extent defined by the model the prior distribution and the data are exchangeable – Wanting to change your prior is like wanting to change some data (c) Stephen Senn 2016 13
  • 14. An example to get you started • You are proposing to estimate the probability  of a binary event – E.g. cure/no cure • You use a uniform prior on  • You now proceed to study 10,000 occurrences • Which does your prior distribution say is more likely? – 10,000 successes – 5,000 successes 5,000 failures, in any order (c) Stephen Senn 2016 14
  • 15. 15 Case Ascertainment • One of the things we learn in statistics is that it matters how we ascertain cases • The selection procedure affects our inferences • We react differently if we learn that the results we are being shown are from a treatment that was one of many if it was chosen randomly or it was chosen as the best observed (c) Stephen Senn 2016
  • 16. 16 A Selection Paradox of Dawid’s • Suppose that we estimate treatment means from a number of treatments in clinical research • We use a standard conjugate prior • Since Bayesian analysis is full conditioned on the data, then for any treatment the posterior mean will not depend on why we have chosen the treatment – At random – Because it gave the largest response See DAWID, A. P. (1994), in Multivariate Analysis and its Applications, eds. T. W. Anderson, K. a.-t. a. Fang, & I. Olkin (c) Stephen Senn 2016
  • 17. A Simulation to Illustrate This • Simulate 10 true means • For each true mean simulate observed value • Now select in one of two ways 1. Randomly choose one member from each group of 10 2. Choose the member with the highest observed mean (c) Stephen Senn 2016 17 2 2 prior mean, 0 prior variance, 1.0 data variance, 4.0 cluster size, 10 number of simulations = 500 = true mean = data mean ˆ posterior meanBayes m m          
  • 18. What the Bayesian theory says (c) Stephen Senn 2016 18     2 2 2 2 2 2 2 2 1 1 ˆ 1 1 ˆ 1 ˆ 0 0 0.2 1 4 Bayes Bayes Bayes m m m m                                      
  • 19. What the simulation shows (c) Stephen Senn 2016 19
  • 20. (c) Stephen Senn 2016 20 Theory says 0.2
  • 21. Does this mean the frequentist intuition is wrong? • Not necessarily • One needs to think carefully about what the prior distribution implies • Actually, even if the prior variance were large the prior distribution would be very informative about two things – Normality – Conditional independence (c) Stephen Senn 2016 21
  • 22. (c) Stephen Senn 2016 22
  • 23. The explanation of the paradox • Having a Normal prior is equivalent to having seen thousands of true means • Furthermore, a priori, the true mean of any value in your sample of ten is exchangeable with any one of these thousands of means • Why should the fact that it is locally the highest have any effect on your Bayesian calibration? • Now let us see what happened when we no longer make the means exchangeable (c) Stephen Senn 2016 23
  • 24. A hierarchical simulation • Simulate cluster mean • Then simulate for cluster members • Simulation run two ways 1. Randomly choose one member from each group of 10 2. Choose the member with the highest observed mean (c) Stephen Senn 2016 24 2 2 2 prior mean of all clusters, 0 prior variance of cluster means, 0.5 within cluster variance, 0.5 data variance, 4.0 cluster size, 10 number of simulations = 500 = true mean = data mean ˆ posterior meanBayes m m            
  • 25. What the simulation shows (c) Stephen Senn 2016 25
  • 26. (c) Stephen Senn 2016 26 The regression equations are now quite different depending on how the means were chosen
  • 27. Lessons • As soon as you replace the conjugate prior with a hierarchical one you get very different results according to selection • Be very careful to establish what your prior implies • True Bayesian inference does not necessarily give you the license to ignore frequentist lessons you might think (c) Stephen Senn 2016 27
  • 28. Historical control • In many indications, the same treatment is often used as a control – Either a placebo – Or a standard treatment • This means that when a new treatment is trialled there will be a lot of information from previous trials on the control being used • Since Bayes is supposed to be a way of synthesizing all information, how would we do this? (c) Stephen Senn 2016 28
  • 29. Problem • Obviously a historical control is not worth the same as a concurrent control • How should we deal with this? • Ask the following question • Given a choice between an infinite number of historical controls and n concurrent controls how large does n have to be before I prefer the latter? (c) Stephen Senn 2016 29
  • 30. Model (frequentist formulation) (c) Stephen Senn 2016 30           2 2 2 2 2 2 2 2 2 2 2 2 * ic* , , , 1 1 1 1 , 2 , 1 1 2 2 Hence, 2 , where n is number of concurrent controls you would p jc i ic i it i it ic it jc it ic it jc it jc it n ic N Y N Y N Var Y Y Var Y Y i j n n n n n n n n                                                  : : : refer to infinitely many historical ones
  • 31. But…. • When you start thinking like this you begin to wonder • Is it really the number of historical control patients that I have that is important? • Or should I really be thinking about the data in some other way? • What do the data really represent? (c) Stephen Senn 2016 31
  • 32. Sacred cow The TARGET study • One of the largest studies ever run in osteoarthritis • 18,000 patients • Randomisation took place in two sub-studies of equal size o Lumiracoxib versus ibuprofen o Lumiracoxib versus naproxen • Purpose to investigate cardiovascular and gastric tolerability of lumiracoxib o That is to say side-effects on the heart and the stomach (c) Stephen Senn 2012 32 (c) Stephen Senn 2016
  • 33. Baseline Demographics Sub-Study 1 Sub Study 2 Demographic Characteristic Lumiracoxib n = 4376 Ibuprofen n = 4397 Lumiracoxib n = 4741 Naproxen n = 4730 Use of low-dose aspirin 975 (22.3) 966 (22.0) 1195 (25.1) 1193 (25.2) History of vascular disease 393 (9.0) 340 (7.7) 588 (12.4) 559 (11.8) Cerebro-vascular disease 69 (1.6) 65 (1.5) 108 (2.3) 107 (2.3) Dyslipidaemias 1030 (23.5) 1025 (23.3) 799 (16.9) 809 (17.1) Nitrate use 105 (2.4) 79 (1.8) 181 (3.8) 165 (3.5) 33(c) Stephen Senn 2016
  • 34. Baseline Chi-square P-values Model Term Demographic Characteristic Sub-study (DF=1) Treatment given Sub- study (DF=2) Treatment (DF=2) Use of low-dose aspirin < 0.0001 0.94 0.0012 History of vascular disease < 0.0001 0.07 <0.0001 Cerebro-vascular disease 0.0002 0.93 0.0208 Dyslipidaemias <0.0001 0.92 <0.0001 Nitrate use < 0.0001 0.10 <0.0001 34(c) Stephen Senn 2016
  • 35. Outcome Variables All four groups Sub-Study 1 Sub Study 2 Outcome Variables Lumiracoxib n = 4376 Ibuprofen n = 4397 Lumiracoxib n = 4741 Naproxen n = 4730 Total of discontinuations 1751 (40.01) 1941 (44.14) 1719 (36.26) 1790 (37.84) CV events 33 (0.75) 32 (0.73) 52 (1.10) 43 (0.91) At least one AE 699 (15.97) 789 (17.94) 710 (14.98) 846 (17.89) Any GI 1855 (42.39) 1851 ( 42.10) 1785 (37.65) 1988 (21.87) Dyspepsia 1230 (28.11) 1205 (27.41) 1037 (21.87) 1119 (23.66) 35(c) Stephen Senn 2016
  • 36. Deviances and P-Values Lumiracoxib only fitting Sub-study Statistic Outcome Variables Deviance P-Value Total of discontinuations 13.61 0.0002 CV events 2.92 0.09 At least one AE 1.73 0.19 Any GI 21.31 <0.0001 Dyspepsia 47.34 < 0.0001 36(c) Stephen Senn 2016
  • 37. A Simple Model An unrealistic balanced trial 𝑛 patients per arm, 𝑐 centres in total with 𝑝 patients per centre 2𝑛 = 𝑝𝑐, 𝑛 = 𝑝𝑐 2 Between-centres variance is 2 within-centre variance is 2. Design Variance of Treatment Contrast Completely randomised 4 𝛾2+𝜎2 𝑐𝑝 Randomised blocks (centre blocks) 4 𝜎2 𝑐𝑝 Cluster randomised 4 𝛾2+ 𝜎2 𝑝 𝑐 When using external controls we have at least the variability of a cluster randomised trial 37 (c) Stephen Senn 2016
  • 38. Lessons from TARGET • If you want to use historical controls you will have to work very hard • You need at least two components of variation in your model o Between centre o Between trial • And possibly a third o Between eras • What seems like a lot of information may not be much • Concurrent control and randomisation seems to work well • Moral for any Bayesian: find out as much as possible about any data you intend to use 38 (c) Stephen Senn 2016
  • 39. That example revisited The question • You are proposing to estimate the probability  of a binary event oE.g. cure/no cure • You use a uniform prior on  • You now proceed to study 10,000 occurrences • Which does your prior distribution say is more likely? o10,000 successes o5,000 successes 5,000 failures, in any order The solution • You started with an ‘uninformative prior • After 10,000 trials the observed proportion must be pretty much what you believe is the true probability • But you said every true probability is equally likely • Therefore 5,000 success in any order is just as likely as 10,000 success (c) Stephen Senn 2016 39
  • 40. Advice • Think hard about any prior distribution • Try to establish the objective content of any prior distribution • Uninformative prior distributions are not appropriate for nuisance parameters • Be prepared to think hierarchically • Check that o The prior distribution states your current belief o No data in any shape of form would cause you to abandon it • If the result seems to contradict frequentist wisdom think carefully why • Develop statistical insight – understand what being Bayesian means • As in any statistically system, ask yourself the question o How did I get to see what I see? 40(c) Stephen Senn 2016
  • 41. (c) Stephen Senn 2016 41 References 1 Senn, S. J. Bayesian, likelihood and frequentist approaches to statistics. Applied Clinical Trials 12, 35-38 (2003). 2 Senn, S. J. Trying to be precise about vagueness. Statistics in Medicine 26, 1417-1430 (2007). 3 Senn, S. A note concerning a selection "Paradox" of Dawid's. American Statistician 62, 206-210, doi:10.1198/000313008x331530 (2008). 4 Senn, S. J. Comment on article by Gelman. Bayesian analysis 3, 459-462 (2008). 5 Senn, S. J. Comment on "Harold Jeffreys's Theory of Probability Revisited". Statistical Science 24, 185-186 (2009). 6 Senn, S. J. You may believe you are a Bayesian but you are probably wrong. Rationality, Markets and Morals 2, 48-66 (2011).