Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our User Agreement and Privacy Policy.

Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our Privacy Policy and User Agreement for details.

Successfully reported this slideshow.

Like this presentation? Why not share!

These slides were presented on November 22 2016 during the Annual Julius Symposium, organised by the Julius Center for Health Sciences and Primary Care, University Medical Hospital Utrecht.

Only a few months ago, the American Statistical Association authoritatively issued an official statement on significance and p-values (American Statistician, 2016, 70:2, 129-133), claiming that the p-value is: “commonly misused and misinterpreted.”

In this presentation I focus on the principles of the ASA statement.

No Downloads

Total views

2,523

On SlideShare

0

From Embeds

0

Number of Embeds

348

Shares

0

Downloads

76

Comments

0

Likes

4

No notes for slide

- 1. On p-values Maarten van Smeden Annual Julius Symposium 2016
- 2. About • statistician by training • phd (2016): diagnostic research in absence gold standard (JC) • post-doc: biostatistics / epidemiological methods (JC)
- 3. About this workshop p-value? ASA statement: why and what? p-value alternatives?
- 4. Go to: pvalue.presenterswall.nl
- 5. Point of departure skeptical whenever I see a p-value
- 6. The term “inference”
- 7. p-value?
- 8. Formally deﬁned by
- 9. The pioneers Ronald Aylmer Fisher (1890 - 1962) Jerzy Neyman (1894-1981) Egon Pearson (1895-1980)
- 10. p-value ≥ α “no effect” p-value < α “effect!” α = .05, unless…
- 11. … the p-value fails “arguably signiﬁcant” (P = 0.07) “direction heading to signiﬁcance” (P = 0.10) “ﬂirting with conventional levels of signiﬁcance” (P > 0.1) “marginally signiﬁcant” (P ≥ 0.1) convenient sample from: https://mchankins.wordpress.com/2013/04/21/still-not-signiﬁcant-2/ listing 509 expressions for non-signiﬁcant results at α = .05 level (24 October 2016)
- 12. + 23!!! supplementary ﬁles Wasserstein & Lazar (2016) The ASA's Statement on p-Values: Context, Process, and Purpose, The American Statistician, 70:2, 129-133
- 13. A few quotes (1) “The ASA has not previously taken positions on speciﬁc matters of statistical practice.” nb. founded in 1839 “Nothing in the ASA statement is new.” from the ASA Statement
- 14. A few quotes (2) “… process was lengthier and more controversial than anticipated.” “… the statement articulates in non-technical terms a few select principles that could improve the conduct or interpretation of quantitative science, according to widespread consensus in the statistical community." from the ASA Statement
- 15. p-value? why?
- 16. Go to pvalue.presenterswall.nl
- 17. Why do we need a statement? ‘“It’s science’s dirtiest secret: The ‘scientiﬁc method’ of testing hypotheses by statistical analysis stands on a ﬂimsy foundation.”’ Quoting Siegfried (2010), Odds Are, It’s Wrong: Science Fails to Face the Shortcomings of Statistics, Science News, 177, 26. from the ASA Statement: Wasserstein & Lazar (2016) The ASA's Statement on p-Values: Context, Process, and Purpose, The American Statistician, 70:2, 129-133
- 18. OK, but why now? “… highly visible discussions over the last few years” “The statistical community has been deeply concerned about issues of reproducibility and replicability …” from the ASA statement
- 19. In popular media http://www.vox.com/2016/3/15/11225162/p-value-simple-deﬁnition-hacking (~ 50 million unique visitors monthly)
- 20. The social sciences
- 21. Drastic measures… NHST = Null hypothesis signiﬁcance testing
- 22. P-value increasingly central in reporting From: Chavalarias et al. JAMA. 2016;315(11):1141-1148, doi:10.1001/jama.2016.1952 Using text-mining >1.6 million abstracts
- 23. In the large (‘big’) data era “With a combination of large datasets, confounding, ﬂexibility in analytical choices …, and superimposed selective reporting bias, using a P < 0.05 threshold to declare “success,” …. means next to nothing.” From ASA supplementary material, response by Ioannidis.
- 24. To summarise: why? • p-values and the P < .05 rule are at the core of inference in today’s science (social, biomedical, …) • there is growing concern that these inference are often wrong • perhaps, if we understand p-values better, we’ll be less often wrong
- 25. p-value? why? what?
- 26. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 27. Statistical model? • every method of statistical inference relies on a web of assumptions which together can be viewed as a ‘statistical model’ • the tested hypothesis is one of these assumptions. Often a ‘zero-effect’ called ‘null hypothesis’
- 28. About assumptions the calculation of p-values always relies on assumptions besides the hypothesis tested. It is easy to ignore/forget those assumptions while analysing. Your assumptions are your windows on the world. Scrub them off every once in a while, or the light won't come in. Alan Alda
- 29. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 30. From a probability point of view p-value*: P(Data|Hypothesis) is not: P(Hypothesis|Data) *Somewhat simpliﬁed, correct notation would be: P(T(X) ≥ x | Hypothesis)
- 31. Does it matter? P(Death|Handgun) = 5% to 20%* P(Handgun|Death) = 0.028%** * from New York Times (http://www.nytimes.com article published: 2008/04/03/) ** from CBS StatLine (concerning deaths and registered gun crimes in 2015 in the Netherlands)
- 32. If there only was a way… P(Data|Hypothesis) P(Hypothesis|Data)
- 33. There is… reverend Thomas Bayes (1702-1761) P(H|D) = P(D|H) P(H) P(D)
- 34. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 35. On bright-line rules “Practices that reduce data analysis or scientiﬁc inference to mechanical “bright-line” rules (such as “p < 0.05”) for justifying scientiﬁc claims or conclusions can lead to erroneous beliefs and poor decision making. A conclusion does not immediately become “true” on one side of the divide and “false” on the other.” from the ASA statement
- 36. If p ~ .05 D Colquhoun (2014). An investigation of the false discovery rate and the misinterpretation of p-values. R.Soc.opensci.1:140216. “If you want to avoid making a fool of yourself very often, do not regard anything greater than p < 0.001 as a demonstration that you have discovered something”
- 37. If p > .05
- 38. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 39. The issue of pre-speciﬁed hypotheses From: http://compare-trials.org/ accessed on November 20 2016
- 40. Ed Yong (2012). Replication studies: Bad copy, Nature. Data credits to: D Fanelli.
- 41. Why is this enormous positivity? If you torture the data long enough, it will confess to anything Ronald Coase besides journal editors requirement for p < .05
- 42. Multiple (potential) comparisons aka - p-hacking - data ﬁshing - data dredging - multiple testing - multiplicity - signiﬁcance chasing - signiﬁcance questing - selective inference - etc.
- 43. Selective reporting “Whenever a researcher chooses what to present based on statistical results, valid interpretation of those results is severely compromised if the reader is not informed of the choice and its basis. Researchers should disclose the number of hypotheses explored during the study, all data collection decisions, all statistical analyses conducted, and all p- values computed. Valid scientiﬁc conclusions based on p- values and related statistics cannot be drawn without at least knowing how many and which analyses were conducted, and how those analyses (including p-values) were selected for reporting.” from the ASA statement
- 44. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 45. About effect size • statistical signiﬁcance does not imply practical importance • to understand practical importance we need information on the effect size • Is the p-value a good measure for effect size?
- 46. Dance of the p-values https://www.youtube.com/watch?v=5OL1RqHrZQ8&t=10s Credits to Professor Geoff Cumming
- 47. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA Statement
- 48. P-values in isolation “Researchers should recognize that a p-value without context or other evidence provides limited information. For example, a p-value near 0.05 taken by itself offers only weak evidence against the null hypothesis. Likewise, a relatively large p-value does not imply evidence in favour of the null hypothesis; many other hypotheses may be equally or more consistent with the observed data. For these reasons, data analysis should not end with the calculation of a p-value when other approaches are appropriate and feasible.” from the ASA statement
- 49. The statement: 6 principles 1. P-values can indicate how incompatible the data are with a speciﬁed statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientiﬁc conclusions and business or policy decisions should not be based only on whether a p-value passes a speciﬁc threshold. 4. Proper inference requires full reporting and transparency. 5. A p-value, or statistical signiﬁcance, does not measure the size of an effect or the importance of a result. 6. By itself, a p-value does not provide a good measure of evidence regarding a model or hypothesis. from the ASA statement
- 50. Agreement reached? “you can believe me that had it been any stronger, then all but one of the statisticians would have resigned.” “If only the rest could have agreed with me, we would have a much stronger statement.” from SlideShare, by Stephen Senn: P Values and the art of herding cats (accessed on Oct 30 2016) Stephen Senn, involved in the ASA statement
- 51. From a practical point of view if you work with p-values (derived from the 6 ASA principles): 1. think carefully about the underlying assumptions 2. avoid statements about the truth of the tested hypothesis 3. avoid strong statements about effect based solely on p < . 05 or absence of effect based solely on p > .05 4. report no. and sequence of analyses; avoid data torture 5. avoid statements about effect size based on p-value 6. if feasible, use additional information from other inferential tools
- 52. p-value? why? what? p-value alternatives?
- 53. Other approaches • Methods that emphasise estimation rather than testing • conﬁdence intervals • prediction intervals • credible intervals • Bayesian methods • Alternative measures of evidence • likelihood ratios • Bayes factors • Other approaches • Decision-theoretic modelling • False discovery rates From ASA statement
- 54. A too short introduction to Bayesian inference Remember Bayes? reverend Thomas Bayes (1702-1761)
- 55. Using Bayes theorem P(θ|D) = P(D|θ) P(θ) P(D) P(θ|D) ∝ P(D|θ) P(θ) “likelihood” “prior distribution” “posterior distribution”
- 56. Rational for Bayesian inference the posterior distribution (θ|D) is “more informative” than the likelihood (D|θ) However: “Proponents of the “Bayesian revolution” should be wary of chasing het another chimera: an apparently universal inference procedure. A better path would be to promote both an understanding of various devices in the “statistical toolbox” and informed judgment to select among these.” Gigerenzer and Marewski (2015), Surrogate Science: The Idol of a Universal Method for Scientiﬁc Inference. Journal of Management
- 57. p-value? why? what? p-value alternatives? some ﬁnal remarks
- 58. The words of the pioneer No scientiﬁc worker has a ﬁxed level of signiﬁcance at which from year to year, and in all circumstances, he rejects hypotheses; he rather gives his mind to each particular case in the light of his evidence and his ideas. Ronald Fisher
- 59. Many initiatives to improve science… see: http://www.scienceintransition.nl/english
- 60. and reduce waste ~ 85% of all health research is being avoidably “wasted” see also: http://blogs.bmj.com/bmj/2016/01/14/paul-glasziou-and-iain-chalmers-is-85-of-health-research-really-wasted/, and: Lancet’s 2014 series on increasing value, reducing waste (incl video’s etc.): http://www.thelancet.com/series/research
- 61. Conclusion • statistical inference is inherently difﬁcult; we should avoid making a fool of ourselves too often • p-values can be useful tools for inference; most often, p- values should not be the ‘star of the inference show’ • bright line rules such as p < .05 give a false sense of scientiﬁc objectivity • like to play around with data? Me too! Think twice before you publish such explorations; if you do, be honest and transparent in reporting
- 62. Some random thoughts • inference is thought as a primarily mathematical or computational problem, it should not. • we should ban the term “signiﬁcant” from scientiﬁc output for describing effects that are accompanied with p < .05. • in applied statistics education, we should invest more time in discussing various forms of inference (e.g., Bayesian inference) and their merits and pitfalls
- 63. Go to: pvalue.presenterswall.nl
- 64. Points for discussion • is there a need for changing the way we do inference? • if so, how and what do we change? • education? • journals? • should we downplay the role of p < .05 in scientiﬁc output?

No public clipboards found for this slide

Be the first to comment