This lecture introduces Bayesian hypothesis testing. It discusses an example comparing HIV infection rates between a treatment and placebo group. A Bayesian analysis is presented that calculates posterior probabilities for the null and alternative hypotheses using prior probabilities and Bayes factors. The lecture outlines general notation for Bayesian testing and discusses issues like choosing prior distributions and testing precise versus imprecise hypotheses. It also discusses interpreting Bayes factors and relates posterior probabilities to p-values in some cases.
The document describes a course on model uncertainty taught in the fall of 2018. It covers topics like statistical and mathematical model uncertainty, Bayesian hypothesis testing and model uncertainty, priors for Bayesian model uncertainty, approximations and computation, model inputs and outputs, model calibration, Gaussian processes, surrogate models, sensitivity analysis, and model discrepancy. The course is taught over 12 weeks by two lecturers and includes weekly topics like introduction to uncertainty, Bayesian analysis, representation of inputs/outputs, calibration, Gaussian processes, surrogate models, sampling techniques, and sensitivity analysis.
This document discusses Bayesian approaches to combining evidence from multiple data sources or models. It recommends combining data probabilistically using Bayes' theorem rather than averaging. It provides an example of combining three data sources on annual rainfall measurements by treating the data sources as independent measurements and deriving the posterior distribution of rainfall amounts given the data. It also discusses challenges that arise when combining dependent data sources or models, and presents examples of hierarchical modeling approaches.
On the vexing dilemma of hypothesis testing and the predicted demise of the B...Christian Robert
The document discusses hypothesis testing from both frequentist and Bayesian perspectives. It introduces the concept of statistical tests as functions that output accept or reject decisions for hypotheses. P-values are presented as a way to quantify uncertainty in these decisions. Bayes' original 1763 paper on Bayesian statistics is summarized, introducing the concept of the posterior distribution. Bayesian hypothesis testing is then discussed, including the optimal Bayes test and the use of Bayes factors to compare hypotheses without requiring prior probabilities on the hypotheses.
This document summarizes a presentation on testing hypotheses as mixture estimation and the challenges of Bayesian testing. The key points are:
1) Bayesian hypothesis testing faces challenges including the dependence on prior distributions, difficulties interpreting Bayes factors, and the inability to use improper priors in most situations.
2) Testing via mixtures is proposed as a paradigm shift that frames hypothesis testing as a model selection problem involving mixture models rather than distinct hypotheses.
3) Traditional Bayesian testing using Bayes factors and posterior probabilities depends strongly on prior distributions and choices that are difficult to justify, while not providing measures of uncertainty around decisions. Alternative approaches are needed to address these issues.
- Approximate Bayesian computation (ABC) is a technique used when the likelihood function is intractable or unavailable. It approximates the Bayesian posterior distribution in a likelihood-free manner.
- ABC works by simulating parameter values from the prior and simulating pseudo-data. Parameter values are accepted if the simulated pseudo-data are "close" to the observed data according to some distance measure and tolerance level.
- ABC originated in population genetics models where genealogies are considered nuisance parameters that cannot be integrated out of the likelihood. It has since been applied to other fields like econometrics for models with complex or undefined likelihoods.
Discussion of Persi Diaconis' lecture at ISBA 2016Christian Robert
This document discusses Monte Carlo methods for numerical integration and estimating normalizing constants. It summarizes several approaches: estimating normalizing constants using samples; reverse logistic regression for estimating constants in mixtures; Xiao-Li's maximum likelihood formulation for Monte Carlo integration; and Persi's probabilistic numerics which provide uncertainties for numerical calculations. The document advocates first approximating the distribution of an integrand before estimating its expectation to incorporate non-parametric information and account for multiple estimators.
This document discusses challenges and recent advances in Approximate Bayesian Computation (ABC) methods. ABC methods are used when the likelihood function is intractable or unavailable in closed form. The core ABC algorithm involves simulating parameters from the prior and simulating data, retaining simulations where the simulated and observed data are close according to a distance measure on summary statistics. The document outlines key issues like scalability to large datasets, assessment of uncertainty, and model choice, and discusses advances such as modified proposals, nonparametric methods, and perspectives that include summary construction in the framework. Validation of ABC model choice and selection of summary statistics remains an open challenge.
The document describes a course on model uncertainty taught in the fall of 2018. It covers topics like statistical and mathematical model uncertainty, Bayesian hypothesis testing and model uncertainty, priors for Bayesian model uncertainty, approximations and computation, model inputs and outputs, model calibration, Gaussian processes, surrogate models, sensitivity analysis, and model discrepancy. The course is taught over 12 weeks by two lecturers and includes weekly topics like introduction to uncertainty, Bayesian analysis, representation of inputs/outputs, calibration, Gaussian processes, surrogate models, sampling techniques, and sensitivity analysis.
This document discusses Bayesian approaches to combining evidence from multiple data sources or models. It recommends combining data probabilistically using Bayes' theorem rather than averaging. It provides an example of combining three data sources on annual rainfall measurements by treating the data sources as independent measurements and deriving the posterior distribution of rainfall amounts given the data. It also discusses challenges that arise when combining dependent data sources or models, and presents examples of hierarchical modeling approaches.
On the vexing dilemma of hypothesis testing and the predicted demise of the B...Christian Robert
The document discusses hypothesis testing from both frequentist and Bayesian perspectives. It introduces the concept of statistical tests as functions that output accept or reject decisions for hypotheses. P-values are presented as a way to quantify uncertainty in these decisions. Bayes' original 1763 paper on Bayesian statistics is summarized, introducing the concept of the posterior distribution. Bayesian hypothesis testing is then discussed, including the optimal Bayes test and the use of Bayes factors to compare hypotheses without requiring prior probabilities on the hypotheses.
This document summarizes a presentation on testing hypotheses as mixture estimation and the challenges of Bayesian testing. The key points are:
1) Bayesian hypothesis testing faces challenges including the dependence on prior distributions, difficulties interpreting Bayes factors, and the inability to use improper priors in most situations.
2) Testing via mixtures is proposed as a paradigm shift that frames hypothesis testing as a model selection problem involving mixture models rather than distinct hypotheses.
3) Traditional Bayesian testing using Bayes factors and posterior probabilities depends strongly on prior distributions and choices that are difficult to justify, while not providing measures of uncertainty around decisions. Alternative approaches are needed to address these issues.
- Approximate Bayesian computation (ABC) is a technique used when the likelihood function is intractable or unavailable. It approximates the Bayesian posterior distribution in a likelihood-free manner.
- ABC works by simulating parameter values from the prior and simulating pseudo-data. Parameter values are accepted if the simulated pseudo-data are "close" to the observed data according to some distance measure and tolerance level.
- ABC originated in population genetics models where genealogies are considered nuisance parameters that cannot be integrated out of the likelihood. It has since been applied to other fields like econometrics for models with complex or undefined likelihoods.
Discussion of Persi Diaconis' lecture at ISBA 2016Christian Robert
This document discusses Monte Carlo methods for numerical integration and estimating normalizing constants. It summarizes several approaches: estimating normalizing constants using samples; reverse logistic regression for estimating constants in mixtures; Xiao-Li's maximum likelihood formulation for Monte Carlo integration; and Persi's probabilistic numerics which provide uncertainties for numerical calculations. The document advocates first approximating the distribution of an integrand before estimating its expectation to incorporate non-parametric information and account for multiple estimators.
This document discusses challenges and recent advances in Approximate Bayesian Computation (ABC) methods. ABC methods are used when the likelihood function is intractable or unavailable in closed form. The core ABC algorithm involves simulating parameters from the prior and simulating data, retaining simulations where the simulated and observed data are close according to a distance measure on summary statistics. The document outlines key issues like scalability to large datasets, assessment of uncertainty, and model choice, and discusses advances such as modified proposals, nonparametric methods, and perspectives that include summary construction in the framework. Validation of ABC model choice and selection of summary statistics remains an open challenge.
The document discusses using random forests for approximate Bayesian computation (ABC) model choice. ABC can be framed as a machine learning problem where simulated datasets are used to learn which model is most appropriate. Random forests are well-suited for this as they can handle many correlated summary statistics without information loss. The random forest predicts the most likely model but not posterior probabilities. Instead, the posterior predictive expected error rate across models is proposed to evaluate model selection performance without unstable probability approximations. An example comparing MA(1) and MA(2) time series models illustrates the approach.
Approximate Bayesian model choice via random forestsChristian Robert
The document describes approximate Bayesian computation (ABC) methods for model choice when likelihoods are intractable. ABC generates parameter-dataset pairs from the prior and retains those where the simulated and observed datasets are similar according to a distance measure on summary statistics. For model choice, ABC approximates posterior model probabilities by the proportion of simulations from each model that are retained. Machine learning techniques can also be used to infer the most likely model directly from the simulated summary statistics.
The document proposes using random forests (RF), a machine learning tool, for approximate Bayesian computation (ABC) model choice rather than estimating model posterior probabilities. RF improves on existing ABC model choice methods by having greater discriminative power among models, being robust to the choice and number of summary statistics, requiring less computation, and providing an error rate to evaluate confidence in the model choice. The authors illustrate the power of the RF-based ABC methodology on controlled experiments and real population genetics datasets.
random forests for ABC model choice and parameter estimationChristian Robert
The document discusses Approximate Bayesian Computation (ABC). It begins by introducing ABC as a likelihood-free method for Bayesian inference when the likelihood function is unavailable or computationally intractable. ABC works by simulating data under different parameter values and accepting simulations that are close to the observed data based on a distance measure.
The document then discusses advances in ABC, including modifying the proposal distribution to increase efficiency, viewing it as a conditional density estimation problem, and including measurement error in the framework. It also discusses the consistency of ABC as the number of simulations increases and sample size grows large. Finally, it discusses applications of ABC to model selection by treating the model index as an additional parameter.
This document discusses approximate Bayesian computation (ABC) techniques for performing Bayesian inference when the likelihood function is not available in closed form. It covers the basic ABC algorithm and discusses challenges with high-dimensional data. It also summarizes recent advances in ABC that incorporate nonparametric regression, reproducing kernel Hilbert spaces, and neural networks to help address these challenges.
This document discusses several perspectives and solutions to Bayesian hypothesis testing. It outlines issues with Bayesian testing such as the dependence on prior distributions and difficulties interpreting Bayesian measures like posterior probabilities and Bayes factors. It discusses how Bayesian testing compares models rather than identifying a single true model. Several solutions to challenges are discussed, like using Bayes factors which eliminate the dependence on prior model probabilities but introduce other issues. The document also discusses testing under specific models like comparing a point null hypothesis to alternatives. Overall it presents both Bayesian and frequentist views on hypothesis testing and some of the open controversies in the field.
This document provides lecture notes on hypothesis testing. It begins with an introduction to hypothesis testing and how it differs from estimation in its hypothetical reasoning approach. It then discusses Fisher's significance testing approach, including defining a test statistic, its sampling distribution under the null hypothesis, and calculating a p-value. It provides examples of applying this approach. Finally, it discusses some weaknesses of Fisher's approach identified by Neyman and Pearson and how their approach improved upon it by introducing the concept of alternative hypotheses and pre-data error probabilities.
1) Likelihood-free Bayesian experimental design is discussed as an intractable likelihood optimization problem, where the goal is to find the optimal design d that minimizes expected loss without using the full posterior distribution.
2) Several Bayesian tools are proposed to make the design problem more Bayesian, including Bayesian non-parametrics, annealing algorithms, and placing a posterior on the design d.
3) Gaussian processes are a default modeling choice for complex unknown functions in these problems, but their accuracy is difficult to assess and they may incur a dimension curse.
This document provides an introduction to Bayesian statistics and machine learning. It discusses key concepts like conditional probability, Bayes' theorem, Bayesian inference, Bayesian model comparison, and Bayesian learning. Conditional probability is fundamental in probability theory and looks at the probability of event A given event B. Bayes' theorem allows updating beliefs with new evidence and can be visualized with diagrams. Bayesian inference involves specifying prior distributions over parameters and updating them based on observed data to obtain posterior distributions. Bayesian models can be compared using Bayes factors, which are ratios of marginal likelihoods. Bayesian learning techniques include Markov chain Monte Carlo methods and hierarchical Bayesian models.
An Introduction to Mis-Specification (M-S) Testingjemille6
This document provides an introduction to mis-specification (M-S) testing, which is a methodology for validating statistical models. It discusses how statistical misspecification can render statistical inference unreliable by distorting nominal error probabilities. As an example, it shows how violating the independence assumption in a normal model can increase the actual type I error rate and reduce power. It argues that model validation through M-S testing is important for ensuring reliable inference, but is often neglected due to misunderstandings. All statistical methods rely on an underlying statistical model, so any misspecification impacts reliability regardless of the method used.
This document summarizes approximate Bayesian computation (ABC) methods. It begins with an overview of ABC, which provides a likelihood-free rejection technique for Bayesian inference when the likelihood function is intractable. The ABC algorithm works by simulating parameters and data until the simulated and observed data are close according to some distance measure and tolerance level. The document then discusses the asymptotic properties of ABC, including consistency of ABC posteriors and rates of convergence under certain assumptions. It also notes relationships between ABC and k-nearest neighbor methods. Examples applying ABC to autoregressive time series models are provided.
The document discusses Hessian matrices in statistics. It begins by introducing the Hessian matrix and describing relevant statistical concepts like maximum likelihood estimation and the likelihood function. It then provides an example of calculating the Hessian matrix for a Gaussian linear regression model estimated using maximum likelihood. The Hessian shows that the MLE solution maximizes the likelihood function and is the minimum. Larger sample sizes improve MLE estimates, as the estimate converges to the true parameter value as the sample size increases.
This document discusses various importance sampling methods for approximating Bayes factors, which are used for Bayesian model selection. It compares regular importance sampling, bridge sampling, harmonic means, mixtures to bridge sampling, and Chib's solution. An example application to probit modeling of diabetes in Pima Indian women is presented to illustrate regular importance sampling. Markov chain Monte Carlo methods like the Metropolis-Hastings algorithm and Gibbs sampling can be used to sample from the probit models.
This document provides an overview of ABC methodology and applications. It begins with examples from population genetics and econometrics that are well-suited for ABC. It then describes the basic ABC algorithm for Bayesian inference using simulation: specifying prior distributions, simulating data under different parameter values, and accepting simulations that best match the observed data. Indirect inference is also discussed as a method for choosing informative summary statistics for ABC. The document traces the origins of ABC to population genetics models from the late 1990s and highlights ongoing contributions from that field to ABC methodology.
- The document provides information about statisticshomeworkhelper.com, a service that offers probability and statistics assignment help. It lists their website, email, and phone number for contacting them.
- It then provides an example of a multi-part statistics problem involving hypothesis testing on coin flips and dice data. It asks the reader to conduct various statistical tests and interpret the results.
- Finally, it lists some additional practice problems involving chi-square tests, ANOVA, and other statistical analyses for the reader to work through.
An introduction to Bayesian Statistics using Pythonfreshdatabos
This document provides an introduction to Bayesian statistics and inference through examples. It begins with an overview of Bayes' Theorem and probability concepts. An example problem about cookies in bowls is used to demonstrate applying Bayes' Theorem to update beliefs based on new data. The document introduces the Pmf class for representing probability mass functions and working through examples numerically. Further examples involving dice and trains reinforce how to build likelihood functions and update distributions. The document concludes with a real-world example of analyzing whether a coin is biased based on spin results.
1. The document provides steps for conducting hypothesis tests using the p-value method, including defining hypotheses, calculating test statistics, finding p-values, and making conclusions.
2. It outlines the procedures for z-tests of means and proportions, t-tests of means, and chi-squared tests of variances. The key differences between these tests are whether the population standard deviation is known or unknown.
3. For each test, the null and alternative hypotheses are stated, the test statistic is calculated using the appropriate formula, the p-value is obtained and used to determine whether to reject or fail to reject the null hypothesis at a given significance level.
The document discusses approximate Bayesian computation (ABC), a simulation-based method for conducting Bayesian inference when the likelihood function is intractable or impossible to compute directly. ABC works by simulating data under different parameter values, and accepting simulations that are close to the observed data according to some distance measure. The document covers the basic ABC algorithm, convergence properties as the tolerance approaches zero, examples of ABC for probit models and MA time series models, and advances such as modifying the proposal distribution to increase efficiency.
This document provides an overview of hypothesis testing basics:
A) Hypothesis testing involves stating a null hypothesis (H0) and alternative hypothesis (Ha) based on a research question. H0 assumes no effect or difference, while Ha claims an effect.
B) A test statistic is calculated from sample data and compared to a theoretical distribution to evaluate H0. For a one-sample z-test with known standard deviation, the test statistic is a z-score.
C) The p-value represents the probability of observing the test statistic or a more extreme value if H0 is true. Small p-values provide evidence against H0. Conventionally, p ≤ 0.05 is considered significant
Bayesian statistics uses probability to represent uncertainty about unknown parameters in statistical models. It differs from classical statistics in that parameters are treated as random variables rather than fixed unknown constants. Bayesian probability represents a degree of belief in an event rather than the physical probability of an event. The Bayes' formula provides a way to update beliefs based on new evidence or data using conditional probability. Bayesian networks are graphical models that compactly represent joint probability distributions over many variables and allow for efficient inference.
The document discusses using random forests for approximate Bayesian computation (ABC) model choice. ABC can be framed as a machine learning problem where simulated datasets are used to learn which model is most appropriate. Random forests are well-suited for this as they can handle many correlated summary statistics without information loss. The random forest predicts the most likely model but not posterior probabilities. Instead, the posterior predictive expected error rate across models is proposed to evaluate model selection performance without unstable probability approximations. An example comparing MA(1) and MA(2) time series models illustrates the approach.
Approximate Bayesian model choice via random forestsChristian Robert
The document describes approximate Bayesian computation (ABC) methods for model choice when likelihoods are intractable. ABC generates parameter-dataset pairs from the prior and retains those where the simulated and observed datasets are similar according to a distance measure on summary statistics. For model choice, ABC approximates posterior model probabilities by the proportion of simulations from each model that are retained. Machine learning techniques can also be used to infer the most likely model directly from the simulated summary statistics.
The document proposes using random forests (RF), a machine learning tool, for approximate Bayesian computation (ABC) model choice rather than estimating model posterior probabilities. RF improves on existing ABC model choice methods by having greater discriminative power among models, being robust to the choice and number of summary statistics, requiring less computation, and providing an error rate to evaluate confidence in the model choice. The authors illustrate the power of the RF-based ABC methodology on controlled experiments and real population genetics datasets.
random forests for ABC model choice and parameter estimationChristian Robert
The document discusses Approximate Bayesian Computation (ABC). It begins by introducing ABC as a likelihood-free method for Bayesian inference when the likelihood function is unavailable or computationally intractable. ABC works by simulating data under different parameter values and accepting simulations that are close to the observed data based on a distance measure.
The document then discusses advances in ABC, including modifying the proposal distribution to increase efficiency, viewing it as a conditional density estimation problem, and including measurement error in the framework. It also discusses the consistency of ABC as the number of simulations increases and sample size grows large. Finally, it discusses applications of ABC to model selection by treating the model index as an additional parameter.
This document discusses approximate Bayesian computation (ABC) techniques for performing Bayesian inference when the likelihood function is not available in closed form. It covers the basic ABC algorithm and discusses challenges with high-dimensional data. It also summarizes recent advances in ABC that incorporate nonparametric regression, reproducing kernel Hilbert spaces, and neural networks to help address these challenges.
This document discusses several perspectives and solutions to Bayesian hypothesis testing. It outlines issues with Bayesian testing such as the dependence on prior distributions and difficulties interpreting Bayesian measures like posterior probabilities and Bayes factors. It discusses how Bayesian testing compares models rather than identifying a single true model. Several solutions to challenges are discussed, like using Bayes factors which eliminate the dependence on prior model probabilities but introduce other issues. The document also discusses testing under specific models like comparing a point null hypothesis to alternatives. Overall it presents both Bayesian and frequentist views on hypothesis testing and some of the open controversies in the field.
This document provides lecture notes on hypothesis testing. It begins with an introduction to hypothesis testing and how it differs from estimation in its hypothetical reasoning approach. It then discusses Fisher's significance testing approach, including defining a test statistic, its sampling distribution under the null hypothesis, and calculating a p-value. It provides examples of applying this approach. Finally, it discusses some weaknesses of Fisher's approach identified by Neyman and Pearson and how their approach improved upon it by introducing the concept of alternative hypotheses and pre-data error probabilities.
1) Likelihood-free Bayesian experimental design is discussed as an intractable likelihood optimization problem, where the goal is to find the optimal design d that minimizes expected loss without using the full posterior distribution.
2) Several Bayesian tools are proposed to make the design problem more Bayesian, including Bayesian non-parametrics, annealing algorithms, and placing a posterior on the design d.
3) Gaussian processes are a default modeling choice for complex unknown functions in these problems, but their accuracy is difficult to assess and they may incur a dimension curse.
This document provides an introduction to Bayesian statistics and machine learning. It discusses key concepts like conditional probability, Bayes' theorem, Bayesian inference, Bayesian model comparison, and Bayesian learning. Conditional probability is fundamental in probability theory and looks at the probability of event A given event B. Bayes' theorem allows updating beliefs with new evidence and can be visualized with diagrams. Bayesian inference involves specifying prior distributions over parameters and updating them based on observed data to obtain posterior distributions. Bayesian models can be compared using Bayes factors, which are ratios of marginal likelihoods. Bayesian learning techniques include Markov chain Monte Carlo methods and hierarchical Bayesian models.
An Introduction to Mis-Specification (M-S) Testingjemille6
This document provides an introduction to mis-specification (M-S) testing, which is a methodology for validating statistical models. It discusses how statistical misspecification can render statistical inference unreliable by distorting nominal error probabilities. As an example, it shows how violating the independence assumption in a normal model can increase the actual type I error rate and reduce power. It argues that model validation through M-S testing is important for ensuring reliable inference, but is often neglected due to misunderstandings. All statistical methods rely on an underlying statistical model, so any misspecification impacts reliability regardless of the method used.
This document summarizes approximate Bayesian computation (ABC) methods. It begins with an overview of ABC, which provides a likelihood-free rejection technique for Bayesian inference when the likelihood function is intractable. The ABC algorithm works by simulating parameters and data until the simulated and observed data are close according to some distance measure and tolerance level. The document then discusses the asymptotic properties of ABC, including consistency of ABC posteriors and rates of convergence under certain assumptions. It also notes relationships between ABC and k-nearest neighbor methods. Examples applying ABC to autoregressive time series models are provided.
The document discusses Hessian matrices in statistics. It begins by introducing the Hessian matrix and describing relevant statistical concepts like maximum likelihood estimation and the likelihood function. It then provides an example of calculating the Hessian matrix for a Gaussian linear regression model estimated using maximum likelihood. The Hessian shows that the MLE solution maximizes the likelihood function and is the minimum. Larger sample sizes improve MLE estimates, as the estimate converges to the true parameter value as the sample size increases.
This document discusses various importance sampling methods for approximating Bayes factors, which are used for Bayesian model selection. It compares regular importance sampling, bridge sampling, harmonic means, mixtures to bridge sampling, and Chib's solution. An example application to probit modeling of diabetes in Pima Indian women is presented to illustrate regular importance sampling. Markov chain Monte Carlo methods like the Metropolis-Hastings algorithm and Gibbs sampling can be used to sample from the probit models.
This document provides an overview of ABC methodology and applications. It begins with examples from population genetics and econometrics that are well-suited for ABC. It then describes the basic ABC algorithm for Bayesian inference using simulation: specifying prior distributions, simulating data under different parameter values, and accepting simulations that best match the observed data. Indirect inference is also discussed as a method for choosing informative summary statistics for ABC. The document traces the origins of ABC to population genetics models from the late 1990s and highlights ongoing contributions from that field to ABC methodology.
- The document provides information about statisticshomeworkhelper.com, a service that offers probability and statistics assignment help. It lists their website, email, and phone number for contacting them.
- It then provides an example of a multi-part statistics problem involving hypothesis testing on coin flips and dice data. It asks the reader to conduct various statistical tests and interpret the results.
- Finally, it lists some additional practice problems involving chi-square tests, ANOVA, and other statistical analyses for the reader to work through.
An introduction to Bayesian Statistics using Pythonfreshdatabos
This document provides an introduction to Bayesian statistics and inference through examples. It begins with an overview of Bayes' Theorem and probability concepts. An example problem about cookies in bowls is used to demonstrate applying Bayes' Theorem to update beliefs based on new data. The document introduces the Pmf class for representing probability mass functions and working through examples numerically. Further examples involving dice and trains reinforce how to build likelihood functions and update distributions. The document concludes with a real-world example of analyzing whether a coin is biased based on spin results.
1. The document provides steps for conducting hypothesis tests using the p-value method, including defining hypotheses, calculating test statistics, finding p-values, and making conclusions.
2. It outlines the procedures for z-tests of means and proportions, t-tests of means, and chi-squared tests of variances. The key differences between these tests are whether the population standard deviation is known or unknown.
3. For each test, the null and alternative hypotheses are stated, the test statistic is calculated using the appropriate formula, the p-value is obtained and used to determine whether to reject or fail to reject the null hypothesis at a given significance level.
The document discusses approximate Bayesian computation (ABC), a simulation-based method for conducting Bayesian inference when the likelihood function is intractable or impossible to compute directly. ABC works by simulating data under different parameter values, and accepting simulations that are close to the observed data according to some distance measure. The document covers the basic ABC algorithm, convergence properties as the tolerance approaches zero, examples of ABC for probit models and MA time series models, and advances such as modifying the proposal distribution to increase efficiency.
This document provides an overview of hypothesis testing basics:
A) Hypothesis testing involves stating a null hypothesis (H0) and alternative hypothesis (Ha) based on a research question. H0 assumes no effect or difference, while Ha claims an effect.
B) A test statistic is calculated from sample data and compared to a theoretical distribution to evaluate H0. For a one-sample z-test with known standard deviation, the test statistic is a z-score.
C) The p-value represents the probability of observing the test statistic or a more extreme value if H0 is true. Small p-values provide evidence against H0. Conventionally, p ≤ 0.05 is considered significant
Bayesian statistics uses probability to represent uncertainty about unknown parameters in statistical models. It differs from classical statistics in that parameters are treated as random variables rather than fixed unknown constants. Bayesian probability represents a degree of belief in an event rather than the physical probability of an event. The Bayes' formula provides a way to update beliefs based on new evidence or data using conditional probability. Bayesian networks are graphical models that compactly represent joint probability distributions over many variables and allow for efficient inference.
hypotesting lecturenotes by Amity universitydeepti .
This document provides an overview of hypothesis testing and the key steps involved:
1. The null and alternative hypotheses are stated, with the null usually claiming "no difference" and the alternative contradicting the null.
2. A test statistic is calculated from the sample data and compared to the distribution assumed by the null hypothesis. For a one-sample z-test, this involves calculating the z-score.
3. The p-value is derived as the probability of obtaining a test statistic at least as extreme as what was observed, assuming the null is true. Small p-values provide strong evidence against the null.
4. Factors like statistical power and sample size requirements are also discussed to ensure
This document provides an overview of hypothesis testing and the steps involved. It discusses:
1) Defining the null and alternative hypotheses based on the research question. The null hypothesis represents "no difference" while the alternative hypothesis claims the null is false.
2) Calculating the test statistic, which is used to test the null hypothesis. For a one-sample z-test, this involves calculating the z-score when the population standard deviation is known.
3) Computing the p-value, which is the probability of observing a test statistic as extreme or more extreme than what was observed, assuming the null hypothesis is true. Small p-values provide strong evidence against the null.
4) Interpre
This document provides an overview of hypothesis testing and the steps involved. It discusses:
1) Defining the null and alternative hypotheses based on the research question. The null hypothesis represents "no difference" while the alternative hypothesis claims the null is false.
2) Calculating the test statistic, which is used to test the null hypothesis. For a one-sample z-test, this involves calculating the z-score when the population standard deviation is known.
3) Computing the p-value, which is the probability of observing a test statistic as extreme or more extreme than what was observed, assuming the null hypothesis is true. Small p-values provide strong evidence against the null.
4) Interpre
This document provides an overview of hypothesis testing and the steps involved. It introduces:
1) The concepts of the null and alternative hypotheses, which are used to frame the research question. The null hypothesis represents "no difference" while the alternative hypothesis claims the null is false.
2) How to calculate the test statistic, which is used to evaluate the null hypothesis based on the sample data. For a one-sample z-test, this involves calculating the z-score.
3) How to determine the p-value, which represents the probability of observing the test statistic or one more extreme, assuming the null hypothesis is true. A small p-value provides evidence against the null.
4)
This document discusses hypothesis testing, including:
1) The objectives are to formulate statistical hypotheses, discuss types of errors, establish decision rules, and choose appropriate tests.
2) Key symbols and concepts are defined, such as the null and alternative hypotheses, Type I and Type II errors, test statistics like z and t, means, variances, sample sizes, and significance levels.
3) The two types of errors in hypothesis testing are discussed. Hypothesis tests can result in correct decisions or two types of errors when the null hypothesis is true or false.
4) Steps in hypothesis testing are outlined, including formulating hypotheses, specifying a significance level, choosing a test statistic, establishing a
BMI (kg/m2)
22.1
23.4
24.8
26.2
27.6
28.9
30.3
31.6
32.9
34.2
35.5
36.8
38.1
39.4
The sample mean is 29.1 kg/m2 and the sample standard
deviation is 4.2 kg/m2. Test the hypothesis that the
population mean BMI is 30 kg/m2 at 5% level of
significance.
This document provides information about statistics and hypothesis testing concepts. It defines key terms like population, sample, parameters, statistics, standard error, random sampling, critical region, acceptance region, one-tailed and two-tailed tests, null and alternative hypotheses, type I and type II errors. It also describes common statistical tests like t-test, F-test, chi-square test and provides their assumptions and uses. Several examples of hypothesis testing problems and their solutions are given to illustrate statistical concepts and procedures.
This document provides an introduction to hypothesis testing including:
1. The 5 steps in a hypothesis test: set up null and alternative hypotheses, define test procedure, collect data, decide whether to reject null hypothesis, interpret results.
2. Large sample tests for the mean involve testing if the population mean is equal to or not equal to a specified value using a test statistic that follows a normal distribution.
3. Type I and Type II errors occur when the decision made based on the hypothesis test does not match the actual truth - a Type I error rejects the null hypothesis when it is true, a Type II error fails to reject the null when it is false. The probability of each error can be minimized by choosing
This document discusses hypothesis testing. It defines key terms like the null hypothesis (Ho), alternative hypothesis (H1), type 1 and type 2 errors, significance level, test statistics, critical values, rejection regions, and one-tailed vs two-tailed tests. It provides examples of how to formulate hypotheses, determine appropriate test statistics, establish critical regions, and make conclusions based on computed test values for both known and unknown population variances with one and two sample tests concerning means.
This document discusses hypothesis testing and the t-test. It covers:
1) The basics of hypothesis testing including null and alternative hypotheses, types of hypotheses, and types of errors.
2) The t-test, which is used for small samples from a normally distributed population. It relies on the t-distribution and the degree of freedom.
3) Applications of the t-test including testing the significance of a single mean, difference between two means, and paired t-tests.
4) When sample sizes are large, the normal distribution can be used instead in Z-tests for similar applications.
Excursion 4 Tour II: Rejection Fallacies: Whose Exaggerating What?jemille6
This document discusses criticisms of p-values and proposes reforms based on Bayesian statistics. It summarizes debates between Fisher and Bayesians regarding p-values exaggerating evidence against the null hypothesis when using certain priors. When a lump prior of 0.5 is given to the null and the remaining 0.5 spread over the alternative, as the sample size increases, a statistically significant result can correspond to a posterior probability for the null that exceeds the prior of 0.5. Reforms are proposed based on likelihood ratios and Bayes factors to define statistical significance in a way more consistent with Bayesian evidence standards.
This document provides information about the t-test and chi-square test. It defines the t-test as a test used to compare the means of two samples when the population standard deviation is unknown. It lists the assumptions of the t-test and provides the formula. An example t-test problem and solution is given. Chi-square is introduced as a test used with categorical and numerical data to test for independence and goodness of fit. The chi-square test statistic, degrees of freedom, and hypothesis testing process are outlined. An example chi-square goodness of fit problem and solution is also provided.
PG STAT 531 Lecture 6 Test of Significance, z TestAashish Patel
The document summarizes key concepts related to tests of significance. It discusses:
1) The difference between population parameters and sample statistics. Parameters describe the population while statistics describe samples.
2) The goal of tests of significance is to determine if an observed difference between a sample and population statistic is statistically significant or likely due to chance. Common tests include z-tests, t-tests, chi-square tests, and F-tests.
3) All tests of significance involve a null hypothesis (H0), which is tested against an alternative hypothesis (Ha). The outcome is either rejecting or failing to reject the null hypothesis based on a significance level like alpha=0.05.
4) Type I
The document summarizes key concepts in hypothesis testing including:
- The null and alternative hypotheses are formulated, with the null hypothesis stating the parameter equals a specific value and the alternative allowing other values.
- There are two types of errors - type I rejects the null when true, type II accepts when false. Tests aim to minimize both.
- The power of a test is the probability it correctly rejects the null when an alternative is true.
- One-tailed tests have critical regions in one tail, two-tailed in both. P-values are used to determine if results are significant.
- Steps of hypothesis testing are outlined along with examples of tests for single and two means/proportions.
1) This document introduces concepts of probability and statistics that will be covered in the course, including defining probability, discrete vs. continuous probability distributions, and how to calculate mean, mode, median, and variance.
2) It discusses the differences between accuracy and precision in measurements and explains common sources of measurement errors like statistical and systematic uncertainties.
3) Key points on experimental results are presented, such as how statistical and systematic errors are combined and various ways of quoting measurements and their uncertainties.
This document provides an overview of hypothesis testing. It defines key terms like the null hypothesis (H0), alternative hypothesis (H1), type I and type II errors, significance level, p-values, and test statistics. It explains the basic steps in hypothesis testing as testing a claim about a population parameter by collecting a sample, determining the appropriate test statistic based on the sampling distribution, and comparing it to critical values to reject or fail to reject the null hypothesis. Examples are provided to demonstrate hypothesis tests for a mean when the population standard deviation is known or unknown.
8. testing of hypothesis for variable & attribute dataHakeem-Ur- Rehman
The document discusses hypothesis testing for continuous variable and attribute data. It begins by defining key concepts in statistical inference like the null and alternative hypotheses. The three types of hypotheses are explained - two-tailed, left-tailed, and right-tailed. The document then discusses hypothesis testing steps including defining the hypotheses, determining the sampling risk of type I and type II errors, calculating the p-value, and making a decision to accept or reject the null hypothesis based on the p-value and significance level. Specific parametric statistical tests are explained like the one sample t-test, two sample t-test, and ANOVA. Examples of each test are provided and how to interpret the results.
Categorical data analysis full lecture note PPT.pptxMinilikDerseh1
This document provides an overview of categorical data analysis techniques. It discusses categorical and quantitative variables, different types of categorical variables, and common distributions for categorical data like binomial and multinomial. Methods for categorical data like chi-square tests, logistic regression, and Poisson regression are presented. Examples are provided to illustrate hypothesis testing, confidence intervals, and likelihood ratio tests for categorical proportions.
Similar to 2018 MUMS Fall Course - Essentials of Bayesian Hypothesis Testing - Jim Berger , September 4, 2018 (20)
Recently, the machine learning community has expressed strong interest in applying latent variable modeling strategies to causal inference problems with unobserved confounding. Here, I discuss one of the big debates that occurred over the past year, and how we can move forward. I will focus specifically on the failure of point identification in this setting, and discuss how this can be used to design flexible sensitivity analyses that cleanly separate identified and unidentified components of the causal model.
I will discuss paradigmatic statistical models of inference and learning from high dimensional data, such as sparse PCA and the perceptron neural network, in the sub-linear sparsity regime. In this limit the underlying hidden signal, i.e., the low-rank matrix in PCA or the neural network weights, has a number of non-zero components that scales sub-linearly with the total dimension of the vector. I will provide explicit low-dimensional variational formulas for the asymptotic mutual information between the signal and the data in suitable sparse limits. In the setting of support recovery these formulas imply sharp 0-1 phase transitions for the asymptotic minimum mean-square-error (or generalization error in the neural network setting). A similar phase transition was analyzed recently in the context of sparse high-dimensional linear regression by Reeves et al.
Many different measurement techniques are used to record neural activity in the brains of different organisms, including fMRI, EEG, MEG, lightsheet microscopy and direct recordings with electrodes. Each of these measurement modes have their advantages and disadvantages concerning the resolution of the data in space and time, the directness of measurement of the neural activity and which organisms they can be applied to. For some of these modes and for some organisms, significant amounts of data are now available in large standardized open-source datasets. I will report on our efforts to apply causal discovery algorithms to, among others, fMRI data from the Human Connectome Project, and to lightsheet microscopy data from zebrafish larvae. In particular, I will focus on the challenges we have faced both in terms of the nature of the data and the computational features of the discovery algorithms, as well as the modeling of experimental interventions.
1) The document presents a statistical modeling approach called targeted smooth Bayesian causal forests (tsbcf) to smoothly estimate heterogeneous treatment effects over gestational age using observational data from early medical abortion regimens.
2) The tsbcf method extends Bayesian additive regression trees (BART) to estimate treatment effects that evolve smoothly over gestational age, while allowing for heterogeneous effects across patient subgroups.
3) The tsbcf analysis of early medical abortion regimen data found the simultaneous administration to be similarly effective overall to the interval administration, but identified some patient subgroups where effectiveness may vary more over gestational age.
Difference-in-differences is a widely used evaluation strategy that draws causal inference from observational panel data. Its causal identification relies on the assumption of parallel trends, which is scale-dependent and may be questionable in some applications. A common alternative is a regression model that adjusts for the lagged dependent variable, which rests on the assumption of ignorability conditional on past outcomes. In the context of linear models, Angrist and Pischke (2009) show that the difference-in-differences and lagged-dependent-variable regression estimates have a bracketing relationship. Namely, for a true positive effect, if ignorability is correct, then mistakenly assuming parallel trends will overestimate the effect; in contrast, if the parallel trends assumption is correct, then mistakenly assuming ignorability will underestimate the effect. We show that the same bracketing relationship holds in general nonparametric (model-free) settings. We also extend the result to semiparametric estimation based on inverse probability weighting.
We develop sensitivity analyses for weak nulls in matched observational studies while allowing unit-level treatment effects to vary. In contrast to randomized experiments and paired observational studies, we show for general matched designs that over a large class of test statistics, any valid sensitivity analysis for the weak null must be unnecessarily conservative if Fisher's sharp null of no treatment effect for any individual also holds. We present a sensitivity analysis valid for the weak null, and illustrate why it is conservative if the sharp null holds through connections to inverse probability weighted estimators. An alternative procedure is presented that is asymptotically sharp if treatment effects are constant, and is valid for the weak null under additional assumptions which may be deemed reasonable by practitioners. The methods may be applied to matched observational studies constructed using any optimal without-replacement matching algorithm, allowing practitioners to assess robustness to hidden bias while allowing for treatment effect heterogeneity.
This document discusses difference-in-differences (DiD) analysis, a quasi-experimental method used to estimate treatment effects. The author notes that while widely applicable, DiD relies on strong assumptions about the counterfactual. She recommends approaches like matching on observed variables between similar populations, thoughtfully specifying regression models to adjust for confounding factors, testing for parallel pre-treatment trends under different assumptions, and considering more complex models that allow for different types of changes over time. The overall message is that DiD requires careful consideration and testing of its underlying assumptions to draw valid causal conclusions.
We present recent advances and statistical developments for evaluating Dynamic Treatment Regimes (DTR), which allow the treatment to be dynamically tailored according to evolving subject-level data. Identification of an optimal DTR is a key component for precision medicine and personalized health care. Specific topics covered in this talk include several recent projects with robust and flexible methods developed for the above research area. We will first introduce a dynamic statistical learning method, adaptive contrast weighted learning (ACWL), which combines doubly robust semiparametric regression estimators with flexible machine learning methods. We will further develop a tree-based reinforcement learning (T-RL) method, which builds an unsupervised decision tree that maintains the nature of batch-mode reinforcement learning. Unlike ACWL, T-RL handles the optimization problem with multiple treatment comparisons directly through a purity measure constructed with augmented inverse probability weighted estimators. T-RL is robust, efficient and easy to interpret for the identification of optimal DTRs. However, ACWL seems more robust against tree-type misspecification than T-RL when the true optimal DTR is non-tree-type. At the end of this talk, we will also present a new Stochastic-Tree Search method called ST-RL for evaluating optimal DTRs.
A fundamental feature of evaluating causal health effects of air quality regulations is that air pollution moves through space, rendering health outcomes at a particular population location dependent upon regulatory actions taken at multiple, possibly distant, pollution sources. Motivated by studies of the public-health impacts of power plant regulations in the U.S., this talk introduces the novel setting of bipartite causal inference with interference, which arises when 1) treatments are defined on observational units that are distinct from those at which outcomes are measured and 2) there is interference between units in the sense that outcomes for some units depend on the treatments assigned to many other units. Interference in this setting arises due to complex exposure patterns dictated by physical-chemical atmospheric processes of pollution transport, with intervention effects framed as propagating across a bipartite network of power plants and residential zip codes. New causal estimands are introduced for the bipartite setting, along with an estimation approach based on generalized propensity scores for treatments on a network. The new methods are deployed to estimate how emission-reduction technologies implemented at coal-fired power plants causally affect health outcomes among Medicare beneficiaries in the U.S.
Laine Thomas presented information about how causal inference is being used to determine the cost/benefit of the two most common surgical surgical treatments for women - hysterectomy and myomectomy.
We provide an overview of some recent developments in machine learning tools for dynamic treatment regime discovery in precision medicine. The first development is a new off-policy reinforcement learning tool for continual learning in mobile health to enable patients with type 1 diabetes to exercise safely. The second development is a new inverse reinforcement learning tools which enables use of observational data to learn how clinicians balance competing priorities for treating depression and mania in patients with bipolar disorder. Both practical and technical challenges are discussed.
The method of differences-in-differences (DID) is widely used to estimate causal effects. The primary advantage of DID is that it can account for time-invariant bias from unobserved confounders. However, the standard DID estimator will be biased if there is an interaction between history in the after period and the groups. That is, bias will be present if an event besides the treatment occurs at the same time and affects the treated group in a differential fashion. We present a method of bounds based on DID that accounts for an unmeasured confounder that has a differential effect in the post-treatment time period. These DID bracketing bounds are simple to implement and only require partitioning the controls into two separate groups. We also develop two key extensions for DID bracketing bounds. First, we develop a new falsification test to probe the key assumption that is necessary for the bounds estimator to provide consistent estimates of the treatment effect. Next, we develop a method of sensitivity analysis that adjusts the bounds for possible bias based on differences between the treated and control units from the pretreatment period. We apply these DID bracketing bounds and the new methods we develop to an application on the effect of voter identification laws on turnout. Specifically, we focus estimating whether the enactment of voter identification laws in Georgia and Indiana had an effect on voter turnout.
This document summarizes a simulation study evaluating causal inference methods for assessing the effects of opioid and gun policies. The study used real US state-level data to simulate the adoption of policies by some states and estimated the effects using different statistical models. It found that with fewer adopting states, type 1 error rates were too high, and most models lacked power. It recommends using cluster-robust standard errors and lagged outcomes to improve model performance. The study aims to help identify best practices for policy evaluation studies.
We study experimental design in large-scale stochastic systems with substantial uncertainty and structured cross-unit interference. We consider the problem of a platform that seeks to optimize supply-side payments p in a centralized marketplace where different suppliers interact via their effects on the overall supply-demand equilibrium, and propose a class of local experimentation schemes that can be used to optimize these payments without perturbing the overall market equilibrium. We show that, as the system size grows, our scheme can estimate the gradient of the platform’s utility with respect to p while perturbing the overall market equilibrium by only a vanishingly small amount. We can then use these gradient estimates to optimize p via any stochastic first-order optimization method. These results stem from the insight that, while the system involves a large number of interacting units, any interference can only be channeled through a small number of key statistics, and this structure allows us to accurately predict feedback effects that arise from global system changes using only information collected while remaining in equilibrium.
We discuss a general roadmap for generating causal inference based on observational studies used to general real world evidence. We review targeted minimum loss estimation (TMLE), which provides a general template for the construction of asymptotically efficient plug-in estimators of a target estimand for realistic (i.e, infinite dimensional) statistical models. TMLE is a two stage procedure that first involves using ensemble machine learning termed super-learning to estimate the relevant stochastic relations between the treatment, censoring, covariates and outcome of interest. The super-learner allows one to fully utilize all the advances in machine learning (in addition to more conventional parametric model based estimators) to build a single most powerful ensemble machine learning algorithm. We present Highly Adaptive Lasso as an important machine learning algorithm to include.
In the second step, the TMLE involves maximizing a parametric likelihood along a so-called least favorable parametric model through the super-learner fit of the relevant stochastic relations in the observed data. This second step bridges the state of the art in machine learning to estimators of target estimands for which statistical inference is available (i.e, confidence intervals, p-values etc). We also review recent advances in collaborative TMLE in which the fit of the treatment and censoring mechanism is tailored w.r.t. performance of TMLE. We also discuss asymptotically valid bootstrap based inference. Simulations and data analyses are provided as demonstrations.
We describe different approaches for specifying models and prior distributions for estimating heterogeneous treatment effects using Bayesian nonparametric models. We make an affirmative case for direct, informative (or partially informative) prior distributions on heterogeneous treatment effects, especially when treatment effect size and treatment effect variation is small relative to other sources of variability. We also consider how to provide scientifically meaningful summaries of complicated, high-dimensional posterior distributions over heterogeneous treatment effects with appropriate measures of uncertainty.
Climate change mitigation has traditionally been analyzed as some version of a public goods game (PGG) in which a group is most successful if everybody contributes, but players are best off individually by not contributing anything (i.e., “free-riding”)—thereby creating a social dilemma. Analysis of climate change using the PGG and its variants has helped explain why global cooperation on GHG reductions is so difficult, as nations have an incentive to free-ride on the reductions of others. Rather than inspire collective action, it seems that the lack of progress in addressing the climate crisis is driving the search for a “quick fix” technological solution that circumvents the need for cooperation.
This document discusses various types of academic writing and provides tips for effective academic writing. It outlines common academic writing formats such as journal papers, books, and reports. It also lists writing necessities like having a clear purpose, understanding your audience, using proper grammar and being concise. The document cautions against plagiarism and not proofreading. It provides additional dos and don'ts for writing, such as using simple language and avoiding filler words. Overall, the key message is that academic writing requires selling your ideas effectively to the reader.
Machine learning (including deep and reinforcement learning) and blockchain are two of the most noticeable technologies in recent years. The first one is the foundation of artificial intelligence and big data, and the second one has significantly disrupted the financial industry. Both technologies are data-driven, and thus there are rapidly growing interests in integrating them for more secure and efficient data sharing and analysis. In this paper, we review the research on combining blockchain and machine learning technologies and demonstrate that they can collaborate efficiently and effectively. In the end, we point out some future directions and expect more researches on deeper integration of the two promising technologies.
In this talk, we discuss QuTrack, a Blockchain-based approach to track experiment and model changes primarily for AI and ML models. In addition, we discuss how change analytics can be used for process improvement and to enhance the model development and deployment processes.
More from The Statistical and Applied Mathematical Sciences Institute (20)
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxEduSkills OECD
Iván Bornacelly, Policy Analyst at the OECD Centre for Skills, OECD, presents at the webinar 'Tackling job market gaps with a skills-first approach' on 12 June 2024
Walmart Business+ and Spark Good for Nonprofits.pdfTechSoup
"Learn about all the ways Walmart supports nonprofit organizations.
You will hear from Liz Willett, the Head of Nonprofits, and hear about what Walmart is doing to help nonprofits, including Walmart Business and Spark Good. Walmart Business+ is a new offer for nonprofits that offers discounts and also streamlines nonprofits order and expense tracking, saving time and money.
The webinar may also give some examples on how nonprofits can best leverage Walmart Business+.
The event will cover the following::
Walmart Business + (https://business.walmart.com/plus) is a new shopping experience for nonprofits, schools, and local business customers that connects an exclusive online shopping experience to stores. Benefits include free delivery and shipping, a 'Spend Analytics” feature, special discounts, deals and tax-exempt shopping.
Special TechSoup offer for a free 180 days membership, and up to $150 in discounts on eligible orders.
Spark Good (walmart.com/sparkgood) is a charitable platform that enables nonprofits to receive donations directly from customers and associates.
Answers about how you can do more with Walmart!"
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Chapter wise All Notes of First year Basic Civil Engineering.pptxDenish Jangid
Chapter wise All Notes of First year Basic Civil Engineering
Syllabus
Chapter-1
Introduction to objective, scope and outcome the subject
Chapter 2
Introduction: Scope and Specialization of Civil Engineering, Role of civil Engineer in Society, Impact of infrastructural development on economy of country.
Chapter 3
Surveying: Object Principles & Types of Surveying; Site Plans, Plans & Maps; Scales & Unit of different Measurements.
Linear Measurements: Instruments used. Linear Measurement by Tape, Ranging out Survey Lines and overcoming Obstructions; Measurements on sloping ground; Tape corrections, conventional symbols. Angular Measurements: Instruments used; Introduction to Compass Surveying, Bearings and Longitude & Latitude of a Line, Introduction to total station.
Levelling: Instrument used Object of levelling, Methods of levelling in brief, and Contour maps.
Chapter 4
Buildings: Selection of site for Buildings, Layout of Building Plan, Types of buildings, Plinth area, carpet area, floor space index, Introduction to building byelaws, concept of sun light & ventilation. Components of Buildings & their functions, Basic concept of R.C.C., Introduction to types of foundation
Chapter 5
Transportation: Introduction to Transportation Engineering; Traffic and Road Safety: Types and Characteristics of Various Modes of Transportation; Various Road Traffic Signs, Causes of Accidents and Road Safety Measures.
Chapter 6
Environmental Engineering: Environmental Pollution, Environmental Acts and Regulations, Functional Concepts of Ecology, Basics of Species, Biodiversity, Ecosystem, Hydrological Cycle; Chemical Cycles: Carbon, Nitrogen & Phosphorus; Energy Flow in Ecosystems.
Water Pollution: Water Quality standards, Introduction to Treatment & Disposal of Waste Water. Reuse and Saving of Water, Rain Water Harvesting. Solid Waste Management: Classification of Solid Waste, Collection, Transportation and Disposal of Solid. Recycling of Solid Waste: Energy Recovery, Sanitary Landfill, On-Site Sanitation. Air & Noise Pollution: Primary and Secondary air pollutants, Harmful effects of Air Pollution, Control of Air Pollution. . Noise Pollution Harmful Effects of noise pollution, control of noise pollution, Global warming & Climate Change, Ozone depletion, Greenhouse effect
Text Books:
1. Palancharmy, Basic Civil Engineering, McGraw Hill publishers.
2. Satheesh Gopi, Basic Civil Engineering, Pearson Publishers.
3. Ketki Rangwala Dalal, Essentials of Civil Engineering, Charotar Publishing House.
4. BCP, Surveying volume 1
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
B. Ed Syllabus for babasaheb ambedkar education university.pdf
2018 MUMS Fall Course - Essentials of Bayesian Hypothesis Testing - Jim Berger , September 4, 2018
1. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Lecture 2: Essentials of Bayesian hypothesis
testing
1
2. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Outline
• An introductory example
• General notation
• Precise and imprecise hypotheses
• Choice of prior distributions for testing
• Paradoxes
• Okham’s Razor
• Multiple hypotheses and sequential testing
• Psychokinesis example
2
4. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Hypotheses and data:
• Alvac had shown no effect (in many studies) as a vaccine against HIV.
• Aidsvax had shown no effect (in many studies) as a vaccine against
HIV.
Question: Would Alvac as a primer and Aidsvax as a booster work?
The Study: Conducted in Thailand with 16,395 individuals from the
general (not high-risk) population:
• 74 HIV cases reported in the 8198 individuals receiving placebos
• 51 HIV cases reported in the 8197 individuals receiving the treatment
4
5. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The test that was performed:
• Let p1 and p2 denote the probability of HIV in the placebo and
treatment populations, respectively. The usual estimates would be
ˆp1 =
74
8198
= .009027, ˆp2 =
51
8197
= .006222 .
• Test H0 : p1 = p2 versus H1 : p1 ≥ p2
• Normal approximation okay, so
z =
ˆp1 − ˆp2
ˆσ{ˆp1−ˆp2}
=
.009027 − .006222
.001359
= 2.06
is approximately N(θ, 1), where θ = (p1 − p2)/(.001359).
We thus test H0 : θ = 0 versus H1 : θ ≥ 0, based on z.
• Observed z = 2.06, so the p-value is (where Z is N(0, 1) and Phi is the
standard normal cdf) P(Z ≥ 2.06) = 1 − Φ(2.06) = 0.02.
• The problem: interpreting this as the odds being 50 to 1 in favor of H1.
5
6. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Bayesian analysis:
• Assign prior probabilities Pr(H0) and Pr(H1).
• Under H1, specify a prior distribution π1(θ) for θ ∈ (0, ∞).
• The prior probability of H0 and observing z is Pr(H0)f(z | 0), where
f(z | 0) is the standard normal density.
• The prior probability of H1 and observing z is Pr(H1)m(z | π1), where
m(z | π1) =
∞
0
f(z | θ)π1(θ)
is the marginal density of z under H1 and the prior π1.
• By Bayes theorem,
Pr(H0 | z) =
Pr(H0)f(z | 0)
Pr(H0)f(z | 0) + Pr(H1)m(z | π1)
=
1
1 + P r(H1)
P r(H0) B10
,
where B10(z) = m(z | π1)/f(z | 0) is the Bayes factor of H1 to H0.
6
7. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The nonincreasing prior π1 most favorable to H1 is π(θ) = Uniform(0, 2.95),
and yields
B10(2.06) =
2.95
0
1√
2π
e−(2.06−θ)2
/2 1
2.95 dθ
1√
2π
e−(2.06−0)2/2
= 5.63 ,
so that
Pr(H0 | z = 2.06) =
1
1 + P r(H1)
P r(H0) × 5.63
,
Here is this posterior probability for various values of Pr(H1):
Pr(H0) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Pr(H0 | 2.06) 0.03 0.06 0.10 0.14 0.20 0.28 0.37 0.50 0.70
Hence the common misinterpretation that p = 0.02 implies odds of 50 in
favor of the alternative is simply very wrong.
7
8. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The following studies
• look at large collections of published studies where 0 < p < 0.05;
• compute a Bayes factor, B01 for each study;
• graph the Bayes factors versus the corresponding p-values.
The first two graphs are for 272 ‘significant’ epidemiological studies with
two different choices of the prior; the third for 50 ‘significant’ meta-analyses
(these three from J.P. Ioannides, Am J Epidemiology, 2008); and the last is
for 314 ecological studies (reported in Elgersma and Green, 2011).
8
10. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
General Notation
• X | θ ∼ f(x | θ).
• To test: H0 : θ ∈ Θ0 vs H1 : θ ∈ Θ1 .
• Prior distribution:
– Prior probabilities Pr(H0) and Pr(H1) of the hypotheses.
– Prior densities (often proper), π0(θ) and π1(θ), on Θ0 and Θ1.
∗ πi(θ) would be a point mass if Θi is a point.
• Marginal likelihoods under the hypotheses:
m(x | Hi) =
Θi
f(x | θ)πi(θ) dθ, i = 0, 1 .
• Bayes factor of H0 to H1:
B01 =
m(x | H0)
m(x | H1)
.
10
12. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Conclusions from posterior probabilities or Bayes factors:
• In some sense, H0 would be accepted if Pr(H0 | x) > Pr(H1 | x) but
there is no real need to state this decision.
• Normally, just the Bayes factor B01 (or B10) is presented because
– it can be combined with differing personal prior odds;
– the ‘default’ Pr(H0) = Pr(H1) can used.
• For interpreting B10, Jeffreys (1961) suggested the scale
B10 Strength of evidence
1:1 to 3:1 Barely worth mentioning
3:1 to 10:1 Substantial
10:1 to 30:1 Strong
30:1 to 100:1 Very strong
> 100:1 Decisive
12
13. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Precise or Imprecise Hypothesis
(often Point Null or One-Sided)
A Key Issue: Is the precise hypothesis being tested plausible?
A precise hypothesis is an hypothesis of lower dimension than the
alternative (e.g. H0 : µ = 0 versus H0 : µ = 0).
A precise hypothesis is plausible if it has a reasonable prior probability of
being true. H0 : there is no Higgs boson particle, is plausible.
Example: Let θ denote the difference in mean treatment effects for cancer
treatments A and B, and test H0 : θ = 0 versus H1 : θ = 0.
Scenario 1: Treatment A = standard chemotherapy
Treatment B = standard chemotherapy + steroids
Scenario 2: Treatment A = standard chemotherapy
Treatment B = a new radiation therapy
H0 : θ = 0 is plausible in Scenario 1, but not in Scenario 2; in the latter
case, instead test H0 : θ < 0 versus H1 : θ > 0.
13
14. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Plausible precise null hypotheses:
• H0 : Gene A is not associated with Disease B.
• H0: There is no psychokinetic effect.
• H0: Vitamin C has no effect on the common cold.
• H0: A new HIV vaccine has no effect.
• H0: Cosmic microwave background radiation is isotropic.
• H0 : Males and females have the same distribution of eye color.
• H0 : Pollutant A does not cause disease B.
Implausible precise null hypotheses:
• H0 : Small mammals are as abundant on livestock grazing land as on
non-grazing land
• H0 : Bird abundance does not depend on the type of forest habitat
they occupy
• H0 : Children of different ages react the same to a given stimulus.
14
15. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Approximating a believable precise hypothesis by an exact
precise null hypothesis
A precise null, like H0 : θ = θ0, is typically never true exactly; rather, it is
used as a surrogate for a ‘real null’
Hǫ
0 : |θ − θ0| < ǫ, ǫ small .
(Even if θ = θ0 in nature, the experiment studying θ will typically have a small
unknown bias, introducing an ǫ.)
Result (Berger and Delampady, 1987 Statistical Science):
Robust Bayesian theory can be used to show that, under reasonable
conditions, if ǫ < 1
4 σˆθ, where σˆθ is the standard error of the estimate of θ,
then
Pr(Hǫ
0 | x) ≈ Pr(H0 | x) .
Note: Typically, σˆθ ≈ c√
n
, where n is the sample size, so for large n the
above condition can be violated, and using a precise null may not be
appropriate, even if the real null is believable.
15
16. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Posterior probabilities can equal p-values in one-sided testing:
Normal Example:
• X | θ ∼ N(x | θ, σ2
)
• One-sided testing
H0 : θ ≤ θ0 vs H1 : θ > θ0
• Choose the usual estimation objective prior π(θ) = 1, for which the
posterior distribution, π(θ | x), can be shown to be N(θ | x, σ2
).
• Posterior probability of H0:
Pr(H0 | x) = Pr(θ ≤ θ0 | x) = Φ
θ0 − x
σ
= 1 − Φ
x − θ0
σ
= Pr(X > x | θ0) = p-value .
• It has been argued (e.g., Berger and Mortera, JASA99) that objective
testing in one-sided cases should use priors more concentrated than
π(θ) = 1, so p-values are still questionable in one-sided testing.
16
17. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Choice of Prior Distributions in Testing
A. Choosing priors for “common parameters” in testing:
Common parameters in densities under two hypotheses are parameters that
have the same role and and are present in both.
Example: If the data xi are i.i.d. N(xi | θ, σ2
), and it is desired to test
H0 : θ = 0 versus H1 : θ = 0 ,
the density under H0 is N(xi | 0, σ2
) and that under H1 is N(xi | θ, σ2
),
so σ2
is a common parameter to both densities and has the same role in
each.
Priors for common parameters: Use the standard ‘objective prior’ for
the common parameters.
Example: For the normal testing problem, the standard objective prior for
the variance is π(σ2
) = 1/σ2
.
17
18. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Example: An example requiring parameter transformation to obtain
‘common parameters’ (Dass and Berger, 2003):
• Vehicle emissions data from McDonald et. al. (1995)
• Data consists of 3 types of emissions, hydrocarbon (HC), carbon
monoxide (CO) and nitrogen oxides (NOx) at 4 different mileage states
0, 4000, 24,000(b) and 24,000(a).
Data for 4,000 miles
HC 0.26 0.48 0.40 0.38 0.31 0.49 0.25 0.23
CO 1.16 1.75 1.64 1.54 1.45 2.59 1.39 1.26
NOx 1.99 1.90 1.89 2.45 1.54 2.01 1.95 2.17
HC 0.39 0.21 0.22 0.45 0.39 0.36 0.44 0.22
CO 2.72 2.23 3.94 1.88 1.49 1.81 2.90 1.16
NOx 1.93 2.58 2.12 1.80 1.46 1.89 1.85 2.21
18
19. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Goal: Based on independent data X = (X1, . . . , Xn), test whether the i.i.d.
Xi follow the Weibull or the lognormal distribution given, respectively, by
H0 : fW (x | β, γ) =
γ
β
x
β
γ−1
exp −
x
β
γ
, x > 0, β > 0, γ > 0.
H1 : fL(x | µ, σ2
) =
1
x
√
2πσ2
exp
−(log x − µ)2
2σ2
, x > 0, σ > 0 .
Both distributions are location-scale distributions in y = log x, i.e., are of
the form
1
σ
g
y − µ
σ
for some density g(·). To see for the Weibull, define y = log(x), µ = log(β),
and σ = 1/γ; then
fW (y | µ, σ) =
1
σ
e(y−µ)/σ
e−e(y−µ)/σ
.
19
20. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Berger, Pericchi and Varshavsky (1998) argue that, for two hypotheses
(models) with the same invariance structure (here location-scale
invariance), one can use the right-Haar prior (usually improper) for both.
Here the right-Haar prior is
πRH
(µ, σ) =
1
σ
dσdµ .
The justification is called predictive matching and goes as follows (related
to arguments in Jeffreys 1961):
• With only two observations (y1, y2), one cannot possibly distinguish
between fW (y | µ, σ) and fL(y | µ, σ) so we should have B01(y1, y2) = 1.
• Lemma:
∞
0
∞
−∞
1
σ
g
y1 − µ
σ
1
σ
g
y2 − µ
σ
πRH
(µ, σ)dµdσ =
1
2|y1 − y2|
for any density g(·), implying B01(y1, y2) = 1 (as desired) for two
location-scale densities when the right-Haar prior is used.
20
21. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Using the right-Haar prior for both models, calculus then yields that the
Bayes factor of H0 to H1 is
B01(X) =
Γ(n)nn
π(n−1)/2
Γ(n − 1/2)
∞
0
v
n
n
i=1
exp
yi − ¯y
syv
−n
dv,
where ¯y = 1
n
n
i=1 yi and s2
y = 1
n
n
i=1(yi − ¯y)2
. Note: B01(X) is also the
classical UMP invariant test statistic.
As an example, consider four of the car emission data sets, each giving the
carbon monoxide emission at a different mileage level.
For testing, H0 : Lognormal versus H1 : Weibull, the results were as
follows:
Data set at Mileage level
0 4000 24,000 30,000
B01 0.404 0.110 0.161 0.410
21
22. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
B. Choosing priors for non-common parameters
If subjective choice is not possible, be aware that
• Vague proper priors are often horrible: for instance, if X ∼ N(x | µ, 1)
and we test H0 : µ = 0 versus H1 : µ = 0 with a Uniform(−c, c) prior
for θ, the Bayes factor is
B01(c) =
f(x | 0)
c
−c
f(x | µ)(2c)−1dµ
≈
2c f(x | 0)
∞
−∞
f(x | µ)dµ
= 2c f(x | 0)
for large c, which depends dramatically on the choice of c.
• Improper priors are problematical, because they are unnormalized;
should we use π(µ) = 1 or π(µ) = 2, yielding
B01 =
f(x | 0)
∞
−∞
f(x | µ)(1)dµ
= f(x | 0) or B01 =
f(x | 0)
∞
−∞
f(x | µ)(2)dµ
=
1
2
f(x | 0) ?
• It is curious here that use of vague proper priors is much worse than
use of objective improper priors (though neither can be justified).
22
23. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Various proposed default priors for non-common parameters
• Conventional priors
– Jeffreys choice
– The ‘robust prior’ and Bayesian t-test
• Priors induced from a single prior
• Intrinsic priors (derived from data or imaginary data)
23
24. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Conventional priors
Jeffreys choices for the normal testing problem:
• Data: X = (X1, X2, . . . , Xn)
• We are testing
H0 : Xi ∼ N(xi | 0, σ2
0) versus H1 : Xi ∼ N(xi | µ, σ2
1).
• We thus seek π0(σ2
0) and π1(µ, σ2
1) = π1(µ | σ2
1)π1(σ2
1).
• Since σ2
0 and σ2
1 are common parameters with the same role, Jeffreys
used the same objective prior π0(σ2
0) = 1/σ2
0 and π1(σ2
1) = 1/σ2
1.
• π1(µ | σ2
1) must be proper (and not vague), since µ only occurs in H1.
Jeffreys argued that it
– should be centered at zero (H0);
– should have scale σ1 (the ‘natural’ scale of the problem);
– should be symmetric around zero;
– should have no moments (more on this later).
The ‘simplest prior’ satisfying these is the Cauchy(µ | 0, σ1) prior,
24
25. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
resulting in
Jeffreys proposal:
π0(σ2
0) =
1
σ2
0
, π1(µ, σ2
1) =
1
πσ1(1 + (µ/σ1)2)
·
1
σ2
1
.
Predictive matching argument for these priors:
For any location scale density 1
σ g(y−µ
σ ) and one observation y
under H1 : µ = 0,
m0(y) =
1
σ
g
y − µ
σ
1
σ2
dσ2
=
2
|y|
;
under H1 : µ = 0 and for any proper prior of the form 1
σ h µ
σ ,
m1(y) =
1
σ
g
y − µ
σ
1
σ
h
µ
σ
1
σ2
dµ dσ2
=
2
|y|
,
so that B01 = 1 for one observation, as should be the case. (Of course, this
doesn’t say that the prior for µ should be Cauchy.)
25
26. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The robust prior and Bayesian t-test
• Computation of B01 for Jeffreys choice of prior requires
one-dimensional numerical integration. (Jeffreys gave a not-very-good
numerical approximation.)
• An alternative is the ‘robust prior’ from Berger (1985) (a generalization
of the Strawderman (1971) prior), to be discussed in later lectures.
– This prior satisfies all desiderata of Jeffreys;
– has identical tails and varies little from the Cauchy prior;
– yields an exact expression for the Bayes factor (Pericchi and Berger)
B01 =
2
n + 1
n − 2
n − 1
t2
1 +
t2
n − 1
− n
2
1 − 1 +
2t2
n2 − 1
−( n
2 −1) −1
,
where t =
√
n¯x/ (xi − ¯x)2/(n − 1) is the usual t-statistic. As t → 0,
B01 → 2(n + 1). For n = 2, this is to be interpreted as
B01 =
2
√
2 t2
√
3(1 + t2) log(1 + 2t2/3)
.
26
27. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Encompassing approach: inducing priors on hypotheses from a
single prior
Sometimes, instead of separately assessing Pr(H0), Pr(H1), π0, π1, it is
possible to start with an overall prior π(θ) and deduce the
Pr(H0), Pr(H1), π0, π1:
Pr(H0) =
Θ0
π(θ) dθ and Pr(H1) =
Θ1
π(θ) dθ
π0(θ) =
1
Pr(H0)
π(θ)1Θ0 (θ) and π1(θ) =
1
Pr(H1)
π(θ)1Θ1 (θ) .
Note: To be sensible, the induced Pr(H0), Pr(H1), π0, π1 must themselves
be sensible.
Example: Intelligence testing:
• X | θ ∼ N(x, | θ, 100), and we observe x = 115.
• To test ‘below average’ versus ’above average’
H0 : θ ≤ 100 vs H1 : θ > 100.
27
28. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
• It is ‘known’ that θ ∼ N(θ | 100, 225).
• induced prior probabilities of hypotheses
Pr(H0) = Pr(θ ≤ 100) = 1
2 = Pr(H1)
• induced densities under each hypothesis:
π0(θ) = 2 N(θ | 100, 225)I(−∞,100)(θ)
π1(θ) = 2 N(θ | 100, 225)I(100,∞)(θ)
• Of course, we would not have needed to formally derive these.
– From the original encompassing prior π(θ), we can derive the
posterior and θ | x = 115 ∼ N(110.39, 69.23).
– Then directly compute the posterior probabilities:
Pr(H0 | x = 115) = Pr(θ ≤ 100 | x = 115) = 0.106
Pr(H1 | x = 115) = Pr(θ > 100 | x = 115) = 0.894
28
29. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Intrinsic priors
Discussion of these can be found in Berger and Pericchi (2001). One
popular such prior, that applies to our testing problem, is the intrinsic
prior defined as follows:
• Let πO
(θ) be a good estimation objective prior (using a constant prior
will almost always work fine), with resulting posterior distribution and
marginal distribution for data x given, respectively, by
πO
(θ | x) = f(x | θ)πO
(θ)/mO
(x), mO
(x) = f(x | θ)πO
(θ) dθ .
• Then the intrinsic prior (which will be proper) is
πI
(θ) = πO
(θ | x∗
)f(x∗
| θ0) dx∗
,
with x∗
= (x∗
1, . . . , x∗
q) being imaginary data of the smallest sample size
q such that mO
(x∗
) < ∞ (this is an imaginary bootstrap construction).
29
30. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
πI
(θ) is often available in closed form, but even if not, computation of the
resulting Bayes factor is often a straightforward numerical exercise.
• The resulting Bayes factor is
B01(x) =
f(x | θ0)
f(x | θ)πI(θ)dθ
=
f(x | θ0)
mO(x | x∗)f(x∗ | θ0)dx∗
.
Example (Higgs Boson Example): Test H0 : θ = 0 versus H0 : θ > 0, based
on Xi ∼ f(xi | θ) = (θ + b) exp{−(θ + b)xi}, where b is known;
• Suppose we choose πO
(θ) = 1/(θ + b) (the more natural square root is
harder to work with).
• A minimal sample size for the resulting posterior to be proper is q = 1.
• Computation then yields πI
(θ) = πO
(θ | x∗
1)f(x∗
1 | 0)dx∗
1 = b/(θ + b)2
.
30
32. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Normal Example (used to illustrate the various “paradoxes”):
• Xi | θ
i.i.d.
∼ N(xi | θ, σ2
), σ2
known.
• Test H0 : θ = 0 versus H1 : θ = 0.
• Can reduce to sufficient statistic ¯x ∼ N(¯x | θ, σ2
/n).
• Prior on H1: π1(θ) = N(θ | 0, v2
0)
• Marginal likelihood under H1: m1(¯x) = N(¯x | 0, v2
0 + σ2
/n).
• posterior probability:
Pr(H0 | ¯x) =
1 +
Pr(H1)
Pr(H0)
1
(2π(v2
0+σ2/n))1/2 exp −1
2
1
v2
0+σ2/n
(¯x)2
1
(2πσ2/n)1/2 exp −1
2
1
σ2/n
(¯x)2
−1
=
1 +
Pr(H1)
Pr(H0)
exp{1
2
z2
[1 + σ2
nv2
0
]−1
}
{1 + nv2
0/σ2}1/2
−1
,
where z =
|¯x|
σ/
√
n
is the usual (frequentist) test statistic for this problem.
32
33. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
An Aside: Comparing Pr(H0 | ¯x) with the p-value for various z and n and
with v2
0 = σ2
(the ‘unit information’ prior):
z p-value n = 5 n = 20 n = 100
1.645 0.1 0.44 0.56 0.72
1.960 0.05 0.33 0.42 0.60
2.576 0.01 0.13 0.16 0.27
33
34. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The Jeffreys-Lindley ‘Paradox’
In the normal testing example, for fixed z and large n,
Pr(H0 | ¯x) =
1 +
Pr(H1)
Pr(H0)
exp{1
2
z2
[1 + σ2
nv2
0
]−1
}
{1 + nv2
0/σ2}1/2
−1
≈ 1 −
Pr(H1)
Pr(H0)
σ
√
n v0
exp{
1
2
z2
} −→ 1 as n → ∞ ,
so that a classical test can strongly reject H0 (which happens when z is
moderately large) and the Bayesian analysis can, at the same time, strongly
support H0 (if n is so large that exp{1
2 z2
}/
√
n is small, even though z is
moderately large); reaching opposite conclusions is the ‘paradox.’
• This is not a paradox in the true sense, since it is just mathematics.
• Robust Bayesian resolution: H0 : θ = 0 is just an approximation to
H0 : |θ| < ǫ, where ǫ can reflect reality or just experimental bias. The
approximation is only accurate when ǫ < σ/(4
√
n) (Berger and Delampady,
1987); thus, for very large n, it is not reasonable to use H0 : θ = 0 as the null
hypothesis, so the ‘paradox’ becomes vacuous.
34
35. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The Jeffreys-Lindley ‘Paradox’and Experimental Bias
Suppose that H0 is truly precise (e.g. 0 psychic effect or ‘no Higgs boson’),
but that the experiment has some bias b ∼ N(b | 0, δ2
). Then
Pr(H0 | ¯x) =
1 +
Pr(H1)
Pr(H0)
exp{1
2
z2
b [1 + (δ2+σ2/n)
v2
0
]−1
}
{1 +
v2
0
δ2+σ2/n
}1/2
−1
≈ 1 −
Pr(H1)
Pr(H0)
δ2 + σ2/n
v0
exp{
1
2
z2
b },
when δ2 + σ2/n is small, and where zb = |¯x|/ δ2 + σ2/n can be thought
of as standard normal under H0 in the presence of the bias. Then
lim
n→∞
Pr(H0 | ¯x) = 1 −
Pr(H1)
Pr(H0)
δ
v0
exp{
(¯x)2
2δ2
}
which does not go to 1. Also
Pr(H0 | ¯x) ≈ 1 −
Pr(H1)
Pr(H0)
δ
v0
exp{
z2
σ2
2nδ2
} ,
for the interesting range of z2
σ2
/[nδ2
].
35
36. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Experimental biases:
Figure 1: Historical record of values of some particle properties published over
time, with quoted error bars (Particle Data Group).
36
37. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Figure 2: Historical record of values of some particle properties published over
time, with quoted error bars (Particle Data Group).
37
38. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The Barlett ‘Paradox’
In the normal testing example, when the prior variance v2
0 is large (i.e., a
vague proper prior is being used),
Pr(H0 | ¯x) =
1 +
Pr(H1)
Pr(H0)
exp{1
2 z2
[1 + σ2
nv2
0
]−1
}
{1 + nv2
0/σ2}1/2
−1
≈ 1 −
Pr(H1)
Pr(H0)
σ
√
n v0
exp{
1
2
z2
},
so that, if v2
0 → ∞, then Pr(H0 | ¯x) → 1, so that proper priors in testing
can not be ‘arbitrarily flat’ (as we saw earlier).
38
39. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Ockham’s Razor
• Attributed to thirteen-century Franciscan monk William of Ockham
(Occam in latin)
“Pluralitas non est ponenda sine necessitate.”
(Plurality must never be posited without necessity.)
“Frustra fit per plura quod potest fieri per pauciora.”
(It is vain to do with more what can be done with fewer.)
• Preferring the simpler of two hypothesis to the more complex when
both agree with data is an old principle in science.
• Regard H0 as simpler than H1 if it makes sharper predictions about
what data will be observed.
• Models are more complex if they have extra adjustable parameters that
allow them to be tweaked to accommodate a wider variety of data.
– “coin is fair” is a simpler model than “coin has unknown bias θ”
– s = a + ut + 1
2 gt2
is simpler than s = a + ut + 1
2 gt2
+ ct3
39
40. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Example: Perihelion of Mercury (with Bill Jefferys)
In the 19th century it was known that there was an unexplained residual
motion of Mercury’s perihelion (the point in its orbit where the planet was
closest to the Sun) in the amount of approximately 43 seconds of arc per
century.
Various hypotheses:
• A planet ‘Vulcan’ close to the sun.
• A ring of matter around the sun.
• Oblateness of the sun.
• Law of gravity is not inverse square but inverse (2 + ǫ).
All these hypotheses had a parameter that could be adjusted to deal with
whatever data on the motion of Mercury existed.
Data in 1920: X = 41.6 where X ∼ N(θ | 22
), θ being the perihelion
advance of Mercury, and the measurement standard deviation is 2.
40
41. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Prior (before data) for gravity model MG: πG(θ) = N(θ | 0, 502
).
• Symmetric about 0 (corresponding to inverse square law).
• Decreasing away from zero; normality is convenient.
• Initially, τ = 50, because a gravity effect which would yield θ > 100
would have had other observed effects.
• We will also consider utilization of classes of priors:
– The class of all N(θ | 0, τ2
) priors, τ > 0.
– The class of all symmetric priors that are nonincreasing in |θ|.
General Relativity (1915) model ME: Predicted θE = 42.9, so no prior
is needed. (Thus this is a ‘simpler’ model.)
Bayes factor of ME to MG = 28.6, strongly favoring General Relativity,
even though the gravity model could fit the data better than General
Relativity. (The computation is an exercise.)
41
42. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Multiple Hypotheses and Sequential Testing
Two nice features of Bayesian testing:
1. Multiple hypotheses can be easily handled.
Example; In a paired difference experiment, Xi is the observed difference in
effect between a subject receiving Treatment 1 and the paired subject
receiving Treatment 2.
Suppose the Xi are i.i.d from the N(xi | θ, 1) density, where
θ = mean effect of Treatment 1 - mean effect of Treatment 2
Standard Testing Formulation:
H0 : θ = 0 (no difference in treatments)
Ha : θ = 0 (a difference exists)
A More Revealing Formulation:
H0 : θ = 0 (no difference)
H1 : θ < 0 ( Treatment 2 is better)
H2 : θ > 0 ( Treatment 1 is better)
42
43. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
2. Interim or sequential analysis does not affect the Bayesian answer.
• In interim or sequential analysis, one periodically looks at the
accumulated data during a study, with the option of stopping the study
and drawing a conclusion at each look.
• In classical statistics, one must increase the error probability that is
used with each look at the data (since each of the analysis stages
increases the probability of having an incorrect rejection).
– Example: In testing whether a normal mean is zero or not (variance
known), and one wants significance at the α = 0.05 level,
∗ If a fixed sample size of 20 was used, reject the null hypothesis if
|¯x20|/
√
20 > 1.96.
∗ If one is going to first take 10 observations and see if rejection is
possible and, if not, take another 10 observations, then one rejects only
if |¯x10|/
√
10 > 2.178 (first stage) or |¯x20|/
√
20 > 2.178 (second stage).
– In the medical literature this is called “spending α for looks at the data.”
• Bayesians do not adjust for looks at the data (more generally called the
stopping rule principle).
43
44. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The three hypothesis testing example, done in a fully sequential
Bayesian way.
• Thus after each observation is taken, the current posterior probability
of each hypothesis is calculated
Prior Distribution:
• Assign H0 and Ha prior probabilities of 1/2 each
• On Ha, assign θ the “default” Normal(0,2) distribution (so that
Pr(H1) = Pr(H2) = 1/4)
(The N(0, 2) prior is just for illustration: Berger and Mortera, 1999, use a
much better intrinsic prior here; see also Barbieri and Liseo, 2006).
Posterior Distribution:
After observing x = (x1, x2, . . . , xn), compute the posterior probabilities of
the various hypotheses, i.e.
Pr(H0 | x), Pr(H1 | x), Pr(H2 | x),
and Pr(Ha | x) = Pr(H0 | x) + Pr(H0 | x)
44
45. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Writing the normal density of the Xi as f(xi | θ, 1), the Bayes factor of H0
to Ha is
Bn =
n
i=1 f(xi | 0, 1)
π(θ) n
i=1 f(xi | θ, 1)dθ
=
√
1 + 2n e−n¯x2
n/(2+ 1
n )
.
The posterior probabilities can then be computed to be (where Φ denotes
the standard normal cdf)
Pr(H0 | x) =
Bn
1 + Bn
Pr(H1 | x) =
Φ −
√
n ¯xn/ 1 + 1
2n
1 + Bn
Pr(H2 | x) = 1 − Pr(H0 | x) − Pr(H1 | x)
45
47. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Comments:
(i) Neither multiple hypothesis nor the sequential aspect caused
difficulties. There is no penalty (e.g. ‘spending α’) for looks at the data
(ii) Quantification of the support for H0 : θ = 0 is direct. At the 3rd
observations, Pr(H0 | x) = .453, at the end, Pr(H0 | x) = .082
(iii) H1 can be ruled out almost immediately
(iv) For testing H0 : θ = 0 versus Ha : θ = 0, the posterior probabilities are
also what are called conditional frequentist error probabilities, as will
be seen in Lecture 2.
– Thus frequentists also don’t have to adjust for looks at the data, if
they use the correct frequentist procedures.
47
48. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
There are two consequences of this result:
1. Use of the Bayes factor gives experimenters the freedom to employ
optional stopping without penalty.
2. There is no harm if ‘undisclosed optional stopping’ is used (common in
some areas of psychology), as long as the Bayes factor is used to assess
significance. In particular, it is a consequence that an experimenter
cannot fool someone through use of undisclosed optional stopping.
The Philosophical Puzzle: How can there be no penalty for interim analysis?
• Bayesian analysis is just probability theory and so cannot be wrong
foundationally.
• The ‘statistician’s client with a grant application example.’
But it is difficult; as Savage (1961) said “When I first heard the stopping rule
principle from Barnard in the early 50’s, I thought it was scandalous that anyone
in the profession could espouse a principle so obviously wrong, even as today I
find it scandalous that anyone could deny a principle so obviously right.”
48
49. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
The reason the Bayes factor does not depend on the stopping rule:
Optional stopping alters the data density to be
τN (x1, x2, . . . , xN )
N
i=1
f(xi | θ, 1) ,
where N is the (random) time at which one stops taking data and
τN (x1, x2, . . . , xN ) gives the probability (often 0 or 1) of stopping sampling.
Then
BN =
τN (x1, x2, . . . , xN ) N
i=1 f(xi | 0, 1)
π(θ)τN (x1, x2, . . . , xN )
N
i=1 f(xi | θ, 1)dθ
=
N
i=1 f(xi | 0, 1)
π(θ)
N
i=1 f(xi | θ, 1)dθ
.
49
50. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Psychokinesis Example
Do people have the ability to perform psychokinesis, affecting objects with
thoughts?
The experiment:
Schmidt, Jahn and Radin (1987) used electronic and
quantum-mechanical random event generators with visual
feedback; the subject with alleged psychokinetic ability tries to
“influence” the generator.
50
51. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Stream of particles
Quantum
Gate
Red light
Green light
Quantum mechanics
implies the particles are
50/50 to go to each light
Tries to make
the particles to
go to red light
ÁÁ
51
52. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Data and model:
• Each “particle” is a Bernoulli trial (red = 1, green = 0)
θ = probability of “1”
n = 104, 490, 000 trials
X = # “successes” (# of 1’s), X ∼ Binomial(n, θ)
x = 52, 263, 470 is the actual observation
To test H0 : θ = 1
2 (subject has no influence)
versus H1 : θ = 1
2 (subject has influence)
• P-value = Pθ= 1
2
(|X − n
2 | ≥ |x − n
2 |) ≈ .0003.
Is there strong evidence against H0 (i.e., strong evidence that the
subject influences the particles) ?
52
53. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Bayesian Analysis: (Jefferys, 1990)
Prior distribution:
Pr(Hi) = prior probability that Hi is true, i = 0, 1;
On H1 : θ = 1
2 , let π(θ) be the prior density for θ.
Subjective Bayes: choose the Pr(Hi) and π(θ) based on personal beliefs
Objective (or default) Bayes: choose
Pr(H0) = Pr(H1) = 1
2
π(θ) = 1 (on 0 < θ < 1)
53
54. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Posterior probability of hypotheses:
Pr(H0|x) =
f(x | θ = 1
2
) Pr(H0)
Pr(H0) f(x | θ = 1
2
) + Pr(H1) f(x | θ)π(θ)dθ
For the objective prior,
B01 = likelihood of observed data under H0
‘average′ likelihood of observed data under H1
=
f(x | θ= 1
2
)
1
0 f(x | θ)π(θ)dθ
≈ 12
Pr(H0 | x = 52, 263, 470) ≈ 0.92 (recall, p-value ≈ .0003)
Posterior density on H1 : θ = 1
2 is
π(θ|x, H1) ∝ π(θ)f(x | θ) ∝ 1 × θx
(1 − θ)n−x
,
the Be(θ | 52, 263, 471 , 52, 226, 531) density.
54
56. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
Choice of the prior density or weight function, π, on { θ : θ = 1
2
}:
Consider πr(θ) = U(θ | 1
2 − r, 1
2 + r) the uniform density on (1
2 − r, 1
2 + r)
Subjective interpretation: r is the largest chance in success probability
that you would expect, given that ESP exists. And you give equal
probability to all θ in the interval (1
2 − r, 1
2 + r) .
Resulting Bayes factor (letting FBe(· | a, b) denote the CDF of the
Beta(a, b) distribution)
B(r) =
f(x | 1/2)
1
0
f(x | θ)πr(θ) dθ
=
n
x
(n + 1) r
2n−1
[FB2 − FB1]−1
where
FB2 = FBe(1
2 + r | x + 1, n − x + 1) and
FB1 = FBe(1
2 − r | x + 1, n − x + 1)
For example, B(0.25) ≈ 6
56
57. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
0.00016 0.00020 0.00024 0.00028
0.0070.0080.0090.010
r
BF
57
58. Model Uncertainty SAMSI – Fall,2018
✬
✫
✩
✪
r = largest increase in success probability that would be expected, given
ESP exists.
the minimum value of B(r) is 1
158 , attained at the minimizing value of
r = .00024
Conclusion: Although the p-value is small (.0003), for typical prior beliefs
the data would provide evidence for the simpler model H0 : no ESP.
Only if one believed a priori that |θ − 1
2 | ≤ .0021, would the evidence
for H1 be at least 20 to 1.
58