1) The document analyzes the relationship between extreme precipitation events in the Gulf of Mexico region and climate change indicators like sea surface temperature and atmospheric CO2 levels.
2) It finds a statistically significant relationship, with extreme precipitation events becoming more likely as Gulf SSTs and CO2 levels increase. Using this relationship, it estimates that Hurricane Harvey in 2017 was a "very likely" 1,000-year rainfall event or rarer when accounting for climate change factors.
3) The analysis uses several statistical methods like generalized extreme value distributions and point process models to analyze extreme precipitation event data at different spatial scales and relate the frequency of extreme events to climate change indicators while accounting for seasonal and other factors.
Presentation of Four Centennial-long Global Gridded Datasets of the Standardi...Agriculture Journal IJOEAR
Abstract— In this article four global gridded datasets of the Standardized Precipitation Index (SPI) are presented. They are computed from four different data sources: UDEL/GEOG/CCR v3.02, GPCC/ v7.0, NOAA-CIRES 20CR v2c and ECMWF ERA-20C each covering more than a century-long period. The SPI is calculated for the most frequently used time windows of 1, 3, 6, and 12 months. UDEL/GEOG/CCR v3.02 and GPCC/ v7.0 are used in the highest native resolution of 0.5×0.5° whilst NOAA-CIRES 20CR v2c and ECMWF ERA-20C are interpolated at 1.5×1.5° and 0.5×0.5° correspondingly. In contrast to some other indices, for example the popular Palmer Drought Severity Index (PDSI), SPI has significant advantages such as simplicity, suitability on variable time scales and robustness rooted in a solid theoretical development. SPI has been selected by the World Meteorological Organization (WMO) as a key indicator for monitoring drought ('Lincoln declaration'). As a result, drought monitoring centres worldwide are effectively exploiting this index and the National Meteorological and Hydrological Services (NMHSs) are encouraged to use it for monitoring meteorological droughts. These facts and the strong conviction of the authors that the free exchange of data and software services are а basis of effective scientific collaboration, are the main motivators to provide these datasets free of charge at ftp://xeo.cfd.meteo.bg/SPI/. The paper briefly presents some possible applications of the SPI data, revealing its suitability for various objective long-term drought studies at any geographical location.
Ten most popular software for prediction of cyclonic stormsMrinmoy Majumder
In recent years the frequency and intensity of cyclones and hurricanes have been increased manifold compared to the last decade. As a result, the necessity for the development of computer models to predict the track, intensity, and time of occurrence of cyclonic storms has increased to avoid loss of life and prevention of damages to public properties. In this presentaion I had tried to highlight the ten most used models in this aspect which are responsible for saving millions of life and their livelihood.
It is based on Journal Paper named
"Mukherjee, M.K.2013, ’Flood Frequency Analysis of River Subernarekha, India, Using Gumbel’s extreme Value Distribution’, IJCER,Vol-3,Issue-7,pp-12-18."
I have studied the journal and make a PPT in the following.
I
Presentation of Four Centennial-long Global Gridded Datasets of the Standardi...Agriculture Journal IJOEAR
Abstract— In this article four global gridded datasets of the Standardized Precipitation Index (SPI) are presented. They are computed from four different data sources: UDEL/GEOG/CCR v3.02, GPCC/ v7.0, NOAA-CIRES 20CR v2c and ECMWF ERA-20C each covering more than a century-long period. The SPI is calculated for the most frequently used time windows of 1, 3, 6, and 12 months. UDEL/GEOG/CCR v3.02 and GPCC/ v7.0 are used in the highest native resolution of 0.5×0.5° whilst NOAA-CIRES 20CR v2c and ECMWF ERA-20C are interpolated at 1.5×1.5° and 0.5×0.5° correspondingly. In contrast to some other indices, for example the popular Palmer Drought Severity Index (PDSI), SPI has significant advantages such as simplicity, suitability on variable time scales and robustness rooted in a solid theoretical development. SPI has been selected by the World Meteorological Organization (WMO) as a key indicator for monitoring drought ('Lincoln declaration'). As a result, drought monitoring centres worldwide are effectively exploiting this index and the National Meteorological and Hydrological Services (NMHSs) are encouraged to use it for monitoring meteorological droughts. These facts and the strong conviction of the authors that the free exchange of data and software services are а basis of effective scientific collaboration, are the main motivators to provide these datasets free of charge at ftp://xeo.cfd.meteo.bg/SPI/. The paper briefly presents some possible applications of the SPI data, revealing its suitability for various objective long-term drought studies at any geographical location.
Ten most popular software for prediction of cyclonic stormsMrinmoy Majumder
In recent years the frequency and intensity of cyclones and hurricanes have been increased manifold compared to the last decade. As a result, the necessity for the development of computer models to predict the track, intensity, and time of occurrence of cyclonic storms has increased to avoid loss of life and prevention of damages to public properties. In this presentaion I had tried to highlight the ten most used models in this aspect which are responsible for saving millions of life and their livelihood.
It is based on Journal Paper named
"Mukherjee, M.K.2013, ’Flood Frequency Analysis of River Subernarekha, India, Using Gumbel’s extreme Value Distribution’, IJCER,Vol-3,Issue-7,pp-12-18."
I have studied the journal and make a PPT in the following.
I
Flood frequency analysis of river kosi, uttarakhand, india using statistical ...eSAT Journals
Abstract In the present study, flood frequency analysis has been applied for river Kosi in Uttarakhand. The river Kosi is an important tributary of Ganga river system, which arising from Koshimool near Kausani, Almora district flows on the western side of the study area and to meet at Ramganga River. The annual flood series analysis has been carried out to estimate the flood quantiles at different return period at Kosi barrage site of river Kosi. The statistical approach provided a significant advantage of estimation of flood at any sites in the homogenous region with very less or no data. In the at –site analysis of annual flood series the Normal, Log normal, Pearson type III, Log Pearson type III, Gumbel and Log Gumbel distribution were applied using method of moments . From the analysis of different goodness of fit tests, it has been found that the Log Gumbel distribution with method of moment as parameters estimation found to be the best-fit distribution for Kosi River and other sites in the region. It is recommended that the regional parameters for Kosi Basin may be used only for primary estimation of flood and should be reviewed when more regional data available. Keywords: Flood Frequency Analysis, River Kosi, Annual Peak Flood discharge, Return Period, Goodness of fit Test.
Climate change is projected to impact drastically in southern African during the 21st century
under low mitigation futures (Niang et al., 2014). African temperatures are projected to rise
rapidly, in the subtropics at least at 1.5 times the global rate of temperature increase (James
and Washington, 2013; Engelbrecht et al., 2015). Moreover, the southern African region is
projected to become generally drier under enhanced anthropogenic forcing (Christensen et
al., 2007; Engelbrecht et al., 2009; James and Washington, 2013; Niang et al., 2014). These
changes in temperature and rainfall patterns will plausibly have a range of impacts in South
Africa, including impacts on energy demand (in terms of achieving human comfort within
buildings and factories), agriculture (e.g. reductions of yield in the maize crop under higher
temperatures and reduced soil moisture), livestock production (e.g. higher cattle mortality as
a result of oppressive temperatures) and water security (through reduced rainfall and
enhanced evapotranspiration) (Engelbrecht et al., 2015).
We want to identify the point(s) in time at which the rate of event occurrences changes, where the number of events is increasing or decreasing in frequency.
We develop a multi-scale streaming anomaly score that takes into account a family of window sizes, making the algorithm scale invariant across a different types of time series with varying pseudo-periodic structure. We explore different aggregation methods of the multi-scale anomaly score to obtain a final anomaly score. We evaluate the performance on the Yahoo! and Numenta Anomaly Benchmark(NAB) datasets.
Changepoint Detection with Bayesian InferenceFrank Kelly
An overview of the application of Bayesian Inference in the detection of changepoints in noisy time series data, applied to three different and diverse domains.
This presentation created and addressed by Jesús Fernandez (University of Cantabria) in the intensive three day course from the BC3, Basque Centre for Climate Change and UPV/EHU (University of the Basque Country) on Climate Change in the Uda Ikastaroak Framework.
The objective of the BC3 Summer School is to offer an updated and multidisciplinary view of the ongoing trends in climate change research. The BC3 Summer School is organized in collaboration with the University of the Basque Country and is a high quality and excellent summer course gathering leading experts in the field and students from top universities and research centres worldwide.
Presentation at the conference Greenmetrics 2016 of the paper "Geographical Load Balancing across Green Datacenters: a Mean Field Analysis" (authors G. Neglia, M. Sereno, G. Bianchi)
Results presented in a Policy Workshop, January 22 in Kingston, Jamaica.
Project goals: (1) Identification of suitable climate proof cultivars for cocoa and tomato to increase resilience of Caribbean agriculture (2) Increase the dialogue between Cocoa and Tomato growers and researchers within the region (3) Inform of national climate compatible policies and risk communication strategies.
Climate model parameterizations of cumulus convection and other clouds that form due to small-scale turbulent eddies are a leading source of uncertainty in predicting the sensitivity of global warming to greenhouse gas increases. Even though we can write down equations governing the physics of cloud formation and fluid motion, these cloud-forming eddies are not resolved by the grid of a climate model, so the subgrid covariability of cloud processes and turbulence must be parameterized. Many approaches are used, all involving numerous subjective assumptions. Even when optimized to match present-day climate, these approaches produce a broad range of predictions about how clouds will change in a future climate.
High resolution models which explicitly simulate the clouds and turbulence on a very fine computational grid more realistically simulate cloud formation compared to observations. But it has proved challenging to translate this skill into better climate model parameterizations.
We will present one naturally stochastic approach for this using a computationally expensive approach called ‘superparameterization’ and then we will lay out a vision for how machine learning could be used to do this translation, which amounts to a form of stochastic coarse-graining. Developing the statistical and computational methods to realize this vision is a good challenge for this SAMSI year.
Flood frequency analysis of river kosi, uttarakhand, india using statistical ...eSAT Journals
Abstract In the present study, flood frequency analysis has been applied for river Kosi in Uttarakhand. The river Kosi is an important tributary of Ganga river system, which arising from Koshimool near Kausani, Almora district flows on the western side of the study area and to meet at Ramganga River. The annual flood series analysis has been carried out to estimate the flood quantiles at different return period at Kosi barrage site of river Kosi. The statistical approach provided a significant advantage of estimation of flood at any sites in the homogenous region with very less or no data. In the at –site analysis of annual flood series the Normal, Log normal, Pearson type III, Log Pearson type III, Gumbel and Log Gumbel distribution were applied using method of moments . From the analysis of different goodness of fit tests, it has been found that the Log Gumbel distribution with method of moment as parameters estimation found to be the best-fit distribution for Kosi River and other sites in the region. It is recommended that the regional parameters for Kosi Basin may be used only for primary estimation of flood and should be reviewed when more regional data available. Keywords: Flood Frequency Analysis, River Kosi, Annual Peak Flood discharge, Return Period, Goodness of fit Test.
Climate change is projected to impact drastically in southern African during the 21st century
under low mitigation futures (Niang et al., 2014). African temperatures are projected to rise
rapidly, in the subtropics at least at 1.5 times the global rate of temperature increase (James
and Washington, 2013; Engelbrecht et al., 2015). Moreover, the southern African region is
projected to become generally drier under enhanced anthropogenic forcing (Christensen et
al., 2007; Engelbrecht et al., 2009; James and Washington, 2013; Niang et al., 2014). These
changes in temperature and rainfall patterns will plausibly have a range of impacts in South
Africa, including impacts on energy demand (in terms of achieving human comfort within
buildings and factories), agriculture (e.g. reductions of yield in the maize crop under higher
temperatures and reduced soil moisture), livestock production (e.g. higher cattle mortality as
a result of oppressive temperatures) and water security (through reduced rainfall and
enhanced evapotranspiration) (Engelbrecht et al., 2015).
We want to identify the point(s) in time at which the rate of event occurrences changes, where the number of events is increasing or decreasing in frequency.
We develop a multi-scale streaming anomaly score that takes into account a family of window sizes, making the algorithm scale invariant across a different types of time series with varying pseudo-periodic structure. We explore different aggregation methods of the multi-scale anomaly score to obtain a final anomaly score. We evaluate the performance on the Yahoo! and Numenta Anomaly Benchmark(NAB) datasets.
Changepoint Detection with Bayesian InferenceFrank Kelly
An overview of the application of Bayesian Inference in the detection of changepoints in noisy time series data, applied to three different and diverse domains.
This presentation created and addressed by Jesús Fernandez (University of Cantabria) in the intensive three day course from the BC3, Basque Centre for Climate Change and UPV/EHU (University of the Basque Country) on Climate Change in the Uda Ikastaroak Framework.
The objective of the BC3 Summer School is to offer an updated and multidisciplinary view of the ongoing trends in climate change research. The BC3 Summer School is organized in collaboration with the University of the Basque Country and is a high quality and excellent summer course gathering leading experts in the field and students from top universities and research centres worldwide.
Presentation at the conference Greenmetrics 2016 of the paper "Geographical Load Balancing across Green Datacenters: a Mean Field Analysis" (authors G. Neglia, M. Sereno, G. Bianchi)
Results presented in a Policy Workshop, January 22 in Kingston, Jamaica.
Project goals: (1) Identification of suitable climate proof cultivars for cocoa and tomato to increase resilience of Caribbean agriculture (2) Increase the dialogue between Cocoa and Tomato growers and researchers within the region (3) Inform of national climate compatible policies and risk communication strategies.
Climate model parameterizations of cumulus convection and other clouds that form due to small-scale turbulent eddies are a leading source of uncertainty in predicting the sensitivity of global warming to greenhouse gas increases. Even though we can write down equations governing the physics of cloud formation and fluid motion, these cloud-forming eddies are not resolved by the grid of a climate model, so the subgrid covariability of cloud processes and turbulence must be parameterized. Many approaches are used, all involving numerous subjective assumptions. Even when optimized to match present-day climate, these approaches produce a broad range of predictions about how clouds will change in a future climate.
High resolution models which explicitly simulate the clouds and turbulence on a very fine computational grid more realistically simulate cloud formation compared to observations. But it has proved challenging to translate this skill into better climate model parameterizations.
We will present one naturally stochastic approach for this using a computationally expensive approach called ‘superparameterization’ and then we will lay out a vision for how machine learning could be used to do this translation, which amounts to a form of stochastic coarse-graining. Developing the statistical and computational methods to realize this vision is a good challenge for this SAMSI year.
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Similar to Climate Extremes Workshop - The Dependence Between Extreme Precipitation and Underlying Indicators of Climate Change - Richard Smith, May 16, 2018
In this project the group members will play with daily rainfall data collected in Gulf coast (535stations in total) from 1949 to 2017. The purposes of this exercise are to:
1) to give students an idea of a typical example of a climate data set (spatio-temporal data) and someassociated scientific questions (e.g. how rainfall extremes vary in space and time and how that mightbe affected by other things like greenhouse gases or temperatures).
2) to get students familiar with data analysis using R including data manipulation, data visualization, and data summary.
3) to introduce some statistical methods (e.g. time series analysis, spatial statistics, extreme value analysis) to analyze this kind of data to "answer" (perform statistical inference) the questions of interest.
Group members: Lin Ge, Jianan Jang, Jessica Robinson, Erin Song, Seth Temple, Adam Wu
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Landslides of any type, and particularly soil slips, pose a great threat in mountainous and steep terrain environ- ments. One of the major triggering mechanisms for slope failures in shallow soils is the build-up of soil pore water pressure resulting in a decrease of effective stress. However, infiltration may have other effects both before and after slope failure. Especially, on steep slopes in shallow soils, soil slips can be triggered by a rapid drop in the apparent cohesion following a decrease in matric suction when a wetting front penetrates into the soil without generating positive pore pressures. These types of failures are very frequent in pre-alpine and alpine landscapes. The key factor for a realistic prediction of rainfall-induced landslides are the interdependence of shear strength and suction and the monitoring of suction changes during the cyclic wetting (due to infiltration) and drying (due to percolation and evaporation) processes. The non-unique relationship between suction and water content, expressed by the Soil Water Retention Curve, results in different values of suction and, therefore, of soil shear strength for the same water content, depending on whether the soil is being wetted (during storms) or dried (during inter-storm periods). We developed a physically based distributed in space and continuous in time model for the simulation of the hydrological triggering of shallow landslides at scales larger than a single slope. In this modeling effort particular weight is given to the modeling of hydrological processes in order to investigate the role of hydrologi- cal triggering mechanisms on soil changes leading to slip occurrences. Specifically, the 3D flow of water and the resulting water balance in the unsaturated and saturated zone is modeled using a Cellular Automata framework. The infinite slope analysis is coupled to the hydrological component of the model for the computation of slope stability. For the computation of the Factor of Safety a unified concept for effective stress under both saturated and unsaturated conditions has been used (Lu Ning and Godt Jonathan, WRR, 2010). A test case of a serious landslide event in Switzerland is investigated to assess the plausibility of the model and to verify its perfomance.
Similar to Climate Extremes Workshop - The Dependence Between Extreme Precipitation and Underlying Indicators of Climate Change - Richard Smith, May 16, 2018 (20)
Recently, the machine learning community has expressed strong interest in applying latent variable modeling strategies to causal inference problems with unobserved confounding. Here, I discuss one of the big debates that occurred over the past year, and how we can move forward. I will focus specifically on the failure of point identification in this setting, and discuss how this can be used to design flexible sensitivity analyses that cleanly separate identified and unidentified components of the causal model.
I will discuss paradigmatic statistical models of inference and learning from high dimensional data, such as sparse PCA and the perceptron neural network, in the sub-linear sparsity regime. In this limit the underlying hidden signal, i.e., the low-rank matrix in PCA or the neural network weights, has a number of non-zero components that scales sub-linearly with the total dimension of the vector. I will provide explicit low-dimensional variational formulas for the asymptotic mutual information between the signal and the data in suitable sparse limits. In the setting of support recovery these formulas imply sharp 0-1 phase transitions for the asymptotic minimum mean-square-error (or generalization error in the neural network setting). A similar phase transition was analyzed recently in the context of sparse high-dimensional linear regression by Reeves et al.
Many different measurement techniques are used to record neural activity in the brains of different organisms, including fMRI, EEG, MEG, lightsheet microscopy and direct recordings with electrodes. Each of these measurement modes have their advantages and disadvantages concerning the resolution of the data in space and time, the directness of measurement of the neural activity and which organisms they can be applied to. For some of these modes and for some organisms, significant amounts of data are now available in large standardized open-source datasets. I will report on our efforts to apply causal discovery algorithms to, among others, fMRI data from the Human Connectome Project, and to lightsheet microscopy data from zebrafish larvae. In particular, I will focus on the challenges we have faced both in terms of the nature of the data and the computational features of the discovery algorithms, as well as the modeling of experimental interventions.
Bayesian Additive Regression Trees (BART) has been shown to be an effective framework for modeling nonlinear regression functions, with strong predictive performance in a variety of contexts. The BART prior over a regression function is defined by independent prior distributions on tree structure and leaf or end-node parameters. In observational data settings, Bayesian Causal Forests (BCF) has successfully adapted BART for estimating heterogeneous treatment effects, particularly in cases where standard methods yield biased estimates due to strong confounding.
We introduce BART with Targeted Smoothing, an extension which induces smoothness over a single covariate by replacing independent Gaussian leaf priors with smooth functions. We then introduce a new version of the Bayesian Causal Forest prior, which incorporates targeted smoothing for modeling heterogeneous treatment effects which vary smoothly over a target covariate. We demonstrate the utility of this approach by applying our model to a timely women's health and policy problem: comparing two dosing regimens for an early medical abortion protocol, where the outcome of interest is the probability of a successful early medical abortion procedure at varying gestational ages, conditional on patient covariates. We discuss the benefits of this approach in other women’s health and obstetrics modeling problems where gestational age is a typical covariate.
Difference-in-differences is a widely used evaluation strategy that draws causal inference from observational panel data. Its causal identification relies on the assumption of parallel trends, which is scale-dependent and may be questionable in some applications. A common alternative is a regression model that adjusts for the lagged dependent variable, which rests on the assumption of ignorability conditional on past outcomes. In the context of linear models, Angrist and Pischke (2009) show that the difference-in-differences and lagged-dependent-variable regression estimates have a bracketing relationship. Namely, for a true positive effect, if ignorability is correct, then mistakenly assuming parallel trends will overestimate the effect; in contrast, if the parallel trends assumption is correct, then mistakenly assuming ignorability will underestimate the effect. We show that the same bracketing relationship holds in general nonparametric (model-free) settings. We also extend the result to semiparametric estimation based on inverse probability weighting.
We develop sensitivity analyses for weak nulls in matched observational studies while allowing unit-level treatment effects to vary. In contrast to randomized experiments and paired observational studies, we show for general matched designs that over a large class of test statistics, any valid sensitivity analysis for the weak null must be unnecessarily conservative if Fisher's sharp null of no treatment effect for any individual also holds. We present a sensitivity analysis valid for the weak null, and illustrate why it is conservative if the sharp null holds through connections to inverse probability weighted estimators. An alternative procedure is presented that is asymptotically sharp if treatment effects are constant, and is valid for the weak null under additional assumptions which may be deemed reasonable by practitioners. The methods may be applied to matched observational studies constructed using any optimal without-replacement matching algorithm, allowing practitioners to assess robustness to hidden bias while allowing for treatment effect heterogeneity.
The world of health care is full of policy interventions: a state expands eligibility rules for its Medicaid program, a medical society changes its recommendations for screening frequency, a hospital implements a new care coordination program. After a policy change, we often want to know, “Did it work?” This is a causal question; we want to know whether the policy CAUSED outcomes to change. One popular way of estimating causal effects of policy interventions is a difference-in-differences study. In this controlled pre-post design, we measure the change in outcomes of people who are exposed to the new policy, comparing average outcomes before and after the policy is implemented. We contrast that change to the change over the same time period in people who were not exposed to the new policy. The differential change in the treated group’s outcomes, compared to the change in the comparison group’s outcomes, may be interpreted as the causal effect of the policy. To do so, we must assume that the comparison group’s outcome change is a good proxy for the treated group’s (counterfactual) outcome change in the absence of the policy. This conceptual simplicity and wide applicability in policy settings makes difference-in-differences an appealing study design. However, the apparent simplicity belies a thicket of conceptual, causal, and statistical complexity. In this talk, I will introduce the fundamentals of difference-in-differences studies and discuss recent innovations including key assumptions and ways to assess their plausibility, estimation, inference, and robustness checks.
We present recent advances and statistical developments for evaluating Dynamic Treatment Regimes (DTR), which allow the treatment to be dynamically tailored according to evolving subject-level data. Identification of an optimal DTR is a key component for precision medicine and personalized health care. Specific topics covered in this talk include several recent projects with robust and flexible methods developed for the above research area. We will first introduce a dynamic statistical learning method, adaptive contrast weighted learning (ACWL), which combines doubly robust semiparametric regression estimators with flexible machine learning methods. We will further develop a tree-based reinforcement learning (T-RL) method, which builds an unsupervised decision tree that maintains the nature of batch-mode reinforcement learning. Unlike ACWL, T-RL handles the optimization problem with multiple treatment comparisons directly through a purity measure constructed with augmented inverse probability weighted estimators. T-RL is robust, efficient and easy to interpret for the identification of optimal DTRs. However, ACWL seems more robust against tree-type misspecification than T-RL when the true optimal DTR is non-tree-type. At the end of this talk, we will also present a new Stochastic-Tree Search method called ST-RL for evaluating optimal DTRs.
A fundamental feature of evaluating causal health effects of air quality regulations is that air pollution moves through space, rendering health outcomes at a particular population location dependent upon regulatory actions taken at multiple, possibly distant, pollution sources. Motivated by studies of the public-health impacts of power plant regulations in the U.S., this talk introduces the novel setting of bipartite causal inference with interference, which arises when 1) treatments are defined on observational units that are distinct from those at which outcomes are measured and 2) there is interference between units in the sense that outcomes for some units depend on the treatments assigned to many other units. Interference in this setting arises due to complex exposure patterns dictated by physical-chemical atmospheric processes of pollution transport, with intervention effects framed as propagating across a bipartite network of power plants and residential zip codes. New causal estimands are introduced for the bipartite setting, along with an estimation approach based on generalized propensity scores for treatments on a network. The new methods are deployed to estimate how emission-reduction technologies implemented at coal-fired power plants causally affect health outcomes among Medicare beneficiaries in the U.S.
Laine Thomas presented information about how causal inference is being used to determine the cost/benefit of the two most common surgical surgical treatments for women - hysterectomy and myomectomy.
We provide an overview of some recent developments in machine learning tools for dynamic treatment regime discovery in precision medicine. The first development is a new off-policy reinforcement learning tool for continual learning in mobile health to enable patients with type 1 diabetes to exercise safely. The second development is a new inverse reinforcement learning tools which enables use of observational data to learn how clinicians balance competing priorities for treating depression and mania in patients with bipolar disorder. Both practical and technical challenges are discussed.
The method of differences-in-differences (DID) is widely used to estimate causal effects. The primary advantage of DID is that it can account for time-invariant bias from unobserved confounders. However, the standard DID estimator will be biased if there is an interaction between history in the after period and the groups. That is, bias will be present if an event besides the treatment occurs at the same time and affects the treated group in a differential fashion. We present a method of bounds based on DID that accounts for an unmeasured confounder that has a differential effect in the post-treatment time period. These DID bracketing bounds are simple to implement and only require partitioning the controls into two separate groups. We also develop two key extensions for DID bracketing bounds. First, we develop a new falsification test to probe the key assumption that is necessary for the bounds estimator to provide consistent estimates of the treatment effect. Next, we develop a method of sensitivity analysis that adjusts the bounds for possible bias based on differences between the treated and control units from the pretreatment period. We apply these DID bracketing bounds and the new methods we develop to an application on the effect of voter identification laws on turnout. Specifically, we focus estimating whether the enactment of voter identification laws in Georgia and Indiana had an effect on voter turnout.
We study experimental design in large-scale stochastic systems with substantial uncertainty and structured cross-unit interference. We consider the problem of a platform that seeks to optimize supply-side payments p in a centralized marketplace where different suppliers interact via their effects on the overall supply-demand equilibrium, and propose a class of local experimentation schemes that can be used to optimize these payments without perturbing the overall market equilibrium. We show that, as the system size grows, our scheme can estimate the gradient of the platform’s utility with respect to p while perturbing the overall market equilibrium by only a vanishingly small amount. We can then use these gradient estimates to optimize p via any stochastic first-order optimization method. These results stem from the insight that, while the system involves a large number of interacting units, any interference can only be channeled through a small number of key statistics, and this structure allows us to accurately predict feedback effects that arise from global system changes using only information collected while remaining in equilibrium.
We discuss a general roadmap for generating causal inference based on observational studies used to general real world evidence. We review targeted minimum loss estimation (TMLE), which provides a general template for the construction of asymptotically efficient plug-in estimators of a target estimand for realistic (i.e, infinite dimensional) statistical models. TMLE is a two stage procedure that first involves using ensemble machine learning termed super-learning to estimate the relevant stochastic relations between the treatment, censoring, covariates and outcome of interest. The super-learner allows one to fully utilize all the advances in machine learning (in addition to more conventional parametric model based estimators) to build a single most powerful ensemble machine learning algorithm. We present Highly Adaptive Lasso as an important machine learning algorithm to include.
In the second step, the TMLE involves maximizing a parametric likelihood along a so-called least favorable parametric model through the super-learner fit of the relevant stochastic relations in the observed data. This second step bridges the state of the art in machine learning to estimators of target estimands for which statistical inference is available (i.e, confidence intervals, p-values etc). We also review recent advances in collaborative TMLE in which the fit of the treatment and censoring mechanism is tailored w.r.t. performance of TMLE. We also discuss asymptotically valid bootstrap based inference. Simulations and data analyses are provided as demonstrations.
We describe different approaches for specifying models and prior distributions for estimating heterogeneous treatment effects using Bayesian nonparametric models. We make an affirmative case for direct, informative (or partially informative) prior distributions on heterogeneous treatment effects, especially when treatment effect size and treatment effect variation is small relative to other sources of variability. We also consider how to provide scientifically meaningful summaries of complicated, high-dimensional posterior distributions over heterogeneous treatment effects with appropriate measures of uncertainty.
Climate change mitigation has traditionally been analyzed as some version of a public goods game (PGG) in which a group is most successful if everybody contributes, but players are best off individually by not contributing anything (i.e., “free-riding”)—thereby creating a social dilemma. Analysis of climate change using the PGG and its variants has helped explain why global cooperation on GHG reductions is so difficult, as nations have an incentive to free-ride on the reductions of others. Rather than inspire collective action, it seems that the lack of progress in addressing the climate crisis is driving the search for a “quick fix” technological solution that circumvents the need for cooperation.
This seminar discussed ways in which to produce professional academic writing, from academic papers to research proposals or technical writing in general.
Machine learning (including deep and reinforcement learning) and blockchain are two of the most noticeable technologies in recent years. The first one is the foundation of artificial intelligence and big data, and the second one has significantly disrupted the financial industry. Both technologies are data-driven, and thus there are rapidly growing interests in integrating them for more secure and efficient data sharing and analysis. In this paper, we review the research on combining blockchain and machine learning technologies and demonstrate that they can collaborate efficiently and effectively. In the end, we point out some future directions and expect more researches on deeper integration of the two promising technologies.
In this talk, we discuss QuTrack, a Blockchain-based approach to track experiment and model changes primarily for AI and ML models. In addition, we discuss how change analytics can be used for process improvement and to enhance the model development and deployment processes.
More from The Statistical and Applied Mathematical Sciences Institute (20)
Operation “Blue Star” is the only event in the history of Independent India where the state went into war with its own people. Even after about 40 years it is not clear if it was culmination of states anger over people of the region, a political game of power or start of dictatorial chapter in the democratic setup.
The people of Punjab felt alienated from main stream due to denial of their just demands during a long democratic struggle since independence. As it happen all over the word, it led to militant struggle with great loss of lives of military, police and civilian personnel. Killing of Indira Gandhi and massacre of innocent Sikhs in Delhi and other India cities was also associated with this movement.
How to Make a Field invisible in Odoo 17Celine George
It is possible to hide or invisible some fields in odoo. Commonly using “invisible” attribute in the field definition to invisible the fields. This slide will show how to make a field invisible in odoo 17.
We all have good and bad thoughts from time to time and situation to situation. We are bombarded daily with spiraling thoughts(both negative and positive) creating all-consuming feel , making us difficult to manage with associated suffering. Good thoughts are like our Mob Signal (Positive thought) amidst noise(negative thought) in the atmosphere. Negative thoughts like noise outweigh positive thoughts. These thoughts often create unwanted confusion, trouble, stress and frustration in our mind as well as chaos in our physical world. Negative thoughts are also known as “distorted thinking”.
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdfTechSoup
In this webinar you will learn how your organization can access TechSoup's wide variety of product discount and donation programs. From hardware to software, we'll give you a tour of the tools available to help your nonprofit with productivity, collaboration, financial management, donor tracking, security, and more.
2024.06.01 Introducing a competency framework for languag learning materials ...Sandy Millin
http://sandymillin.wordpress.com/iateflwebinar2024
Published classroom materials form the basis of syllabuses, drive teacher professional development, and have a potentially huge influence on learners, teachers and education systems. All teachers also create their own materials, whether a few sentences on a blackboard, a highly-structured fully-realised online course, or anything in between. Despite this, the knowledge and skills needed to create effective language learning materials are rarely part of teacher training, and are mostly learnt by trial and error.
Knowledge and skills frameworks, generally called competency frameworks, for ELT teachers, trainers and managers have existed for a few years now. However, until I created one for my MA dissertation, there wasn’t one drawing together what we need to know and do to be able to effectively produce language learning materials.
This webinar will introduce you to my framework, highlighting the key competencies I identified from my research. It will also show how anybody involved in language teaching (any language, not just English!), teacher training, managing schools or developing language learning materials can benefit from using the framework.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
Synthetic Fiber Construction in lab .pptxPavel ( NSTU)
Synthetic fiber production is a fascinating and complex field that blends chemistry, engineering, and environmental science. By understanding these aspects, students can gain a comprehensive view of synthetic fiber production, its impact on society and the environment, and the potential for future innovations. Synthetic fibers play a crucial role in modern society, impacting various aspects of daily life, industry, and the environment. ynthetic fibers are integral to modern life, offering a range of benefits from cost-effectiveness and versatility to innovative applications and performance characteristics. While they pose environmental challenges, ongoing research and development aim to create more sustainable and eco-friendly alternatives. Understanding the importance of synthetic fibers helps in appreciating their role in the economy, industry, and daily life, while also emphasizing the need for sustainable practices and innovation.
How to Create Map Views in the Odoo 17 ERPCeline George
The map views are useful for providing a geographical representation of data. They allow users to visualize and analyze the data in a more intuitive manner.
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Cambridge International AS A Level Biology Coursebook - EBook (MaryFosbery J...
Climate Extremes Workshop - The Dependence Between Extreme Precipitation and Underlying Indicators of Climate Change - Richard Smith, May 16, 2018
1. The Dependence Between Extreme
Precipitation and Underlying
Indicators of Climate Change
Richard L Smith
Departments of STOR and Biostatistics, University
of North Carolina at Chapel Hill
Joint work with Ken Kunkel
North Carolina Institute for Climate Studies and Department of
Marine, Earth, Atmospheric Sciences , North Carolina State University
SAMSI Climate Extremes Workshop
May 16, 2018
1
1
2. Motivation
• 2017 featured the largest rain storm in history in the Gulf of
Mexico
• It also featured the highest mean annual (July–June) sea sur-
face temperature in history in the Gulf of Mexico (SSTGM)
• Here, we shall demonstrate a relationship among the fre-
quency of extreme storms, SSTGM and atmospheric CO2
• This enables us to answer three questions:
– How extreme was Harvey, taking SSTGM and CO2 into
account?
– To what extent can extreme events be “attributed” to
human influence?
– What are the probabilities of similar events in the future?
2
3. Other references on Hurricane Harvey:
Van Oldenborgh et al, Environmental Research Letters, 2017
Risser and Wehner, GRL, 2017
Emanuel, PNAS, 2017
Some of what I present is included in the book chapter:
Hammerling, Katzfuss and Smith, Climate Change Detection and
Attribution, to appear in Chapman and Hall Handbook of Envi-
ronmental Statistics (A. Gelfand, M. Fuentes, J. Hoeting and R.
Smith, eds)
3
4. Handbook of
Environmental and
Ecological Statistics
HandbookofEnvironmental
andEcologicalStatistics
Edited by
Alan E. Gelfand
Montse Fuentes
Jennifer A. Hoeting
Richard L. Smith
Gelfand
Fuentes
Hoeting
Smith
K27284
w w w . c r c p r e s s . c o m
copy to come
Chapman & Hall/CRC
Handbooks of Modern
Statistical Methods
Chapman & Hall/CRC
Handbooks of Modern
Statistical Methods
Statistics
K27284_cover.indd All Pages 5/11/18 11:44 AM
4
5. −100 −95 −90 −85 −80
202224262830 Region Used for Gulf of Mexico SSTs:
21 to 29 degrees N, 83 to 97 degrees W
Longitude
Latitude
5
7. To begin, I’d like to make some basic points about the use of a
Generalized Extreme Value (GEV) distribution to estimate the
probability of an extreme event such as the precipitation event
associated with Hurricane Harvey
These will be illustrated using data on a single station (instead
of spatial aggregates). Specifically, I look at the precipitation
series at Houston Hobby Airport, 1949–2017.
7
19. Alternatively, we may represent the posterior distribution in terms
of its median (0.00077) and its 17th and 83rd percentiles (0.000036,
0.0029).
The exceedance probability for the 7-day Harvey precipitation
event is likely between 0.000036 and 0.0029.
In IPCC-speak:
• Likely means probability at least 0.66
• Very Likely means probability at least 0.9
• Virtually certain means probability at least 0.99
But it is not clear whether they intend these probabilities to be
interpreted in Bayesian fashion!
19
20. GEV Parameters With Covariates I.
We extend the preceding analyses by including two covariates:
SSTt is Gulf of Mexico annual mean SST in year t, expressed as
the deviation from 26oC
CO2t is global mean CO2 in year t, scaled as 0.01(CO2t − 350).
20
24. Bayesian Estimation of Exceedance Probabilities
The preceding model was refitted using a Bayesian analysis with
a flat prior.
Exceedance probabilities for the 2017 event were calculated.
Because we are interested in long-term climatic features rather
than short-term weather events, we replaced SSTt by the fitted
curve under a 4-DF spline (same as in the previous picture)
Computed posterior median probabilities and the 17th and 83rd
percentiles of the posterior density
Result: For 2017, posterior median exceedance probability is
0.0019 (return value 525 years) with a likely range from 0.00022
to 0.00685 (return values 145 to 4472 years).
24
26. Analysis of Extreme Gulf Storms Data
(Ken Kunkel)
1. Define a grid of 1
3-degree by 1
3-degree cells covering longi-
tudes 80–100oW and latitudes 25–35oN (large blue box on
figure). This represents an approximate area of 10,000 km2.
2. Within the grid, consider all possible 2-degree by 2-degree
boxes (all boxes like the red box in the figure).
3. Compute daily precipitation for 1949-present as a simple av-
erage of all stations in each box. All boxes that are wholly
or partly over water are not included in this analysis.
4. For each grid box, identify top 5-day precipitation totals.
5. Pool everything together and identify the top 100 events
for 1949–2017 across the entire region, ignoring those that
overlap in time or space with larger event.
6. Rank and plot these.
7. Also did same analysis on other grid sizes from 1o to 3o.
26
28. 0 20 40 60 80 100
10152025
Plot of Ranked Precipitation Events
Index
PrecipationTotal Harvey
(Computed from 2o gridboxes)
28
29. 1950 1960 1970 1980 1990 2000 2010
10152025
Plot of Precipitation Events by Year
Year
PrecipationTotal
Harvey
(Computed from 2o gridboxes)
29
30. 0 1 2 3 4 5
05101520
Exponential QQ Plot Based on Exceedances
(Line through first n−1 points)
Reduced Value
ExcessOverSmallestValue
Harvey
(Computed from 2o gridboxes)
30
31. We get very similar plots from other
gridbox sizes between 1o and 3o
31
32. Summaries for Five Grid Sizes
Grid Size Largest Mean of next 10 Ratio
1.0 40.87 23.17 1.76
1.5 31.56 16.66 1.89
2.0 27.99 13.59 2.06
2.5 19.52 11.07 1.76
3.0 15.67 9.40 1.67
32
33. Conclusion from Initial Data Plots
1. Harvey is clearly very extreme when computed over 2o × 2o
grid boxes, but the other grid box sizes also show Harvey as
extreme
2. There is some evidence that extreme events were increasing
over time (prior to 2017) but this is hard to judge visually
33
34. Outline of Analysis Method
1. Point process approach: Smith (1989), Chapter 7 of Coles
(2001): represents exceedances over a threshold as a two-
dimensional point process, each point corresponding to the
time and level of an extreme event
2. Threshold defined by k’th largest event, for some k ≤ 100
3. Equivalent to the approach based on the Generalized Pareto
Distribution but more flexible for modeling of covariates
34
35. Point Process Approach
Assume two-dimensional plot of exceedance times and values
above a threshold u forms a nonhomogeneous Poisson process
with
Λ(A) = (t2 − t1)Ψ(y; µ, σ, ξ),
Ψ(y; µ, σ, ξ) = 1 + ξ
y − µ
σ
−1/ξ
provided 1 + ξ(y − µ)/σ > 0.
35
36. In practice:
• Group observations by month
• Threshold u taken as k’th order statistic, for any k ≤ 100.
• Assume nt values over u in month t, at yi,t, i = 1, ..., nt
• GEV parameters for month t are µt, σt, ξt
• Contribution to likelihood for month t is
exp
−
1
12
1 + ξt
u − µt
σt
−1/ξt
·
nt
i=1
1
σt
1 + ξt
yi,t − µt
σt
−1/ξt−1
.
• Parameters µt, σt, ξt may be taken as constants or functions
of time t and/or SST St
36
37. Recode:
T months, N exceedances, yi ≥ u in month ti, i = 1, ..., N.
h =
1
12
T
t=1
1 + ξt
u − µt
σt
−1/ξt
+
N
i=1
log σti +
1
ξti
+ 1 log 1 + ξti
yi − µti
σti
Minimize h based on:
• Covariance matrix X
• n1 covariates in µt, columns i1,j, j = 1, .., n1
• n2 covariates in log σt, columns i2,j, j = 1, .., n2
• µt = θ1 +
n1
j=1 θ3+jxt,i1,j
• log σt = θ2 +
n2
j=1 θ3+n1+jxt,i2,j
• ξt = θ3
37
38. Selection of covariates:
• µt includes terms cos 2πti
12 , sin 2πti
12 , where ti is month of
i’th event, to account for seasonal variation
• Either µt or log σt may include terms based on lagged SSTGM
(from HADISST) database and on annual global CO2 levels
from the RCP database (https://tntcat.iiasa.ac.at/RcpDb;
also used by Wehner and Risser in their paper)
• The SSTGM data are averaged over lagged monthly anoma-
lies from 1 to 6 months before the event in question, after
trial and error to determine the best combination of lags
• Two model selection criteria:
(a) AIC — usual definition
(b) Minimum p-value — select the model that has the
smallest p-value when tested against the model in which
SSTGM and CO2 are omitted entirely
38
39. print this first
“Best Model” Results, p-value Criterion
Size of k p-value Covariates in µt Covariates in log σt
Grid SSTGM CO2 SSTGM CO2
1o 40 0.009 Y N N N
1o 70 0.004 Y N N N
1o 98 0.001 Y N N N
1.5o 40 0.00008 N N Y Y
1.5o 70 0.001 Y Y N N
1.5o 98 0.006 Y Y N Y
2o 40 0.007 Y Y N N
2o 70 0.001 Y Y N N
2o 97 0.006 Y Y N Y
2.5o 40 0.0007 N Y N N
2.5o 70 0.013 N Y N Y
2.5o 98 0.012 N Y N Y
3o 40 0.011 N∗ Y N N
3o 70 0.013 N Y N N
3o 98 0.008 Y Y N Y
∗ only one that is different under AIC criterion
39
41. Bayesian Analysis
1. Uses ”Adaptive Metropolis-Hastings” approach (Haario et
al., 2001) but with some tweaks
(a) 250,000 iterations, saved every tenth iteration
(b) Different lengths of initial warm-up period
(c) Different starting conditions for MCMC
(d) Even though AMH uses an automatic choice of step size
for the trial steps of the parameter updates, I have found
it useful to reduce this under some circumstances
2. Results:
(a) Generally good mixing assessed by eye
(b) Even a “problem case” for which ˆξ = −1 under MLE per-
formed reasonably, though the ACF of the MCMC output
decayed more slowly
41
42. Summary of Observational Data Analysis
1. “Point process approach” to modeling of extreme events —
used threshold exceedances but allows covariates to be in-
cluded in GEV parameters µi, σi, ξi.
2. Covariates include monthly SSTGM (averaged over lags 1–6)
and annual CO2, as well as seasonal terms
3. Fitted to five different grid cells and three values of the num-
ber of included order statistics k based on data from 1949–
2016
4. Maximum likelihood and Bayesian methods — convergence
fine for k = 70 or 99 but unclear for k = 40
5. Conclusion: There is a clear “climate change” signal in the
distribution of extreme events. The rest of the paper will pin
down the nature of that effect and its influence on present
and future extreme event probabilities
42
43. Climate Model Data
1. Our analyses of Gulf Storms show dependence on SSTs and
CO2
2. This raises two questions for which we would like to use
climate model data:
(a) Detection and attribution: how much has the probability
of an extreme increased as a result of global warming?
(Technique: compare predictions based on models that
do or do not include the greenhouse gas component)
(b) Projecting future extreme events: how much will the prob-
ability of an extreme event change between now and, say
2080?
(c) The latter question relies on defining a particular climate
scenario (rcp 8.5)
43
44. Outline plan
1. Access to CMIP5 dataset
(a) 41 climate models
(b) Between 1 and 10 ensemble members for each model
(c) Use three scenarios: historical, historical natural, rcp8.5
(d) For each combination of month, climate model and sce-
nario, calculate Gulf of Mexico average, defined as mean
SST over 21-28 N, 83-97 W (237,016 total observations)
2. Use model output to estimate:
(a) Differences between natural-forcings and all-forcings SSTs
for 2017
(b) Differences between SSTs for 2017 and 2080
3. These will be combined with observational data model to
estimate probabilities of extreme events
44
45. Detailed method
1. Combine historical (all forcings) and rcp 8.5 data: time frame
1949-2100
2. Same for more runs with natural forcings only, time frame
1949-2019 (but most series stop at 2005)
3. Combine monthly data into July-June annual averages
4. For replicates of a single model, average over all replicates
keeping track of number of replicates (weights for subsequent
regressions)
5. Fit linear regression with autoregression and weights
6. Use model runs to estimate (a) differences between 2017
SSTs under natural and historical models, (b) differences
between 2017 and 2080 SSTs, with standard errors for each
7. There are 19 climate models for which both sets of calcula-
tions are possible. The variability among climate models is a
first indicator of climate-model uncertainty
45
47. Conclusions from Climate Model Output
1. All models show positive values for both variables plotted
2. Considerable intra-model uncertainty (horizontal and vertical
widths of each rectangle represent 95% confidence intervals)
3. By the inter-model uncertainty is even greater
4. Chose four models to use as examples for ongoing analysis:
CCSM4, GISS-E2-R, HadGEM2-ES, IPSL-CM5A-LR
47
48. Analysis
For each size of cell (1, 1.5, 2, 2.5, 3):
• Compute probability of exceeding August 2017 level during
2017, based on data 1949-2016
• Three scenarios:
– 2017 calculation (“climate” not “weather” — used long-
term trends to establish SST and CO2 values for 2017)
– Compare exceedance probabilities for 2017 under anthro-
pogenic forcings v. 2017 under natural forcings; compute
relative risk or FAR
– Compare exceedance probabilities for 2080 under rcp8.5
v. 2017 under same; compute relative risk or FAR
• Express all results in terms of posterior probabilities and cred-
ible intervals to capture uncertainty
48
55. Effect of Varying the Climate Model
• Previous results were based on CCSM4 — low projections of
future SSTs compared with other models
• I also tried the same calculations using the HadGEM2-ES
which modeled larger trends
• As might be expected, those analyses that used SSTGM as
a covariate showed even larger increases in projected proba-
bilities through 2080
• A problem with all of these analyses: none of the climate
models shows very good agreement between the observed
and simulated SSTGMs.
55
57. For the last part of the talk I return to my opening example
(Hammerling et al. 2018) where we took a different approach
to reconciling the model-based and observational constructions
of SSTGM.
Future plan is to do this for Ken Kunkel’s data as well.
57
58. Climate Model Data I.
Data from CMIP5: used to calculate annual SST means over
the Gulf of Mexico
• Historical all-forcings data up to 2005 or 2012
• Historical natural-forcings data up to 2005 or 2012
• Future forcings data under the RCP 8.5 scenario
• All model runs have been converted to anomalies
• Four climate models; also computed average over the four
models
58
59. Climate Model Data II.
The model Gulf of Mexico SSTs do not follow the observa-
tional data very closely so, in order to use the regression model
fitted previously to observational SSTs, we proceed as follows.
The observational SSTs for 1949–2017 are regressed on two
covariates: first, the difference between historical-forcings and
natural-forcings climate model runs, and, second, the natural-
forcings climate model runs on their own. The two components
together are then used to define the “all forcings” signal and
the second component on its own is used to define the “natural
forcings” signal. Both components are represented via smooth-
ing splines to give a smooth signal. This exercise is repeated for
each of the four climate models and also with all four models
averaged to give the curves in the next figure.
59
61. Climate Model Data III.
This exercise was repeated to obtain future projections of Gulf
of Mexico SST up to 2080; see Figure. Since there are no
natural-forcings projections over this time period, only the RCP
8.5 values are shown.
61
63. Calculations of Exceedance Probabilities I.
We now repeat the calculation of the probability of a Harvey-sized
event under the circumstances, (a) for 2017 under all forcings,
(b) for 2017 under natural forcings, (c) for 2080 under RCP 8.5.
The calculation is repeated for all four climate models and for the
average over the four models; we used the same posterior density
output as before to obtain Bayesian posterior curves. Finally, we
took the ratio of (a) to (b) (relative risk for 2017 under the
all-forcungs and natural-forcings scenario), and the ratio of (c)
to (a) (relative risk for a Harvey-sized event in 2080 compared
with 2017). The results are in Table:
63
64. Calculations of Exceedance Probabilities II.
Model Present Future
Lower Mid Upper Lower Mid Upper
CCSM4 1.5 2.0 3.2 9.0 26.2 133
GISS-E2-R 1.8 2.5 4.8 13.5 43.5 244
HadGEM2-ES 1.6 2.1 3.5 23.6 73.3 415
IPSL-CM5A-LR 1.5 2.0 3.3 10.8 33.8 186
Combined 1.7 2.4 4.4 14.3 46.0 254
Relative risks. The columns labelled “Present” refer to relative
risks for the 2017 event under an all-forcings scenario versus
a natural-forcings scenario, computed under four climate mod-
els and with all four models combined. Lower, mid and upper
bounds correspond to the 17th, 50th and 83rd percentiles of the
posterior distribution. The columns labelled “Future” are relative
risks for such an event in 2080 against 2017; same conventions
regarding climate models and percentiles.
64
65. Summary
For the combined-model results, the relative risk of the Harvey
precipitation under all-forcings versus natural-forcings scenarios
is estimated as 2.4, “likely” between 1.7 and 4.4. For all five
sets of model results, the lower bound exceeds 1, proving that
it’s “likely” that anthropogenic conditions affected Harvey. This
is consistent with earlier results reported by van Oldenburgh et
al. (2017), Risser and Wehner (2017) and Emanuel (2017).
For the relative risks of a Harvey-sized event in 2080 against
2017, the posterior means range from 26 to 73, with “likely”
bounds ranging from 9 to 415. Evidently, the uncertainty range
for future projections is very wide. Given that Emanuel (2017)
obtained an estimated relative risk of 18 by completely different
methods, there seems to be some agreement that a drastic rise
in the frequency of this type of event is to be expected.
65