This document summarizes a paper on Bayesian measures of model complexity and fit. It discusses using the posterior expected residual information, denoted pD, as a Bayesian measure of model complexity that accounts for the number of effective parameters. pD can be used to compare complex hierarchical models by balancing measures of fit and complexity. It is defined as the deviation of the estimated residual information from the true residual information. The paper also notes some observations about pD, such as that it is not invariant to transformations and depends on choices like the prior and estimator. pD can indicate conflicts between prior and data if negative.
Pattern-based classification of demographic sequencesDmitrii Ignatov
We have proposed prefix-based gapless sequential patterns for classification of demographic sequences. In comparison to black-box machine learning techniques, this one provides interpretable patterns suitable for treatment by professional demographers. As for the language, we have used Pattern Structures as an extension of Formal Concept Analysis for the case of complex data like sequences, graphs, intervals, etc.
Hybrid Block Method for the Solution of First Order Initial Value Problems of...iosrjce
Method of collocation of the differential system and interpolation of the approximate solution which
is a combination of power series and exponential function at some selected grid and off-grid points to generate
a linear multistep method which is implemented in block method is considered in this paper. The basic
properties of the block method which include; consistency, convergence and stability interval is verified. The
method is tested on some numerical experiments and found to have better stability condition and better
approximation than the existing methods
Optimization of Mechanical Design Problems Using Improved Differential Evolut...IDES Editor
Differential Evolution (DE) is a novel evolutionary
approach capable of handling non-differentiable, non-linear
and multi-modal objective functions. DE has been consistently
ranked as one of the best search algorithm for solving global
optimization problems in several case studies. This paper
presents an Improved Constraint Differential Evolution
(ICDE) algorithm for solving constrained optimization
problems. The proposed ICDE algorithm differs from
unconstrained DE algorithm only in the place of initialization,
selection of particles to the next generation and sorting the
final results. Also we implemented the new idea to five versions
of DE algorithm. The performance of ICDE algorithm is
validated on four mechanical engineering problems. The
experimental results show that the performance of ICDE
algorithm in terms of final objective function value, number
of function evaluations and convergence time.
Spectral Element Methods in Large Eddy Simulationgaurav dhir
Lower Order Finite Volume based RANS simulations have become an industry standard in Computational Fluid Dynamics. However, there's a major interest among academics to advance the state of the art by employing spectral element methods in Large Eddy Simulation. In this presentation, we study major issues which inhibit spectral element methods from becoming a mainstream CFD method
Pattern-based classification of demographic sequencesDmitrii Ignatov
We have proposed prefix-based gapless sequential patterns for classification of demographic sequences. In comparison to black-box machine learning techniques, this one provides interpretable patterns suitable for treatment by professional demographers. As for the language, we have used Pattern Structures as an extension of Formal Concept Analysis for the case of complex data like sequences, graphs, intervals, etc.
Hybrid Block Method for the Solution of First Order Initial Value Problems of...iosrjce
Method of collocation of the differential system and interpolation of the approximate solution which
is a combination of power series and exponential function at some selected grid and off-grid points to generate
a linear multistep method which is implemented in block method is considered in this paper. The basic
properties of the block method which include; consistency, convergence and stability interval is verified. The
method is tested on some numerical experiments and found to have better stability condition and better
approximation than the existing methods
Optimization of Mechanical Design Problems Using Improved Differential Evolut...IDES Editor
Differential Evolution (DE) is a novel evolutionary
approach capable of handling non-differentiable, non-linear
and multi-modal objective functions. DE has been consistently
ranked as one of the best search algorithm for solving global
optimization problems in several case studies. This paper
presents an Improved Constraint Differential Evolution
(ICDE) algorithm for solving constrained optimization
problems. The proposed ICDE algorithm differs from
unconstrained DE algorithm only in the place of initialization,
selection of particles to the next generation and sorting the
final results. Also we implemented the new idea to five versions
of DE algorithm. The performance of ICDE algorithm is
validated on four mechanical engineering problems. The
experimental results show that the performance of ICDE
algorithm in terms of final objective function value, number
of function evaluations and convergence time.
Spectral Element Methods in Large Eddy Simulationgaurav dhir
Lower Order Finite Volume based RANS simulations have become an industry standard in Computational Fluid Dynamics. However, there's a major interest among academics to advance the state of the art by employing spectral element methods in Large Eddy Simulation. In this presentation, we study major issues which inhibit spectral element methods from becoming a mainstream CFD method
Enhancing Partition Crossover with Articulation Points Analysisjfrchicanog
This is the presentation of the paper entitled "Enhancing Partition Crossover with Articulation Points Analysis" at the ECOM track in gECCO 2018 (Kyoto). This paper was awarded with a "Best Paper Award"
Since the advent of the horseshoe priors for regularization, global-local shrinkage methods have proved to be a fertile ground for the development of Bayesian theory and methodology in machine learning. They have achieved remarkable success in computation, and enjoy strong theoretical support. Much of the existing literature has focused on the linear Gaussian case. The purpose of the current talk is to demonstrate that the horseshoe priors are useful more broadly, by reviewing both methodological and computational developments in complex models that are more relevant to machine learning applications. Specifically, we focus on methodological challenges in horseshoe regularization in nonlinear and non-Gaussian models; multivariate models; and deep neural networks. We also outline the recent computational developments in horseshoe shrinkage for complex models along with a list of available software implementations that allows one to venture out beyond the comfort zone of the canonical linear regression problems.
Self-organizing Network for Variable Clustering and Predictive ModelingHui Yang
Rapid advancement of sensing and information technology brings the big data, which presents a gold mine of the 21st century to advance knowledge discovery. However, big data also brings significant challenges for data-driven decision making. In particular, it is common that a large number of variables (or predictors, features) underlie the big data. Complex interdependence structures among variables challenge the traditional framework of predictive modeling. This paper presents a new methodology of self-organizing network for variable clustering and predictive modeling. Specifically, we developed a new approach, namely nonlinear coupling analysis to measure nonlinear interdependence structures among variables. Further, all the variables are embedded as nodes in a complex network. Nonlinear-coupling forces move these nodes to derive a self-organizing topology of network. As such, variables are clustered as sub-network communities in the space. Experimental results demonstrated that the proposed method not only outperforms traditional variable clustering algorithms such as hierarchical clustering and oblique principal component analysis, but also effectively identify interdependent structures among variables and further improves the performance of predictive modeling. The proposed new methodology of self-organizing variable clustering is generally applicable for data-driven decision making in many disciplines that involve a large number of highly-redundant variables.
On solving fuzzy delay differential equationusing bezier curves IJECEIAES
In this article, we plan to use Bezier curves method to solve linear fuzzy delay differential equations. A Bezier curves method is presented and modified to solve fuzzy delay problems taking the advantages of the fuzzy set theory properties. The approximate solution with different degrees is compared to the exact solution to confirm that the linear fuzzy delay differential equations process is accurate and efficient. Numerical example is explained and analyzed involved first order linear fuzzy delay differential equations to demonstrate these proper features of this proposed problem.
Multimodal Residual Networks for Visual QAJin-Hwa Kim
Deep neural networks continue to advance the state-of-the-art of image recognition tasks with various methods. However, applications of these methods to multimodality remain limited. We present Multimodal Residual Networks (MRN) for the multimodal residual learning of visual question-answering, which extends the idea of the deep residual learning. Unlike the deep residual learning, MRN effectively learns the joint representation from vision and language information. The main idea is to use element-wise multiplication for the joint residual mappings exploiting the residual learning of the attentional models in recent studies. Various alternative models introduced by multimodality are explored based on our study. We achieve the state-of-the-art results on the Visual QA dataset for both Open-Ended and Multiple-Choice tasks. Moreover, we introduce a novel method to visualize the attention effect of the joint representations for each learning block using back-propagation algorithm, even though the visual features are collapsed without spatial information.
Medical pathology images are visually evaluated by experts for disease diagnosis, but the connectionbetween image features and the state of the cells in an image is typically unknown. To understand thisrelationship, we describe a multimodal modeling and inference framework that estimates shared latentstructure of joint gene expression levels and medical image features. The method is built aroundprobabilistic canonical correlation analysis (PCCA), which is jointly fit to image embeddings that are learnedusing convolutional neural networks and linear embeddings of paired gene expression data. We finallydiscuss a set of theoretical and empirical challenges in domain adaptation settings arising from genomics data.(based on work in collab with Gregory Gundersen and Barbara E. Engelhardt)
Penalty Function Method For Solving Fuzzy Nonlinear Programming Problempaperpublications3
Abstract: In this work, the fuzzy nonlinear programming problem (FNLPP) has been developed and their result have also discussed. The numerical solutions of crisp problems and have been compared and the fuzzy solution and its effectiveness have also been presented and discussed. The penalty function method has been developed and mixed with Nelder and Mend’s algorithm of direct optimization problem solutionhave been used together to solve this FNLPP.
Keyword:Fuzzy set theory, fuzzy numbers, decision making, nonlinear programming, Nelder and Mend’s algorithm, penalty function method.
PhD Dissertation Talk, 22 April 2011
----
The main topic of this thesis addresses the important problem of mining numerical data, and especially gene expression data. These data characterize the behaviour of thousand of genes in various biological situations (time, cell, etc.).
A difficult task consists in clustering genes to obtain classes of genes with similar behaviour, supposed to be involved together within a biological process.
Accordingly, we are interested in designing and comparing methods in the field of knowledge discovery from biological data. We propose to study how the conceptual classification method called Formal Concept Analysis (FCA) can handle the problem of extracting interesting classes of genes. For this purpose, we have designed and experimented several original methods based on an extension of FCA called pattern structures. Furthermore, we show that these methods can enhance decision making in agronomy and crop sanity in the vast formal domain of information fusion.
Study of relevancy, diversity, and novelty in recommender systemsChemseddine Berbague
In the next slides, we present our approach to tackling the conflicting recommendation quality in recommender systems using a genetic-based clustering algorithm. In our approach, we studied the users' tendencies toward diversity and proposed a pairwise similarity measure to amount it. Later, we used the new similarity within a fitness function to create overlapped clusters and to recommend balanced recommendations in terms of diversity and relevancy.
Enhancing Partition Crossover with Articulation Points Analysisjfrchicanog
This is the presentation of the paper entitled "Enhancing Partition Crossover with Articulation Points Analysis" at the ECOM track in gECCO 2018 (Kyoto). This paper was awarded with a "Best Paper Award"
Since the advent of the horseshoe priors for regularization, global-local shrinkage methods have proved to be a fertile ground for the development of Bayesian theory and methodology in machine learning. They have achieved remarkable success in computation, and enjoy strong theoretical support. Much of the existing literature has focused on the linear Gaussian case. The purpose of the current talk is to demonstrate that the horseshoe priors are useful more broadly, by reviewing both methodological and computational developments in complex models that are more relevant to machine learning applications. Specifically, we focus on methodological challenges in horseshoe regularization in nonlinear and non-Gaussian models; multivariate models; and deep neural networks. We also outline the recent computational developments in horseshoe shrinkage for complex models along with a list of available software implementations that allows one to venture out beyond the comfort zone of the canonical linear regression problems.
Self-organizing Network for Variable Clustering and Predictive ModelingHui Yang
Rapid advancement of sensing and information technology brings the big data, which presents a gold mine of the 21st century to advance knowledge discovery. However, big data also brings significant challenges for data-driven decision making. In particular, it is common that a large number of variables (or predictors, features) underlie the big data. Complex interdependence structures among variables challenge the traditional framework of predictive modeling. This paper presents a new methodology of self-organizing network for variable clustering and predictive modeling. Specifically, we developed a new approach, namely nonlinear coupling analysis to measure nonlinear interdependence structures among variables. Further, all the variables are embedded as nodes in a complex network. Nonlinear-coupling forces move these nodes to derive a self-organizing topology of network. As such, variables are clustered as sub-network communities in the space. Experimental results demonstrated that the proposed method not only outperforms traditional variable clustering algorithms such as hierarchical clustering and oblique principal component analysis, but also effectively identify interdependent structures among variables and further improves the performance of predictive modeling. The proposed new methodology of self-organizing variable clustering is generally applicable for data-driven decision making in many disciplines that involve a large number of highly-redundant variables.
On solving fuzzy delay differential equationusing bezier curves IJECEIAES
In this article, we plan to use Bezier curves method to solve linear fuzzy delay differential equations. A Bezier curves method is presented and modified to solve fuzzy delay problems taking the advantages of the fuzzy set theory properties. The approximate solution with different degrees is compared to the exact solution to confirm that the linear fuzzy delay differential equations process is accurate and efficient. Numerical example is explained and analyzed involved first order linear fuzzy delay differential equations to demonstrate these proper features of this proposed problem.
Multimodal Residual Networks for Visual QAJin-Hwa Kim
Deep neural networks continue to advance the state-of-the-art of image recognition tasks with various methods. However, applications of these methods to multimodality remain limited. We present Multimodal Residual Networks (MRN) for the multimodal residual learning of visual question-answering, which extends the idea of the deep residual learning. Unlike the deep residual learning, MRN effectively learns the joint representation from vision and language information. The main idea is to use element-wise multiplication for the joint residual mappings exploiting the residual learning of the attentional models in recent studies. Various alternative models introduced by multimodality are explored based on our study. We achieve the state-of-the-art results on the Visual QA dataset for both Open-Ended and Multiple-Choice tasks. Moreover, we introduce a novel method to visualize the attention effect of the joint representations for each learning block using back-propagation algorithm, even though the visual features are collapsed without spatial information.
Medical pathology images are visually evaluated by experts for disease diagnosis, but the connectionbetween image features and the state of the cells in an image is typically unknown. To understand thisrelationship, we describe a multimodal modeling and inference framework that estimates shared latentstructure of joint gene expression levels and medical image features. The method is built aroundprobabilistic canonical correlation analysis (PCCA), which is jointly fit to image embeddings that are learnedusing convolutional neural networks and linear embeddings of paired gene expression data. We finallydiscuss a set of theoretical and empirical challenges in domain adaptation settings arising from genomics data.(based on work in collab with Gregory Gundersen and Barbara E. Engelhardt)
Penalty Function Method For Solving Fuzzy Nonlinear Programming Problempaperpublications3
Abstract: In this work, the fuzzy nonlinear programming problem (FNLPP) has been developed and their result have also discussed. The numerical solutions of crisp problems and have been compared and the fuzzy solution and its effectiveness have also been presented and discussed. The penalty function method has been developed and mixed with Nelder and Mend’s algorithm of direct optimization problem solutionhave been used together to solve this FNLPP.
Keyword:Fuzzy set theory, fuzzy numbers, decision making, nonlinear programming, Nelder and Mend’s algorithm, penalty function method.
PhD Dissertation Talk, 22 April 2011
----
The main topic of this thesis addresses the important problem of mining numerical data, and especially gene expression data. These data characterize the behaviour of thousand of genes in various biological situations (time, cell, etc.).
A difficult task consists in clustering genes to obtain classes of genes with similar behaviour, supposed to be involved together within a biological process.
Accordingly, we are interested in designing and comparing methods in the field of knowledge discovery from biological data. We propose to study how the conceptual classification method called Formal Concept Analysis (FCA) can handle the problem of extracting interesting classes of genes. For this purpose, we have designed and experimented several original methods based on an extension of FCA called pattern structures. Furthermore, we show that these methods can enhance decision making in agronomy and crop sanity in the vast formal domain of information fusion.
Study of relevancy, diversity, and novelty in recommender systemsChemseddine Berbague
In the next slides, we present our approach to tackling the conflicting recommendation quality in recommender systems using a genetic-based clustering algorithm. In our approach, we studied the users' tendencies toward diversity and proposed a pairwise similarity measure to amount it. Later, we used the new similarity within a fitness function to create overlapped clusters and to recommend balanced recommendations in terms of diversity and relevancy.
La Procuraduría General de Justicia comparte esta información sobre cómo prevenir un secuestro virtual, extorsión telefónica y chantajes.
Es importante conocer esta información que puede ser de gran utilidad para toda la población.
Slides of pattern recognition Course of Professor Zohreh Azimifar at Shiraz University.
اسلاید های درس شناسایی آماری الگو استاد زهره عظیمی فر در دانشگاه شیراز.
45Reliability Demonstration Testing for Discrete-Type Software Products Based...idescitation
Reliability demonstration testing for software
products is performed for the purpose of examining whether
the specified reliability is realized in the software after the
development process is completed. This study proposes a model
of reliability demonstration testing for discrete-type software
such as software for numerical calculations. The number of
input data sets for test and acceptance number of input data
sets causing software failures in the test are designed based
on variation distance. This model has less parameters to be
prespecified than the statistical model.
Scaling Multinomial Logistic Regression via Hybrid ParallelismParameswaran Raman
Distributed algorithms in machine learning follow two main paradigms: data parallel, where the data is distributed across multiple workers and model parallel, where the model parameters are partitioned across multiple workers. The main limitation of the first approach is that the model parameters need to be replicated on every machine. This is problematic when the number of parameters is very large, and hence cannot fit in a single machine. The drawback of the latter approach is that the data needs to be replicated on each machine. Such replications limit the scalability of machine learning algorithms, since in several real-world tasks it is observed that the data and model sizes typically grow hand in hand. In this talk, I will present Hybrid-Parallelism, a new paradigm that partitions both, the data as well as the model parameters simultaneously in a completely de-centralized manner. As a result, each worker only needs access to a subset of the data and a subset of the parameters while performing parameter updates. Next, I will present a case-study showing how to apply these ideas to reformulate Multinomial Logistic Regression to achieve Hybrid Parallelism (DSMLR: Doubly-Separable Multinomial Logistic Regression). Finally, I will demonstrate the versatility of DS-MLR under various scenarios in data and model parallelism, through an empirical study consisting of real-world datasets.
To describe the dynamics taking place in networks that structurally change over time, we propose an approach to search for attributes whose value changes impact the topology of the graph. In several applications, it appears that the variations of a group of attributes are often followed by some structural changes in the graph that one may assume they generate. We formalize the triggering pattern discovery problem as a method jointly rooted in sequence mining and graph analysis. We apply our approach on three real-world dynamic graphs of different natures - a co-authoring network, an airline network, and a social bookmarking system - assessing the relevancy of the triggering pattern mining approach.
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...NTNU
The introduction of expert knowledge when learning Bayesian Networks from data is known to be an excellent approach to boost the performance of automatic learning methods, specially when the data is scarce. Previous approaches for this problem based on Bayesian statistics introduce the expert knowledge modifying the prior probability distributions. In this study, we propose a new methodology based on Monte Carlo simulation which starts with non-informative priors and requires knowledge from the expert a posteriori, when the simulation ends. We also explore a new Importance Sampling method for Monte Carlo simulation and the definition of new non-informative priors for the structure of the network. All these approaches are experimentally validated with five standard Bayesian networks.
Read more:
http://link.springer.com/chapter/10.1007%2F978-3-642-14049-5_70
In this talk we will describe a methodology to handle the causality to make inference on common-cause failure in a situation of missing data. The data are collected in the form of contingency table but the available information are only the numbers of CCF of different orders and the numbers of failure due to a given cause. Therefore only the margins of the contingency table are observed; thefrequencies in each cell are unknown. Assuming a Poisson model for the count, we suggest a Bayesian approach and we use the inverse Bayes formula (IBF) combined with a Metropolis-Hastings algorithm to make inference on the rate of occurrence for the different combination cause, order. The performance of the resulting algorithm is evaluated through simulations. A comparison is made with results obtained from the _-composition approach to deal with causality suggested by Zheng et al. (2013).
The main goal of calibration is usually to improve the predictive performance of the simulator but the values of the parameters in the model may also be of intrinsic scientific interest in their own right. As an example of the latter we will discuss CO2 retrievals from the the Orbiting Carbon Observatory 2 (OCO-2). In order to make appropriate use of observations of the physical system it is important to recognize model discrepancy, the difference between reality and the simulator output. We illustrate through a simple example that an analysis that does not account for model discrepancy may lead to biased and over-confident parameter estimates and predictions. The challenge with incorporating model discrepancy in statistical inverse problems is being confounded with calibration parameters, which will only be resolved with meaningful priors. For our simple example, we model the model-discrepancy via a Gaussian process and demonstrate that through accounting for model discrepancy our prediction within the range of data is correct. We will then discuss the effect of model discrepancy in CO2 retrievals. This is joint work with Anthony O'Hagan, University of Sheffield, and Jonathan Hobbs and Amy Braverman at the Jet Propulsion Laboratory.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Essentials of Automations: Optimizing FME Workflows with Parameters
Model complexity
1. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measures of model complexity
and fit
by D. J. Spiegelhalter, N. G. Best, B. P. Carlin and A. van der
Linde, 2002
presented by Ilaria Masiani
TSI-EuroBayes student
Université Paris Dauphine
Reading seminar on Classics, October 21, 2013
Ilaria Masiani
October 21, 2013
2. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Presentation of the paper
Bayesian measures of model complexity and fit by David J.
Spiegelhalter, Nicola G. Best, Bradley P. Carlin and
Angelika van der Linde
Published in 2002 for J. Royal Statistical Society, series B,
vol.64, Part 4, pp. 583-639
Ilaria Masiani
October 21, 2013
3. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
4. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
5. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Introduction
Model comparison:
measure of fit (ex. deviance statistic)
complexity (n. of free parameters in the model)
=⇒Trade-off of these two quantities
Ilaria Masiani
October 21, 2013
6. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Some of usual model comparison criterion:
ˆ
Akaike information criterion: AIC= −2log{p(y |θ)} + 2p
Bayesian information criterion:
ˆ
BIC= −2log{p(y |θ)} + plog(n)
The problem: both require to know p
Sometimes not clearly defined, e.g., complex hierarchical
models
Ilaria Masiani
October 21, 2013
7. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
=⇒This paper suggests Bayesian measures of complexity and
fit that can be combined to compare complex models.
Ilaria Masiani
October 21, 2013
8. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Observations on pD
9. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Complexity reflects the ’difficulty in estimation’.
Measure of complexity may depend on:
prior information
observed data
Ilaria Masiani
October 21, 2013
Observations on pD
10. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
True model
’All models are wrong, but some are useful’
Box (1976)
Ilaria Masiani
October 21, 2013
Observations on pD
11. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
True model
pt (Y ) ’true’ distribution of unobserved future data Y
θt ’pseudotrue’ parameter value
p(Y |θt ) likelihood specified by θt
Ilaria Masiani
October 21, 2013
12. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Residual information
residual information in data y conditional on θ:
−2log{p(y |θ)}
up to a multiplicative constant (Kullback and Leibler, 1951)
˜
estimator θ(y ) of θt
excess of the true over the estimated residual information:
˜
˜
dΘ {y , θt , θ(y )} = −2log{p(y |θt )} + 2log[p{y |θ(y )}]
Ilaria Masiani
October 21, 2013
13. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Residual information
residual information in data y conditional on θ:
−2log{p(y |θ)}
up to a multiplicative constant (Kullback and Leibler, 1951)
˜
estimator θ(y ) of θt
excess of the true over the estimated residual information:
˜
˜
dΘ {y , θt , θ(y )} = −2log{p(y |θt )} + 2log[p{y |θ(y )}]
Ilaria Masiani
October 21, 2013
14. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Residual information
residual information in data y conditional on θ:
−2log{p(y |θ)}
up to a multiplicative constant (Kullback and Leibler, 1951)
˜
estimator θ(y ) of θt
excess of the true over the estimated residual information:
˜
˜
dΘ {y , θt , θ(y )} = −2log{p(y |θt )} + 2log[p{y |θ(y )}]
Ilaria Masiani
October 21, 2013
15. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Outline
1
Introduction
2
Complexity of a Bayesian model
Bayesian measure of model complexity
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Observations on pD
16. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Bayesian measure of model complexity
unknown θt replaced by random variable θ
˜
dΘ {y , θ, θ(y )} estimated by its posterior expectation w.r.t.
p(θ|y ) :
˜
˜
pD {y , Θ, θ(y )} = Eθ|y [dΘ {y , θ, θ(y )}]
˜
= Eθ|y [−2log{p(y |θ)}] + 2log[p{y |θ(y )}]
pD proposal as the effective number of parameters w.r.t.
model with focus Θ
Ilaria Masiani
October 21, 2013
17. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Bayesian measure of model complexity
unknown θt replaced by random variable θ
˜
dΘ {y , θ, θ(y )} estimated by its posterior expectation w.r.t.
p(θ|y ) :
˜
˜
pD {y , Θ, θ(y )} = Eθ|y [dΘ {y , θ, θ(y )}]
˜
= Eθ|y [−2log{p(y |θ)}] + 2log[p{y |θ(y )}]
pD proposal as the effective number of parameters w.r.t.
model with focus Θ
Ilaria Masiani
October 21, 2013
18. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Bayesian measure of model complexity
unknown θt replaced by random variable θ
˜
dΘ {y , θ, θ(y )} estimated by its posterior expectation w.r.t.
p(θ|y ) :
˜
˜
pD {y , Θ, θ(y )} = Eθ|y [dΘ {y , θ, θ(y )}]
˜
= Eθ|y [−2log{p(y |θ)}] + 2log[p{y |θ(y )}]
pD proposal as the effective number of parameters w.r.t.
model with focus Θ
Ilaria Masiani
October 21, 2013
19. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Effective number of parameters
˜
¯
tipically θ(y ) = E(θ|y ) = θ.
f (y ) fully specified standardizing term, function of the data
Then
Definition
¯
pD = D(θ) − D(θ)
where
D(θ) = −2log{p(y |θ)} + 2log{f (y )}
is the ’Bayesian deviance’.
Ilaria Masiani
October 21, 2013
(1)
20. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Effective number of parameters
˜
¯
tipically θ(y ) = E(θ|y ) = θ.
f (y ) fully specified standardizing term, function of the data
Then
Definition
¯
pD = D(θ) − D(θ)
where
D(θ) = −2log{p(y |θ)} + 2log{f (y )}
is the ’Bayesian deviance’.
Ilaria Masiani
October 21, 2013
(1)
21. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Outline
1
Introduction
2
Complexity of a Bayesian model
Observations on pD
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Observations on pD
22. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Observations on pD
1
¯
(1) can be rewritten as D(θ) = D(θ) + pD =⇒ measure of
’adeguacy’
3
pD depends on: data, choice of focus Θ, prior info, choice
˜
of θ(y ) =⇒ lack of invariance to tranformations
˜
using θ(y ) = E(θ|y ), pD ≥ 0 for any log-concave likelihood
in θ (Jensen’s inequality) =⇒ negative pD s indicate conflict
between prior and data
4
pD easily calculated after a MCMC run
2
Ilaria Masiani
October 21, 2013
23. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Observations on pD
1
¯
(1) can be rewritten as D(θ) = D(θ) + pD =⇒ measure of
’adeguacy’
3
pD depends on: data, choice of focus Θ, prior info, choice
˜
of θ(y ) =⇒ lack of invariance to tranformations
˜
using θ(y ) = E(θ|y ), pD ≥ 0 for any log-concave likelihood
in θ (Jensen’s inequality) =⇒ negative pD s indicate conflict
between prior and data
4
pD easily calculated after a MCMC run
2
Ilaria Masiani
October 21, 2013
24. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Observations on pD
1
¯
(1) can be rewritten as D(θ) = D(θ) + pD =⇒ measure of
’adeguacy’
3
pD depends on: data, choice of focus Θ, prior info, choice
˜
of θ(y ) =⇒ lack of invariance to tranformations
˜
using θ(y ) = E(θ|y ), pD ≥ 0 for any log-concave likelihood
in θ (Jensen’s inequality) =⇒ negative pD s indicate conflict
between prior and data
4
pD easily calculated after a MCMC run
2
Ilaria Masiani
October 21, 2013
25. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Bayesian measure of model
complexity
Observations on pD
Observations on pD
1
¯
(1) can be rewritten as D(θ) = D(θ) + pD =⇒ measure of
’adeguacy’
3
pD depends on: data, choice of focus Θ, prior info, choice
˜
of θ(y ) =⇒ lack of invariance to tranformations
˜
using θ(y ) = E(θ|y ), pD ≥ 0 for any log-concave likelihood
in θ (Jensen’s inequality) =⇒ negative pD s indicate conflict
between prior and data
4
pD easily calculated after a MCMC run
2
Ilaria Masiani
October 21, 2013
26. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
pD for exponential family likelihoods
27. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
pD for approximately normal likelihoods
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
pD for exponential family likelihoods
28. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
Negligible prior informations
ˆ
ˆ
Assume θ|y ∼ N(θ, −Lθ ), then expanding D(θ) around θ
ˆ
ˆ
ˆ
ˆ
D(θ) ≈ D(θ) − (θ − θ)T Lθ (θ − θ)
ˆ
ˆ
≈ D(θ) + χ2
p
=⇒
ˆ
pD = Eθ|y {D(θ)} − D(θ) ≈ p
Ilaria Masiani
October 21, 2013
(2)
29. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
Negligible prior informations
ˆ
ˆ
Assume θ|y ∼ N(θ, −Lθ ), then expanding D(θ) around θ
ˆ
ˆ
ˆ
ˆ
D(θ) ≈ D(θ) − (θ − θ)T Lθ (θ − θ)
ˆ
ˆ
≈ D(θ) + χ2
p
=⇒
ˆ
pD = Eθ|y {D(θ)} − D(θ) ≈ p
Ilaria Masiani
October 21, 2013
(2)
30. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
pD for normal likelihoods
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
pD for exponential family likelihoods
31. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
General hierarchical normal model (know variance)
y ∼ N(A1 θ, C1 )
θ ∼ N(A2 φ, C2 )
¯
Then θ|y is normal with mean θ = Vb and covariance V .
=⇒
pD = tr (−L V )
−1
where −L = AT C1 A1 is the Fisher information.
1
In this case, pD is invariant to affine tranformations of θ.
Ilaria Masiani
October 21, 2013
32. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
General hierarchical normal model (know variance)
y ∼ N(A1 θ, C1 )
θ ∼ N(A2 φ, C2 )
¯
Then θ|y is normal with mean θ = Vb and covariance V .
=⇒
pD = tr (−L V )
−1
where −L = AT C1 A1 is the Fisher information.
1
In this case, pD is invariant to affine tranformations of θ.
Ilaria Masiani
October 21, 2013
33. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
General hierarchical normal model (know variance)
y ∼ N(A1 θ, C1 )
θ ∼ N(A2 φ, C2 )
¯
Then θ|y is normal with mean θ = Vb and covariance V .
=⇒
pD = tr (−L V )
−1
where −L = AT C1 A1 is the Fisher information.
1
In this case, pD is invariant to affine tranformations of θ.
Ilaria Masiani
October 21, 2013
34. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
In normal models:
ˆ
y = Hy , with H hat matrix (that projects the data onto the
−1
fitted values) =⇒ H = A1 VAT C1
1
Then
pD = tr (H)
tr (H) = sum of leverages (influence of each observation
on its fitted value)
Ilaria Masiani
October 21, 2013
35. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
Conjugate normal-gamma model (unknow precision τ )
y ∼ N(A1 θ, τ −1 C1 )
θ ∼ N(A2 φ, τ −1 C2 )
¯ τ ˆ
τ
pD = tr (H) + q(θ)(¯ − τ ) − n{log(τ ) − log(ˆ)}
−1
where q(θ) = (y − A1 θ)T C1 (y − A1 θ).
It can be shown that for large n the choice of parameterization
of τ will make little difference to pD .
Ilaria Masiani
October 21, 2013
36. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
Conjugate normal-gamma model (unknow precision τ )
y ∼ N(A1 θ, τ −1 C1 )
θ ∼ N(A2 φ, τ −1 C2 )
¯ τ ˆ
pD = tr (H) + q(θ)(¯ − τ ) − n{log(τ ) − log(ˆ)}
τ
−1
where q(θ) = (y − A1 θ)T C1 (y − A1 θ).
It can be shown that for large n the choice of parameterization
of τ will make little difference to pD .
Ilaria Masiani
October 21, 2013
37. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
Conjugate normal-gamma model (unknow precision τ )
y ∼ N(A1 θ, τ −1 C1 )
θ ∼ N(A2 φ, τ −1 C2 )
¯ τ ˆ
pD = tr (H) + q(θ)(¯ − τ ) − n{log(τ ) − log(ˆ)}
τ
−1
where q(θ) = (y − A1 θ)T C1 (y − A1 θ).
It can be shown that for large n the choice of parameterization
of τ will make little difference to pD .
Ilaria Masiani
October 21, 2013
38. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
pD for exponential family likelihoods
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
pD for exponential family likelihoods
39. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
One-parameter exponential family
Definition
Assume to have p groups of observations, each of ni
observations in group i has same distribution.
For jth observation in ith group:
log{p(yij |θi , φ)} = wi {yij θi − b(θi )}/φ + c(yij , φ)
where
µi = E(Yij |θi , φ) = b (θi )
V (Yij |θi , φ) = b (θi )φ/wi
wi constant.
Ilaria Masiani
October 21, 2013
40. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
One-parameter exponential family
¯
If Θ focus, bi = Eθi |y {b(θi )}, then the contribution of ith group to
the effective number of parameters:
Θ
¯
¯
pDi = 2ni wi {bi − b(θi )}/φ
=⇒ lack of invariance of pD to reparametrization
Ilaria Masiani
October 21, 2013
41. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
pD for approximately normal
likelihoods
pD for normal likelihoods
pD for exponential family likelihoods
One-parameter exponential family
¯
If Θ focus, bi = Eθi |y {b(θi )}, then the contribution of ith group to
the effective number of parameters:
Θ
¯
¯
pDi = 2ni wi {bi − b(θi )}/φ
=⇒ lack of invariance of pD to reparametrization
Ilaria Masiani
October 21, 2013
42. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
43. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Sampling theory diagnostics for lack of Bayesian fit
Eθ|y {D(θ)} = D(θ) measure of fit or ’adeguacy’
If the model is true
¯
EY (D) = EY [Eθ|y {D(θ)}]
= Eθ (EY |θ [−2log
p(Y |θ)
])
ˆ
p{Y |θ(Y )}
≈ Eθ [EY |θ (χ2 )]
p
= Eθ (p) = p
For one-parameter exponential family p = n, then
¯
EY (D) ≈ n
Ilaria Masiani
October 21, 2013
44. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Sampling theory diagnostics for lack of Bayesian fit
Eθ|y {D(θ)} = D(θ) measure of fit or ’adeguacy’
If the model is true
¯
EY (D) = EY [Eθ|y {D(θ)}]
= Eθ (EY |θ [−2log
p(Y |θ)
])
ˆ
p{Y |θ(Y )}
≈ Eθ [EY |θ (χ2 )]
p
= Eθ (p) = p
For one-parameter exponential family p = n, then
¯
EY (D) ≈ n
Ilaria Masiani
October 21, 2013
45. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Sampling theory diagnostics for lack of Bayesian fit
Eθ|y {D(θ)} = D(θ) measure of fit or ’adeguacy’
If the model is true
¯
EY (D) = EY [Eθ|y {D(θ)}]
= Eθ (EY |θ [−2log
p(Y |θ)
])
ˆ
p{Y |θ(Y )}
≈ Eθ [EY |θ (χ2 )]
p
= Eθ (p) = p
For one-parameter exponential family p = n, then
¯
EY (D) ≈ n
Ilaria Masiani
October 21, 2013
46. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Bayesian criteria for model
comparison
47. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
Definition of the problem
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Bayesian criteria for model
comparison
48. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
Model comparison: the problem
Yrep = independent replicate data set
˜
˜
L(Y , θ) = loss in assigning to data Y a probability p(Y |θ)
˜
L(y , θ(y )) = ’apparent’ loss repredicting the observed y
˜
˜
˜
EYrep |θt [L{y , θ(y )}] = L{y , θ(y )} + cΘ {y , θt , θ(y )}
˜
where cΘ is the ’optimism’ associated with the estimator θ(y )
(Efron, 1986)
Ilaria Masiani
October 21, 2013
49. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
Model comparison: the problem
Yrep = independent replicate data set
˜
˜
L(Y , θ) = loss in assigning to data Y a probability p(Y |θ)
˜
L(y , θ(y )) = ’apparent’ loss repredicting the observed y
˜
˜
˜
EYrep |θt [L{y , θ(y )}] = L{y , θ(y )} + cΘ {y , θt , θ(y )}
˜
where cΘ is the ’optimism’ associated with the estimator θ(y )
(Efron, 1986)
Ilaria Masiani
October 21, 2013
50. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
˜
Assuming L(Y , θ) = −2log{p(Y |θ)},
to estimate cΘ :
1
Classical approach: attempts to estimate the sampling
expectation of cΘ
2
Bayesian approach: direct calculation of the posterior
expectation of cΘ
Ilaria Masiani
October 21, 2013
51. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
˜
Assuming L(Y , θ) = −2log{p(Y |θ)},
to estimate cΘ :
1
Classical approach: attempts to estimate the sampling
expectation of cΘ
2
Bayesian approach: direct calculation of the posterior
expectation of cΘ
Ilaria Masiani
October 21, 2013
52. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
˜
Assuming L(Y , θ) = −2log{p(Y |θ)},
to estimate cΘ :
1
Classical approach: attempts to estimate the sampling
expectation of cΘ
2
Bayesian approach: direct calculation of the posterior
expectation of cΘ
Ilaria Masiani
October 21, 2013
53. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
Classical criteria for model comparison
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Bayesian criteria for model
comparison
54. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
Expected optimism: π(θt ) = EY |θt [cΘ {Y , θt , θ(Y )}]
All criteria for models comparison based on minimizing
ˆ
˜
˜
EYrep |θt [L{Yrep , θ(y )}] = L{y , θ(y )} + π (θt )
ˆ
Efron (1986) π(θt ) for the log-loss function: πE (θt ) ≈ 2p
Considered as corresponding to a plug-in estimate of fit +
twice the effective number of parameters in the model
Ilaria Masiani
October 21, 2013
55. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
Expected optimism: π(θt ) = EY |θt [cΘ {Y , θt , θ(Y )}]
All criteria for models comparison based on minimizing
ˆ
˜
˜
EYrep |θt [L{Yrep , θ(y )}] = L{y , θ(y )} + π (θt )
ˆ
Efron (1986) π(θt ) for the log-loss function: πE (θt ) ≈ 2p
Considered as corresponding to a plug-in estimate of fit +
twice the effective number of parameters in the model
Ilaria Masiani
October 21, 2013
56. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
Expected optimism: π(θt ) = EY |θt [cΘ {Y , θt , θ(Y )}]
All criteria for models comparison based on minimizing
ˆ
˜
˜
EYrep |θt [L{Yrep , θ(y )}] = L{y , θ(y )} + π (θt )
ˆ
Efron (1986) π(θt ) for the log-loss function: πE (θt ) ≈ 2p
Considered as corresponding to a plug-in estimate of fit +
twice the effective number of parameters in the model
Ilaria Masiani
October 21, 2013
57. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
˜
Expected optimism: π(θt ) = EY |θt [cΘ {Y , θt , θ(Y )}]
All criteria for models comparison based on minimizing
ˆ
˜
˜
EYrep |θt [L{Yrep , θ(y )}] = L{y , θ(y )} + π (θt )
ˆ
Efron (1986) π(θt ) for the log-loss function: πE (θt ) ≈ 2p
Considered as corresponding to a plug-in estimate of fit +
twice the effective number of parameters in the model
Ilaria Masiani
October 21, 2013
58. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
Bayesian criteria for model comparison
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Bayesian criteria for model
comparison
59. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
AIME: identify models that best explain the observed data
but
with the expectation that they minimize uncertainty about
observations generated in the same way
Ilaria Masiani
October 21, 2013
60. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
Deviance information criterion (DIC)
Definition
¯
DIC = D(θ) + 2pD
¯
= D + pD
Classical estimate of fit + twice the effective number of
parameters
Also a Bayesian measure of fit, penalized by complexity pD
Ilaria Masiani
October 21, 2013
61. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Definition of the problem
Classical criteria for model
comparison
Bayesian criteria for model
comparison
DIC and AIC
ˆ
Akaike information criterion=⇒ AIC= 2p − 2log{p(y |θ)}
ˆ
θ =MLE
From result (2): pD ≈ p in models with negligible prior
¯
information =⇒ DIC≈ 2p + D(θ)
Ilaria Masiani
October 21, 2013
62. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
Six-cities study
63. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
Spatial distribution of lip cancer in Scotland
7
Conclusion
Ilaria Masiani
October 21, 2013
Six-cities study
64. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Data on the rates of lip cancer in 56 districts in Scotland
(Clayton and Kaldor, 1987; Breslow and Clayton, 1993)
yi observed numbers of cases for each county i
Ei expected numbers of cases for each county i
Ai list for each county of its ni adjacent counties
yi ∼ Pois(exp{θi }Ei )
exp{θi } underlying true area-specific relative risk of lip cancer
Ilaria Masiani
October 21, 2013
65. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Data on the rates of lip cancer in 56 districts in Scotland
(Clayton and Kaldor, 1987; Breslow and Clayton, 1993)
yi observed numbers of cases for each county i
Ei expected numbers of cases for each county i
Ai list for each county of its ni adjacent counties
yi ∼ Pois(exp{θi }Ei )
exp{θi } underlying true area-specific relative risk of lip cancer
Ilaria Masiani
October 21, 2013
66. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Candidate models for θi
Model 1:
θi = α0
Model 2:
θi = α0 + γi
(exchangeable random effect)
Model 3:
θi = α0 + δi
(spatial random effect)
Model 4:
θi = α0 + γi + δi
Model 5:
θi = αi
Ilaria Masiani
(pooled)
(exchang.+ spatial effects)
(saturated)
October 21, 2013
67. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Priors
α0 improper uniform prior
αi (i = 1, ..., 56) normal priors with large variance
γi ∼ N(0, λ−1 )
γ
δi |δi ∼ N
1
ni
j∈Ai
δj , ni1 δ
λ
with
56
i=1 δi
=0
conditional autoregressive prior (Besag, 1974)
λγ , λδ ∼ Gamma(0.5, 0.0005)
Ilaria Masiani
October 21, 2013
68. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Saturated deviance
[yi log{yi /exp(θi )Ei } − {yi − exp(θi )Ei }]
D(θ) = 2
i
(McCullagh and Nelder, 1989, pg 34)
obtained by taking as standardizing factor:
ˆ
−2log{f (y )} = −2 i log{p(yi |θi )} = 208.0
Ilaria Masiani
October 21, 2013
69. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Results
For each model, two independent chains of MCMC (WinBUGS)
for 15000 iterations each (burn-in after 5000 it.)
Deviance summaries using three alternative parameterizations
(mean, canonical, median).
Ilaria Masiani
October 21, 2013
70. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Deviance calculations
¯
D mean of the posterior samples of the saturated deviance
D(¯) by plugging the posterior mean of µi = exp(θi )Ei into
µ
the saturated deviance
¯
D(θ) by plugging the posterior means of α0 , αi , γi , δi into
the linear predictor θi
D(med) by plugging the posterior median of θi into the
saturated deviance
Ilaria Masiani
October 21, 2013
71. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Deviance calculations
¯
D mean of the posterior samples of the saturated deviance
D(¯) by plugging the posterior mean of µi = exp(θi )Ei into
µ
the saturated deviance
¯
D(θ) by plugging the posterior means of α0 , αi , γi , δi into
the linear predictor θi
D(med) by plugging the posterior median of θi into the
saturated deviance
Ilaria Masiani
October 21, 2013
72. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Deviance calculations
¯
D mean of the posterior samples of the saturated deviance
D(¯) by plugging the posterior mean of µi = exp(θi )Ei into
µ
the saturated deviance
¯
D(θ) by plugging the posterior means of α0 , αi , γi , δi into
the linear predictor θi
D(med) by plugging the posterior median of θi into the
saturated deviance
Ilaria Masiani
October 21, 2013
73. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Deviance calculations
¯
D mean of the posterior samples of the saturated deviance
D(¯) by plugging the posterior mean of µi = exp(θi )Ei into
µ
the saturated deviance
¯
D(θ) by plugging the posterior means of α0 , αi , γi , δi into
the linear predictor θi
D(med) by plugging the posterior median of θi into the
saturated deviance
Ilaria Masiani
October 21, 2013
74. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Observations on pD s results
Ilaria Masiani
October 21, 2013
Six-cities study
75. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Observations on pD s results
From result (2): pD ≈ p
pooled model 1: pD = 1.0
saturated model 5: pD from 52.8 to 55.9
models 3-4 with spatial random effects: pD around 31
model 2 with only exchangeable random effects: pD
around 43
Ilaria Masiani
October 21, 2013
76. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Comparison of DIC
Ilaria Masiani
October 21, 2013
Six-cities study
77. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Comparison of DIC
DIC subject to Monte Carlo sampling error (function of
stochastic quantities)
Either of models 3 or 4 is superior to the others
Models 2 and 5 are superior to model 1
Ilaria Masiani
October 21, 2013
78. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
¯
Absolute measure of fit: compare D with n = 56
All models (except pooled model 1) adequate overall fit to the
data =⇒ comparison essentially based on pD s
Ilaria Masiani
October 21, 2013
79. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
¯
Absolute measure of fit: compare D with n = 56
All models (except pooled model 1) adequate overall fit to the
data =⇒ comparison essentially based on pD s
Ilaria Masiani
October 21, 2013
80. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
Six-cities study
7
Conclusion
Ilaria Masiani
October 21, 2013
Six-cities study
81. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Subset of data from the six-cities study: longitudinal study of
health effects of air pollution (Fitzmaurice and Laird, 1993)
yij repeated binary measurement of the wheezing status of
child i at time j (1, yes; 0, no), i = 1, ..., I, j = 1, ..., J
I = 537 children living in Stuebenville, Ohio
J = 4 time points
aij age of child i in years at measurement point j (7, 8, 9,
10 years)
si smoking status of child i’s mother (1, yes; 0, no)
Ilaria Masiani
October 21, 2013
82. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Subset of data from the six-cities study: longitudinal study of
health effects of air pollution (Fitzmaurice and Laird, 1993)
yij repeated binary measurement of the wheezing status of
child i at time j (1, yes; 0, no), i = 1, ..., I, j = 1, ..., J
I = 537 children living in Stuebenville, Ohio
J = 4 time points
aij age of child i in years at measurement point j (7, 8, 9,
10 years)
si smoking status of child i’s mother (1, yes; 0, no)
Ilaria Masiani
October 21, 2013
83. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Conditional response model
Yij ∼ Bernoulli(pij )
pij = Pr(Yij = 1) = g −1 (µij )
µij = β0 + β1 zij1 + β2 zij2 + β3 zij3 + bi
¯
zijk = xijk − x ..k , k = 1, 2, 3
xij1 = aij , xij2 = si , xij3 = aij si
bi individual-specific random effects: bi ∼ N(0, λ−1 )
Ilaria Masiani
October 21, 2013
84. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Conditional response model
Yij ∼ Bernoulli(pij )
pij = Pr(Yij = 1) = g −1 (µij )
µij = β0 + β1 zij1 + β2 zij2 + β3 zij3 + bi
¯
zijk = xijk − x ..k , k = 1, 2, 3
xij1 = aij , xij2 = si , xij3 = aij si
bi individual-specific random effects: bi ∼ N(0, λ−1 )
Ilaria Masiani
October 21, 2013
85. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Conditional response model
Yij ∼ Bernoulli(pij )
pij = Pr(Yij = 1) = g −1 (µij )
µij = β0 + β1 zij1 + β2 zij2 + β3 zij3 + bi
¯
zijk = xijk − x ..k , k = 1, 2, 3
xij1 = aij , xij2 = si , xij3 = aij si
bi individual-specific random effects: bi ∼ N(0, λ−1 )
Ilaria Masiani
October 21, 2013
86. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Conditional response model
Yij ∼ Bernoulli(pij )
pij = Pr(Yij = 1) = g −1 (µij )
µij = β0 + β1 zij1 + β2 zij2 + β3 zij3 + bi
¯
zijk = xijk − x ..k , k = 1, 2, 3
xij1 = aij , xij2 = si , xij3 = aij si
bi individual-specific random effects: bi ∼ N(0, λ−1 )
Ilaria Masiani
October 21, 2013
87. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Model choice: link function g(·)
Model 1:
g(pij ) = logit(pij ) = log{pij /(1 − pij )}
Model 2:
g(pij ) = probit(pij ) = Φ−1 (pij )
Model 3:
g(pij ) = cloglog(pij ) = log{−log(1 − pij )}
Ilaria Masiani
October 21, 2013
88. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Priors and deviance form
βk flat priors
λ ∼ Gamma(0.001, 0.001)
D = −2
{yij log(pij ) + (1 − yij )log(1 − pij )}
i,j
Ilaria Masiani
October 21, 2013
89. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Spatial distribution of lip cancer
Six-cities study
Results
Gibbs sampler for 5000 iterations (burn-in after 1000 it.)
Deviance summaries for canonical and mean
parameterizations.
Ilaria Masiani
October 21, 2013
90. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Outline
1
Introduction
2
Complexity of a Bayesian model
3
Forms for pD
4
Diagnostics for fit
5
Model comparison criterion
6
Examples
7
Conclusion
Ilaria Masiani
October 21, 2013
91. Introduction
Complexity
Forms for pD
Diagnostics for fit
Model comparison criterion
Examples
Conclusion
Conclusion
pD may not be invariant to the chosen parametrization
Similarities to frequentist measures but based on
expectations w.r.t. parameters, in place of sampling
expectations
DIC viewed as a Bayesian analogue of AIC, similar
justification but wider applicability
Involves Monte Carlo sampling and negligible analytic work
Ilaria Masiani
October 21, 2013
92. Appendix
References
References I
McCullagh, P. and Nelder, J.
Generalized Linear Models.
2nd edn. London: Chapman and Hall, 1989.
Besag, J.
Spatial interaction and the statistical analysis of lattice
systems.
J. R. Statist. Soc., series B, 36, 192-236, 1974.
Clayton, D.G. and Kaldor, J.
Empirical Bayes estimates of age-standardised relative risk
for use in disease mapping.
Biometrics, 43, 671-681, 1987.
Ilaria Masiani
October 21, 2013
93. Appendix
References
References II
Efron, B.
How biased is the apparent error rate of a prediction rule?
J. Ann. Statistic. Ass., 81, 461-470, 1986.
Fitzmaurice, G. and Laird, N.
A likelihood-based method for analysing longitudinal binary
responses.
Biometrika, 80, 141-151, 1993.
Kullback, S. and Leibler, R.A.
On information and sufficienty.
Ann. Math. Statist., 22, 79-86, 1951.
Ilaria Masiani
October 21, 2013
94. Appendix
References
References III
Spiegelhalter, D.J., Best, N.G., Carlin, B.P. and van der
Linde, A.
Bayesian measures of model complexity and fit.
J. Royal Statistical Society, series B, vol.64, Part 4, pp.
583-639, 2002.
Ilaria Masiani
October 21, 2013