International Conference on Monte Carlo techniques
Closing conference of thematic cycle
Paris July 5-8th 2016
Campus les cordeliers
Jere Koskela's slides
International Conference on Monte Carlo techniques
Closing conference of thematic cycle
Paris July 5-8th 2016
Campus les Cordeliers
Slides of Richard Everitt's presentation
International Conference on Monte Carlo techniques
Closing conference of thematic cycle
Paris July 5-8th 2016
Campus les cordeliers
Jere Koskela's slides
International Conference on Monte Carlo techniques
Closing conference of thematic cycle
Paris July 5-8th 2016
Campus les Cordeliers
Slides of Richard Everitt's presentation
A review of one of the most popular methods of clustering, a part of what is know as unsupervised learning, K-Means. Here, we go from the basic heuristic used to solve the NP-Hard problem to an approximation algorithm K-Centers. Additionally, we look at variations coming from the Fuzzy Set ideas. In the future, we will add more about On-Line algorithms in the line of Stochastic Gradient Ideas...
Master Thesis on the Mathematial Analysis of Neural NetworksAlina Leidinger
Master Thesis submitted on June 15, 2019 at TUM's chair of Applied Numerical Analysis (M15) at the Mathematics Department.The project was supervised by Prof. Dr. Massimo Fornasier. The thesis took a detailed look at the existing mathematical analysis of neural networks focusing on 3 key aspects: Modern and classical results in approximation theory, robustness and Scattering Networks introduced by Mallat, as well as unique identification of neural network weights. See also the one page summary available on Slideshare.
Universal Approximation Theorem
Here, we prove that the perceptron multi-layer can approximate all continuous functions in the hypercube [0,1]. For this, we used the Cybenko proof... I tried to include the basic in topology and mathematical analysis to make the slides more understandable. However, they still need some work to be done. In addition, I am a little bit rusty in my mathematical analysis, so I am still not so convinced with my linear functional I defined for the proof...!!! Back to the Rudin and Apostol!!! So expect changes in the future.
We apply tensor train (TT) data format to solve an elliptic PDE with uncertain coefficients. We reduce complexity and storage from exponential to linear. Post-processing in TT format is also provided.
A review of one of the most popular methods of clustering, a part of what is know as unsupervised learning, K-Means. Here, we go from the basic heuristic used to solve the NP-Hard problem to an approximation algorithm K-Centers. Additionally, we look at variations coming from the Fuzzy Set ideas. In the future, we will add more about On-Line algorithms in the line of Stochastic Gradient Ideas...
Master Thesis on the Mathematial Analysis of Neural NetworksAlina Leidinger
Master Thesis submitted on June 15, 2019 at TUM's chair of Applied Numerical Analysis (M15) at the Mathematics Department.The project was supervised by Prof. Dr. Massimo Fornasier. The thesis took a detailed look at the existing mathematical analysis of neural networks focusing on 3 key aspects: Modern and classical results in approximation theory, robustness and Scattering Networks introduced by Mallat, as well as unique identification of neural network weights. See also the one page summary available on Slideshare.
Universal Approximation Theorem
Here, we prove that the perceptron multi-layer can approximate all continuous functions in the hypercube [0,1]. For this, we used the Cybenko proof... I tried to include the basic in topology and mathematical analysis to make the slides more understandable. However, they still need some work to be done. In addition, I am a little bit rusty in my mathematical analysis, so I am still not so convinced with my linear functional I defined for the proof...!!! Back to the Rudin and Apostol!!! So expect changes in the future.
We apply tensor train (TT) data format to solve an elliptic PDE with uncertain coefficients. We reduce complexity and storage from exponential to linear. Post-processing in TT format is also provided.
CHN and Swap Heuristic to Solve the Maximum Independent Set ProblemIJECEIAES
We describe a new approach to solve the problem to find the maximum independent set in a given Graph, known also as Max-Stable set problem (MSSP). In this paper, we show how Max-Stable problem can be reformulated into a linear problem under quadratic constraints, and then we resolve the QP result by a hybrid approach based Continuous Hopfeild Neural Network (CHN) and Local Search. In a manner that the solution given by the CHN will be the starting point of the local search. The new approach showed a good performance than the original one which executes a suite of CHN runs, at each execution a new leaner constraint is added into the resolved model. To prove the efficiency of our approach, we present some computational experiments of solving random generated problem and typical MSSP instances of real life problem.
Sequential quasi-Monte Carlo (SQMC) is a quasi-Monte Carlo (QMC) version of sequential Monte Carlo (or particle filtering), a popular class of Monte Carlo techniques used to carry out inference in state space models. In this talk I will first review the SQMC methodology as well as some theoretical results. Although SQMC converges faster than the usual Monte Carlo error rate its performance deteriorates quickly as the dimension of the hidden variable increases. However, I will show with an example that SQMC may perform well for some "high" dimensional problems. I will conclude this talk with some open problems and potential applications of SQMC in complicated settings.
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
ICML 2021 tutorial on random matrix theory and machine learning.
Part 3 covers: 1. Motivation: Average-case versus worst-case in high dimensions 2. Algorithm halting times (runtimes) 3. Outlook
Research internship on optimal stochastic theory with financial application u...Asma Ben Slimene
This is a presntation of my second year intership on optimal stochastic theory and how we can apply it on some financial application then how we can solve such problems using finite differences methods!
Enjoy it !
Presentation on stochastic control problem with financial applications (Merto...Asma Ben Slimene
This is an introductory to optimal stochastic control theory with two applications in finance: Merton portfolio problem and Investement/consumption problem with numerical results using finite differences approach
My talk at the International Conference on Monte Carlo Methods and Applications (MCM2032) related to advances in mathematical aspects of stochastic simulation and Monte Carlo methods at Sorbonne Université June 28, 2023, about my recent works (i) "Numerical Smoothing with Hierarchical Adaptive Sparse Grids and Quasi-Monte Carlo Methods for Efficient Option Pricing" (link: https://doi.org/10.1080/14697688.2022.2135455), and (ii) "Multilevel Monte Carlo with Numerical Smoothing for Robust and Efficient Computation of Probabilities and Densities" (link: https://arxiv.org/abs/2003.05708).
Similar to Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passing - Presentation (20)
Predicting rainfall using ensemble of ensemblesVarad Meru
The Paper was done in a group of three for the class project of CS 273: Introduction to Machine Learning at UC Irvine. The group members were Prolok Sundaresan, Varad Meru, and Prateek Jain.
Regression is an approach for modeling the relationship between data X and the dependent variable y. In this report, we present our experiments with multiple approaches, ranging from Ensemble of Learning to Deep Learning Networks on the weather modeling data to predict the rainfall. The competition was held on the online data science competition portal ‘Kaggle’. The results for weighted ensemble of learners gave us a top-10 ranking, with the testing root-mean-squared error being 0.5878.
Generating Musical Notes and Transcription using Deep LearningVarad Meru
Music has always been the most followed art form, and lot of research had gone into understanding it. In recent years, deep learning approaches for building unsupervised hierarchical representations from unlabeled data have gained significant interest. Progress in fields, such as image processing and natural language processing, has been substantial, but to my knowledge, methods on auditory data for learning representations have not been studied extensively. In this project I try to use two methods for generating music from range of musical inputs such as MIDI to complex WAV formats. I use RNN-RBMs and CDBN to explore music.
Kakuro: Solving the Constraint Satisfaction ProblemVarad Meru
This work was done as a part of the project for the course CS 271: Introduction to Artificial Intelligence (http://www.ics.uci.edu/~kkask/Fall-2014%20CS271/index.html), taught in Fall 2014.
CS295 Week5: Megastore - Providing Scalable, Highly Available Storage for Int...Varad Meru
Slides created as a part of CS 295's week 5 on Transactions and Systems.
CS 295 (Cloud Computing and BigData) at UCI - https://sites.google.com/site/cs295cloudcomputing/
Cassandra - A Decentralized Structured Storage SystemVarad Meru
Slides created as a part of CS 295's week 4 on NoSQL Basics.
CS 295 (Cloud Computing and BigData) at UCI - https://sites.google.com/site/cs295cloudcomputing/
Slides created as a part of CS 295's week 1 on Cloud Computing Basics.
CS 295 (Cloud Computing and BigData) at UCI - https://sites.google.com/site/cs295cloudcomputing/
Live Wide-Area Migration of Virtual Machines including Local Persistent State.Varad Meru
Slides created as a part of CS 295's week 2 on Virtualization in cloud.
CS 295 (Cloud Computing and BigData) at UCI - https://sites.google.com/site/cs295cloudcomputing/
Machine Learning and Apache Mahout : An IntroductionVarad Meru
An Introductory presentation on Machine Learning and Apache Mahout. I presented it at the BigData Meetup - Pune Chapter's first meetup (http://www.meetup.com/Big-Data-Meetup-Pune-Chapter/).
K-Means, its Variants and its ApplicationsVarad Meru
This presentation was given by our project group at the Lead College competition at Shivaji University. Our project got the 1st Prize. We focused mainly on Rough K-Means and build a Social-Network-Recommender System based on Rough K-Means.
The Members of the Project group were -
Mansi Kulkarni,
Nikhil Ingole,
Prasad Mohite,
Varad Meru
Vishal Bhavsar.
Wonderful Experience !!!
Introduction to Mahout and Machine LearningVarad Meru
This presentation gives an introduction to Apache Mahout and Machine Learning. It presents some of the important Machine Learning algorithms implemented in Mahout. Machine Learning is a vast subject; this presentation is only a introductory guide to Mahout and does not go into lower-level implementation details.
This article got published in the Software Developer's Journal's February Edition.
It describes the use of MapReduce paradigm to design Clustering algorithms and explain three algorithms using MapReduce.
- K-Means Clustering
- Canopy Clustering
- MinHash Clustering
I gave a series of Seminars at the following colleges in Solapur.
1. Walchand Institute of Technology, Solapur.
2. Brahmdevdada Mane Institute of Technology, Solapur.
3. Orchid College of Engineering & Technology, Solapur.
4. SVERI's College of Engineering, Pandharpur.
It focussed on what 'BigData' is and how the next generation of professionals should be ready the BigData revolution
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passing - Presentation
1. Outline Introduction Background STC Framework Experimental Results Conclusion
Subproblem-Tree Calibration: A Unified Approach
to Max-Product Message Passing
Varad Meru, Prolok Sundaresan
Department of Computer Science,
Donald Bren School of Information and Computer Science,
UC Irvine
December 10th, 2014
Citation: Wang, Huayan, and Koller Daphne. ”Subproblem-tree
calibration: A unified approach to max-product message passing.” In
Proceedings of the 30th International Conference on Machine Learning
(ICML-13), pp. 190-198. 2013.
2. Outline Introduction Background STC Framework Experimental Results Conclusion
Outline
Introduction
Background
MAP: Maximum a posteriori estimation.
LP relaxation, and dual decomposition
Bethe cluster graphs
STC Framework
Subproblem multi-graph and subproblem trees
Max-consistency and dual-optimal on trees
The STC algorithm
Fixed-point characterization
Choosing allocation weights
General primal solutions
Experimental Results
Conclusion
3. Outline Introduction Background STC Framework Experimental Results Conclusion
Introduction I
MAP-MRF : Finding the most probable assignments for MRFs
(MPE)
NP-Hard
Large family of methods based on solving a dual problem of an
LP relaxation.
Recent Advances.
Convergent version of these algorithms can be interpreted as
block coordinate descent (BCD) in the dual.
Variants operate on small blocks - Max-product linear
programming algorithm (MPLP), max-sum diffusion (MSD)
and Tree-weighted max-product message passing (TRW-S).
Given block of dual-variables: enforce some consistency
constraint over the block.
Observation
Difficulties in generalizing these methods arise due to strong
consistency constraint - which are sufficient but not necessary.
4. Outline Introduction Background STC Framework Experimental Results Conclusion
Introduction II
Aim
Dual-optimality can be established on a much broader choices
of the dual objective.
Deriving a “unified” message passing algorithms in an arbitrary
dual-decomposition.
Properties of the Resulted Algorithm (subproblem-tree
calibration, or STC)
Message passing on graph-object (subproblem multi-graph, or
SMG)
Subsumes MPLP, MSD, and TRW-S
Achieves dual-optimality on blocks with flexible choices.
5. Outline Introduction Background STC Framework Experimental Results Conclusion
MAP Inference
MAP Inference problem over X and graph strcuture
G = {V, E} can be formulated as
maximize
X
Θ(X)
Where Θ(X) = α∈A θα(Xα); A is the set of MRF cliques.
xi ∈ V al(Xi) and x = x1:N
6. Outline Introduction Background STC Framework Experimental Results Conclusion
LP relaxation, Dual decomposition I
Large family of MAP inference methods based on solving
Linear Programming (LP) relaxation
maximize
µ∈M
Θ · µ
Where µ = {µi(xi), µij(xi, xj)|∀i, xi, (i, j), (xi, xj)}; Θ is all
MRF parameters {θi, θij} concatenated in same ordering as µ
A decomposition of Θ(X) into subproblems c ∈ C,
parameterized by {Θc}
∀x,
c∈C
Θc
(x|c) = Θ(x)
Where x|c denotes restricting the joint assignment to the
scope of subproblem c.
7. Outline Introduction Background STC Framework Experimental Results Conclusion
LP relaxation, Dual decomposition II
Enforcing constraint by expressing reparameterization in terms
of messages
Θc
= Θc
+
c :Xc∩X c=∅
δc →c(Xc ∩ X c)
where the messages satisfy δc →c = −δc→c
Each subproblem has its own copy of variables Xc
8. Outline Introduction Background STC Framework Experimental Results Conclusion
Bethe cluster (region) Graph I
Bipartite structure: one layer of “factor” nodes and one layer
of small (usually unary) nodes.
Restricted Design due to historical concern of satisfying the
’running intersection property’.
D( δf→i ) =
i
max
Xi
θi
(Xi
) +
f
max
Xf
Θf
(Xf
)
where the messages are only defined between the two layers
(Bipartite structure).
The dual (mentioned earlier) becomes more restricted due to
the requirement of satisfying the running intersection property.
9. Outline Introduction Background STC Framework Experimental Results Conclusion
Bethe cluster (region) Graph II
(a) Markov Random
Field
(b) Cluster Graph (not Bethe
Cluster)
11. Outline Introduction Background STC Framework Experimental Results Conclusion
SMG and subproblem-tree I
Subproblem Multi-Graph/Tree
Given C, the subproblem multi-graph (SMG) G = (V, E) has
one node for each c ∈ C and one edge between c and c for each
tuple (c,c ,ϕ), where ϕ ∈ V ∪ E is shared by c and c . A
subproblem multi-graph (SMG) is a tree T ⊂ G
If we include all unary subproblems into the decomposition,
we would get a SMG similar to Fig: (c) but with extra edges
among the non-unary subproblems.
So a tree in the Bethe cluster graph (which we call a Bethe
tree) is also a subproblem tree by definition.
13. Outline Introduction Background STC Framework Experimental Results Conclusion
SMG and subproblem-tree III
For each SMG edge (c, c , ϕ) ∈ E, we have messages
δc →c = −δc→c . Therefore the block (of dual variables)
associated with subproblem tree T is given by:
BT
= {δc →c(Xϕ) : (c, c , ϕ) ∈ T }. (1)
14. Outline Introduction Background STC Framework Experimental Results Conclusion
Max-consistency and dual-optimal trees I
Given a block BT associated with some subproblem tree T , we
want to achieve dual-optimal w.r.t. that block
Dual-optimal on T
The subproblem potentials Θc
are dual-optimal on T if we can not
further decrease the dual objective by changing messages in BT .
Message passing algorithm achieves dual-optimality by
enforcing some Consistency Constraint.
We first identify constraint that is equivalent to dual-optimal on T .
Assignments agree on T
Assignments to all subproblems {xc}c∈T agree on T , denoted as
xc ∼ T , if for ∀(c, c , ϕ) ∈ T , we have xc
ϕ = xc
ϕ .
15. Outline Introduction Background STC Framework Experimental Results Conclusion
Max-consistency and dual-optimal trees II
Weak max-consistency on T
{Θc
}c∈T satisfies weak max-consistency if
c∈T
max
Xc
Θc
(Xc
) = max
{Xc}∼T
c∈T
Θc
(Xc
)
Maximizing each subproblem independently gets to the same
optimal value as maximizing them while requiring the
assignments to agree on the tree.
Let Mc
ϕ be the (log)-max-marginal of c on ϕ, then
Mc
ϕ(xϕ) = max
Xc|ϕ=xϕ
Θc
(Xc
)
if ϕ = (i, j) ∈ E, Xc|ϕ = xϕ means Xc
i = xi and Xc
j = xj
16. Outline Introduction Background STC Framework Experimental Results Conclusion
Max-consistency and dual-optimal trees III
Strong max-consistency on T
{Θc
}c∈T satisfies strong max-consistency if
Mc
ϕ = Mc
ϕ ∀(c, c , ϕ) ∈ T
The relations among these consistency constraints are:
Proposition 1.
For any Bethe tree T ,
MPLP max-consistency =⇒ Weak max-consistency
For any subproblem tree T (including Bethe trees),
Strong max-consistency =⇒ Weak max-consistency.
Weak max-consistency ⇐⇒ Dual-optimal on T .
17. Outline Introduction Background STC Framework Experimental Results Conclusion
Subproblem tree calibration algorithm I
Algorithm calibrates a subproblem-tree by an upstream pass
and a downstream pass
Both update subproblem potentials “in place” without storing
any message.
(a) MRF (b) SMG (c) Spanning
Tree of SMG
Figure 2: Flow of the Algorithm: Start with (a) to generate (b) and
randomly selected (c) and ”Calibrate”
18. Outline Introduction Background STC Framework Experimental Results Conclusion
Subproblem tree calibration algorithm II
Algorithm -
1. Given MRF (left figure)
2. Split into subproblems (dual decomposition)
3. Build a multi-graph with a node for each subproblem (middle
figure)
4. Repeat
a. Randomly choose a subproblem-tree (right figure)
b. “Calibrate” the tree by max-product / min-sum message
passing
Properties
1 Each tree calibration is a block coordinate descent step for the
dual problem.
2 The “block” corresponds to all edges in the subproblem-tree.
3 Subsumes MPLP, TRW-S, and max-sum diffusion as special
cases.
4 Handles larger and more flexible “blocks” than these methods.
20. Outline Introduction Background STC Framework Experimental Results Conclusion
Choosing allocation weights
After STC, for each subproblem c
max
Xc
Θc
(Xc
) = ac · max
{X¯c}∼T
¯c∈T
Θ¯c
(X¯c
)
The downstream pass allocate “energy” to all subproblems
according to their allocation weights.
”Energy” = negative lograrithm of the probabilities. Helps in
the case of very small values to avoid numerical underflow as
well as making the computations easier to handle - moving
from max-product to max-summations.
21. Outline Introduction Background STC Framework Experimental Results Conclusion
General Primal solution
Given subproblem potentials, solutions to the original MAP
inference problem can be constructed in different ways
Visit the variables (in the original MRF) in some ordering, for
example, X, X, . . . XN . And for Xi we choose the
assignment:
xi = arg max
c:i∈scope(c)
max
XcXi
Θc
(Xc
|Xj = xj, ∀j < i)
Visiting each Xi, we choose its assignment to maximize the
sum of all max-marginals from all subproblems covering Xi.
Fix Xi = xi in all subproblems.
22. Outline Introduction Background STC Framework Experimental Results Conclusion
Experimental MAP inference tasks I
1 The protein design benchmark
20 largest problems from that dataset
Number of Variables - 101 to 180
Number of Edges - 1973 to 3005
Variable Cardinality - 154
2 Synthetic 20-by-20 grid
Potentials from N(0, 1)
Variable Cardinality - 100
3 ”Object detection” task from PIC-2011
37 problem instances
Number of Variables - 60 / problem instance
Number of Edges - 1770 / problem instance
Variable Cardinality - 11 - 21
23. Outline Introduction Background STC Framework Experimental Results Conclusion
Experimental MAP inference tasks II
We observe that different methods tend to “converge” to different
dual objectives, Even though the dual objectives in each plot
should have exactly the same optimal value.
25. Outline Introduction Background STC Framework Experimental Results Conclusion
Conclusion
Two dimensions of flexibility in designing a message passing
algorithm for MAP inference:
Choosing blocks to update
Choosing a dual state on a plateau in each BCD step.
STC algorithm can be applied with extreme flexibility in these
choices.
Finding Principled and adaptive strategies in making these
choices will help design much more powerful message passing
algorithms.