This document summarizes a tutorial on distributed constraint handling and optimization. It discusses:
1) Distributed constraint reasoning, where a set of agents must come to an agreement about actions to jointly find the best solution.
2) Example applications that can be modeled as distributed constraint optimization problems (DCOPs) including graph coloring, meeting scheduling, and target tracking.
3) Complete algorithms for solving DCOPs exactly, focusing on decentralized search-based approaches like ADOPT and dynamic programming approaches like DPOP.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/bdti/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Shehrzad Qureshi, Senior Engineer at BDTI, presents the "Demystifying Deep Neural Networks" tutorial at the May 2017 Embedded Vision Summit.
What are deep neural networks, and how do they work? In this talk, Qureshi provides an introduction to deep convolutional neural networks (CNNs), which have recently demonstrated impressive success on a wide range of vision tasks. Without using a lot of complex math, he introduces the basics of CNNs. He explores the differences between shallow and deep networks, and explains why deep learning has only recently become prevalent. He examines the different types of layers used in contemporary CNN designs and illustrates why networks composed of these layers are well suited to vision tasks.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/bdti/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Shehrzad Qureshi, Senior Engineer at BDTI, presents the "Demystifying Deep Neural Networks" tutorial at the May 2017 Embedded Vision Summit.
What are deep neural networks, and how do they work? In this talk, Qureshi provides an introduction to deep convolutional neural networks (CNNs), which have recently demonstrated impressive success on a wide range of vision tasks. Without using a lot of complex math, he introduces the basics of CNNs. He explores the differences between shallow and deep networks, and explains why deep learning has only recently become prevalent. He examines the different types of layers used in contemporary CNN designs and illustrates why networks composed of these layers are well suited to vision tasks.
Ant colony search and heuristic techniques for optimal dispatch of energy sou...Beniamino Murgante
Ant colony search and heuristic techniques for optimal dispatch of energy sources in micro-grids - Eleonora Riva Sanseverino – University of Palermo (Italy)
Intelligent Analysis of Environmental Data (S4 ENVISA Workshop 2009)
High-performance graph analysis is unlocking knowledge in computer security, bioinformatics, social networks, and many other data integration areas. Graphs provide a convenient abstraction for many data problems beyond linear algebra. Some problems map directly to linear algebra. Others, like community detection, look eerily similar to sparse linear algebra techniques. And then there are algorithms that strongly resist attempts at making them look like linear algebra. This talk will cover recent results with an emphasis on streaming graph problems where the graph changes and results need updated with minimal latency. We’ll also touch on issues of sensitivity and reliability where graph analysis needs to learn from numerical analysis and linear algebra.
PR-317: MLP-Mixer: An all-MLP Architecture for VisionJinwon Lee
Computer Vision 분야에서 CNN은 과연 살아남을 수 있을까요?
안녕하세요 TensorFlow Korea 논문 읽기 모임 PR-12의 317번째 논문 리뷰입니다.
이번에는 Google Research, Brain Team의 MLP-Mixer: An all-MLP Architecture for Vision을 리뷰해보았습니다.
Attention의 공격도 버거운데 이번에는 MLP(Multi-Layer Perceptron)의 공격입니다.
MLP만을 사용해서 Image Classification을 하는데 성능도 좋고 속도도 빠르고....
구조를 간단히 소개해드리면 ViT(Vision Transformer)의 self-attention 부분을 MLP로 변경하였습니다.
MLP block 2개를 사용하여 하나는 patch(token)들 간의 연산을 하는데 사용하고, 하나는 patch 내부 연산을 하는데 사용합니다.
사실 MLP를 사용하긴 했지만 논문에도 언급되어 있듯이, 이 부분을 일종의 convolution이라고 볼 수 있는데요...
그래도 transformer 기반의 network이 가질 수밖에 없는 quadratic complexity를 linear로 낮춰주고
convolution의 inductive bias 거의 없이 아주아주 simple한 구조를 활용하여 이렇게 좋은 성능을 보여준 점이 멋집니다.
반면에 역시나 data를 많이 써야 한다거나, MLP의 한계인 fixed length의 input만 받을 수 있다는 점은 단점이라고 생각하는데요,
이 연구를 시작으로 MLP도 다시한번 조명받는 계기가 되면 좋을 것 같네요
비슷한 시점에 나온 비슷한 연구들도 마지막에 간략하게 소개하였습니다.
재미있게 봐주세요. 감사합니다!
논문링크: https://arxiv.org/abs/2105.01601
영상링크: https://youtu.be/KQmZlxdnnuY
https://github.com/telecombcn-dl/dlmm-2017-dcu
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.
TensorFlow Korea 논문읽기모임 PR12 243째 논문 review입니다
이번 논문은 RegNet으로 알려진 Facebook AI Research의 Designing Network Design Spaces 입니다.
CNN을 디자인할 때, bottleneck layer는 정말 좋을까요? layer 수는 많을 수록 높은 성능을 낼까요? activation map의 width, height를 절반으로 줄일 때(stride 2 혹은 pooling), channel을 2배로 늘려주는데 이게 최선일까요? 혹시 bottleneck layer가 없는 게 더 좋지는 않은지, 최고 성능을 내는 layer 수에 magic number가 있는 건 아닐지, activation이 절반으로 줄어들 때 channel을 2배가 아니라 3배로 늘리는 게 더 좋은건 아닌지?
이 논문에서는 하나의 neural network을 잘 design하는 것이 아니라 Auto ML과 같은 기술로 좋은 neural network을 찾을 수 있는 즉 좋은 neural network들이 살고 있는 좋은 design space를 design하는 방법에 대해서 얘기하고 있습니다. constraint이 거의 없는 design space에서 human-in-the-loop을 통해 좋은 design space로 그 공간을 좁혀나가는 방법을 제안하였는데요, EfficientNet보다 더 좋은 성능을 보여주는 RegNet은 어떤 design space에서 탄생하였는지 그리고 그 과정에서 우리가 당연하게 여기고 있었던 design choice들이 잘못된 부분은 없었는지 아래 동영상에서 확인하실 수 있습니다~
영상링크: https://youtu.be/bnbKQRae_u4
논문링크: https://arxiv.org/abs/2003.13678
Distance-based bias in model-directed optimization of additively decomposable...Martin Pelikan
For many optimization problems it is possible to define a distance metric between problem variables that correlates with the likelihood and strength of interactions between the variables. For example, one may define a metric so that the dependencies between variables that are closer to each other with respect to the metric are expected to be stronger than the dependencies between variables that are further apart. The purpose of this paper is to describe a method that combines such a problem-specific distance metric with information mined from probabilistic models obtained in previous runs of estimation of distribution algorithms with the goal of solving future problem instances of similar type with increased speed, accuracy and reliability. While the focus of the paper is on additively decomposable problems and the hierarchical Bayesian optimization algorithm, it should be straightforward to generalize the approach to other model-directed optimization techniques and other problem classes. Compared to other techniques for learning from experience put forward in the past, the proposed technique is both more practical and more broadly applicable.
Recurrent Neural Networks have shown to be very powerful models as they can propagate context over several time steps. Due to this they can be applied effectively for addressing several problems in Natural Language Processing, such as Language Modelling, Tagging problems, Speech Recognition etc. In this presentation we introduce the basic RNN model and discuss the vanishing gradient problem. We describe LSTM (Long Short Term Memory) and Gated Recurrent Units (GRU). We also discuss Bidirectional RNN with an example. RNN architectures can be considered as deep learning systems where the number of time steps can be considered as the depth of the network. It is also possible to build the RNN with multiple hidden layers, each having recurrent connections from the previous time steps that represent the abstraction both in time and space.
A comprehensive tutorial on Convolutional Neural Networks (CNN) which talks about the motivation behind CNNs and Deep Learning in general, followed by a description of the various components involved in a typical CNN layer. It explains the theory involved with the different variants used in practice and also, gives a big picture of the whole network by putting everything together.
Next, there's a discussion of the various state-of-the-art frameworks being used to implement CNNs to tackle real-world classification and regression problems.
Finally, the implementation of the CNNs is demonstrated by implementing the paper 'Age ang Gender Classification Using Convolutional Neural Networks' by Hassner (2015).
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8Hakky St
This is the documentation of the study-meeting in lab.
Tha book title is "Hands-On Machine Learning with Scikit-Learn and TensorFlow" and this is the chapter 8.
Ant colony search and heuristic techniques for optimal dispatch of energy sou...Beniamino Murgante
Ant colony search and heuristic techniques for optimal dispatch of energy sources in micro-grids - Eleonora Riva Sanseverino – University of Palermo (Italy)
Intelligent Analysis of Environmental Data (S4 ENVISA Workshop 2009)
High-performance graph analysis is unlocking knowledge in computer security, bioinformatics, social networks, and many other data integration areas. Graphs provide a convenient abstraction for many data problems beyond linear algebra. Some problems map directly to linear algebra. Others, like community detection, look eerily similar to sparse linear algebra techniques. And then there are algorithms that strongly resist attempts at making them look like linear algebra. This talk will cover recent results with an emphasis on streaming graph problems where the graph changes and results need updated with minimal latency. We’ll also touch on issues of sensitivity and reliability where graph analysis needs to learn from numerical analysis and linear algebra.
PR-317: MLP-Mixer: An all-MLP Architecture for VisionJinwon Lee
Computer Vision 분야에서 CNN은 과연 살아남을 수 있을까요?
안녕하세요 TensorFlow Korea 논문 읽기 모임 PR-12의 317번째 논문 리뷰입니다.
이번에는 Google Research, Brain Team의 MLP-Mixer: An all-MLP Architecture for Vision을 리뷰해보았습니다.
Attention의 공격도 버거운데 이번에는 MLP(Multi-Layer Perceptron)의 공격입니다.
MLP만을 사용해서 Image Classification을 하는데 성능도 좋고 속도도 빠르고....
구조를 간단히 소개해드리면 ViT(Vision Transformer)의 self-attention 부분을 MLP로 변경하였습니다.
MLP block 2개를 사용하여 하나는 patch(token)들 간의 연산을 하는데 사용하고, 하나는 patch 내부 연산을 하는데 사용합니다.
사실 MLP를 사용하긴 했지만 논문에도 언급되어 있듯이, 이 부분을 일종의 convolution이라고 볼 수 있는데요...
그래도 transformer 기반의 network이 가질 수밖에 없는 quadratic complexity를 linear로 낮춰주고
convolution의 inductive bias 거의 없이 아주아주 simple한 구조를 활용하여 이렇게 좋은 성능을 보여준 점이 멋집니다.
반면에 역시나 data를 많이 써야 한다거나, MLP의 한계인 fixed length의 input만 받을 수 있다는 점은 단점이라고 생각하는데요,
이 연구를 시작으로 MLP도 다시한번 조명받는 계기가 되면 좋을 것 같네요
비슷한 시점에 나온 비슷한 연구들도 마지막에 간략하게 소개하였습니다.
재미있게 봐주세요. 감사합니다!
논문링크: https://arxiv.org/abs/2105.01601
영상링크: https://youtu.be/KQmZlxdnnuY
https://github.com/telecombcn-dl/dlmm-2017-dcu
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.
TensorFlow Korea 논문읽기모임 PR12 243째 논문 review입니다
이번 논문은 RegNet으로 알려진 Facebook AI Research의 Designing Network Design Spaces 입니다.
CNN을 디자인할 때, bottleneck layer는 정말 좋을까요? layer 수는 많을 수록 높은 성능을 낼까요? activation map의 width, height를 절반으로 줄일 때(stride 2 혹은 pooling), channel을 2배로 늘려주는데 이게 최선일까요? 혹시 bottleneck layer가 없는 게 더 좋지는 않은지, 최고 성능을 내는 layer 수에 magic number가 있는 건 아닐지, activation이 절반으로 줄어들 때 channel을 2배가 아니라 3배로 늘리는 게 더 좋은건 아닌지?
이 논문에서는 하나의 neural network을 잘 design하는 것이 아니라 Auto ML과 같은 기술로 좋은 neural network을 찾을 수 있는 즉 좋은 neural network들이 살고 있는 좋은 design space를 design하는 방법에 대해서 얘기하고 있습니다. constraint이 거의 없는 design space에서 human-in-the-loop을 통해 좋은 design space로 그 공간을 좁혀나가는 방법을 제안하였는데요, EfficientNet보다 더 좋은 성능을 보여주는 RegNet은 어떤 design space에서 탄생하였는지 그리고 그 과정에서 우리가 당연하게 여기고 있었던 design choice들이 잘못된 부분은 없었는지 아래 동영상에서 확인하실 수 있습니다~
영상링크: https://youtu.be/bnbKQRae_u4
논문링크: https://arxiv.org/abs/2003.13678
Distance-based bias in model-directed optimization of additively decomposable...Martin Pelikan
For many optimization problems it is possible to define a distance metric between problem variables that correlates with the likelihood and strength of interactions between the variables. For example, one may define a metric so that the dependencies between variables that are closer to each other with respect to the metric are expected to be stronger than the dependencies between variables that are further apart. The purpose of this paper is to describe a method that combines such a problem-specific distance metric with information mined from probabilistic models obtained in previous runs of estimation of distribution algorithms with the goal of solving future problem instances of similar type with increased speed, accuracy and reliability. While the focus of the paper is on additively decomposable problems and the hierarchical Bayesian optimization algorithm, it should be straightforward to generalize the approach to other model-directed optimization techniques and other problem classes. Compared to other techniques for learning from experience put forward in the past, the proposed technique is both more practical and more broadly applicable.
Recurrent Neural Networks have shown to be very powerful models as they can propagate context over several time steps. Due to this they can be applied effectively for addressing several problems in Natural Language Processing, such as Language Modelling, Tagging problems, Speech Recognition etc. In this presentation we introduce the basic RNN model and discuss the vanishing gradient problem. We describe LSTM (Long Short Term Memory) and Gated Recurrent Units (GRU). We also discuss Bidirectional RNN with an example. RNN architectures can be considered as deep learning systems where the number of time steps can be considered as the depth of the network. It is also possible to build the RNN with multiple hidden layers, each having recurrent connections from the previous time steps that represent the abstraction both in time and space.
A comprehensive tutorial on Convolutional Neural Networks (CNN) which talks about the motivation behind CNNs and Deep Learning in general, followed by a description of the various components involved in a typical CNN layer. It explains the theory involved with the different variants used in practice and also, gives a big picture of the whole network by putting everything together.
Next, there's a discussion of the various state-of-the-art frameworks being used to implement CNNs to tackle real-world classification and regression problems.
Finally, the implementation of the CNNs is demonstrated by implementing the paper 'Age ang Gender Classification Using Convolutional Neural Networks' by Hassner (2015).
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8Hakky St
This is the documentation of the study-meeting in lab.
Tha book title is "Hands-On Machine Learning with Scikit-Learn and TensorFlow" and this is the chapter 8.
The Comprehensive Product Platform Planning (CP3) framework presents a flexible mathematical model of the platform planning process, which allows (i) the formation of sub-families of products, and (ii) the simultaneous identification and quantification of plat- form/scaling design variables. The CP3 model is founded on a generalized commonality matrix that represents the product platform plan, and yields a mixed binary-integer non- linear programming problem. In this paper, we develop a methodology to reduce the high dimensional binary integer problem to a more tractable integer problem, where the com- monality matrix is represented by a set of integer variables. Subsequently, we determine the feasible set of values for the integer variables in the case of families with 3 − 7 kinds of products. The cardinality of the feasible set is found to be orders of magnitude smaller than the total number of unique combinations of the commonality variables. In addition, we also present the development of a generalized approach to Mixed-Discrete Non-Linear Optimization (MDNLO) that can be implemented through standard non-gradient based op- timization algorithms. This MDNLO technique is expected to provide a robust and compu- tationally inexpensive optimization framework for the reduced CP3 model. The generalized approach to MDNLO uses continuous optimization as the primary search strategy, how- ever, evaluates the system model only at the feasible locations in the discrete variable space.
Improving Hardware Efficiency for DNN ApplicationsChester Chen
Speaker: Dr. Hai (Helen) Li is the Clare Boothe Luce Associate Professor of Electrical and Computer Engineering and Co-director of the Duke Center for Evolutionary Intelligence at Duke University
In this talk, I will introduce a few recent research spotlights by the Duke Center for Evolutionary Intelligence. The talk will start with the structured sparsity learning (SSL) method which attempts to learn a compact structure from a bigger DNN to reduce computation cost. It generates a regularized structure with high execution efficiency. Our experiments on CPU, GPU, and FPGA platforms show on average 3~5 times speedup of convolutional layer computation of AlexNet. Then, the implementation and acceleration of DNN applications on mobile computing systems will be introduced. MoDNN is a local distributed system which partitions DNN models onto several mobile devices to accelerate computations. ApesNet is an efficient pixel-wise segmentation network, which understands road scenes in real-time, and has achieved promising accuracy. Our prospects on the adoption of emerging technology will also be given at the end of this talk, offering the audiences an alternative thinking about the future evolution and revolution of modern computing systems.
Similar to T12 Distributed search and constraint handling (20)
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
Biological screening of herbal drugs: Introduction and Need for
Phyto-Pharmacological Screening, New Strategies for evaluating
Natural Products, In vitro evaluation techniques for Antioxidants, Antimicrobial and Anticancer drugs. In vivo evaluation techniques
for Anti-inflammatory, Antiulcer, Anticancer, Wound healing, Antidiabetic, Hepatoprotective, Cardio protective, Diuretics and
Antifertility, Toxicity studies as per OECD guidelines
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
Executive Directors Chat Leveraging AI for Diversity, Equity, and InclusionTechSoup
Let’s explore the intersection of technology and equity in the final session of our DEI series. Discover how AI tools, like ChatGPT, can be used to support and enhance your nonprofit's DEI initiatives. Participants will gain insights into practical AI applications and get tips for leveraging technology to advance their DEI goals.
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Dr. Vinod Kumar Kanvaria
Exploiting Artificial Intelligence for Empowering Researchers and Faculty,
International FDP on Fundamentals of Research in Social Sciences
at Integral University, Lucknow, 06.06.2024
By Dr. Vinod Kumar Kanvaria
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...Levi Shapiro
Letter from the Congress of the United States regarding Anti-Semitism sent June 3rd to MIT President Sally Kornbluth, MIT Corp Chair, Mark Gorenberg
Dear Dr. Kornbluth and Mr. Gorenberg,
The US House of Representatives is deeply concerned by ongoing and pervasive acts of antisemitic
harassment and intimidation at the Massachusetts Institute of Technology (MIT). Failing to act decisively to ensure a safe learning environment for all students would be a grave dereliction of your responsibilities as President of MIT and Chair of the MIT Corporation.
This Congress will not stand idly by and allow an environment hostile to Jewish students to persist. The House believes that your institution is in violation of Title VI of the Civil Rights Act, and the inability or
unwillingness to rectify this violation through action requires accountability.
Postsecondary education is a unique opportunity for students to learn and have their ideas and beliefs challenged. However, universities receiving hundreds of millions of federal funds annually have denied
students that opportunity and have been hijacked to become venues for the promotion of terrorism, antisemitic harassment and intimidation, unlawful encampments, and in some cases, assaults and riots.
The House of Representatives will not countenance the use of federal funds to indoctrinate students into hateful, antisemitic, anti-American supporters of terrorism. Investigations into campus antisemitism by the Committee on Education and the Workforce and the Committee on Ways and Means have been expanded into a Congress-wide probe across all relevant jurisdictions to address this national crisis. The undersigned Committees will conduct oversight into the use of federal funds at MIT and its learning environment under authorities granted to each Committee.
• The Committee on Education and the Workforce has been investigating your institution since December 7, 2023. The Committee has broad jurisdiction over postsecondary education, including its compliance with Title VI of the Civil Rights Act, campus safety concerns over disruptions to the learning environment, and the awarding of federal student aid under the Higher Education Act.
• The Committee on Oversight and Accountability is investigating the sources of funding and other support flowing to groups espousing pro-Hamas propaganda and engaged in antisemitic harassment and intimidation of students. The Committee on Oversight and Accountability is the principal oversight committee of the US House of Representatives and has broad authority to investigate “any matter” at “any time” under House Rule X.
• The Committee on Ways and Means has been investigating several universities since November 15, 2023, when the Committee held a hearing entitled From Ivory Towers to Dark Corners: Investigating the Nexus Between Antisemitism, Tax-Exempt Universities, and Terror Financing. The Committee followed the hearing with letters to those institutions on January 10, 202
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
1. Distributed Constraint Handling
and Optimization
Alessandro Farinelli1 Alex Rogers2 Meritxell Vinyals1
1 Computer Science Department
University of Verona, Italy
2 Agents, Interaction and Complexity Group
School of Electronics and Computer Science
University of Southampton,UK
Tutorial EASSS 2012 Valencia
https://sites.google.com/site/
easss2012optimization/
4. Constraints
• Pervade our everyday lives
• Are usually perceived as elements that limit solutions to the
problems we face
5. Constraints
From a computational point of view, they:
• Reduce the space of possible solutions
• Encode knowledge about the problem at hand
• Are key components for efficiently solving hard problems
6. Constraint Processing
Many different disciplines deal with hard computational problems that
can be made tractable by carefully considering the constraints that
define the structure of the problem.
Planning Operational Automated Reasoning Computer
Scheduling Research Decision Theory Vision
7. Constraint Processing in Multi-Agent Systems
Focus on how constraint processing can be used to address
optimization problems in Multi-Agent Systems (MAS) where:
A set of agents must come to some agreement, typically via some
form of negotiation, about which action each agent should take in
order to jointly obtain the best solution for the whole system.
M2
M1
A1 A2 A2 A3
8. Distributed Constraint Optimization Problems (DCOPs)
We will consider Distributed Constraint Optimization Problems (DCOP)
where:
Each agent negotiates locally with just a subset of other agents
(usually called neighbors) that are those that can directly influence
his/her behavior.
M2
M1
A2
A1 A3
9. Distributed Constraint Optimization Problems (DCOPs)
After reading this chapter, you will understand:
• The mathematical formulation of a DCOP
• The main exact solution techniques for DCOPs
• Key differences, benefits and limitations
• The main approximate solution techniques for DCOPs
• Key differences, benefits and limitations
• The quality guarantees these approach provide:
• Types of quality guarantees
• Frameworks and techniques
11. Constraint Networks
A constraint network N is formally defined as a tuple X, D,C where:
• X = {x1 , . . . , xn } is a set of discrete variables;
• D = {D1 , . . . , Dn } is a set of variable domains, which enumerate
all possible values of the corresponding variables; and
• C = {C1 , . . . ,Cm } is a set of constraints; where a constraint Ci is
defined on a subset of variables Si ⊆ X which comprise the
scope of the constraint
• r = |Si | is the arity of the constraint
• Two types: hard or soft
12. Hard constraints
• A hard constraint Cih is a relation Ri that enumerates all the valid
joint assignments of all variables in the scope of the constraint.
Ri ⊆ Di1 × . . . × Dir
Ri xj xk
0 1
1 0
13. Soft constraints
• A soft constraint Cis is a function Fi that maps every possible joint
assignment of all variables in the scope to a real value.
Fi : Di1 × . . . × Dir → ℜ
Fi xj xk
2 0 0
0 0 1
0 1 0
1 1 1
14. Binary Constraint Networks
x1
• Binary constraint networks are those where:
F1,3
,2
• Each constraint (soft or hard) is defined
F1
over two variables.
• Every constraint network can be mapped to
a binary constraint network x2 x3
4
F1,
• requires the addition of variables and
constraints
F2,4
• may add complexity to the model
• They can be represented by a constraint
graph x4
15. Different objectives, different problems
• Constraint Satisfaction Problem (CSP)
• Objective: find an assignment for all the variables in the network
that satisfies all constraints.
• Constraint Optimization Problem (COP)
• Objective: find an assignment for all the variables in the network
that satisfies all constraints and optimizes a global function.
• Global function = aggregation (typically sum) of local functions.
F(x) = ∑i Fi (xi )
16. Distributed Constraint Reasoning
A1
When operating in a
decentralized context:
• a set of agents control
variables
A2
• agents interact to find a A4
solution to the constraint
network
A3
17. Distributed Constraint Reasoning
Two types of decentralized problems:
• distributed CSP (DCSP)
• distributed COP (DCOP)
Here, we focus on DCOPs.
18. Distributed Constraint Optimization Problem (DCOP)
A DCOP consists of a constraint network N = X, D,C and a set of
agents A = {A1 , . . . , Ak } where each agent:
• controls a subset of the variables Xi ⊆ X
• is only aware of constraints that involve variable it controls
• communicates only with its neighbours
19. Distributed Constraint Optimization Problem (DCOP)
• Agents are assumed to be fully cooperative
• Goal: find the assignment that optimizes the global function, not
their local local utilities.
• Solving a COP is NP-Hard and DCOP is as hard as COP.
22. Real World Applications
Many standard benchmark problems in computer science can be
modeled using the DCOP framework:
• graph coloring
As can many real world applications:
• human-agent organizations (e.g. meeting scheduling)
• sensor networks and robotics (e.g. target tracking)
24. Graph coloring
• Popular benchmark
• Simple formulation
• Complexity controlled with few parameters:
• Number of available colors
• Number of nodes
• Density (#nodes/#constraints)
• Many versions of the problem:
• CSP, MaxCSP, COP
25. Graph coloring - CSP
• Nodes can take k colors
• Any two adjacent nodes should have different colors
• If it happens this is a conflict
Yes! No!
27. Graph coloring - COP
• Different weights to violated constraints
• Preferences for different colors
0 -2
-1
-3 -2
-1
28. Graph coloring - DCOP
• Each node:
• controlled by one agent
• Each agent:
• Preferences for different colors
• Communicates with its direct neighbours in the graph
-1
A1 A3 • A1 and A2 exchange
-3 -2 preferences and conflicts
• A3 and A4 do not
A2 A4 communicate
-1
33. Meeting Scheduling
Two meetings that share a participant cannot overlap
Window: 15:00-18:00
Duration: 2h
Window: 15:00-17:00
Duration: 1h
34. DCOP formalization for the meeting scheduling problem
• A set of agents representing participants
• A set of variables representing meeting starting times according
to a participant.
• Hard Constraints:
• Starting meeting times across different agents are equal
• Meetings for the same agent are non-overlapping.
• Soft Constraints:
• Represent agent preferences on meeting starting times.
Objective: find a valid schedule for the meeting while maximizing the
sum of individuals’ preferences.
37. Target Tracking
A set of sensors tracking a set of targets in order to provide an
accurate estimate of their positions.
T4
T1
T3
T2
Crucial for surveillance and monitoring applications.
38. Target Tracking
Sensors can have different sensing modalities that impact on the
accuracy of the estimation of the targets’ positions.
MODES
T4
MODES MODES
T1 T3
MODES T2
40. DCOP formalization for the target tracking problem
• Agents represent sensors
• Variables encode the different sensing modalities of each sensor
• Constraints
• relate to a specific target
• represent how sensor modalities impacts on the tracking
performance
• Objective:
• Maximize coverage of the environment
• Provide accurate estimations of potentially dangerous targets
42. Complete Algorithms
U
D Always find an optimal solution
D Exhibit an exponentially increasing coordination overhead
Very limited scalability on general problems.
46. ADOPT
ADOPT (Asynchronous Distributed OPTimization) [Modi et al., 2005]:
• Distributed backtrack search using a best-first strategy
• Best value based on local information:
• Lower/upper bound estimates of each possible value of its
variable
• Backtrack thresholds used to speed up the search of previously
explored solutions.
• Termination conditions that check if the bound interval is less than
a given valid error bound (0 if optimal)
47. ADOPT by example
4 variables (4 agents): x1 , x2 , x3 , x4 with D = {0, 1}
4 identical cost functions
x1 Fi, j xi xj
2 0 0
F1,3
,2
F1
0 0 1
0 1 0
x2 x3
1 1 1
4
F1,
F2,4
Goal: find a variable assignment with minimal cost
x4 Solution: x1 = 1, x2 = 0, x3 = 0 and x4 = 1
giving total cost 1.
48. DFS arrangement
• Before executing ADOPT, agents must be arranged in a depth
first search (DFS) tree.
• DFS trees have been frequently used in optimization because
they have two interesting properties:
• Agents in different branches of the tree do not share any
constraints;
• Every constraint network admits a DFS tree.
49. ADOPT by example
A1 (root)
x1
→
←
Fi, j xi xj
ent
par
par
ent ,
2 0 0
F1,3
,2
,
chil
hil d
F1
0 0 1
d→
←c
0 1 0
x2 x3 1 1 1 A2 A3
4
F1,
→
arent
F2,4
→
il d , p
← ch
x4 DFS arrangement
A4
50. Cost functions
The local cost function for an agent Ai (δ (xi )) is the sum of the values
of constraints involving only higher neighbors in the DFS.
52. Initialization
Each agent initially chooses a random value for their variables and
initialize the lower and upper bounds to zero and infinity respectively.
x1 = 0, LB = 0,UB = ∞ A1
x2 = 0, LB = 0,UB = ∞ A2 A3 x3 = 0, LB = 0,UB = ∞
x4 = 0, LB = 0,UB = ∞ A4
53. ADOPT by example
Value messages are sent by an agent to all its neighbors that are
lower in the DFS tree
x1 = 0 A1
← =0
x 1−−
−
−
−−
=→
x1
0 A1 sends three value
message to A2 , A3 and
x2 = 0 A2 A3 x3 = 0
←−0
A4 informing them that its
−−
x1 =
←−0
−−
current value is 0.
x2 =
x4 = 0 A4
54. ADOPT by example
Current Context: a partial variable assignment maintained by each
agent that records the assignment of all higher neighbours in the DFS.
A1
• Updated by each VALUE
message
c2 : {x1 = 0} A2 A3
• If current context is not
c3 : {x1 = 0} compatible with some child
context, the latter is re-initialized
A4 (also the child bound)
c4 : {x1 = 0, x2 = 0}
55. ADOPT by example
Each agent Ai sends a cost message to its parent A p
A1
, c2 ]
[0, 0
Each cost message reports:
[0, ∞
,
c 3]
• The minimum lower bound (LB)
• The maximum upper bound (UB)
A2 A3
• The context (ci )
4] c
[0, 0,
[LB,UP, ci ]
A4
56. Lower bound computation
Each agent evaluates for each possible value of its variable:
• its local cost function with respect to the current context
• adding all the compatible lower bound messages received from
children.
Analogous computation for upper bounds
57. ADOPT by example
Consider the lower bound in the cost message sent by A4 :
A1 • Recall that A4 local cost function is:
δ (x1 , x2 , x4 ) = F1,4 (x1 , x4 ) + F2,4 (x2 , x4 )
• Restricted to the current context
c4 = {(x1 = 0, x2 = 0)}:
λ (0, 0, x4 ) = F1,4 (0, x4 ) + F2,4 (0, x4 ).
A2 A3 • For x4 = 0:
λ (0, 0, 0) = F1,4 (0, 0) + F2,4 (0, 0) = 2 + 2 = 4.
4] c
• For x4 = 1:
[0, 0,
λ (0, 0, 1) = F1,4 (0, 1) + F2,4 (0, 1) = 0 + 0 = 0.
A4 Then the minimum lower bound across variable
values is LB = 0.
58. ADOPT by example
Each agent asynchronously chooses the value of its variable that
minimizes its lower bound.
A2 computes for each possible value of its
A1
variable its local function restricted to the
, c2 ]
current context c2 = {(x1 = 0)}
[0, 2
(λ (0, x2 ) = F1,2 (0, x2 )) and adding lower
bound message from A4 (lb).
x2 = 0 → 1 A2 A3 • For x2 = 0: LB(x2 = 0) = λ (0, x2 =
0) + lb(x2 = 0) = 2 + 0 = 2.
1
x2 =
• For x2 = 1: LB(x2 = 1) = λ (0, x2 =
1) + 0 = 0 + 0 = 0.
A4
A2 changes its value to x2 = 1 with LB = 0.
59. Backtrack thresholds
The search strategy is based on lower bounds
Problem
• Values abandoned before proven to be
suboptimal
• Lower/upper bounds only stored for the
current context
Solution
• Backtrack thresholds: used to speed up
the search of previously explored
solutions.
60. ADOPT by example
x1 = 0 → 1 → 0
A1
A1 changes its value and the context with
x1 = 0 is visited again.
• Reconstructing from scratch is inefficient
A2 A3
• Remembering solutions is expensive
A4
61. Backtrack thresholds
Solution: Backtrack thresholds
• Lower bound previously determined by children
• Polynomial space
• Control backtracking to efficiently search
• Key point: do not change value until LB(currentvalue)> threshold
62. A child agent will not change its variable value so long as cost is less
than the backtrack threshold given to it by its parent.
LB(x1 = 0) = 1 A1
t (x 1 =
1
2
0) =
0) =
t (x1 =
1
2
1 1
LB(x2 = 0) > 2 ? A2 A3 LB(x3 = 0) > 2 ?
A4
63. Rebalance incorrect threshold
How to correctly subdivide threshold among children?
• Parent distributes the accumulated bound among children
• Arbitrarily/Using some heuristics
• Correct subdivision as feedback is received from children
• LB < t(CONT EXT )
• t(CONT EXT ) = ∑Ci t(CONT EXT ) + δ
64. Backtrack Threshold Computation
A1
(2) 2 t
1
=1
= 0) =
1 (x 1 =
(1) LB
1
• When A1 receives a new lower bound
x
0) = 0
(2) t (
from A2 rebalances thresholds
• A1 resends threshold messages to A2
A2 A3 and A3
A4
65. ADOPT extensions
• BnB-ADOPT [Yeoh et al., 2008] reduces computation time by
using depth-first search with branch and bound strategy
• [Ali et al., 2005] suggest the use of preprocessing techniques for
guiding ADOPT search and show that this can result in a
consistent increase in performance.
67. DPOP
DPOP (Dynamic Programming Optimization Protocol) [Petcu and
Faltings, 2005]:
• Based on the dynamic programming paradigm.
• Special case of Bucket Tree Elimination Algorithm (BTE)
[Dechter, 2003].
68. DPOP by example
x1 x1
Fi, j xi xj
← P2 , child →
F1,3
,2
F1
→
← PP3
2 0 0
4
il d , PP
,4
0 0 1
F2,3
F1
, pseud
x2 x3 0 1 0
ud och
1 1 1
ochil d
x2
← pse
,4
F2
→
←
→
=>
P3,
4
,P
chi
ld
chi
ld
x4
DFS arrangement
←
→
Objective: find assignment x4 x3
with maximal value
69. DPOP phases
Given a DFS tree structure, DPOP runs in two phases:
• Util propagation: agents exchange util messages up the tree.
• Aim: aggregate all info so that root agent can choose optimal
value
• Value propagation: agents exchange value messages down the
tree.
• Aim: propagate info so that all agents can make their choice given
choices of ancestors
70. Sepi : set of agents preceding Ai in the pseudo-tree order that are
connected with Ai or with a descendant of Ai .
x1 Sep1 = 0
/
← P2 , child →
→
← PP3
4
il d , PP
, pseud
ud och
ochil d
x2 Sep2 = {x1 }
← pse
→
←
→
P3,
4
,P
chi
ld
chi
ld
←
→
Sep4 = {x1 , x2 } x4 x3 Sep3 = {x1 , x2 }
71. Util message
The Util message Ui→ j that agent Ai sends to its parent A j can be
computed as:
Ui→ j (Sepi ) = max Uk→i ⊗ Fi,p
xi Ak ∈Ci A p ∈Pi ∪PPi
Size exponential All incoming messages Shared constraints with
in Sepi from children parents/pseudoparents
The ⊗ operator is a join operator that sums up functions with different
but overlapping scores consistently.
74. Value message
Keeping fixed the value of parent/pseudoparents, finds the value that
maximizes the computed cost function in the util phase:
∗
xi = arg max ∑ U j→i (xi , x∗ ) +
p ∑ Fi, j (xi , x∗ )
j
xi
A j ∈Ci A j ∈Pi ∪PPi
where x∗ = A j ∈Pi ∪PPi {x∗ } is the set of optimal values for Ai ’s parent
p j
and pseudoparents received from Ai ’s parent.
Propagates this value through children down the tree:
∗ ∗
Vi→ j = {xi = xi } ∪ {xs = xs }
xs ∈Sepi ∩Sep j
76. DPOP extensions
• MB-DPOP [Petcu and Faltings, 2007] trades-off message size
against the number of messages.
• A-DPOP trades-off message size against solution quality [Petcu
and Faltings, 2005(2)].
77. Conclusions
• Constraint processing
• exploit problem structure to solve hard problems efficiently
• DCOP framework
• applies constraint processing to solve decision making problems
in Multi-Agent Systems
• increasingly being applied within real world problems.
78. References I
• [Modi et al., 2005] P. J. Modi, W. Shen, M. Tambe, and M.Yokoo. ADOPT: Asynchronous
distributed constraint optimization with quality guarantees. Artificial Intelligence Jour- nal,
(161):149-180, 2005.
• [Yeoh et al., 2008] W. Yeoh, A. Felner, and S. Koenig. BnB-ADOPT: An asynchronous
branch-and-bound DCOP algorithm. In Proceedings of the Seventh International Joint
Conference on Autonomous Agents and Multiagent Systems, pages 591Ð598, 2008.
• [Ali et al., 2005] S. M. Ali, S. Koenig, and M. Tambe. Preprocessing techniques for
accelerating the DCOP algorithm ADOPT. In Proceedings of the Fourth International Joint
Conference on Autonomous Agents and Multiagent Systems, pages 1041Ð1048, 2005.
• [Petcu and Faltings, 2005] A. Petcu and B. Faltings. DPOP: A scalable method for
multiagent constraint opti- mization. In Proceedings of the Nineteenth International Joint
Conference on Arti- ficial Intelligence, pages 266-271, 2005.
• [Dechter, 2003] R. Dechter. Constraint Processing. Morgan Kaufmann, 2003.
79. References II
• [Petcu and Faltings, 2005(2)] A. Petcu and B. Faltings. A-DPOP: Approximations in
distributed optimization. In Principles and Practice of Constraint Programming, pages
802-806, 2005.
• [Petcu and Faltings, 2007] A. Petcu and B. Faltings. MB-DPOP: A new memory-bounded
algorithm for distributed optimization. In Proceedings of the Twentieth International Joint
Confer- ence on Artificial Intelligence, pages 1452-1457, 2007.
• [S. Fitzpatrick and L. Meetrens, 2003] S. Fitzpatrick and L. Meetrens. Distributed Sensor
Networks: A multiagent perspective, chapter Distributed coordination through anarchic
optimization, pages 257- 293. Kluwer Academic, 2003.
• [R. T. Maheswaran et al., 2004] R. T. Maheswaran, J. P. Pearce, and M. Tambe.
Distributed algorithms for DCOP: A graphical game-based approach. In Proceedings of
the Seventeenth International Conference on Parallel and Distributed Computing
Systems, pages 432-439, 2004.
82. Why Approximate Algorithms
• Motivations
– Often optimality in practical applications is not achievable
– Fast good enough solutions are all we can have
• Example – Graph coloring
– Medium size problem (about 20 nodes, three colors per
node)
– Number of states to visit for optimal solution in the worst
case 3^20 = 3 billions of states
• Key problem
– Provides guarantees on solution quality
83. Exemplar Application: Surveillance
• Event Detection
– Vehicles passing on a road
• Energy Constraints
– Sense/Sleep modes
– Recharge when sleeping
• Coordination
– Activity can be detected
by single sensor duty cycle
– Roads have different time
traffic loads Good Schedule
• Aim [Rogers et al. 10] Bad Schedule
– Focus on road with more
traffic load Heavy traffic road small road
85. Guarantees on solution quality
• Key Concept: bound the optimal solution
– Assume a maximization problem
– optimal solution, a solution
–
– percentage of optimality
• [0,1]
• The higher the better
– approximation ratio
• >= 1
• The lower the better
– is the bound
86. Types of Guarantees
Instance-specific Accuracy: high alpha
Generality: less use of
Bounded Max- instance specific knowledge
Sum
DaCSA
Accuracy
Instance-generic
No guarantees
K-optimality
MGM-1,
T-optimality
DSA-1,
Max-Sum Region Opt.
Generality
87. Centralized Local Greedy approaches
• Greedy local search
– Start from random solution
– Do local changes if global solution improves
– Local: change the value of a subset of variables, usually one
-1 -1 -1
-4
-1
0
0 -2
-1 -1
-2
0
88. Centralized Local Greedy approaches
• Problems
– Local minima
– Standard solutions: RandomWalk, Simulated Annealing
-1 -1
-2
-1 -1
-1 -1 -1 -1 -1 -1
89. Distributed Local Greedy approaches
• Local knowledge
• Parallel execution:
– A greedy local move might be harmful/useless
– Need coordination
-1 -1 -1
-4
-1
0 0 -2
-2 0 0 -2
-2
-4
90. Distributed Stochastic Algorithm
• Greedy local search with activation probability to
mitigate issues with parallel executions
• DSA-1: change value of one variable at time
• Initialize agents with a random assignment and
communicate values to neighbors
• Each agent:
– Generates a random number and execute only if rnd less
than activation probability
– When executing changes value maximizing local gain
– Communicate possible variable change to neighbors
92. DSA-1: discussion
• Extremely “cheap” (computation/communication)
• Good performance in various domains
– e.g. target tracking [Fitzpatrick Meertens 03, Zhang et al. 03],
– Shows an anytime property (not guaranteed)
– Benchmarking technique for coordination
• Problems
– Activation probablity must be tuned [Zhang et al. 03]
– No general rule, hard to characterise results across domains
93. Maximum Gain Message (MGM-1)
• Coordinate to decide who is going to move
– Compute and exchange possible gains
– Agent with maximum (positive) gain executes
• Analysis [Maheswaran et al. 04]
– Empirically, similar to DSA
– More communication (but still linear)
– No Threshold to set
– Guaranteed to be monotonic (Anytime behavior)
94. MGM-1: Example
-1 -1
0 -2 -1 -1
-1 -1 0 -2
G = -2
G=0
G=0 G=2
95. Local greedy approaches
• Exchange local values for variables
– Similar to search based methods (e.g. ADOPT)
• Consider only local information when maximizing
– Values of neighbors
• Anytime behaviors
• Could result in very bad solutions
96. Max-sum
Agents iteratively computes local functions that depend
only on the variable they control
X1 X2
Choose arg max
X4 X3 Shared constraint
All incoming
messages except x2
All incoming
messages
97. Factor Graph and GDL
• Factor Graph
– [Kschischang, Frey, Loeliger 01]
– Computational framework to represent factored computation
– Bipartite graph, Variable - Factor
H ( X1, X 2 , X 3 ) = H ( X1) + H ( X 2 | X1) + H ( X 3 | X1 )
H ( X 2 | X1)
x1 x2 H ( X1) x1 x2
x3 H ( X 3 | X1)
x3
98. Max-Sum on acyclic graphs
• Max-sum Optimal on acyclic
graphs
– Different branches are H ( X 2 | X1)
independent
H ( X1) x1 x2
– Each agent can build a correct
estimation of its contribution to the
global problem (z functions)
• Message equations very similar
to Util messages in DPOP x3
– GDL generalizes DPOP [Vinyals
et al. 2010a] H ( X 3 | X1)
sum up info from other nodes
local maximization step
99. (Loopy) Max-sum Performance
• Good performance on loopy networks [Farinelli et al. 08]
– When it converges very good results
• Interesting results when only one cycle [Weiss 00]
– We could remove cycle but pay an exponential price (see
DPOP)
– Java Library for max-sum http://code.google.com/p/jmaxsum/
100. Max-Sum for low power devices
• Low overhead
– Msgs number/size
• Asynchronous computation
– Agents take decisions whenever new messages arrive
• Robust to message loss
106. Quality guarantees for approx.
techniques
• Key area of research
• Address trade-off between guarantees and
computational effort
• Particularly important for many real world applications
– Critical (e.g. Search and rescue)
– Constrained resource (e.g. Embedded devices)
– Dynamic settings
107. Instance-generic guarantees
Instance-specific
Bounded Max- Characterise solution quality without
Sum running the algorithm
DaCSA
Accuracy
Instance-generic
No guarantees
K-optimality
MGM-1,
T-optimality
DSA-1,
Max-Sum Region Opt.
Generality
108. K-Optimality framework
• Given a characterization of solution gives bound on
solution quality [Pearce and Tambe 07]
• Characterization of solution: k-optimal
• K-optimal solution:
– Corresponding value of the objective function can not be
improved by changing the assignment of k or less
variables.
110. Bounds for K-Optimality
For any DCOP with non-negative rewards [Pearce and Tambe 07]
Number of agents Maximum arity of constraints
K-optimal solution
Binary Network (m=2):
111. K-Optimality Discussion
• Need algorithms for computing k-optimal solutions
– DSA-1, MGM-1 k=1; DSA-2, MGM-2 k=2 [Maheswaran et al. 04]
– DALO for generic k (and t-optimality) [Kiekintveld et al. 10]
• The higher k the more complex the computation
(exponential)
Percentage of Optimal:
• The higher k the better
• The higher the number of
agents the worst
112. Trade-off between generality and solution
quality
• K-optimality based on worst case analysis
• assuming more knowledge gives much better bounds
• Knowledge on structure [Pearce and Tambe 07]
113. Trade-off between generality and
solution quality
• Knowledge on reward [Bowring et al. 08]
• Beta: ratio of least minimum reward to the maximum
114. Off-Line Guarantees: Region
Optimality
• k-optimality: use size as a criterion for optimality
• t-optimality: use distance to a central agent in the
constraint graph
• Region Optimality: define regions based on general
criteria (e.g. S-size bounded distance) [Vinyals et al 11]
• Ack: Meritxell Vinyals
3-size regions 1-distance regions C regions
x0 x0
x1 x2 x1 x2
x3 x3
x0 x0
x1 x2 x1 x2
x3 x3
x0 x1 x2 x3 x0 x1 x2 x3
115. Size-Bounded Distance
• Region optimality can explore new
regions: s-size bounded distance
• One region per agent, largest t- 3-size bounded distance
distance group whose size is less
than s x0 x0
x1 x2 x1 x2
• S-Size-bounded distance
x3 x3
– C-DALO extension of DALO for general
regions t=1 t=0
– Can provides better bounds and x0 x0
keep under control size and x1 x2 x1 x2
number of regions x3 x3
t=0 t=1
116. Max-Sum and Region Optimality
• Can use region optimality to provide bounds for Max-
sum [Vinyals et al 10b]
• Upon convergence Max-Sum is optimal on SLT regions of
the graph [Weiss 00]
• Single Loops and Trees (SLT): all groups of agents whose
vertex induced subgraph contains at most one cycle
x0
x0 x0
x1 x2 x1 x2
x1 x2 x3 x3
x0 x0
x3 x1 x2 x1 x2
x3 x3
117. Bounds for Max-Sum
• Complete: same as
3-size optimality
• bipartite
• 2D grids
119. Instance-specific guarantees
Instance-specific
Bounded Max- Characterise solution quality after/while
Sum running the algorithm
DaCSA
Accuracy
Instance-generic
No guarantees
K-optimality
MGM-1,
T-optimality
DSA-1,
Max-Sum Region Opt.
Generality
120. Bounded Max-Sum
Aim: Remove cycles from Factor Graph avoiding
exponential computation/communication (e.g. no junction tree)
Key Idea: solve a relaxed problem instance [Rogers et al.11]
X1 F2 X3 X1 F2 X3
Build Spanning tree
F1 X2 F3 F1 X2 F3
Compute Bound
Run Max-Sum
X1 X2 X3
Optimal solution on tree
121. Factor Graph Annotation
• Compute a weight for
each edge
– maximum possible impact X1 F2 X3
of the variable on the w21 w23
function
w11 w22 w33
w12 w32
F1 X2 F3
122. Factor Graph Modification
• Build a Maximum
Spanning Tree
– Keep higher weights
X1 F2 X3
• Cut remaining
w21 w23
dependencies
– Compute
• Modify functions w11 w22 w33
• Compute bound w12 w32
F1 X2 F3
W = w22 + w23
123. Results: Random Binary Network
Optimal
Bound is significant Approx.
– Approx. ratio is Lower Bound
typically 1.23 (81 %) Upper Bound
Comparison with k-optimal
with knowledge on
reward structure
Much more accurate less
general
124. Discussion
• Discussion with other data-dependent techniques
– BnB-ADOPT [Yeoh et al 09]
• Fix an error bound and execute until the error bound is met
• Worst case computation remains exponential
– ADPOP [Petcu and Faltings 05b]
• Can fix message size (and thus computation) or error bound and
leave the other parameter free
• Divide and coordinate [Vinyals et al 10]
– Divide problems among agents and negotiate agreement
by exchanging utility
– Provides anytime quality guarantees
125. Summary
• Approximation techniques crucial for practical applications:
surveillance, rescue, etc.
• DSA, MGM, Max-Sum heuristic approaches
– Low coordination overhead, acceptable performance
– No guarantees (convergence, solution quality)
• Instance generic guarantees:
– K-optimality framework
– Loose bounds for large scale systems
• Instance specific guarantees
– Bounded max-sum, ADPOP, BnB-ADOPT
– Performance depend on specific instance
126. References I
DOCPs for MRS
• [Delle Fave et al 12] A methodology for deploying the max-sum algorithm and a case study on
unmanned aerial vehicles. In, IAAI 2012
• [Taylor et al. 11] Distributed On-line Multi-Agent Optimization Under Uncertainty: Balancing
Exploration and Exploitation, Advances in Complex Systems
MGM
• [Maheswaran et al. 04] Distributed Algorithms for DCOP: A Graphical Game-Based Approach,
PDCS-2004
DSA
• [Fitzpatrick and Meertens 03] Distributed Coordination through Anarchic Optimization,
Distributed Sensor Networks: a multiagent perspective.
• [Zhang et al. 03] A Comparative Study of Distributed Constraint algorithms, Distributed
Sensor Networks: a multiagent perspective.
Max-Sum
• [Stranders at al 09] Decentralised Coordination of Mobile Sensors Using the Max-Sum
Algorithm, AAAI 09
• [Rogers et al. 10] Self-organising Sensors for Wide Area Surveillance Using the Max-sum
Algorithm, LNCS 6090 Self-Organizing Architectures
• [Farinelli et al. 08] Decentralised coordination of low-power embedded devices using the
max-sum algorithm, AAMAS 08
127. References II
Instance-based Approximation
• [Yeoh et al. 09] Trading off solution quality for faster computation in DCOP search algorithms,
IJCAI 09
• [Petcu and Faltings 05b] A-DPOP: Approximations in Distributed Optimization, CP 2005
• [Rogers et al. 11] Bounded approximate decentralised coordination via the max-sum
algorithm, Artificial Intelligence 2011.
Instance-generic Approximation
• [Vinyals et al 10b] Worst-case bounds on the quality of max-product fixed-points, NIPS 10
• [Vinyals et al 11] Quality guarantees for region optimal algorithms, AAMAS 11
• [Pearce and Tambe 07] Quality Guarantees on k-Optimal Solutions for Distributed Constraint
Optimization Problems, IJCAI 07
• [Bowring et al. 08] On K-Optimal Distributed Constraint Optimization Algorithms: New
Bounds and Algorithms, AAMAS 08
• [Weiss 00] Correctness of local probability propagation in graphical models with loops, Neural
Computation
• [Kiekintveld et al. 10] Asynchronous Algorithms for Approximate Distributed Constraint
Optimization with Quality Bounds, AAMAS 10