Many multi-objective optimisation problems incorporate computationally or financially expensive objective functions. State-of-the-art algorithms therefore construct surrogate model(s) of the parameter space to objective functions mapping to guide the choice of the next solution to expensively evaluate. Starting from an initial set of solutions, an infill criterion — a surrogate-based indicator of quality — is extremised to determine which solution to evaluate next, until the budget of expensive evaluations is exhausted. Many successful infill criteria are dependent on multi-dimensional integration, which may result in infill criteria that are themselves impractically expensive. We propose a computationally cheap infill criterion based on the minimum probability of improvement over the estimated Pareto set. We also present a range of set-based scalarisation methods modelling hypervolume contribution, dominance ratio and distance measures. These permit the use of straightforward expected improvement as a cheap infill criterion. We investigated the performance of these novel strategies on standard multi-objective test problems, and compared them with the popular SMS-EGO and ParEGO methods. Unsurprisingly, our experiments show that the best strategy is problem dependent, but in many cases a cheaper strategy is at least as good as more expensive alternatives.
Preprint repository: https://ore.exeter.ac.uk/repository/handle/10871/27157
Dr Chris Drovandi (QUT) - Bayesian Indirect Inference Using a Parametric Auxi...QUT_SEF
Dr Chris Drovandi, statistical science lecturer at QUT, discusses a general framework for likelihood-free Bayesian inference problems called Bayesian Indirect Likelihood (BIL). There is a focus on some specific instances of the BIL framework that use in some way a parametric auxiliary model, which is an alternative model that possesses a tractable likelihood function. These methods are referred to as parametric Bayesian Indirect Inference (pBII) methods.
One class of pBII methods uses the score or parameter of the auxiliary model to form summary statistics for ABC. A different class (called parametric BIL, pBIL) uses the likelihood (either at the full-data or summary statistic level) of the auxiliary model as a replacement to the likelihood of the true model.
The theoretical aspects of pBIL are explored and contrasted against the ABC approach that uses summary statistics formed from the auxiliary model. The theoretical results and the performance of the methods will be demonstrated on examples of varying complexity, including estimating the parameters of a stochastic process for macroparasite population evolution.
This is joint work with Prof. Tony Pettitt, Dr Anthony Lee and Leah South.
Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Opti...Fabian Pedregosa
Short presentation of the paper "Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Optimization"
https://arxiv.org/abs/1707.06468
Introduction to Bayesian modelling and inference with Pyro for meetup group. Part of the presentation is a hands on, with some examples available here: https://github.com/ahmadsalim/2019-meetup-pyro-intro
Dr Chris Drovandi (QUT) - Bayesian Indirect Inference Using a Parametric Auxi...QUT_SEF
Dr Chris Drovandi, statistical science lecturer at QUT, discusses a general framework for likelihood-free Bayesian inference problems called Bayesian Indirect Likelihood (BIL). There is a focus on some specific instances of the BIL framework that use in some way a parametric auxiliary model, which is an alternative model that possesses a tractable likelihood function. These methods are referred to as parametric Bayesian Indirect Inference (pBII) methods.
One class of pBII methods uses the score or parameter of the auxiliary model to form summary statistics for ABC. A different class (called parametric BIL, pBIL) uses the likelihood (either at the full-data or summary statistic level) of the auxiliary model as a replacement to the likelihood of the true model.
The theoretical aspects of pBIL are explored and contrasted against the ABC approach that uses summary statistics formed from the auxiliary model. The theoretical results and the performance of the methods will be demonstrated on examples of varying complexity, including estimating the parameters of a stochastic process for macroparasite population evolution.
This is joint work with Prof. Tony Pettitt, Dr Anthony Lee and Leah South.
Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Opti...Fabian Pedregosa
Short presentation of the paper "Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Optimization"
https://arxiv.org/abs/1707.06468
Introduction to Bayesian modelling and inference with Pyro for meetup group. Part of the presentation is a hands on, with some examples available here: https://github.com/ahmadsalim/2019-meetup-pyro-intro
Learning to discover monte carlo algorithm on spin ice manifoldKai-Wen Zhao
The global update Monte Carlo sampler can be discovered naturally by trained machine using policy gradient method on topologically constrained environment.
SAT based planning for multiagent systemsRavi Kuril
Multi-agent Classical planning using SAT approach. This document describes the approach and discusses all the experiments and the respective results. I have considered State of the art tools for comparison purpose. Implementation code can be found on GitHub link https://github.com/ravikuril/SATbasedClassicalPlanning . For more Information contact me on ravikuril.du.or@gmail.com
This is the presentation of the paper "Quasi-Optimal Recombination Operator" presented in EvoCOP 2019 (Best paper session). The paper is available in LNCS with doi: https://doi.org/10.1007/978-3-030-16711-0_9
Asynchronous Stochastic Optimization, New Analysis and AlgorithmsFabian Pedregosa
As datasets continue to increase in size and multi-core computer architectures are developed, asynchronous parallel optimization algorithms become more and more essential to the field of Machine Learning. In this talk I will describe two of our recent contributions to this topic. First, we highlight an important technical issue present in a large fraction of the recent convergence proofs for asynchronous parallel optimization algorithms and propose a new framework that resolves it [1]. Second, we propose a novel asynchronous variant of SAGA, a stochastic method that combines the low cost per iteration of SGD with the fast convergence rates of gradient descent [2]
[1] Leblond, R., Pedregosa, F., & Lacoste-Julien, S. (2018). Improved asynchronous parallel optimization analysis for stochastic incremental methods. arXiv:1801.03749, https://arxiv.org/pdf/1801.03749.pdf
[2] Pedregosa, F., Leblond, R., & Lacoste-Julien, S. (2017). Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Optimization. In Advances in Neural Information Processing Systems, http://papers.nips.cc/paper/6611-breaking-the-nonsmooth-barrier-a-scalable-parallel-method-for-composite-optimization.pdf
Kernelization algorithms for graph and other structure modification problemsAnthony Perez
Thesis defense on November 14th, 2011, in Montpellier.
Jury:
Stéphane Bessy, Bruno Durand, Frédéric Havet, Rolf Niedermeier, Christophe Paul & Ioan Todinca.
Query Rewriting and Optimization for Ontological DatabasesGiorgio Orsi
Ontological queries are evaluated against a knowledge base consisting of an extensional database and an ontology (i.e., a set of logical assertions and constraints that derive new intensional knowledge from the extensional database), rather than directly on the extensional database. The evaluation and optimization of such queries is an intriguing new problem for database research. In this article, we discuss two important aspects of this problem: query rewriting and query optimization. Query rewriting consists of the compilation
of an ontological query into an equivalent first-order query against the underlying extensional database.
We present a novel query rewriting algorithm for rather general types of ontological constraints that is well suited for practical implementations. In particular, we show how a conjunctive query against a knowledge base, expressed using linear and sticky existential rules, that is, members of the recently introduced Datalog+/- family of ontology languages, can be compiled into a union of conjunctive queries (UCQ) against the underlying database. Ontological query optimization, in this context, attempts to improve this rewriting process soas to produce possibly small and cost-effective UCQ rewritings for an input query.
We provide a review of the recent literature on statistical risk bounds for deep neural networks. We also discuss some theoretical results that compare the performance of deep ReLU networks to other methods such as wavelets and spline-type methods. The talk will moreover highlight some open problems and sketch possible new directions.
To describe the dynamics taking place in networks that structurally change over time, we propose an approach to search for attributes whose value changes impact the topology of the graph. In several applications, it appears that the variations of a group of attributes are often followed by some structural changes in the graph that one may assume they generate. We formalize the triggering pattern discovery problem as a method jointly rooted in sequence mining and graph analysis. We apply our approach on three real-world dynamic graphs of different natures - a co-authoring network, an airline network, and a social bookmarking system - assessing the relevancy of the triggering pattern mining approach.
Presentation at OM-2017, the Twelfth International Workshop on Ontology Matching collocated with the 16th International Semantic Web Conference ISWC-2017, October 21st, 2017, Vienna, Austria
This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227).
Applying Linear Optimization Using GLPKJeremy Chen
A brief introduction to linear optimization with a focus on applying it with the high-quality open-source solver GLPK.
Originally prepared for an intra-department sharing session.
Learning to discover monte carlo algorithm on spin ice manifoldKai-Wen Zhao
The global update Monte Carlo sampler can be discovered naturally by trained machine using policy gradient method on topologically constrained environment.
SAT based planning for multiagent systemsRavi Kuril
Multi-agent Classical planning using SAT approach. This document describes the approach and discusses all the experiments and the respective results. I have considered State of the art tools for comparison purpose. Implementation code can be found on GitHub link https://github.com/ravikuril/SATbasedClassicalPlanning . For more Information contact me on ravikuril.du.or@gmail.com
This is the presentation of the paper "Quasi-Optimal Recombination Operator" presented in EvoCOP 2019 (Best paper session). The paper is available in LNCS with doi: https://doi.org/10.1007/978-3-030-16711-0_9
Asynchronous Stochastic Optimization, New Analysis and AlgorithmsFabian Pedregosa
As datasets continue to increase in size and multi-core computer architectures are developed, asynchronous parallel optimization algorithms become more and more essential to the field of Machine Learning. In this talk I will describe two of our recent contributions to this topic. First, we highlight an important technical issue present in a large fraction of the recent convergence proofs for asynchronous parallel optimization algorithms and propose a new framework that resolves it [1]. Second, we propose a novel asynchronous variant of SAGA, a stochastic method that combines the low cost per iteration of SGD with the fast convergence rates of gradient descent [2]
[1] Leblond, R., Pedregosa, F., & Lacoste-Julien, S. (2018). Improved asynchronous parallel optimization analysis for stochastic incremental methods. arXiv:1801.03749, https://arxiv.org/pdf/1801.03749.pdf
[2] Pedregosa, F., Leblond, R., & Lacoste-Julien, S. (2017). Breaking the Nonsmooth Barrier: A Scalable Parallel Method for Composite Optimization. In Advances in Neural Information Processing Systems, http://papers.nips.cc/paper/6611-breaking-the-nonsmooth-barrier-a-scalable-parallel-method-for-composite-optimization.pdf
Kernelization algorithms for graph and other structure modification problemsAnthony Perez
Thesis defense on November 14th, 2011, in Montpellier.
Jury:
Stéphane Bessy, Bruno Durand, Frédéric Havet, Rolf Niedermeier, Christophe Paul & Ioan Todinca.
Query Rewriting and Optimization for Ontological DatabasesGiorgio Orsi
Ontological queries are evaluated against a knowledge base consisting of an extensional database and an ontology (i.e., a set of logical assertions and constraints that derive new intensional knowledge from the extensional database), rather than directly on the extensional database. The evaluation and optimization of such queries is an intriguing new problem for database research. In this article, we discuss two important aspects of this problem: query rewriting and query optimization. Query rewriting consists of the compilation
of an ontological query into an equivalent first-order query against the underlying extensional database.
We present a novel query rewriting algorithm for rather general types of ontological constraints that is well suited for practical implementations. In particular, we show how a conjunctive query against a knowledge base, expressed using linear and sticky existential rules, that is, members of the recently introduced Datalog+/- family of ontology languages, can be compiled into a union of conjunctive queries (UCQ) against the underlying database. Ontological query optimization, in this context, attempts to improve this rewriting process soas to produce possibly small and cost-effective UCQ rewritings for an input query.
We provide a review of the recent literature on statistical risk bounds for deep neural networks. We also discuss some theoretical results that compare the performance of deep ReLU networks to other methods such as wavelets and spline-type methods. The talk will moreover highlight some open problems and sketch possible new directions.
To describe the dynamics taking place in networks that structurally change over time, we propose an approach to search for attributes whose value changes impact the topology of the graph. In several applications, it appears that the variations of a group of attributes are often followed by some structural changes in the graph that one may assume they generate. We formalize the triggering pattern discovery problem as a method jointly rooted in sequence mining and graph analysis. We apply our approach on three real-world dynamic graphs of different natures - a co-authoring network, an airline network, and a social bookmarking system - assessing the relevancy of the triggering pattern mining approach.
Presentation at OM-2017, the Twelfth International Workshop on Ontology Matching collocated with the 16th International Semantic Web Conference ISWC-2017, October 21st, 2017, Vienna, Austria
This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227).
Applying Linear Optimization Using GLPKJeremy Chen
A brief introduction to linear optimization with a focus on applying it with the high-quality open-source solver GLPK.
Originally prepared for an intra-department sharing session.
The granting process of all credit institutions rejects applicants who seem risky regarding the repayment of their debt. A credit score is calculated and associated with a cut-off value beneath which an applicant is rejected. Developing a new scorecard, i.e. a correspondence table between a client's characteristics and his/her score, implies having a learning dataset in which the response variable good/bad borrower is known, so that rejects are de facto excluded from the learning process.
It might have deep consequences on the scorecard relevance as the learning population has been financed and considered good by the previous model. Previous works from the literature in this matter consisted mostly in empirical methods to exploit rejected applicants' data in the scorecard development process and experiments. We propose a rational criterion to evaluate the quality of a scoring model. We review each existing method in light of our criterion and dig out their implicit mathematical hypotheses.
We show that up to now, no Reject Inference method can guarantee to provide an improved credit scorecard. To support these theoretical findings, we added experiments on simulated and real data from the french branch of Crédit Agricole Consumer Finance.
These slides were presented as part of a talk I gave at the 49èmes Journées de Statistique, 29th of May, in Avignon, France.
Enhancing Partition Crossover with Articulation Points Analysisjfrchicanog
This is the presentation of the paper entitled "Enhancing Partition Crossover with Articulation Points Analysis" at the ECOM track in gECCO 2018 (Kyoto). This paper was awarded with a "Best Paper Award"
Surrogate models emulate expensive computer simulations. The objective is to approximate a function, $f$, of $d$ variables to a given tolerance, $\varepsilon$, using as few function values as possible, preferably $O(d)$. We explain how tractability theory provides lower bounds on the number of function values required for any possible method. We also propose method for sampling $f$ and approximating $f$ that achieves this objective and the kind of underlying structure that $f$ must have for success.
The retrieval algorithms in remote sensing generally involve complex physical forward models that are nonlinear and computationally expensive to evaluate. Statistical emulation provides an alternative with cheap computation and can be used to calibrate model parameters and to improve computational efficiency of the retrieval algorithms. We introduce a framework of combining dimension reduction of input and output spaces and Gaussian process emulation technique. The functional principal component analysis (FPCA) is chosen to reduce to the output space of thousands of dimensions by orders of magnitude. In addition, instead of making restrictive assumptions regarding the correlation structure of the high-dimensional input space, we identity and exploit the most important directions of this space and thus construct a Gaussian process emulator with feasible computation. We will present preliminary results obtained from applying our method to OCO-2 data, and discuss how our framework can be generalized in distributed systems.
"An Evaluation of Models for Runtime Approximation in Link Discovery" as presented in the IEEE/WIC/ACM WI, August 25th, 2017, held in Leipzig, Germany.
This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227).
Self-Adapting Large Neighborhood Search: Application to single-mode schedulin...Philippe Laborie
Providing robust scheduling algorithms that can solve a large variety of scheduling problems with good performance is one of the biggest challenge of practical schedulers today. In this paper we present a robust scheduling algorithm based on Self-Adapting Large Neighborhood Search and apply it to a large panel of single-mode scheduling problems. The approach combines Large Neighborhood Search with a portfolio of neighborhoods and completion strategies together with Machine Learning techniques to converge on the most efficient neighborhoods and completion strategies for the problem being solved. The algorithm is evaluated on a set of 21 scheduling benchmarks, most of which are well established in the scheduling community. Despite the generality of the approach, for 17 benchmarks out of 21, its mean relative distance to state-of-the-art problem specific algorithms is less than 4%. It even outperforms state-of-the-art problem-specific algorithms on 7 benchmarks clearly showing that our algorithm offers a valuable compromise between robustness and performance.
How Reliable is Duality Theory in Empirical Work?contenidos-ort
Coautores: Francisco Rosas (Universidad ORT Uruguay) and Sergio H. Lence(Iowa State University).
2016 Agricultural and Applied Economics Association (AAEA) Annual Meetings. July 2016, Boston, MA.
La teoría de dualidad, que establece una relación entre la función de beneficios de una firma competitiva y su tecnología de producción, ha sido utilizado por ejemplo para estimar elasticidades.
En este estudio se pone en manifiesto problemas de precisión de dicha teoría en algunas aplicaciones prácticas debido a importantes sesgos en la estimaciones de parámetros conocidos de una función de producción.
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Chiheb Ben Hammouda
Conference talk at the SIAM Conference on Financial Mathematics and Engineering, held in virtual format, June 1-4 2021, about our recently published work "Hierarchical adaptive sparse grids and quasi-Monte Carlo for option pricing under the rough Bergomi model".
- Link of the paper: https://www.tandfonline.com/doi/abs/10.1080/14697688.2020.1744700
Individualized treatment rules (ITR) assign treatments according to different patients' characteristics. Despite recent advances on the estimation of ITRs, much less attention has been given to uncertainty assessments for the estimated rules. We propose a hypothesis testing procedure for the estimated ITRs from a general framework that directly optimizes overall treatment bene t equipped with sparse penalties. Specifically, we construct a local test for testing low dimensional components of high-dimensional linear decision rules. The procedure can apply to observational studies by taking into account the additional variability from the estimation of propensity score. Theoretically, our test extends the decorrelated score test proposed in Nang and Liu (2017, Ann. Stat.) and is valid no matter whether model selection consistency for the true parameters holds or not. The proposed methodology is illustrated with numerical studies and a real data example on electronic health records of patients with Type-II Diabetes.
Nec 602 unit ii Random Variables and Random processDr Naim R Kidwai
The presentation explains concept of Probability, random variable, statistical averages, correlation, sum of random Variables, Central Limit Theorem,
random process, classification of random processes, power spectral density, multiple random processes.
Similar to Alternative Infill Strategies for Expensive Multi-Objective Optimisation (20)
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Literature Review Basics and Understanding Reference Management.pptxDr Ramhari Poudyal
Three-day training on academic research focuses on analytical tools at United Technical College, supported by the University Grant Commission, Nepal. 24-26 May 2024
Online aptitude test management system project report.pdfKamal Acharya
The purpose of on-line aptitude test system is to take online test in an efficient manner and no time wasting for checking the paper. The main objective of on-line aptitude test system is to efficiently evaluate the candidate thoroughly through a fully automated system that not only saves lot of time but also gives fast results. For students they give papers according to their convenience and time and there is no need of using extra thing like paper, pen etc. This can be used in educational institutions as well as in corporate world. Can be used anywhere any time as it is a web based application (user Location doesn’t matter). No restriction that examiner has to be present when the candidate takes the test.
Every time when lecturers/professors need to conduct examinations they have to sit down think about the questions and then create a whole new set of questions for each and every exam. In some cases the professor may want to give an open book online exam that is the student can take the exam any time anywhere, but the student might have to answer the questions in a limited time period. The professor may want to change the sequence of questions for every student. The problem that a student has is whenever a date for the exam is declared the student has to take it and there is no way he can take it at some other time. This project will create an interface for the examiner to create and store questions in a repository. It will also create an interface for the student to take examinations at his convenience and the questions and/or exams may be timed. Thereby creating an application which can be used by examiners and examinee’s simultaneously.
Examination System is very useful for Teachers/Professors. As in the teaching profession, you are responsible for writing question papers. In the conventional method, you write the question paper on paper, keep question papers separate from answers and all this information you have to keep in a locker to avoid unauthorized access. Using the Examination System you can create a question paper and everything will be written to a single exam file in encrypted format. You can set the General and Administrator password to avoid unauthorized access to your question paper. Every time you start the examination, the program shuffles all the questions and selects them randomly from the database, which reduces the chances of memorizing the questions.
Water billing management system project report.pdfKamal Acharya
Our project entitled “Water Billing Management System” aims is to generate Water bill with all the charges and penalty. Manual system that is employed is extremely laborious and quite inadequate. It only makes the process more difficult and hard.
The aim of our project is to develop a system that is meant to partially computerize the work performed in the Water Board like generating monthly Water bill, record of consuming unit of water, store record of the customer and previous unpaid record.
We used HTML/PHP as front end and MYSQL as back end for developing our project. HTML is primarily a visual design environment. We can create a android application by designing the form and that make up the user interface. Adding android application code to the form and the objects such as buttons and text boxes on them and adding any required support code in additional modular.
MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software. It is a stable ,reliable and the powerful solution with the advanced features and advantages which are as follows: Data Security.MySQL is free open source database that facilitates the effective management of the databases by connecting them to the software.
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
KuberTENes Birthday Bash Guadalajara - K8sGPT first impressionsVictor Morales
K8sGPT is a tool that analyzes and diagnoses Kubernetes clusters. This presentation was used to share the requirements and dependencies to deploy K8sGPT in a local environment.
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesChristina Lin
Traditionally, dealing with real-time data pipelines has involved significant overhead, even for straightforward tasks like data transformation or masking. However, in this talk, we’ll venture into the dynamic realm of WebAssembly (WASM) and discover how it can revolutionize the creation of stateless streaming pipelines within a Kafka (Redpanda) broker. These pipelines are adept at managing low-latency, high-data-volume scenarios.
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Alternative Infill Strategies for Expensive Multi-Objective Optimisation
1. Alternative Infill Strategies
for Expensive Multi-Objective Optimisation
Alma Rahat
Richard Everson
Jonathan Fieldsend
Department of Computer Science
University of Exeter
United Kingdom
Supported by Engineering and Physical Sciences Research Council (EPSRC), UK
Genetic and Evolutionary Computation Conference (GECCO), Berlin
18 July 2017
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 1 / 12
2. Expensive Optimisation Problems
x = (cheese, . . . )
ingredients vector
f (x)
taste
bake cake
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 2 / 12
3. Expensive Optimisation Problems
x = (cheese, . . . )
ingredients vector
f (x)
taste
bake cake
Expensive (computationally and/or financially) function evaluations.
Limited budget on function evaluations.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 2 / 12
4. Expensive Optimisation Problems
x = (cheese, . . . )
ingredients vector
f (x)
taste
bake cake
Expensive (computationally and/or financially) function evaluations.
Limited budget on function evaluations.
Analytical model and gradients may not be available.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 2 / 12
5. Expensive Optimisation Problems
x = (cheese, . . . )
ingredients vector
f (x)
taste
bake cake
Expensive (computationally and/or financially) function evaluations.
Limited budget on function evaluations.
Analytical model and gradients may not be available.
Solution: surrogate-assisted optimisation.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 2 / 12
6. Efficient Global Optimisation (EGO)
x
f(x)
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
7. Efficient Global Optimisation (EGO)
x
f(x)
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
8. Efficient Global Optimisation (EGO)
x
p(ˆf|D)
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
9. Efficient Global Optimisation (EGO)
x
p(ˆf|D)
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
10. Efficient Global Optimisation (EGO)
p(ˆf|D)EI(x)
x
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
11. Efficient Global Optimisation (EGO)
p(ˆf|D)EI(x)
x
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Sub-problem: maxx EI(x)
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
12. Efficient Global Optimisation (EGO)
p(ˆf|D)EI(x)
x
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Sub-problem: maxx EI(x)
Repeat until budget is
exhausted
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
13. Efficient Global Optimisation (EGO)
p(ˆf|D)EI(x)
x
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Sub-problem: maxx EI(x)
Repeat until budget is
exhausted
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
14. Efficient Global Optimisation (EGO)
p(ˆf|D)EI(x)
x
Initial samples (e.g. Latin
Hypercube):
D = {(xi , f (xi ))}
Fit a Gaussian process
(GP) model: p(ˆf (x)|D)
Define infill criterion:
expected improvement,
EI(x)
Sub-problem: maxx EI(x)
Repeat until budget is
exhausted (10 FEs)
Infill criterion is a surrogate based measure of utility.
Computation time for the infill criterion matters.
1 sec/evaluation × 100000 evaluations ≈ 1.15 days
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 3 / 12
15. Multi-Objective EGO
Multi-Surrogate Approaches
Model each function independently.
Infill criterion: S-metric, Expected Hypervolume Improvment, etc.
x Expensive Problem
f1(x)
f2(x)
p(ˆf1|D)
p(ˆf2|D)
Infill
Criterion
Mono-Surrogate Approaches
Model scalarised function, e.g. ParEGO (augmented Chebyshev).
Infill criterion: expected improvement in scalarised function.
x Expensive Problem
f1(x)
f2(x)
p(ˆg|D)g(x) Infill
Criterion
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 4 / 12
16. Multi-Objective EGO
Multi-Surrogate Approaches (effective but more expensive)
Model each function independently.
Infill criterion: S-metric, Expected Hypervolume Improvment, etc.
x Expensive Problem
f1(x)
f2(x)
p(ˆf1|D)
p(ˆf2|D)
Infill
Criterion
Mono-Surrogate Approaches (cheap but less effective)
Model scalarised function, e.g. ParEGO (augmented Chebyshev).
Infill criterion: expected improvement in scalarised function.
x Expensive Problem
f1(x)
f2(x)
p(ˆg|D)g(x) Infill
Criterion
Goal: cheap multi-surrogate infill criterion or effective scalarisation.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 4 / 12
17. Minimum Probability of Improvement (MPoI)
p(ˆf2|D)
p(ˆf1|D)
(ˆµ1(xi ), ˆµ2(xi ))
ˆσ1(xi )
ˆσ2(xi )
Multi-Surrogates: multi-variate
predictive distribution.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 5 / 12
18. Minimum Probability of Improvement (MPoI)
(ˆµ1(xj ), ˆµ2(xj ))
p(ˆf2|D)
p(ˆf1|D)
(ˆµ1(xi ), ˆµ2(xi ))
ˆσ1(xi )
ˆσ2(xi )
Multi-Surrogates: multi-variate
predictive distribution.
Probability of dominance.
P(xj xi ) =
M
m=1 P(ˆfm(xj ) < ˆfm(xi ))
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 5 / 12
19. Minimum Probability of Improvement (MPoI)
(ˆµ1(xj ), ˆµ2(xj ))
p(ˆf2|D)
p(ˆf1|D)
(ˆµ1(xi ), ˆµ2(xi ))
ˆσ1(xi )
ˆσ2(xi )
Multi-Surrogates: multi-variate
predictive distribution.
Probability of dominance.
P(xj xi ) =
M
m=1 P(ˆfm(xj ) < ˆfm(xi ))
Probability of improvement.
P(xi xj or xi ||xj ) = 1 − P(xj xi )
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 5 / 12
20. Minimum Probability of Improvement (MPoI)
(ˆµ1(xj ), ˆµ2(xj ))
p(ˆf2|D)
p(ˆf1|D)
(ˆµ1(xi ), ˆµ2(xi ))
ˆσ1(xi )
ˆσ2(xi )
Multi-Surrogates: multi-variate
predictive distribution.
Probability of dominance.
P(xj xi ) =
M
m=1 P(ˆfm(xj ) < ˆfm(xi ))
Probability of improvement.
P(xi xj or xi ||xj ) = 1 − P(xj xi )
Multi-Surrogate Infill Criterion.
Minimum probability of improvement over Pareto set P∗.
minx∈P∗ 1 − P(x xi )
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 5 / 12
21. Minimum Probability of Improvement (MPoI)
0.00 0.25 0.50 0.75 1.00
f1(x)
0.0
0.2
0.4
0.6
0.8
1.0
f2(x)
0.00 0.13 0.27 0.40 0.53 0.67 0.80 0.93
Minimum Probability of Improvement (MPoI)
Multi-Surrogates: multi-variate
predictive distribution.
Probability of dominance.
P(xj xi ) =
M
m=1 P(ˆfm(xj ) < ˆfm(xi ))
Probability of improvement.
P(xi xj or xi ||xj ) = 1 − P(xj xi )
Multi-Surrogate Infill Criterion.
Minimum probability of improvement over Pareto set P∗.
minx∈P∗ 1 − P(x xi )
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 5 / 12
31. Minimum Signed Distance (MSD)
f2(x)
f1(x)
Mono-surrogate approach:
set-based scalarisation function.
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 8 / 12
32. Minimum Signed Distance (MSD)
f2(x)
f1(x)
Mono-surrogate approach:
set-based scalarisation function.
Scalarisation using a distance
measure.
gd (x, X) = minx ∈P∗ d(x, x )
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 8 / 12
33. Minimum Signed Distance (MSD)
0.00 0.25 0.50 0.75 1.00
f1(x)
0.0
0.2
0.4
0.6
0.8
1.0
f2(x)
-1.69 -1.35 -1.00 -0.66 -0.31 0.04 0.38 0.73
Minimum Signed Distance (MSD)
Mono-surrogate approach:
set-based scalarisation function.
Scalarisation using a distance
measure.
gd (x, X) = minx ∈P∗ d(x, x )
Mono-Surrogate Scalarisation.
Distance measure:
d(x, x ) = M
m=1 fm(x) − fm(x ).
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 8 / 12
34. Experiment Setup
65 initial samples.
Budget: 250 function evaluations.
Infill criteria optimisation:
Optimiser: Bipop-CMA-ES.
Budget: 20000 function
evaluations per dimension.
Statistical tests:
11 simulation runs.
Matched samples.
Friedman test to determine if a
difference exists.
Wilcoxon Rank Sum test with
Bonferroni correction to identify
winner.
Mann-Whitney-U test to compare
with Latin Hypercube Samples.
Problem Parameters Objectives
n M
DTLZ1 6 3
DTLZ2 6 3
DTLZ5 6 6
DTLZ7 6 4
WFG1 6 2
WFG2 6 2
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 9 / 12
37. Performance Comparison: Computation Time
2 3 4 5 6
Number of Objectives
10−4
10−3
10−2
10−1
100
Time(seconds)
|P∗
| = 10
2 3 4 5 6
Number of Objectives
10−4
10−3
10−2
10−1
100
Time(seconds)
|P∗
| = 50
2 3 4 5 6
Number of Objectives
10−4
10−3
10−2
10−1
100
Time(seconds)
|P∗
| = 100
Mono-Surrogate
MPoI
SMS-EGO
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 11 / 12
38. Summary
5.1 5.2 5.3 5.4
Hypervolume
msd2(2)
msd3(7)
msd4(1)
msd5(3)
msd7(5)
msd8(0)
optSAF(5)
SMSEGO(2)
DTLZ2
97 98 99
Hypervolume
msd2(1)
msd3(5)
msd4(0)
msd5(2)
msd7(2)
msd8(0)
optSAF(0)
SMSEGO(6)
UF1
Fast alternative strategies
perform as well as SMS-EGO
in half the test problems and
outperform ParEGO.
Overall rank: SMS-EGO,
HypI, DomRank, MPoI,
MSD, ParEGO, LHS.
Performance is problem
dependent.
Current and Future work
Choosing the best infill
strategy from all available
strategies during
optimisation.
Python code available at: https://bitbucket.org/arahat/gecco-2017
Rahat, Everson and Fieldsend Expensive Multi-Objective Optimisation GECCO, Berlin, 18 July 2017 12 / 12