The document discusses how machine learning and economics can inform each other. It describes how markets can be viewed as decentralized algorithms that solve complex tasks. Recommendation systems are discussed as an example of machine learning success, but it is noted that recommending the same items to everyone is problematic. As an alternative, the document proposes creating markets with recommendation systems on both sides, like a market connecting diners and restaurants. The rest of the document discusses several specific examples at the intersection of machine learning and economics research, like bandits that compete and finding Nash equilibria with gradient-based algorithms.
This document discusses perspectives on artificial intelligence and challenges in the field of intelligent infrastructure (II). It describes four generations of machine learning and notes that current AI solutions are often developed for human-imitative problems rather than intelligence augmentation or II problems. Creating markets by blending statistics, economics, and computer science may help solve challenges involving large numbers of linked decisions and resource scarcity. Data flows can enable both load balancing and economic value when producers and consumers are connected in markets. While ML has advanced, robust and scalable solutions to modern data problems remain challenging.
A presentation at the Open University, Milton Keynes, 2003. The paper presents three different examples of simulation: An agent-based model of adaptive behaviour in oligopoly, a learning model of consumption and lifestyle and a preliminary attempt to model social mobility processes.
Artificial Intelligence and Economic Theories: Skynet in the marketTshilidzi Marwala
This book theoretically and practically updates major economic ideas such as demand and supply, rational choice and expectations, bounded rationality, behavioral economics, information asymmetry, pricing, efficient market hypothesis, game theory, mechanism design, portfolio theory, causality and financial engineering in the age of significant advances in man-machine systems. The advent of artificial intelligence has changed many disciplines such as engineering, social science and economics. Artificial intelligence is a computational technique which is inspired by natural intelligence concepts such as the swarming of birds, the working of the brain and the pathfinding of the ants.
Artificial Intelligence and Economic Theory: Skynet in the Market analyses the impact of artificial intelligence on economic theories, a subject that has not been studied. It also introduces new economic theories and these are rational counterfactuals and rational opportunity costs. These ideas are applied to diverse areas such as modelling of the stock market, credit scoring, HIV and interstate conflict. Artificial intelligence ideas used in this book include neural networks, particle swarm optimization, simulated annealing, fuzzy logic and genetic algorithms. It, furthermore, explores ideas in causality including Granger as well as the Pearl causality models.
This document discusses computer simulation in economics and social science. It outlines the ideology of simulation using heterogeneous agents situated within complex, dynamic environments. It presents the agent-based approach as an alternative to aggregate models. Several case studies are described, including pricing under oligopoly, lifestyle emergence among pensioners, and social mobility. The document argues that simulation is a useful tool that can inspire new theories, integrate research, and help address issues around micro-macro links.
Machine learning models are trained with a certain amount of labeled data and will use it to make predictions on unseen data. Based on this data, machines define a set of rules that they apply to all datasets, helping them provide consistent and accurate results. No need to worry about human error or innate bias.
"The greater promise of Big Data lies not in doing old things in slightly new ways. Instead, it lies in doing new things that were previously not possible. One major class of new things is adding intelligence to large-scale systems. In this session I will present a survey of how machine learning can be applied to real-life situations without having to get a PhD in advanced mathematics. These systems can be built today from open source components to increase business revenues by understanding what customers need and want. I will provide real world examples of best practices and pitfalls in machine learning including practical ways to build maintainable, high performance systems." - Ted Dunning
Value Element Analysis - Identifying markets and communities a new approach - Owen Nicholson
The document discusses identifying the right markets for a new product. It recommends following an iterative process of:
1. Identifying the key value elements of the product.
2. Hypothesizing what types of individuals would find each value element appealing.
3. Identifying communities or markets where those individuals would be found.
4. Performing an attractiveness analysis on the potential markets using a technique called SMART.
5. Engaging with the top markets identified by the analysis to test hypotheses.
6. Updating assumptions and repeating the process as needed to increase confidence in market selection.
A worked example of applying this process to a hypothetical "see-through toaster
This document discusses perspectives on artificial intelligence and challenges in the field of intelligent infrastructure (II). It describes four generations of machine learning and notes that current AI solutions are often developed for human-imitative problems rather than intelligence augmentation or II problems. Creating markets by blending statistics, economics, and computer science may help solve challenges involving large numbers of linked decisions and resource scarcity. Data flows can enable both load balancing and economic value when producers and consumers are connected in markets. While ML has advanced, robust and scalable solutions to modern data problems remain challenging.
A presentation at the Open University, Milton Keynes, 2003. The paper presents three different examples of simulation: An agent-based model of adaptive behaviour in oligopoly, a learning model of consumption and lifestyle and a preliminary attempt to model social mobility processes.
Artificial Intelligence and Economic Theories: Skynet in the marketTshilidzi Marwala
This book theoretically and practically updates major economic ideas such as demand and supply, rational choice and expectations, bounded rationality, behavioral economics, information asymmetry, pricing, efficient market hypothesis, game theory, mechanism design, portfolio theory, causality and financial engineering in the age of significant advances in man-machine systems. The advent of artificial intelligence has changed many disciplines such as engineering, social science and economics. Artificial intelligence is a computational technique which is inspired by natural intelligence concepts such as the swarming of birds, the working of the brain and the pathfinding of the ants.
Artificial Intelligence and Economic Theory: Skynet in the Market analyses the impact of artificial intelligence on economic theories, a subject that has not been studied. It also introduces new economic theories and these are rational counterfactuals and rational opportunity costs. These ideas are applied to diverse areas such as modelling of the stock market, credit scoring, HIV and interstate conflict. Artificial intelligence ideas used in this book include neural networks, particle swarm optimization, simulated annealing, fuzzy logic and genetic algorithms. It, furthermore, explores ideas in causality including Granger as well as the Pearl causality models.
This document discusses computer simulation in economics and social science. It outlines the ideology of simulation using heterogeneous agents situated within complex, dynamic environments. It presents the agent-based approach as an alternative to aggregate models. Several case studies are described, including pricing under oligopoly, lifestyle emergence among pensioners, and social mobility. The document argues that simulation is a useful tool that can inspire new theories, integrate research, and help address issues around micro-macro links.
Machine learning models are trained with a certain amount of labeled data and will use it to make predictions on unseen data. Based on this data, machines define a set of rules that they apply to all datasets, helping them provide consistent and accurate results. No need to worry about human error or innate bias.
"The greater promise of Big Data lies not in doing old things in slightly new ways. Instead, it lies in doing new things that were previously not possible. One major class of new things is adding intelligence to large-scale systems. In this session I will present a survey of how machine learning can be applied to real-life situations without having to get a PhD in advanced mathematics. These systems can be built today from open source components to increase business revenues by understanding what customers need and want. I will provide real world examples of best practices and pitfalls in machine learning including practical ways to build maintainable, high performance systems." - Ted Dunning
Value Element Analysis - Identifying markets and communities a new approach - Owen Nicholson
The document discusses identifying the right markets for a new product. It recommends following an iterative process of:
1. Identifying the key value elements of the product.
2. Hypothesizing what types of individuals would find each value element appealing.
3. Identifying communities or markets where those individuals would be found.
4. Performing an attractiveness analysis on the potential markets using a technique called SMART.
5. Engaging with the top markets identified by the analysis to test hypotheses.
6. Updating assumptions and repeating the process as needed to increase confidence in market selection.
A worked example of applying this process to a hypothetical "see-through toaster
The document discusses artificial intelligence and expert systems. It defines AI as creating intelligent machines that can learn, think, perceive, analyze and act like humans. Expert systems are described as computer programs that mimic human experts by applying specialized knowledge and reasoning skills to solve complex problems. Examples of early expert systems like DENDRAL and MYCIN are provided. The roles, uses, and typical structure of expert systems are outlined.
- The document outlines a workshop on using choice modelling as a tool for environmental decision making.
- Choice modelling involves presenting respondents with choices between scenarios that involve environmental trade-offs and measuring their preferences.
- The workshop covers what choice modelling is, how to conduct a choice modelling study, and recent Australian applications of the method.
Utah Code Camp 2014 - Learning from Data by Thomas HollowayThomas Holloway
The document provides an overview of machine learning and artificial intelligence goals including deduction, reasoning, problem solving, knowledge representation, planning, learning, natural language processing, motion and manipulation, perception, social intelligence and creativity. It discusses different machine learning techniques like supervised learning, unsupervised learning, reinforcement learning and developmental learning. It also covers topics like linear regression, logistic regression, neural networks, overfitting, regularization and more.
Foundations of Competitive Strategy (Part I)ProfessorDeitz
Powerpoint slides to accompany lecture on Foundations of Competitive Strategy.
Review history of economic thought. Links mainstream (neoclassical) economic thought to predominant strategy frameworks, such as Porter's 5 forces model and Value Chain. Critique of mainstream approaches, introduction to evolutionary economics and contemporary strategy frameworks.
The document describes rational choice theory and how it can be used to explain individual decision-making and social outcomes. It outlines the central assumptions of rational choice theory, which are that decision-makers have logically consistent goals and choose the best available option given those goals. Experimental methods and laboratory experiments are also discussed as ways to test rational choice theory by studying individual behavior in controlled settings with financial incentives.
Dr. Stephen Koontz - Thinning Cash Fed Cattle Trade: How Thin is Too Thin & W...John Blue
Thinning Cash Fed Cattle Trade: How Thin is Too Thin & What to Do About It? - Dr. Stephen Koontz, Colorado State University, from the 2014 Iowa Cattle Industry Convention, December 8 - 10, 2014, Des Moines IA, USA
More presentations at http://www.trufflemedia.com/agmedia/conference/2014-iowa-cattle-industry-convention
In 3 sentences:
The document summarizes research on developing a demand-led market for third sector support infrastructure and capacity building in the UK. It discusses efforts to shift from a supplier-led model to one where frontline organizations directly purchase support from approved providers using grants or vouchers. Key challenges involved developing tools to help organizations judge provider quality in a complex market with many singular offerings.
Animal disease control and value chain practices: Incorporating economics and...ILRI
Presented by Karl M. Rich, ILRI, at the 5th Food Safety and Zoonoses Symposium of Asia Pacific, Global Health Institute 2018, Chiang Mai, Thailand, 6-7 July 2018
The document discusses various topics related to e-marketing, including:
1) E-marketing strategies such as product development, pricing, branding, and website promotion.
2) Types of e-markets and advertising approaches like placement and search engine positioning.
3) Consumer behavior concepts like the consumer decision making process and sources of problem recognition.
4) Economics topics such as consumer choice, utility, and preferences.
Recommender systems are a technological proxy for social recommendations that suggest similar items or ideas to users based on their tastes. They aim to automate word-of-mouth recommendations by analyzing user preferences and behavior patterns to suggest new content. Recommender systems are commonly used on e-commerce sites to improve sales by suggesting additional items customers may like. They can use collaborative filtering by comparing users, content-based filtering by analyzing item content, or knowledge-based approaches. Maintaining accurate recommender systems over time with large amounts of user data presents challenges.
These slides talk about a general approach towards the trade-off between selection and fulfillment in an e-commerce recommendation system. This is also an useful tutorial for beginners who want to explore the domain of the applications of data science and machine learning in e-commerce. These slides introduces the use of Python,R and D3.js in data analytics and data science. For any questions/suggestions/comments please email me at akansha.tamu@gmail.com.
Future of AI-powered automation in businessLouis Dorard
Starting from examples of current use cases of AI in business and in everyday life, we'll see what the future holds and we'll mention questions to address when giving autonomy to intelligent machines. We'll also aim at demystifying how AI works, in particular how machines can use data to automatically learn business rules and actions to perform in different contexts.
This document discusses recommendation systems and how to develop them. It begins by introducing the speaker and an overview of the topics to be covered. It then explains what recommendation systems are and different types including search, content-based, and collaborative filtering. It discusses drawbacks like cold start problems and sparsity and ways to address them. The document concludes with tips for refining recommendation models like normalization, capturing trends, and temporal factors.
Data Con LA 2022 - Real world consumer segmentationData Con LA
Jaysen Gillespie, Head of Analytics and Data Science at RTB House
1. Shopkick has over 30M downloads, but the userbase is very heterogeneous. Anecdotal evidence indicated a wide variety of users for whom the app holds long-term appeal.
2. Marketing and other teams challenged Analytics to get beyond basic summary statistics and develop a holistic segmentation of the userbase.
3. Shopkick's data science team used SQL and python to gather data, clean data, and then perform a data-driven segmentation using a k-means algorithm.
4. Interpreting the results is more work -- and more fun -- than running the algo itself. We'll discuss how we transform from ""segment 1"", ""segment 2"", etc. to something that non-analytics users (Marketing, Operations, etc.) could actually benefit from.
5. So what? How did team across Shopkick change their approach given what Analytics had discovered.
Artificial Intelligence and The ComplexityHendri Karisma
This document discusses the complexity of artificial intelligence and machine learning. It notes that complexity arises from big data's volume, variety, velocity and veracity, as well as from knowledge representation, unlabeled data, feature engineering, hardware limitations, and the stack of methods and technologies used. High performance computing techniques like in-memory data fabrics and GPU machines can help address these complexities. Topological data analysis is also mentioned as a technique that can help with complexity through properties like coordinate and deformation invariance and compressed representations.
The document provides an overview of recommender systems. It discusses the typical architecture of recommender systems and describes three main types: collaborative filtering systems, content-based systems, and knowledge-based systems. It also covers paradigms like collaborative filtering, content-based, knowledge-based, and hybrid recommender systems. The document then focuses on collaborative filtering techniques like user-based nearest neighbor collaborative filtering and item-based collaborative filtering. It also discusses latent factor models, matrix factorization approaches, and context-based recommender systems.
This is the first lecture of the AI course offered by me at PES University, Bangalore. In this presentation we discuss the different definitions of AI, the notion of Intelligent Agents, distinguish an AI program from a complex program such as those that solve complex calculus problems (see the integration example) and look at the role of Machine Learning and Deep Learning in the context of AI. We also go over the course scope and logistics.
1. Machine learning is a set of techniques that use data to build models that can make predictions without being explicitly programmed.
2. There are two main types of machine learning: supervised learning, where the model is trained on labeled examples, and unsupervised learning, where the model finds patterns in unlabeled data.
3. Common machine learning algorithms include linear regression, logistic regression, decision trees, support vector machines, naive Bayes, k-nearest neighbors, k-means clustering, and random forests. These can be used for regression, classification, clustering, and dimensionality reduction.
Walk Through a Real World ML Production ProjectBill Liu
Success in productionizing ML models is difficult to achieve due to tools, processes and operational procedures. In this session, we demonstrate how data scientists and ML engineers collaborate and efficiently deploy models to production with the Wallaroo platform.
Using a real world scenario we will click down into the ML production journey that Data Scientists and ML engineers go through to take ML models into production. In this session you will learn:
The current pain points and blockers to production
The 2 persona roles in the ML production process. Data Scientist (DS) and ML Engineer
How the ML engineer creates a workspace in Wallaroo, and invites the DS to collaborate
How the DS uploads and deploys models to WL performing simple validation checks on output
How the ML Engineer can check model health (inference speed, etc)
How the DS checks logs, looks for anomalies
How the DS switches model in the pipeline
Speakers: Nina Zumel, Martin Bald
Redefining MLOps with Model Deployment, Management and Observability in Produ...Bill Liu
Tech talk: https://www.aicamp.ai/event/eventdetails/W2022052410
What happens after your machine learning models are deployed in production? How do you make sure that your model performance does not degrade as data and the world change?
The constantly changing data creates challenges for data scientists and engineering teams on how to detect which models have been affected and how to get their ML applications up and running seamlessly.
In this session we will take a deep dive into the new ML model monitoring and drift detection technology. We will discuss:
- How to track the ongoing accuracy of their models in production
- How to immediately detect drift before it causes significant damage to the business
- How to locate the cause of model drifting in live environments.
We will also discuss how data scientists and ML engineers can collaborate effectively using their respective tools to identify issues and take the necessary actions with a live demo and a real world use case.
Speaker: Younes Amar, Head of Product Wallaroo AI.
Resources: https://docs.wallaroo.ai/
More Related Content
Similar to AISF19 - On Blending Machine Learning with Microeconomics
The document discusses artificial intelligence and expert systems. It defines AI as creating intelligent machines that can learn, think, perceive, analyze and act like humans. Expert systems are described as computer programs that mimic human experts by applying specialized knowledge and reasoning skills to solve complex problems. Examples of early expert systems like DENDRAL and MYCIN are provided. The roles, uses, and typical structure of expert systems are outlined.
- The document outlines a workshop on using choice modelling as a tool for environmental decision making.
- Choice modelling involves presenting respondents with choices between scenarios that involve environmental trade-offs and measuring their preferences.
- The workshop covers what choice modelling is, how to conduct a choice modelling study, and recent Australian applications of the method.
Utah Code Camp 2014 - Learning from Data by Thomas HollowayThomas Holloway
The document provides an overview of machine learning and artificial intelligence goals including deduction, reasoning, problem solving, knowledge representation, planning, learning, natural language processing, motion and manipulation, perception, social intelligence and creativity. It discusses different machine learning techniques like supervised learning, unsupervised learning, reinforcement learning and developmental learning. It also covers topics like linear regression, logistic regression, neural networks, overfitting, regularization and more.
Foundations of Competitive Strategy (Part I)ProfessorDeitz
Powerpoint slides to accompany lecture on Foundations of Competitive Strategy.
Review history of economic thought. Links mainstream (neoclassical) economic thought to predominant strategy frameworks, such as Porter's 5 forces model and Value Chain. Critique of mainstream approaches, introduction to evolutionary economics and contemporary strategy frameworks.
The document describes rational choice theory and how it can be used to explain individual decision-making and social outcomes. It outlines the central assumptions of rational choice theory, which are that decision-makers have logically consistent goals and choose the best available option given those goals. Experimental methods and laboratory experiments are also discussed as ways to test rational choice theory by studying individual behavior in controlled settings with financial incentives.
Dr. Stephen Koontz - Thinning Cash Fed Cattle Trade: How Thin is Too Thin & W...John Blue
Thinning Cash Fed Cattle Trade: How Thin is Too Thin & What to Do About It? - Dr. Stephen Koontz, Colorado State University, from the 2014 Iowa Cattle Industry Convention, December 8 - 10, 2014, Des Moines IA, USA
More presentations at http://www.trufflemedia.com/agmedia/conference/2014-iowa-cattle-industry-convention
In 3 sentences:
The document summarizes research on developing a demand-led market for third sector support infrastructure and capacity building in the UK. It discusses efforts to shift from a supplier-led model to one where frontline organizations directly purchase support from approved providers using grants or vouchers. Key challenges involved developing tools to help organizations judge provider quality in a complex market with many singular offerings.
Animal disease control and value chain practices: Incorporating economics and...ILRI
Presented by Karl M. Rich, ILRI, at the 5th Food Safety and Zoonoses Symposium of Asia Pacific, Global Health Institute 2018, Chiang Mai, Thailand, 6-7 July 2018
The document discusses various topics related to e-marketing, including:
1) E-marketing strategies such as product development, pricing, branding, and website promotion.
2) Types of e-markets and advertising approaches like placement and search engine positioning.
3) Consumer behavior concepts like the consumer decision making process and sources of problem recognition.
4) Economics topics such as consumer choice, utility, and preferences.
Recommender systems are a technological proxy for social recommendations that suggest similar items or ideas to users based on their tastes. They aim to automate word-of-mouth recommendations by analyzing user preferences and behavior patterns to suggest new content. Recommender systems are commonly used on e-commerce sites to improve sales by suggesting additional items customers may like. They can use collaborative filtering by comparing users, content-based filtering by analyzing item content, or knowledge-based approaches. Maintaining accurate recommender systems over time with large amounts of user data presents challenges.
These slides talk about a general approach towards the trade-off between selection and fulfillment in an e-commerce recommendation system. This is also an useful tutorial for beginners who want to explore the domain of the applications of data science and machine learning in e-commerce. These slides introduces the use of Python,R and D3.js in data analytics and data science. For any questions/suggestions/comments please email me at akansha.tamu@gmail.com.
Future of AI-powered automation in businessLouis Dorard
Starting from examples of current use cases of AI in business and in everyday life, we'll see what the future holds and we'll mention questions to address when giving autonomy to intelligent machines. We'll also aim at demystifying how AI works, in particular how machines can use data to automatically learn business rules and actions to perform in different contexts.
This document discusses recommendation systems and how to develop them. It begins by introducing the speaker and an overview of the topics to be covered. It then explains what recommendation systems are and different types including search, content-based, and collaborative filtering. It discusses drawbacks like cold start problems and sparsity and ways to address them. The document concludes with tips for refining recommendation models like normalization, capturing trends, and temporal factors.
Data Con LA 2022 - Real world consumer segmentationData Con LA
Jaysen Gillespie, Head of Analytics and Data Science at RTB House
1. Shopkick has over 30M downloads, but the userbase is very heterogeneous. Anecdotal evidence indicated a wide variety of users for whom the app holds long-term appeal.
2. Marketing and other teams challenged Analytics to get beyond basic summary statistics and develop a holistic segmentation of the userbase.
3. Shopkick's data science team used SQL and python to gather data, clean data, and then perform a data-driven segmentation using a k-means algorithm.
4. Interpreting the results is more work -- and more fun -- than running the algo itself. We'll discuss how we transform from ""segment 1"", ""segment 2"", etc. to something that non-analytics users (Marketing, Operations, etc.) could actually benefit from.
5. So what? How did team across Shopkick change their approach given what Analytics had discovered.
Artificial Intelligence and The ComplexityHendri Karisma
This document discusses the complexity of artificial intelligence and machine learning. It notes that complexity arises from big data's volume, variety, velocity and veracity, as well as from knowledge representation, unlabeled data, feature engineering, hardware limitations, and the stack of methods and technologies used. High performance computing techniques like in-memory data fabrics and GPU machines can help address these complexities. Topological data analysis is also mentioned as a technique that can help with complexity through properties like coordinate and deformation invariance and compressed representations.
The document provides an overview of recommender systems. It discusses the typical architecture of recommender systems and describes three main types: collaborative filtering systems, content-based systems, and knowledge-based systems. It also covers paradigms like collaborative filtering, content-based, knowledge-based, and hybrid recommender systems. The document then focuses on collaborative filtering techniques like user-based nearest neighbor collaborative filtering and item-based collaborative filtering. It also discusses latent factor models, matrix factorization approaches, and context-based recommender systems.
This is the first lecture of the AI course offered by me at PES University, Bangalore. In this presentation we discuss the different definitions of AI, the notion of Intelligent Agents, distinguish an AI program from a complex program such as those that solve complex calculus problems (see the integration example) and look at the role of Machine Learning and Deep Learning in the context of AI. We also go over the course scope and logistics.
1. Machine learning is a set of techniques that use data to build models that can make predictions without being explicitly programmed.
2. There are two main types of machine learning: supervised learning, where the model is trained on labeled examples, and unsupervised learning, where the model finds patterns in unlabeled data.
3. Common machine learning algorithms include linear regression, logistic regression, decision trees, support vector machines, naive Bayes, k-nearest neighbors, k-means clustering, and random forests. These can be used for regression, classification, clustering, and dimensionality reduction.
Similar to AISF19 - On Blending Machine Learning with Microeconomics (20)
Walk Through a Real World ML Production ProjectBill Liu
Success in productionizing ML models is difficult to achieve due to tools, processes and operational procedures. In this session, we demonstrate how data scientists and ML engineers collaborate and efficiently deploy models to production with the Wallaroo platform.
Using a real world scenario we will click down into the ML production journey that Data Scientists and ML engineers go through to take ML models into production. In this session you will learn:
The current pain points and blockers to production
The 2 persona roles in the ML production process. Data Scientist (DS) and ML Engineer
How the ML engineer creates a workspace in Wallaroo, and invites the DS to collaborate
How the DS uploads and deploys models to WL performing simple validation checks on output
How the ML Engineer can check model health (inference speed, etc)
How the DS checks logs, looks for anomalies
How the DS switches model in the pipeline
Speakers: Nina Zumel, Martin Bald
Redefining MLOps with Model Deployment, Management and Observability in Produ...Bill Liu
Tech talk: https://www.aicamp.ai/event/eventdetails/W2022052410
What happens after your machine learning models are deployed in production? How do you make sure that your model performance does not degrade as data and the world change?
The constantly changing data creates challenges for data scientists and engineering teams on how to detect which models have been affected and how to get their ML applications up and running seamlessly.
In this session we will take a deep dive into the new ML model monitoring and drift detection technology. We will discuss:
- How to track the ongoing accuracy of their models in production
- How to immediately detect drift before it causes significant damage to the business
- How to locate the cause of model drifting in live environments.
We will also discuss how data scientists and ML engineers can collaborate effectively using their respective tools to identify issues and take the necessary actions with a live demo and a real world use case.
Speaker: Younes Amar, Head of Product Wallaroo AI.
Resources: https://docs.wallaroo.ai/
These days, training of the Machine Learning models at the device Edge is still a risky endeavor. It is frequently considered a purely academic subject with little value for real-life product development.
In her talk, Vera will challenge this misconception, talk about the advantages of learning at the Edge and guide you through the Edge learning decision-making framework and design principles.
https://www.aicamp.ai/event/eventdetails/W2021102210
Attention Is All You Need.
With these simple words, the Deep Learning industry was forever changed. Transformers were initially introduced in the field of Natural Language Processing to enhance language translation, but they demonstrated astonishing results even outside language processing. In particular, they recently spread in the Computer Vision community, advancing the state-of-the-art on many vision tasks. But what are Transformers? What is the mechanism of self-attention, and do we really need it? How did they revolutionize Computer Vision? Will they ever replace convolutional neural networks?
These and many other questions will be answered during the talk.
In this tech talk, we will discuss:
- A piece of history: Why did we need a new architecture?
- What is self-attention, and where does this concept come from?
- The Transformer architecture and its mechanisms
- Vision Transformers: An Image is worth 16x16 words
- Video Understanding using Transformers: the space + time approach
- The scale and data problem: Is Attention what we really need?
- The future of Computer Vision through Transformers
Speaker: Davide Coccomini, Nicola Messina
Website: https://www.aicamp.ai/event/eventdetails/W2021101110
Deep AutoViML For Tensorflow Models and MLOps WorkflowsBill Liu
deep_autoviml is a powerful new deep learning library with a very simple design goal: Make it as easy as possible for novices and experts alike to experiment with and build tensorflow.keras preprocessing pipelines and models in as few lines of code as possible.
deep_autoviml will enable data scientists, ML engineers and data engineers to fast prototype tensorflow models and data pipelines for MLOps workflows using the latest TF 2.4+ and keras preprocessing layers. You can now upload your saved model to any Cloud provider and make predictions out of the box since all the data preprocessing layers are attached to the model itself!
In this webinar, we will discuss the problems that deep_AutoViML can solve, its architecture design and demo how to build powerful TF.Keras models on structured data, NLP and Image data domains.
https://www.aicamp.ai/event/eventdetails/W2021080918
Metaflow: The ML Infrastructure at NetflixBill Liu
Metaflow was started at Netflix to answer a pressing business need: How to enable an organization of data scientists, who are not software engineers by training, build and deploy end-to-end machine learning workflows and applications independently. We wanted to provide the best possible user experience for data scientists, allowing them to focus on parts they like (modeling using their favorite off-the-shelf libraries) while providing robust built-in solutions for the foundational infrastructure: data, compute, orchestration, and versioning.
Today, the open-source Metaflow powers hundreds of business-critical ML projects at Netflix and other companies from bioinformatics to real estate.
In this talk, you will learn about:
- What to expect from a modern ML infrastructure stack.
- Using Metaflow to boost the productivity of your data science organization, based on lessons learned from Netflix.
- Deployment strategies for a full stack of ML infrastructure that plays nicely with your existing systems and policies.
https://www.aicamp.ai/event/eventdetails/W2021080510
Daria Baidakova is the Director of Educational Programs at Toloka, an open crowdsourcing platform. She manages crowdsourcing courses and organizes tutorials and hackathons. Toloka provides labeled data via crowdsourcing which is a missing pillar of AI. It offers data annotation, content collection, and business decision making through crowdsourcing. Toloka manages quality through controls at the project level, performer selection, and result aggregation rather than direct management of individuals.
Building large scale transactional data lake using apache hudiBill Liu
Data is a critical infrastructure for building machine learning systems. From ensuring accurate ETAs to predicting optimal traffic routes, providing safe, seamless transportation and delivery experiences on the Uber platform requires reliable, performant large-scale data storage and analysis. In 2016, Uber developed Apache Hudi, an incremental processing framework, to power business critical data pipelines at low latency and high efficiency, and helps distributed organizations build and manage petabyte-scale data lakes.
In this talk, I will describe what is APache Hudi and its architectural design, and then deep dive to improving data operations by providing features such as data versioning, time travel.
We will also go over how Hudi brings kappa architecture to big data systems and enables efficient incremental processing for near real time use cases.
Speaker: Satish Kotha (Uber)
Apache Hudi committer and Engineer at Uber. Previously, he worked on building real time distributed storage systems like Twitter MetricsDB and BlobStore.
website: https://www.aicamp.ai/event/eventdetails/W2021043010
Deep Reinforcement Learning and Its ApplicationsBill Liu
What is the most exciting AI news in recent years? AlphaGo!
What are key techniques for AlphaGo? Deep learning and reinforcement learning (RL)!
What are application areas for deep RL? A lot! In fact, besides games, deep RL has been making tremendous achievements in diverse areas like recommender systems and robotics.
In this talk, we will introduce deep reinforcement learning, present several applications, and discuss issues and potential solutions for successfully applying deep RL in real life scenarios.
https://www.aicamp.ai/event/eventdetails/W2021042818
Big Data and AI in Fighting Against COVID-19Bill Liu
This document discusses how big data and AI can help fight Covid-19. It describes supercomputers being used for scientific research on Covid-19. An open data lake has been created containing various Covid-19 datasets for analysis. Natural language processing and BERT are being used to answer scientific questions from the Covid-19 literature by generating summaries and highlighting relevant text passages. Challenges are being conducted on the Covid-19 Open Research Dataset to further advance research.
Highly-scalable Reinforcement Learning RLlib for Real-world ApplicationsBill Liu
This document discusses Ray and RLlib, a reinforcement learning library built on Ray. It provides three key points:
1. Ray is a framework for building distributed applications and services with shared memory abstraction. It allows ML workloads to scale beyond a single machine.
2. RLlib is a scalable reinforcement learning library that uses Ray. It supports a wide range of algorithms and execution models. This allows for easy implementation and comparison of RL techniques.
3. The Ray community is growing rapidly and provides resources like tutorials and Slack support to help users adopt Ray and RLlib for their distributed Python and reinforcement learning applications.
Build computer vision models to perform object detection and classification w...Bill Liu
event: https://learn.xnextcon.com/event/eventdetails/W20042918
video:
description: Computer Vision has received significant attention over the recent years, both within academia, and industry. As the state-of-the-art rapidly improves, the art-of-the-possible follows , offering innovative forms of computer vision applications for different scenarios.
In this talk, Ramine will cover the background and development of computer vision, and demonstrate how to use AWS to build robust, computer vision models to perform object detection and classification.
Key Takeaways:
Understand the history of Computer Vision
Learn how to use Amazon SageMaker to build and Deploy Computer Vision Models
How to orchestrate multiple models for implementing a real-world use case
Causal Inference in Data Science and Machine LearningBill Liu
Event: https://learn.xnextcon.com/event/eventdetails/W20042010
Video: https://www.youtube.com/channel/UCj09XsAWj-RF9kY4UvBJh_A
Modern machine learning techniques are able to learn highly complex associations from data, which has led to amazing progress in computer vision, NLP, and other predictive tasks. However, there are limitations to inference from purely probabilistic or associational information. Without understanding causal relationships, ML models are unable to provide actionable recommendations, perform poorly in new, but related environments, and suffer from a lack of interpretability.
In this talk, I provide an introduction to the field of causal inference, discuss its importance in addressing some of the current limitations in machine learning, and provide some real-world examples from my experience as a data scientist at Brex.
https://learn.xnextcon.com/event/eventdetails/W20040610
This talk explains how to practically bring the power of convolutional neural networks and deep learning to memory and power-constrained devices like smartphones. You will learn various strategies to circumvent obstacles and build mobile-friendly shallow CNN architectures that significantly reduce the memory footprint and therefore make them easier to store on a smartphone;
The talk also dives into how to use a family of model compression techniques to prune the network size for live image processing, enabling you to build a CNN version optimized for inference on mobile devices. Along the way, you will learn practical strategies to preprocess your data in a manner that makes the models more efficient in the real world.
Weekly #105: AutoViz and Auto_ViML Visualization and Machine LearningBill Liu
https://learn.xnextcon.com/event/eventdetails/W20040310
I will describe what is available in terms of Open Source and Proprietary tools for automating Data Science tasks and introduce 2 new tools: one to visualize any sized data set with one click, another: to try multiple ML models and techniques with a single call. I will provide the Github Repos for both for free in the talk.
Travel is a force for good economically and socially. Expedia Group is building towards an "AI-First World" of travel where AI is used throughout the entire travel planning, booking, and experience process. They are developing massive multi-dimensional AI models and an AI platform to deliver highly personalized and targeted experiences. Overcoming challenges like data quality, scalability, and control will be key to realizing an AI-First vision where travel is optimized for each individual user.
AISF19 - Unleash Computer Vision at the EdgeBill Liu
This document discusses the key drivers enabling computer vision at the edge, including new machine learning approaches, optimized model architectures, hardware innovations, and improved software tools. It describes how machine learning has advanced computer vision by enabling end-to-end learning without predefined features. Edge-optimized models like GoogleNet and ShuffleNet are discussed. The proliferation of cameras, embedded processors, and AI accelerators is enabling computer vision everywhere. Open-source tools like OpenCV and frameworks like TensorFlow are supporting development, along with platforms to speed application creation.
AISF19 - Building Scalable, Kubernetes-Native ML/AI Pipelines with TFX, KubeF...Bill Liu
This document discusses modern machine learning pipelines and popular open source tools to build them. It defines key characteristics of ML pipelines like experiment tracking, hyperparameter optimization, distributed execution, and metadata/data versioning. Popular tools covered are KubeFlow for Kubernetes+TensorFlow, Airflow for data and feature engineering, MLflow for experiment tracking, and TensorFlow Extended (TFX) libraries. The document demonstrates these tools and argues that while the field is emerging, simplicity is important and one should only use necessary components of different tools.
This document discusses elastic distributed deep learning training at scale on-premises and in the cloud. It introduces the architecture of elastic distributed training, which combines high performance synchronization techniques like distributed data parallel with session scheduling and elastic scaling to provide flexibility. This allows training jobs to automatically scale up and down resources based on policies while maintaining high performance. It aims to make distributed training transparent to frameworks like TensorFlow and PyTorch.
Monthly AI Tech Talks in Toronto 2019-08-28
https://www.meetup.com/aittg-toronto
The talk will cover the end-to-end details including contextual and linguistic feature extraction, vectorization, n-grams, topic modeling, named entity resolution which are based on concepts from mathematics, information retrieval and natural language processing. We will also be diving into more advanced feature engineering strategies such as word2vec, GloVe and fastText that leverage deep learning models.
In addition, attendees will learn how to combine NLP features with numeric and categorical features and analyze the feature importance from the resulting models.
The following libraries will be used to demonstrate the aforementioned feature engineering techniques: spaCy, Gensim, fasText and Keras in Python.
https://www.meetup.com/aittg-toronto/events/261940480/
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Keywords: AI, Containeres, Kubernetes, Cloud Native
Event Link: https://meine.doag.org/events/cloudland/2024/agenda/#agendaId.4211
"Choosing proper type of scaling", Olena SyrotaFwdays
Imagine an IoT processing system that is already quite mature and production-ready and for which client coverage is growing and scaling and performance aspects are life and death questions. The system has Redis, MongoDB, and stream processing based on ksqldb. In this talk, firstly, we will analyze scaling approaches and then select the proper ones for our system.
"$10 thousand per minute of downtime: architecture, queues, streaming and fin...Fwdays
Direct losses from downtime in 1 minute = $5-$10 thousand dollars. Reputation is priceless.
As part of the talk, we will consider the architectural strategies necessary for the development of highly loaded fintech solutions. We will focus on using queues and streaming to efficiently work and manage large amounts of data in real-time and to minimize latency.
We will focus special attention on the architectural patterns used in the design of the fintech system, microservices and event-driven architecture, which ensure scalability, fault tolerance, and consistency of the entire system.
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor IvaniukFwdays
At this talk we will discuss DDoS protection tools and best practices, discuss network architectures and what AWS has to offer. Also, we will look into one of the largest DDoS attacks on Ukrainian infrastructure that happened in February 2022. We'll see, what techniques helped to keep the web resources available for Ukrainians and how AWS improved DDoS protection for all customers based on Ukraine experience
ScyllaDB is making a major architecture shift. We’re moving from vNode replication to tablets – fragments of tables that are distributed independently, enabling dynamic data distribution and extreme elasticity. In this keynote, ScyllaDB co-founder and CTO Avi Kivity explains the reason for this shift, provides a look at the implementation and roadmap, and shares how this shift benefits ScyllaDB users.
Introducing BoxLang : A new JVM language for productivity and modularity!Ortus Solutions, Corp
Just like life, our code must adapt to the ever changing world we live in. From one day coding for the web, to the next for our tablets or APIs or for running serverless applications. Multi-runtime development is the future of coding, the future is to be dynamic. Let us introduce you to BoxLang.
Dynamic. Modular. Productive.
BoxLang redefines development with its dynamic nature, empowering developers to craft expressive and functional code effortlessly. Its modular architecture prioritizes flexibility, allowing for seamless integration into existing ecosystems.
Interoperability at its Core
With 100% interoperability with Java, BoxLang seamlessly bridges the gap between traditional and modern development paradigms, unlocking new possibilities for innovation and collaboration.
Multi-Runtime
From the tiny 2m operating system binary to running on our pure Java web server, CommandBox, Jakarta EE, AWS Lambda, Microsoft Functions, Web Assembly, Android and more. BoxLang has been designed to enhance and adapt according to it's runnable runtime.
The Fusion of Modernity and Tradition
Experience the fusion of modern features inspired by CFML, Node, Ruby, Kotlin, Java, and Clojure, combined with the familiarity of Java bytecode compilation, making BoxLang a language of choice for forward-thinking developers.
Empowering Transition with Transpiler Support
Transitioning from CFML to BoxLang is seamless with our JIT transpiler, facilitating smooth migration and preserving existing code investments.
Unlocking Creativity with IDE Tools
Unleash your creativity with powerful IDE tools tailored for BoxLang, providing an intuitive development experience and streamlining your workflow. Join us as we embark on a journey to redefine JVM development. Welcome to the era of BoxLang.
Essentials of Automations: Exploring Attributes & Automation ParametersSafe Software
Building automations in FME Flow can save time, money, and help businesses scale by eliminating data silos and providing data to stakeholders in real-time. One essential component to orchestrating complex automations is the use of attributes & automation parameters (both formerly known as “keys”). In fact, it’s unlikely you’ll ever build an Automation without using these components, but what exactly are they?
Attributes & automation parameters enable the automation author to pass data values from one automation component to the next. During this webinar, our FME Flow Specialists will cover leveraging the three types of these output attributes & parameters in FME Flow: Event, Custom, and Automation. As a bonus, they’ll also be making use of the Split-Merge Block functionality.
You’ll leave this webinar with a better understanding of how to maximize the potential of automations by making use of attributes & automation parameters, with the ultimate goal of setting your enterprise integration workflows up on autopilot.
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
Lee Barnes - Path to Becoming an Effective Test Automation Engineer.pdfleebarnesutopia
So… you want to become a Test Automation Engineer (or hire and develop one)? While there’s quite a bit of information available about important technical and tool skills to master, there’s not enough discussion around the path to becoming an effective Test Automation Engineer that knows how to add VALUE. In my experience this had led to a proliferation of engineers who are proficient with tools and building frameworks but have skill and knowledge gaps, especially in software testing, that reduce the value they deliver with test automation.
In this talk, Lee will share his lessons learned from over 30 years of working with, and mentoring, hundreds of Test Automation Engineers. Whether you’re looking to get started in test automation or just want to improve your trade, this talk will give you a solid foundation and roadmap for ensuring your test automation efforts continuously add value. This talk is equally valuable for both aspiring Test Automation Engineers and those managing them! All attendees will take away a set of key foundational knowledge and a high-level learning path for leveling up test automation skills and ensuring they add value to their organizations.
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving
Manufacturing custom quality metal nameplates and badges involves several standard operations. Processes include sheet prep, lithography, screening, coating, punch press and inspection. All decoration is completed in the flat sheet with adhesive and tooling operations following. The possibilities for creating unique durable nameplates are endless. How will you create your brand identity? We can help!
GlobalLogic Java Community Webinar #18 “How to Improve Web Application Perfor...GlobalLogic Ukraine
Під час доповіді відповімо на питання, навіщо потрібно підвищувати продуктивність аплікації і які є найефективніші способи для цього. А також поговоримо про те, що таке кеш, які його види бувають та, основне — як знайти performance bottleneck?
Відео та деталі заходу: https://bit.ly/45tILxj
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
AI in the Workplace Reskilling, Upskilling, and Future Work.pptxSunil Jagani
Discover how AI is transforming the workplace and learn strategies for reskilling and upskilling employees to stay ahead. This comprehensive guide covers the impact of AI on jobs, essential skills for the future, and successful case studies from industry leaders. Embrace AI-driven changes, foster continuous learning, and build a future-ready workforce.
Read More - https://bit.ly/3VKly70
[OReilly Superstream] Occupy the Space: A grassroots guide to engineering (an...Jason Yip
The typical problem in product engineering is not bad strategy, so much as “no strategy”. This leads to confusion, lack of motivation, and incoherent action. The next time you look for a strategy and find an empty space, instead of waiting for it to be filled, I will show you how to fill it in yourself. If you’re wrong, it forces a correction. If you’re right, it helps create focus. I’ll share how I’ve approached this in the past, both what works and lessons for what didn’t work so well.
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation F...AlexanderRichford
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation Functions to Prevent Interaction with Malicious QR Codes.
Aim of the Study: The goal of this research was to develop a robust hybrid approach for identifying malicious and insecure URLs derived from QR codes, ensuring safe interactions.
This is achieved through:
Machine Learning Model: Predicts the likelihood of a URL being malicious.
Security Validation Functions: Ensures the derived URL has a valid certificate and proper URL format.
This innovative blend of technology aims to enhance cybersecurity measures and protect users from potential threats hidden within QR codes 🖥 🔒
This study was my first introduction to using ML which has shown me the immense potential of ML in creating more secure digital environments!
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
3. Where is Intelligence Found on Earth?
• Brains and Minds
University of California, Berkeley
4. Brains and Minds
• But our understanding of how intelligence arises in
brains and minds is very primitive; we’re still at the
stage of metaphor
University of California, Berkeley
5. Brains and Minds
• But our understanding of how intelligence arises in
brains and minds is very primitive; we’re still at the
stage of metaphor
• Moreover, to build intelligent systems it is neither
necessary nor sufficient that the components of the
system exhibit human-like intelligence
University of California, Berkeley
6. Brains and Minds
• But our understanding of how intelligence arises in
brains and minds is very primitive; we’re still in the
stage of metaphor
• Moreover, to build intelligent systems it is neither
necessary nor sufficient that the components of the
system exhibit human-like intelligence
• Are there other examples of intelligence worth
mimicking in computer algorithms on the planet?
University of California, Berkeley
7. Where is Intelligence Found on Earth?
• Brains and Minds
• Markets
University of California, Berkeley
8. Markets
• Markets can be viewed as decentralized algorithms
• They accomplish complex tasks like bringing the
necessary goods into a city day in and day out
• They are adaptive (accommodating change in physical
or social structure), robust (working rain or shine),
scalable (working in small villages and big cities), and
they can have a very long lifetime
– indeed, they can work for decades or centuries
– if we’re looking for principles for lifelong adaptation, we should be
considering markets as intelligent systems!
• Of course, markets aren’t perfect, which simply means
that there are research opportunities
University of California, Berkeley
9. Examples at the Interface of ML and Econ
• Multi-way markets in which the individual agents need to
explore to learn their preferences
• Large-scale multi-way markets in which agents view other
sides of the market via recommendation systems
• Inferential methods for mitigating information asymmetries
• Latent variable inference in game theory
• Data collection in strategic settings
• Information sharing, free riding
• The goal is to discover new principles to build healthy (e.g.,
fair) learning-based markets that are stabilized over long
stretches of time
University of California, Berkeley
10. AI (aka Machine Learning) Successes
• First Generation (‘90-’00): the backend
– e.g., fraud detection, search, supply-chain management
• Second Generation (‘00-’10): the human side
– e.g., recommendation systems, commerce, social media
• Third Generation (‘10-now): pattern recognition
– e.g., speech recognition, computer vision, translation
• Fourth Generation (emerging): markets
– not just one agent making a decision or sequence of
decisions
– but a huge interconnected web of data, agents, decisions
University of California, Berkeley
11. Decisions
• It’s not just a matter of a threshold
• Real-world decisions with consequences
– counterfactuals, provenance, relevance, dialog
• Sets of decisions across a network
– false-discovery rate (instead of precision/recall/accuracy)
• Sets of decisions across a network over time
– streaming, asynchronous decisions (cf. Zrnic, Ramdas & Jordan,
Asynchronous online testing of multiple hypotheses, arXiv, 2019)
• Decisions when there is scarcity and competition
– need for an economic perspective
12. Consider Classical Recommendation
Systems
• A record is kept of each customer’s purchases
• Customers are “similar” if they buy similar sets of
items
• Items are “similar” are they are bought together by
multiple customers
• Recommendations are made on the basis of these
similarities
• These systems have become a commodity
13. Multiple Decisions with Competition
• Suppose that recommending a certain movie is a good
business decision (e.g., because it’s very popular)
• Is it OK to recommend the same movie to everyone?
• Is it OK to recommend the same book to everyone?
• Is it OK to recommend the same restaurant to
everyone?
• Is it OK to recommend the same street to every driver?
• Is it OK to recommend the same stock purchase to
everyone?
14. The Alternative: Create a Market
• A two-way market between consumers and producers
– based on recommendation systems on both sides
• E.g., diners are one side of the market, and restaurants
on the other side
• E.g., drivers are one side of the market, and street
segments on the other side
• This isn’t just classical microeconomics; the use of
recommendation systems is key
University of California, Berkeley
15. The Alternative: Create a Market
• A two-way market between consumers and producers
– based on recommendation systems on both sides
• E.g., diners are one side of the market, and restaurants
on the other side
• E.g., drivers are one side of the market, and street
segments on the other side
• This isn’t just classical microeconomics; the use of
recommendation systems is key
University of California, Berkeley
16. Machine Learning Meets Economics
• Bandits that compete (with Lydia Liu and Horia Mania)
– how to explore and exploit when others may also be attempting to
explore and exploit
• Finding Nash equilibria with gradient-based algorithms in
high-dimensional action spaces (with Eric Mazumdar)
– avoiding saddle points that are not Nash equilibria
• Local Stackelberg equilibria in nonconvex-nonconcave
games (with Chi Jin and Praneeth Netrapalli)
– and how does this relate to SGD?
• Asynchronous, online false-discovery rate control (with
Tijana Zrnic and Aaditya Ramdas)
– the real-world A/B testing problem
University of California, Berkeley
18. • MABs offer a natural platform to understand exploration /
exploitation trade-offs
1
2
3
.
Multi-Armed Bandits
19. Buyers / Demand Sellers / Supply
1 > 3 > 2
Suppose we have a market in which the participants have
preferences:
2 > 3 > 1
1 > 2 > 3
1 > 2 > 3
3 > 1 > 2
2 > 1 > 3
Gale and Shapley introduced this problem in 1962 and proposed
a celebrated algorithm that always finds a stable match
In this algorithm one side of the market iteratively makes
proposals to the other side
Matching Markets
20. What if the participants in the market do not know their
preferences a priori, but observe noisy utilities through
repeated interactions?
Matching Markets Meet Learning
Now the participants have an exploration/exploitation problem,
in the context of other participants
22. • We conceive of a bandit market: agents on one side, arms on
the other side.
Agents get noisy rewards when they pull arms.
Arms have preferences over agents (these
preferences can also express agents’ skill
levels)
When multiple agents pull the same arm only
the most preferred agent gets a reward.
Bandit Markets
23. Then it is natural to define the regret of agent i up to time n
as:
Mean reward of
stable match
Reward at time t
Minimizing this regret is natural. It says that agents should
expect rewards as good as their stable match in hindsight.
Regret in Bandit Markets
24. Gale-Shapley upper confidence bounds (GS-UCB):
• Agents rank arms according to upper confidence bounds
for the mean rewards.
• Agents submit rankings to a matching platform.
• The platform uses these rankings to run the Gale-Shapley
algorithm to match agents and arms.
• Agents receive rewards and update upper confidence
bounds.
• Repeat.
Regret-Minimizing Algorithm
25. Theorem (informal): If there are N agents and K arms and
GS-UCB is run, the regret of agent i satisfies
Reward gap of possibly other agents.
• In other words, if the bear decides to explore more, the human
might have higher regret.
• See paper for refinements of this bound and further discussion of
exploration-exploitation trade-offs in this setting.
• Finally, we note that GS-UCB is incentive compatible. No single
agent has an incentive to deviate from the method.
Theorem
27. Why zero-sum games?
• A lot of recent interest in finding the local Nash equilibria of zero-sum
continuous games.
• e.g. adversarial learning, training GANs, robust reinforcement learning
Zero-Sum Games
28. Issues with gradient-play in 2-player zero-sum games
Not all locally
asymptotically
stable equilibria
are differential
Nash equilibria!
Simultaneous Gradient
descent (simGD) has
two main issues:
• Convergence to limit
cycles.
• Convergence to non-
Nash fixed points.
2
29. Recent algorithms do not solve these problems
2
9
On finding local Nash equilibria (and only local Nash equilibria) in zero-sum continuous games. E. Mazumdar, M. Jordan, S. Sastry NeurIPS
Behaviors in Zero-sum Games
Algorithm
Avoid some
local Nash
eq.
Converge to
non-Nash
eq.
Oscillate
around eq.
Converge to
limit cycles
Gradient
play
Symplectic
Gradient
Adjustment
Consensus
Optimization
30. New algorithm for gradient-based learning in zero-sum
games:
3
0
Local Symplectic Surgery:
On finding local Nash equilibria (and only local Nash equilibria) in zero-sum continuous games. E. Mazumdar, M. Jordan, S. Sastry NeurIPS
By cancelling out the symplectic part of the vector field
around critical points, the only equilibria to which this method
can converge are the local Nash equilibria of the game.
31. 31On finding local Nash equilibria (and only local Nash equilibria) in zero-sum continuous games. E. Mazumdar, M. Jordan, S. Sastry NeurIPS
GAN experiments on a test for mode-collapse
LSS
(~50,000 iters)
simGD
(~200,000 iters)
Consensus opt.
(~100,000 iters)
Curvature exploitation
(~15,000 iters)
Ground truth is a mixture of 16 Gaussians used to test for mode collapse, with covariance 0.005 I2
Generator and discriminator are Tanh neural networks with 4 hidden layers of 100 neurons each.
32. 3
2
On finding local Nash equilibria (and only local Nash equilibria) in zero-sum continuous games. E. Mazumdar, M. Jordan, S. Sastry NeurIPS
LSS seems to converge to better solutions
LSS
(~50,000
iters)
simGD
(~200,000
iters)
Consensus
(~100,000
iters)
Curvature
(~15,000 iters)
From an initialization where
simGD quickly converges to an
incorrect distribution, LSS
recovers the ground truth.
• Consensus optimization, without
careful hyper-parameter tuning
gives rise to new non-Nash
equilibria, which results in worse
performance.
• Local curvature exploitation is
hard to implement in high
dimensional settings due to the
need to find eigenvalue/eigenvector
pairs of the diagonal blocks of J at
each iteration
33. What is Local Optimality in Nonconvex-Nonconcave
Minimax Optimization?
Chi Jin1, Praneeth Netrapalli2, Michael I. Jordan1
1
University of California, Berkeley. 2
Microsoft Research, India.
36. Algorithms and Questions
Gradient Descent Ascent (GDA):
xt+1 = xt − ηx f (xt ).
yt+1 = yt + ηy f (yt ).
−2 0 2
x1
−2
−1
0
1
2
x2
Fundamental questions:
• What “optimal” points should we find?
• What are the points that GDA converges to? (if it converges)
37. Nash Equilibrium
Point (x , y ) is a Nash equilibrium if:
• y is a maximum of f (x , ·); x is a minimum of f (·, y ).
Convex-concave case: GDA converges to a Nash equilibrium.
38. Local Nash Equilibrium
Nonconvex-nonconcave case: It’s NP-hard to find Nash equilibrium.
Find a local Nash equilibrium—point (x , y ) such that
• y is a local maximum of f (x , ·); x is a local minimum of f (·, y ).
Are local Nash equilibria what we want in GAN/adversarial training?
39. Simultaneous vs. Sequential
simultaneous
vs.
sequential
Nash equilibria come from simultaneous games—both act simultaneously.
GAN/adversarial training are sequential games—one leader, one follower.
For nonconvex-nonconcave f , which player acts first is crucial, since in general
min
x∈Rd
max
y∈Rd
f (x, y) = max
y∈Rd
min
x∈Rd
f (x, y)
40. Minimax Point (Stackelberg Equilibrium)
Point (x , y ) is a Minimax Point (or Stackelberg Equilibrium) if:
• y is a maximum of f (x , ·);
• x is a minimum of φ(·) where φ(x) = maxy f (x, y).
[Leader always prepares for the best follower.]
Minimax points always exist even for nonconvex-nonconcave functions.
41. A New Notion of Local Optimality
[This Work] Point (x , y ) is a local minimax point if:
• y is a local maximum of f (x , ·);
• ∃ 0 > 0, so that x is a local minimum of g (·) for any ≤ 0,
where g (x) = maxy: y−y ≤ f (x, y).
42. Limit Points of GDA
[DP18, MR18]: The stable limit points of GDA need not to be local Nash!
Theorem [This Work]
When learning rate ηy /ηx → ∞, the stable limit points of GDA are exactly
local minimax points up to some degenerate points.
γ-GDA
Local
Minimax
(∞-GDA)
Local
Maximin
Local
Nash