This document provides an overview of transfer learning. It begins by comparing traditional machine learning to transfer learning, noting that transfer learning aims to transfer knowledge across different but related domains or tasks. It discusses settings of transfer learning including inductive, transductive, and unsupervised. Approaches covered include instance transfer, feature representation transfer, model transfer, and relational knowledge transfer. It also notes negative transfer can occur if domains are too dissimilar. The document provides examples of various transfer learning approaches and discusses motivation, assumptions, and goals.
In some applications, the output of the system is a sequence of actions. In such a case, a single action is not important
game playing where a single move by itself is not that important.in the case of the agent acts on its environment, it receives some evaluation of its action (reinforcement),
but is not told of which action is the correct one to achieve its goal
In this talk we discuss about the aplicação of Reinforcement Learning to Games. Recently, OpenAI created an algorithm capable of beating a human team in DOTA, considered a game with great amount of complexity and strategy. In this talk, we'll evaluate the role Reinforcement Learning plays in the world of games, taking a look at some of main achievements and how they look like in terms of implementation. We'll also take a look at some of the history of AI applied to games and how things evolved over time.
Reinforcement learning is a machine learning technique that involves trial-and-error learning. The agent learns to map situations to actions by trial interactions with an environment in order to maximize a reward signal. Deep Q-networks use reinforcement learning and deep learning to allow agents to learn complex behaviors directly from high-dimensional sensory inputs like pixels. DQN uses experience replay and target networks to stabilize learning from experiences. DQN has achieved human-level performance on many Atari 2600 games.
Data Analysis: Evaluation Metrics for Supervised Learning Models of Machine L...Md. Main Uddin Rony
This document discusses various machine learning evaluation metrics for supervised learning models. It covers classification, regression, and ranking metrics. For classification, it describes accuracy, confusion matrix, log-loss, and AUC. For regression, it discusses RMSE and quantiles of errors. For ranking, it explains precision-recall, precision-recall curves, F1 score, and NDCG. The document provides examples and visualizations to illustrate how these metrics are calculated and used to evaluate model performance.
Transfer learning aims to improve learning in a target domain by leveraging knowledge from a related source domain. It is useful when the target domain has limited labeled data. There are several approaches, including instance-based approaches that reweight or resample source instances, and feature-based approaches that learn a transformation to align features across domains. Spectral feature alignment is one technique that builds a graph of correlations between pivot features shared across domains and domain-specific features, then applies spectral clustering to derive new shared features.
This document discusses intelligent agents and their environments. It covers:
1) Intelligent agents are entities that perceive their environment through sensors and act upon the environment through actuators. They map percept sequences to actions.
2) A rational agent should select actions that are expected to maximize its performance measure given the percept sequence and its prior knowledge. Performance measures evaluate how well the agent solves its task.
3) Agent environments can have different properties such as being fully or partially observable, deterministic or stochastic, episodic or sequential, static or dynamic, discrete or continuous, and single-agent or multi-agent. The simplest is fully observable, deterministic, etc. but most real environments are more complex.
1. The document discusses various machine learning classification algorithms including neural networks, support vector machines, logistic regression, and radial basis function networks.
2. It provides examples of using straight lines and complex boundaries to classify data with neural networks. Maximum margin hyperplanes are used for support vector machine classification.
3. Logistic regression is described as useful for binary classification problems by using a sigmoid function and cross entropy loss. Radial basis function networks can perform nonlinear classification with a kernel trick.
An introduction to Machine Learning (and a little bit of Deep Learning)Thomas da Silva Paula
25-min talk about Machine Learning and a little bit of Deep Learning. Starts with some basic definitions (Supervised and Unsupervised Learning). Then, neural networks basic functionality is explained, ending up in Deep Learning and Convolutional Neural Networks.
Machine Learning Meetup that happened in Porto Alegre, Brazil.
In some applications, the output of the system is a sequence of actions. In such a case, a single action is not important
game playing where a single move by itself is not that important.in the case of the agent acts on its environment, it receives some evaluation of its action (reinforcement),
but is not told of which action is the correct one to achieve its goal
In this talk we discuss about the aplicação of Reinforcement Learning to Games. Recently, OpenAI created an algorithm capable of beating a human team in DOTA, considered a game with great amount of complexity and strategy. In this talk, we'll evaluate the role Reinforcement Learning plays in the world of games, taking a look at some of main achievements and how they look like in terms of implementation. We'll also take a look at some of the history of AI applied to games and how things evolved over time.
Reinforcement learning is a machine learning technique that involves trial-and-error learning. The agent learns to map situations to actions by trial interactions with an environment in order to maximize a reward signal. Deep Q-networks use reinforcement learning and deep learning to allow agents to learn complex behaviors directly from high-dimensional sensory inputs like pixels. DQN uses experience replay and target networks to stabilize learning from experiences. DQN has achieved human-level performance on many Atari 2600 games.
Data Analysis: Evaluation Metrics for Supervised Learning Models of Machine L...Md. Main Uddin Rony
This document discusses various machine learning evaluation metrics for supervised learning models. It covers classification, regression, and ranking metrics. For classification, it describes accuracy, confusion matrix, log-loss, and AUC. For regression, it discusses RMSE and quantiles of errors. For ranking, it explains precision-recall, precision-recall curves, F1 score, and NDCG. The document provides examples and visualizations to illustrate how these metrics are calculated and used to evaluate model performance.
Transfer learning aims to improve learning in a target domain by leveraging knowledge from a related source domain. It is useful when the target domain has limited labeled data. There are several approaches, including instance-based approaches that reweight or resample source instances, and feature-based approaches that learn a transformation to align features across domains. Spectral feature alignment is one technique that builds a graph of correlations between pivot features shared across domains and domain-specific features, then applies spectral clustering to derive new shared features.
This document discusses intelligent agents and their environments. It covers:
1) Intelligent agents are entities that perceive their environment through sensors and act upon the environment through actuators. They map percept sequences to actions.
2) A rational agent should select actions that are expected to maximize its performance measure given the percept sequence and its prior knowledge. Performance measures evaluate how well the agent solves its task.
3) Agent environments can have different properties such as being fully or partially observable, deterministic or stochastic, episodic or sequential, static or dynamic, discrete or continuous, and single-agent or multi-agent. The simplest is fully observable, deterministic, etc. but most real environments are more complex.
1. The document discusses various machine learning classification algorithms including neural networks, support vector machines, logistic regression, and radial basis function networks.
2. It provides examples of using straight lines and complex boundaries to classify data with neural networks. Maximum margin hyperplanes are used for support vector machine classification.
3. Logistic regression is described as useful for binary classification problems by using a sigmoid function and cross entropy loss. Radial basis function networks can perform nonlinear classification with a kernel trick.
An introduction to Machine Learning (and a little bit of Deep Learning)Thomas da Silva Paula
25-min talk about Machine Learning and a little bit of Deep Learning. Starts with some basic definitions (Supervised and Unsupervised Learning). Then, neural networks basic functionality is explained, ending up in Deep Learning and Convolutional Neural Networks.
Machine Learning Meetup that happened in Porto Alegre, Brazil.
Dr. Subrat Panda gave an introduction to reinforcement learning. He defined reinforcement learning as dealing with agents that must sense and act upon their environment to receive delayed scalar feedback in the form of rewards. He described key concepts like the Markov decision process framework, value functions, Q-functions, exploration vs exploitation, and extensions like deep reinforcement learning. He listed several real-world applications of reinforcement learning and resources for learning more.
Model-based reinforcement learning techniques were presented that use learned models to improve upon model-free deep reinforcement learning. Several papers augmented deep networks with model-based components like planners or simulators to leverage predictions and reduce sample complexity. Techniques included using model rollouts to augment state representations, learning abstract state representations to simplify value prediction, and optimizing policies on ensemble models. While model-based methods show promise in addressing deep RL limitations, challenges remain in learning accurate models and developing policies robust to model errors.
This document provides an introduction to deep reinforcement learning. It begins with an overview of reinforcement learning and its key characteristics such as using reward signals rather than supervision and sequential decision making. The document then covers the formulation of reinforcement learning problems using Markov decision processes and the typical components of an RL agent including policies, value functions, and models. It discusses popular RL algorithms like Q-learning, deep Q-networks, and policy gradient methods. The document concludes by outlining some potential applications of deep reinforcement learning and recommending further educational resources.
The document discusses graph data mining and provides the following key points:
1. It outlines topics in graph data mining including frequent subgraph mining, graph indexing, similarity search, classification, and clustering.
2. Frequent subgraph mining aims to discover subgraphs that occur frequently in a graph database based on a minimum support threshold.
3. Graph indexing and similarity search techniques aim to enable efficient subgraph search in large graph databases by indexing substructures.
Problem solving
Problem formulation
Search Techniques for Artificial Intelligence
Classification of AI searching Strategies
What is Search strategy ?
Defining a Search Problem
State Space Graph versus Search Trees
Graph vs. Tree
Problem Solving by Search
Transfer learning aims to improve learning outcomes for a target task by leveraging knowledge from a related source task. It does this by influencing the target task's assumptions based on what was learned from the source task. This can allow for faster and better generalized learning in the target task. However, there is a risk of negative transfer where performance decreases. To avoid this, methods examine task similarity and reject harmful source knowledge, or generate multiple mappings between source and target to identify the best match. The goal of transfer learning is to start higher, learn faster, and achieve better overall performance compared to learning the target task without transfer.
The document discusses recurrent neural networks (RNNs) and long short-term memory (LSTM) networks. It provides details on the architecture of RNNs including forward and back propagation. LSTMs are described as a type of RNN that can learn long-term dependencies using forget, input and output gates to control the cell state. Examples of applications for RNNs and LSTMs include language modeling, machine translation, speech recognition, and generating image descriptions.
An introduction to reinforcement learningJie-Han Chen
This document provides an introduction and overview of reinforcement learning. It begins with a syllabus that outlines key topics such as Markov decision processes, dynamic programming, Monte Carlo methods, temporal difference learning, deep reinforcement learning, and active research areas. It then defines the key elements of reinforcement learning including policies, reward signals, value functions, and models of the environment. The document discusses the history and applications of reinforcement learning, highlighting seminal works in backgammon, helicopter control, Atari games, Go, and dialogue generation. It concludes by noting challenges in the field and prominent researchers contributing to its advancement.
This document provides an overview of machine learning, including definitions, types, steps, and applications. It defines machine learning as a field that gives computers the ability to learn without being explicitly programmed. The document outlines the main types of machine learning as supervised learning, unsupervised learning, semi-supervised learning, and reinforcement learning. It also describes the typical steps in a machine learning process as gathering data, preparing data, choosing a model, training, evaluation, and prediction. Examples of machine learning applications discussed include prediction, image recognition, natural language processing, and personal assistants. Popular machine learning languages and packages are also listed.
I.ITERATIVE DEEPENING DEPTH FIRST SEARCH(ID-DFS) II.INFORMED SEARCH IN ARTIFI...vikas dhakane
Artificial Intelligence: Introduction, Typical Applications. State Space Search: Depth Bounded
DFS, Depth First Iterative Deepening. Heuristic Search: Heuristic Functions, Best First Search,
Hill Climbing, Variable Neighborhood Descent, Beam Search, Tabu Search. Optimal Search: A
*
algorithm, Iterative Deepening A*
, Recursive Best First Search, Pruning the CLOSED and OPEN
Lists
Hello~! :)
While studying the Sutton-Barto book, the traditional textbook for Reinforcement Learning, I created PPT about the Multi-armed Bandits, a Chapter 2.
If there are any mistakes, I would appreciate your feedback immediately.
Thank you.
The document discusses local search algorithms for optimization problems, including hill climbing, simulated annealing, and Tabu search. Hill climbing performs a local search by iteratively moving to neighbor states with improved cost until a local optimum is reached. Simulated annealing allows some "bad" moves with decreasing probability to help escape local optima. Tabu search uses a tabu list to avoid getting stuck in cycles and encourages exploring new regions of the search space. These local search methods are suitable for problems where the solution is the goal state itself rather than the path to get there.
Data Science, Machine Learning and Neural NetworksBICA Labs
Lecture briefly overviewing state of the art of Data Science, Machine Learning and Neural Networks. Covers main Artificial Intelligence technologies, Data Science algorithms, Neural network architectures and cloud computing facilities enabling the whole stack.
This document summarizes a machine learning workshop on feature selection. It discusses typical feature selection methods like single feature evaluation using metrics like mutual information and Gini indexing. It also covers subset selection techniques like sequential forward selection and sequential backward selection. Examples are provided showing how feature selection improves performance for logistic regression on large datasets with more features than samples. The document outlines the workshop agenda and provides details on when and why feature selection is important for machine learning models.
Reinforcement Learning (RL) approaches to deal with finding an optimal reward based policy to act in an environment (Charla en Inglés)
However, what has led to their widespread use is its combination with deep neural networks (DNN) i.e., deep reinforcement learning (Deep RL). Recent successes on not only learning to play games but also superseding humans in it and academia-industry research collaborations like for manipulation of objects, locomotion skills, smart grids, etc. have surely demonstrated their case on a wide variety of challenging tasks.
With application spanning across games, robotics, dialogue, healthcare, marketing, energy and many more domains, Deep RL might just be the power that drives the next generation of Artificial Intelligence (AI) agents!
Machine learning involves developing systems that can learn from data and experience. The document discusses several machine learning techniques including decision tree learning, rule induction, case-based reasoning, supervised and unsupervised learning. It also covers representations, learners, critics and applications of machine learning such as improving search engines and developing intelligent tutoring systems.
Deep learning - Conceptual understanding and applicationsBuhwan Jeong
This document provides an overview of deep learning, including conceptual understanding and applications. It defines deep learning as a deep and wide artificial neural network. It describes key concepts in artificial neural networks like signal transmission between neurons, graphical models, linear/logistic regression, weights/biases/activation, and backpropagation. It also discusses popular deep learning applications and techniques like speech recognition, natural language processing, computer vision, representation learning using restricted Boltzmann machines and autoencoders, and deep network architectures.
Stuart russell and peter norvig artificial intelligence - a modern approach...Lê Anh Đạt
This document provides publishing information for the book "Artificial Intelligence: A Modern Approach". It lists the editorial staff and production team, including the Vice President and Editorial Director, Editor-in-Chief, Executive Editor, and others. It also provides copyright information, acknowledging that the content is protected and requires permission for reproduction. Finally, it is dedicated to the authors' families and includes a preface giving an overview of the book.
Flavours of Physics Challenge: Transfer Learning approachAlexander Rakhlin
Presentation for "Heavy Flavour Data Mining workshop", February 18-19, University of Zurich. I discuss the solution that won Physics Prize of Flavours of Physics challenge organized by CERN, Yandex, Intel at Kaggle.
This document presents a method for self-taught clustering that uses unlabeled auxiliary data to improve clustering performance on a target dataset. The method performs co-clustering on both the target and auxiliary data in a common feature space in order to learn a representation that is consistent between the two datasets. The algorithm iteratively chooses cluster assignments for target instances, auxiliary instances, and features to optimize an objective function that minimizes information loss. Experimental results on image datasets demonstrate that the self-taught clustering approach can enhance clustering performance by utilizing irrelevant auxiliary data.
Dr. Subrat Panda gave an introduction to reinforcement learning. He defined reinforcement learning as dealing with agents that must sense and act upon their environment to receive delayed scalar feedback in the form of rewards. He described key concepts like the Markov decision process framework, value functions, Q-functions, exploration vs exploitation, and extensions like deep reinforcement learning. He listed several real-world applications of reinforcement learning and resources for learning more.
Model-based reinforcement learning techniques were presented that use learned models to improve upon model-free deep reinforcement learning. Several papers augmented deep networks with model-based components like planners or simulators to leverage predictions and reduce sample complexity. Techniques included using model rollouts to augment state representations, learning abstract state representations to simplify value prediction, and optimizing policies on ensemble models. While model-based methods show promise in addressing deep RL limitations, challenges remain in learning accurate models and developing policies robust to model errors.
This document provides an introduction to deep reinforcement learning. It begins with an overview of reinforcement learning and its key characteristics such as using reward signals rather than supervision and sequential decision making. The document then covers the formulation of reinforcement learning problems using Markov decision processes and the typical components of an RL agent including policies, value functions, and models. It discusses popular RL algorithms like Q-learning, deep Q-networks, and policy gradient methods. The document concludes by outlining some potential applications of deep reinforcement learning and recommending further educational resources.
The document discusses graph data mining and provides the following key points:
1. It outlines topics in graph data mining including frequent subgraph mining, graph indexing, similarity search, classification, and clustering.
2. Frequent subgraph mining aims to discover subgraphs that occur frequently in a graph database based on a minimum support threshold.
3. Graph indexing and similarity search techniques aim to enable efficient subgraph search in large graph databases by indexing substructures.
Problem solving
Problem formulation
Search Techniques for Artificial Intelligence
Classification of AI searching Strategies
What is Search strategy ?
Defining a Search Problem
State Space Graph versus Search Trees
Graph vs. Tree
Problem Solving by Search
Transfer learning aims to improve learning outcomes for a target task by leveraging knowledge from a related source task. It does this by influencing the target task's assumptions based on what was learned from the source task. This can allow for faster and better generalized learning in the target task. However, there is a risk of negative transfer where performance decreases. To avoid this, methods examine task similarity and reject harmful source knowledge, or generate multiple mappings between source and target to identify the best match. The goal of transfer learning is to start higher, learn faster, and achieve better overall performance compared to learning the target task without transfer.
The document discusses recurrent neural networks (RNNs) and long short-term memory (LSTM) networks. It provides details on the architecture of RNNs including forward and back propagation. LSTMs are described as a type of RNN that can learn long-term dependencies using forget, input and output gates to control the cell state. Examples of applications for RNNs and LSTMs include language modeling, machine translation, speech recognition, and generating image descriptions.
An introduction to reinforcement learningJie-Han Chen
This document provides an introduction and overview of reinforcement learning. It begins with a syllabus that outlines key topics such as Markov decision processes, dynamic programming, Monte Carlo methods, temporal difference learning, deep reinforcement learning, and active research areas. It then defines the key elements of reinforcement learning including policies, reward signals, value functions, and models of the environment. The document discusses the history and applications of reinforcement learning, highlighting seminal works in backgammon, helicopter control, Atari games, Go, and dialogue generation. It concludes by noting challenges in the field and prominent researchers contributing to its advancement.
This document provides an overview of machine learning, including definitions, types, steps, and applications. It defines machine learning as a field that gives computers the ability to learn without being explicitly programmed. The document outlines the main types of machine learning as supervised learning, unsupervised learning, semi-supervised learning, and reinforcement learning. It also describes the typical steps in a machine learning process as gathering data, preparing data, choosing a model, training, evaluation, and prediction. Examples of machine learning applications discussed include prediction, image recognition, natural language processing, and personal assistants. Popular machine learning languages and packages are also listed.
I.ITERATIVE DEEPENING DEPTH FIRST SEARCH(ID-DFS) II.INFORMED SEARCH IN ARTIFI...vikas dhakane
Artificial Intelligence: Introduction, Typical Applications. State Space Search: Depth Bounded
DFS, Depth First Iterative Deepening. Heuristic Search: Heuristic Functions, Best First Search,
Hill Climbing, Variable Neighborhood Descent, Beam Search, Tabu Search. Optimal Search: A
*
algorithm, Iterative Deepening A*
, Recursive Best First Search, Pruning the CLOSED and OPEN
Lists
Hello~! :)
While studying the Sutton-Barto book, the traditional textbook for Reinforcement Learning, I created PPT about the Multi-armed Bandits, a Chapter 2.
If there are any mistakes, I would appreciate your feedback immediately.
Thank you.
The document discusses local search algorithms for optimization problems, including hill climbing, simulated annealing, and Tabu search. Hill climbing performs a local search by iteratively moving to neighbor states with improved cost until a local optimum is reached. Simulated annealing allows some "bad" moves with decreasing probability to help escape local optima. Tabu search uses a tabu list to avoid getting stuck in cycles and encourages exploring new regions of the search space. These local search methods are suitable for problems where the solution is the goal state itself rather than the path to get there.
Data Science, Machine Learning and Neural NetworksBICA Labs
Lecture briefly overviewing state of the art of Data Science, Machine Learning and Neural Networks. Covers main Artificial Intelligence technologies, Data Science algorithms, Neural network architectures and cloud computing facilities enabling the whole stack.
This document summarizes a machine learning workshop on feature selection. It discusses typical feature selection methods like single feature evaluation using metrics like mutual information and Gini indexing. It also covers subset selection techniques like sequential forward selection and sequential backward selection. Examples are provided showing how feature selection improves performance for logistic regression on large datasets with more features than samples. The document outlines the workshop agenda and provides details on when and why feature selection is important for machine learning models.
Reinforcement Learning (RL) approaches to deal with finding an optimal reward based policy to act in an environment (Charla en Inglés)
However, what has led to their widespread use is its combination with deep neural networks (DNN) i.e., deep reinforcement learning (Deep RL). Recent successes on not only learning to play games but also superseding humans in it and academia-industry research collaborations like for manipulation of objects, locomotion skills, smart grids, etc. have surely demonstrated their case on a wide variety of challenging tasks.
With application spanning across games, robotics, dialogue, healthcare, marketing, energy and many more domains, Deep RL might just be the power that drives the next generation of Artificial Intelligence (AI) agents!
Machine learning involves developing systems that can learn from data and experience. The document discusses several machine learning techniques including decision tree learning, rule induction, case-based reasoning, supervised and unsupervised learning. It also covers representations, learners, critics and applications of machine learning such as improving search engines and developing intelligent tutoring systems.
Deep learning - Conceptual understanding and applicationsBuhwan Jeong
This document provides an overview of deep learning, including conceptual understanding and applications. It defines deep learning as a deep and wide artificial neural network. It describes key concepts in artificial neural networks like signal transmission between neurons, graphical models, linear/logistic regression, weights/biases/activation, and backpropagation. It also discusses popular deep learning applications and techniques like speech recognition, natural language processing, computer vision, representation learning using restricted Boltzmann machines and autoencoders, and deep network architectures.
Stuart russell and peter norvig artificial intelligence - a modern approach...Lê Anh Đạt
This document provides publishing information for the book "Artificial Intelligence: A Modern Approach". It lists the editorial staff and production team, including the Vice President and Editorial Director, Editor-in-Chief, Executive Editor, and others. It also provides copyright information, acknowledging that the content is protected and requires permission for reproduction. Finally, it is dedicated to the authors' families and includes a preface giving an overview of the book.
Flavours of Physics Challenge: Transfer Learning approachAlexander Rakhlin
Presentation for "Heavy Flavour Data Mining workshop", February 18-19, University of Zurich. I discuss the solution that won Physics Prize of Flavours of Physics challenge organized by CERN, Yandex, Intel at Kaggle.
This document presents a method for self-taught clustering that uses unlabeled auxiliary data to improve clustering performance on a target dataset. The method performs co-clustering on both the target and auxiliary data in a common feature space in order to learn a representation that is consistent between the two datasets. The algorithm iteratively chooses cluster assignments for target instances, auxiliary instances, and features to optimize an objective function that minimizes information loss. Experimental results on image datasets demonstrate that the self-taught clustering approach can enhance clustering performance by utilizing irrelevant auxiliary data.
Using web images to learn concept classifiers for videos is challenging due to differences in data distributions between image and video domains. Transfer learning from images to videos can result in either positive or negative transfer depending on factors like the number of positive examples, type of concept (object vs. event), and degree of mismatch between source and target data distributions. The study found that semantic pooling of additional web images helped improve classifiers for 22 out of 46 concepts when the target domain had less than 100 positive examples. Events were more difficult to transfer compared to objects and scenes. Concepts with lower maximum mean discrepancy between source and target data distributions also tended to benefit more from transfer learning. While transfer is not always beneficial, semantic pooling provides a practical strategy for divers
This document summarizes research on transfer defect learning to improve cross-project defect prediction. It presents Transfer Component Analysis (TCA) as a state-of-the-art transfer learning technique that maps data from source and target projects into a shared feature space to make their distributions more similar. It then proposes TCA+ which augments TCA with data normalization and decision rules to select the optimal normalization method based on characteristics of the source and target datasets. Experimental results on two cross-project defect prediction datasets show that TCA+ significantly outperforms traditional cross-project prediction and basic TCA.
The Blue Brain Project aims to create a digital reconstruction of the brain through reverse engineering mammalian brain circuitry. It was founded in 2005 by the Brain and Mind Institute of the École Polytechnique Fédérale de Lausanne in Switzerland with $20 million in initial funding. While the Blue Brain Project could help address various health issues, it still requires significant additional research, time, and funding to fully map the human brain.
This document provides an overview of software defect prediction approaches from the 1970s to the present. It discusses early approaches using simple metrics like lines of code and complexity metrics. It then covers the development of prediction models using machine learning techniques like regression and classification. More recent topics discussed include just-in-time prediction models, practical applications in industry, using historical metrics from software repositories, addressing noise in data, and the feasibility of cross-project prediction. The document outlines challenges and opportunities for future work in the field of software defect prediction.
Learn to Build an App to Find Similar Images using Deep Learning- Piotr TeterwakPyData
This document discusses using deep learning and deep features to build an app that finds similar images. It begins with an overview of deep learning and how neural networks can learn complex patterns in data. The document then discusses how pre-trained neural networks can be used as feature extractors for other domains through transfer learning. This reduces data and tuning requirements compared to training new deep learning models. The rest of the document focuses on building an image similarity service using these techniques, including training a model with GraphLab Create and deploying it as a web service with Dato Predictive Services.
This document discusses transfer learning and domain adaptation techniques for training deep learning models when limited labeled data is available. It describes using pre-trained networks to extract features, fine-tuning networks on related tasks, and unsupervised domain adaptation methods. Transfer learning can outperform training from scratch by leveraging knowledge gained from large labeled datasets.
Artificial Neural Network Seminar - Google BrainRawan Al-Omari
it's our seminar in artificial neural network course, at F.I.T.E, AI Dept.
it's about Google Brain project, and who they using neural network in building it .
actually it's a very interesting project they work on it .
for more information about this project :
http://nyti.ms/T5E71e
http://imatge-upc.github.io/telecombcn-2016-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.
Transfer of learning refers to applying knowledge learned in one context to new contexts. Positive transfer occurs when learning in one context improves performance in another, while negative transfer happens when prior learning hurts new performance. Near transfer is between similar contexts, while far transfer bridges more dissimilar situations. Theories of transfer include identical elements (common aspects facilitate learning), generalization (experiences apply broadly), and configuration (perceived similarities enable transfer). Factors like ability, nature, attitudes, teaching style, and understanding impact how well transfer succeeds. Motivation directs attention toward goals and depends on individual characteristics and experiences.
This document discusses the Blue Brain project, which aims to create a virtual human brain through detailed biological simulations. It describes how the Blue Brain is being constructed by scanning brain tissue to measure neuronal properties, translating these into mathematical models, and running large-scale simulations. The Blue Brain would function similarly to a natural brain, processing inputs, interpreting them through neuronal states, and producing outputs. It could allow human intelligence and memory to be uploaded and stored digitally. The Blue Brain may one day help treat conditions like memory loss and allow life-like artificial intelligence.
This document discusses the Blue Brain project, which aims to create a virtual human brain through computer simulation. The Blue Brain would function similarly to a human brain by taking inputs, interpreting information, storing memories, and generating outputs. Researchers plan to upload an actual human brain's contents into the Blue Brain by using nanobots to scan neuronal connections and activity in the body and transfer that data to the virtual brain. The Blue Brain could help keep knowledge and intelligence after death and allow people with disabilities to regain abilities. It may take 30 years to fully develop the technology needed.
This document summarizes a seminar presentation on brain fingerprinting technology. Brain fingerprinting uses EEG to measure electrical brain wave responses, specifically the P300 wave, to stimuli presented on a computer in order to determine if individuals have hidden information stored in their brains. It works by presenting probes, targets, and irrelevant stimuli and analyzing the brain's differential response. There are four phases: evidence collection, brain evidence collection, computer analysis, and determining guilt or innocence. Unlike polygraph tests, it does not rely on physiological responses but on cognitive brain responses. Case studies showed it correctly identified information stored in a murder suspect's brain and its potential use in identifying trained terrorists.
Blue brain enables humans to give new dimensions to science and technology and make enormous development in making the best possible enlightenment to the present scenario.the details can be seen by going though the power point presentation
Brain fingerprinting is a technique developed by Lawrence Farwell that uses electroencephalography (EEG) to detect electrical brainwave responses called MERMERs that are elicited when a person recognizes familiar stimuli. It works by measuring the brain's response when a subject is exposed to words or images related to a crime. If the brainwave patterns match those that would be expected from someone familiar with the crime details, it suggests the person has knowledge of the crime. Brain fingerprinting has been used to help solve criminal cases and evaluate brain functioning, though further research with larger samples is still needed to fully validate its accuracy and capabilities.
This document provides an overview of transfer learning. It begins by comparing traditional machine learning to transfer learning, noting that transfer learning aims to transfer knowledge across different but related domains or tasks. It discusses why transfer learning is useful when labeled data is limited. It then outlines various settings of transfer learning like inductive, transductive, and unsupervised. It reviews common approaches to transfer learning like instance transfer, feature representation transfer, model transfer, and relational knowledge transfer. It also notes the potential issue of negative transfer. In closing, the document provides a high-level survey of the concepts and techniques in the field of transfer learning.
This document provides an outline and overview of Yoshua Bengio's 2012 tutorial on representation learning. The key points covered include:
1) The tutorial will cover motivations for representation learning, algorithms such as probabilistic models and auto-encoders, and analysis and practical issues.
2) Representation learning aims to automatically learn good representations of data rather than relying on handcrafted features. Learning representations can help address challenges like exploiting unlabeled data and the curse of dimensionality.
3) Deep learning algorithms attempt to learn multiple levels of increasingly complex representations, with the goal of developing more abstract, disentangled representations that generalize beyond local patterns in the data.
This document introduces a method called "co-curricular learning" that dynamically combines clean-data selection and domain-data selection for neural machine translation. It applies an EM-style optimization procedure to refine the "co-curriculum." Experimental results on two domains demonstrate the effectiveness of the method and properties of the data scheduled by the co-curriculum.
(1) Learning visual representations for unfamiliar environments is challenging due to domain shift between training and test data distributions. (2) The paper proposes learning asymmetric transformations to map target domain data to the source domain in order to address this domain shift problem. (3) The key aspects of the approach include learning nonlinear kernel-based transformations between domains in a regularized manner and evaluating its ability to generalize to novel target classes not seen during training.
This document discusses approaches and methods for text classification. It outlines rule-based classification, statistical machine learning approaches like decision trees, k-nearest neighbors, naive Bayes, hidden Markov models, and support vector machines. It also discusses recent deep learning methods like convolutional neural networks, recurrent neural networks, bidirectional LSTMs, hierarchical attention networks, and more for text classification without feature engineering. The document provides examples of how each method has been applied and highlights their strengths and limitations.
Dixon developed a theory of five knowledge transfer mechanisms based on who receives the knowledge, the nature and type of task, and type of knowledge transferred. The mechanisms include serial transfer of tacit knowledge between the same teams, near transfer of explicit knowledge for similar routine tasks, and far transfer which makes tacit knowledge from non-routine tasks available to other teams. These different transfer types are important for organizations to effectively share and apply knowledge across situations, teams, and projects.
This document provides an outline for a presentation on machine learning and deep learning. It begins with an introduction to machine learning basics and types of learning. It then discusses what deep learning is and why it is useful. The main components and hyperparameters of deep learning models are explained, including activation functions, optimizers, cost functions, regularization methods, and tuning. Basic deep neural network architectures like convolutional and recurrent networks are described. An example application of relation extraction is provided. The document concludes by listing additional deep learning topics.
AI&BigData Lab 2016. Александр Баев: Transfer learning - зачем, как и где.GeeksLab Odessa
4.6.16 AI&BigData Lab
Upcoming events: goo.gl/I2gJ4H
Поговорим об одной из базовых практических техник обучения нейронных сетей - предобучение, finetuning, transfer learning. В каких случаях применять, какие модели использовать, где их брать и как адаптировать.
PPT : [26 Feb]
Source target transformation
Brain segmentation
Existing models
What are disadvantages and disadvantages in the adversial stretching the networks
What changes do we need to perform
Resources:
https://www.v7labs.com/blog/domain-adaptation-guide
Some Standard Datasets:
Amazon, Webcam, and DSLR are has three distinct domains
ML technique that focuses on training models on a source domain and then adapting them to perform well on a target domain, where the source and target domains may have different distributions. Real world datasets often suffer from domain shift, where data distributions differ due to factors like sensor changes, environment variations, or user demographics.
Types of Domain Adaptation:
Supervised Domain Adaptation(SDA) : The model has labels in both the source and target domain
Unsupervised Domain Adaptation(UDA) : Only labeled data from the source domain is available during training.
Semi Supervised Domain Adaptation(SSDA) : A combination of labeled source domain data and limited labeled target domain data are used during training.
Multi-Source Domain Adaptation: Extends domain adaptation to multiple source domains. The model adapts to the target domain using knowledge from multiple related sources.
Adaptation Techniques:
Feature level adaptation : Transforming feature from both domains to a common space, reducing domain discrepancy.
Instance level adaptation :
Model level adaptation:
This adaptation can involve modifying the architecture, parameters, or other components of the model to reduce the impact of domain shift.
Classifier adaptation:
Generative models:
In the context of domain adaptation, generative models can be used to generate synthetic data in the target domain that is similar to the actual target domain data. This synthetic data can be combined with the source domain data during training to improve the model's ability to generalize to the target domain.
Transfer learning:
Domain adaptation is often a part of transfer learning, where knowledge gained from one task or domain is applied to another.
Applications:
Object Recognition, nlp, sentimental analysis
Challenges: dealing with the heterogeneity between domains, selecting suitable adaptation techniques, addressing the limited availability of labeled data in the target domain are some common challenges.
How it works:
Training a model on the source domain.
Extracting knowledge about the data distribution.
Adapting the model to the target domain by aligning features, modifying the classifier, or learning a shared representation.
Fine-tuning the model on the target domain (optional, depending on the technique).
Disadvantages of Adversarial Domain Adaptation:
1. Computational cost: Training two competing neural networks simultaneously can be expensive.
Self-supervised learning: Leverages unlabeled data from the target domain to learn generalizable representations without relying solely on adversarial training.
Main single agent machine learning algorithmsbutest
This document summarizes several machine learning algorithms and their potential applications to multi-agent systems. It describes algorithms such as decision trees, neural networks, Bayesian methods, reinforcement learning, inductive logic programming, case-based reasoning, support vector machines, and genetic algorithms. For each algorithm, it provides a brief description and discusses any existing or potential work applying the algorithm to multi-agent domains.
Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distrib...MLAI2
While tasks could come with varying the number of instances and classes in realistic settings, the existing meta-learning approaches for few-shot classification assume that number of instances per task and class is fixed. Due to such restriction, they learn to equally utilize the meta-knowledge across all the tasks, even when the number of instances per task and class largely varies. Moreover, they do not consider distributional difference in unseen tasks, on which the meta-knowledge may have less usefulness depending on the task relatedness. To overcome these limitations, we propose a novel meta-learning model that adaptively balances the effect of the meta-learning and task-specific learning within each task. Through the learning of the balancing variables, we can decide whether to obtain a solution by relying on the meta-knowledge or task-specific learning. We formulate this objective into a Bayesian inference framework and tackle it using variational inference. We validate our Bayesian Task-Adaptive Meta-Learning (Bayesian TAML) on two realistic task- and class-imbalanced datasets, on which it significantly outperforms existing meta-learning approaches. Further ablation study confirms the effectiveness of each balancing component and the Bayesian learning framework.
This document presents a lab seminar on semi-supervised learning. It begins with background on semi-supervised learning and examples of applications. It then discusses common semi-supervised learning methods like EM with generative models, co-training, transductive SVMs, and graph-based methods. Next, it covers assumptions of semi-supervised learning, noting the utility of unlabeled data depends on problem structure matching model assumptions. Finally, it proposes future work on multi-edge graph-based semi-supervised learning.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
A Comprehensive Guide to DeFi Development Services in 2024Intelisync
DeFi represents a paradigm shift in the financial industry. Instead of relying on traditional, centralized institutions like banks, DeFi leverages blockchain technology to create a decentralized network of financial services. This means that financial transactions can occur directly between parties, without intermediaries, using smart contracts on platforms like Ethereum.
In 2024, we are witnessing an explosion of new DeFi projects and protocols, each pushing the boundaries of what’s possible in finance.
In summary, DeFi in 2024 is not just a trend; it’s a revolution that democratizes finance, enhances security and transparency, and fosters continuous innovation. As we proceed through this presentation, we'll explore the various components and services of DeFi in detail, shedding light on how they are transforming the financial landscape.
At Intelisync, we specialize in providing comprehensive DeFi development services tailored to meet the unique needs of our clients. From smart contract development to dApp creation and security audits, we ensure that your DeFi project is built with innovation, security, and scalability in mind. Trust Intelisync to guide you through the intricate landscape of decentralized finance and unlock the full potential of blockchain technology.
Ready to take your DeFi project to the next level? Partner with Intelisync for expert DeFi development services today!
5th LF Energy Power Grid Model Meet-up SlidesDanBrown980551
5th Power Grid Model Meet-up
It is with great pleasure that we extend to you an invitation to the 5th Power Grid Model Meet-up, scheduled for 6th June 2024. This event will adopt a hybrid format, allowing participants to join us either through an online Mircosoft Teams session or in person at TU/e located at Den Dolech 2, Eindhoven, Netherlands. The meet-up will be hosted by Eindhoven University of Technology (TU/e), a research university specializing in engineering science & technology.
Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
What to expect
For the upcoming meetup we are organizing, we have an exciting lineup of activities planned:
-Insightful presentations covering two practical applications of the Power Grid Model.
-An update on the latest advancements in Power Grid -Model technology during the first and second quarters of 2024.
-An interactive brainstorming session to discuss and propose new feature requests.
-An opportunity to connect with fellow Power Grid Model enthusiasts and users.
Digital Banking in the Cloud: How Citizens Bank Unlocked Their MainframePrecisely
Inconsistent user experience and siloed data, high costs, and changing customer expectations – Citizens Bank was experiencing these challenges while it was attempting to deliver a superior digital banking experience for its clients. Its core banking applications run on the mainframe and Citizens was using legacy utilities to get the critical mainframe data to feed customer-facing channels, like call centers, web, and mobile. Ultimately, this led to higher operating costs (MIPS), delayed response times, and longer time to market.
Ever-changing customer expectations demand more modern digital experiences, and the bank needed to find a solution that could provide real-time data to its customer channels with low latency and operating costs. Join this session to learn how Citizens is leveraging Precisely to replicate mainframe data to its customer channels and deliver on their “modern digital bank” experiences.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Freshworks Rethinks NoSQL for Rapid Scaling & Cost-EfficiencyScyllaDB
Freshworks creates AI-boosted business software that helps employees work more efficiently and effectively. Managing data across multiple RDBMS and NoSQL databases was already a challenge at their current scale. To prepare for 10X growth, they knew it was time to rethink their database strategy. Learn how they architected a solution that would simplify scaling while keeping costs under control.
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Trusted Execution Environment for Decentralized Process MiningLucaBarbaro3
Presentation of the paper "Trusted Execution Environment for Decentralized Process Mining" given during the CAiSE 2024 Conference in Cyprus on June 7, 2024.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/temporal-event-neural-networks-a-more-efficient-alternative-to-the-transformer-a-presentation-from-brainchip/
Chris Jones, Director of Product Management at BrainChip , presents the “Temporal Event Neural Networks: A More Efficient Alternative to the Transformer” tutorial at the May 2024 Embedded Vision Summit.
The expansion of AI services necessitates enhanced computational capabilities on edge devices. Temporal Event Neural Networks (TENNs), developed by BrainChip, represent a novel and highly efficient state-space network. TENNs demonstrate exceptional proficiency in handling multi-dimensional streaming data, facilitating advancements in object detection, action recognition, speech enhancement and language model/sequence generation. Through the utilization of polynomial-based continuous convolutions, TENNs streamline models, expedite training processes and significantly diminish memory requirements, achieving notable reductions of up to 50x in parameters and 5,000x in energy consumption compared to prevailing methodologies like transformers.
Integration with BrainChip’s Akida neuromorphic hardware IP further enhances TENNs’ capabilities, enabling the realization of highly capable, portable and passively cooled edge devices. This presentation delves into the technical innovations underlying TENNs, presents real-world benchmarks, and elucidates how this cutting-edge approach is positioned to revolutionize edge AI across diverse applications.
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
Skybuffer AI: Advanced Conversational and Generative AI Solution on SAP Busin...Tatiana Kojar
Skybuffer AI, built on the robust SAP Business Technology Platform (SAP BTP), is the latest and most advanced version of our AI development, reaffirming our commitment to delivering top-tier AI solutions. Skybuffer AI harnesses all the innovative capabilities of the SAP BTP in the AI domain, from Conversational AI to cutting-edge Generative AI and Retrieval-Augmented Generation (RAG). It also helps SAP customers safeguard their investments into SAP Conversational AI and ensure a seamless, one-click transition to SAP Business AI.
With Skybuffer AI, various AI models can be integrated into a single communication channel such as Microsoft Teams. This integration empowers business users with insights drawn from SAP backend systems, enterprise documents, and the expansive knowledge of Generative AI. And the best part of it is that it is all managed through our intuitive no-code Action Server interface, requiring no extensive coding knowledge and making the advanced AI accessible to more users.
1. A Survey on
Transfer Learning
Sinno Jialin Pan
Department of Computer Science and Engineering
The Hong Kong University of Science and Technology
Joint work with Prof. Qiang Yang
2. Transfer Learning? (DARPA 05)
Transfer Learning (TL):
The ability of a system to recognize and apply knowledge and
skills learned in previous tasks to novel tasks (in new domains)
It is motivated by human learning. People can often transfer knowledge
learnt previously to novel situations
Chess Checkers
Mathematics Computer Science
Table Tennis Tennis
3. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
4. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
5. Traditional ML vs. TL
(P. Langley 06)
Traditional ML in Transfer of learning
multiple domains across domains
training items
training items
test items
test items
Humans can learn in many domains. Humans can also transfer from one
domain to other domains.
6. Traditional ML vs. TL
Learning Process of Learning Process of
Traditional ML Transfer Learning
training items training items
Learning System Learning System Learning System
Knowledge Learning System
7. Notation
Domain:
It consists of two components: A feature space , a marginal distribution
In general, if two domains are different, then they may have different feature
spaces
or different marginal distributions.
Task:
Given a specific domain and label space , for each in the domain, to
predict its corresponding label
In general, if two tasks are different, then they may have different label spaces or
different conditional distributions
8. Notation
For simplicity, we only consider at most two domains and two tasks.
Source domain:
Task in the source domain:
Target domain:
Task in the target domain
9. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
10. Why Transfer Learning?
In some domains, labeled data are in short supply.
In some domains, the calibration effort is very expensive.
In some domains, the learning process is time consuming.
How to extract knowledge learnt from related domains to help
learning in a target domain with a few labeled data?
How to extract knowledge learnt from related domains to speed up
learning in a target domain?
Transfer learning techniques may help!
11. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
12. Settings of Transfer Learning
Transfer learning settings Labeled data in Labeled data in Tasks
a source domain a target domain
Inductive Transfer Learning Classification
× √
Regression
√ √ …
Transductive Transfer Learning Classification
√ ×
Regression
…
Unsupervised Transfer Learning Clustering
× ×
…
13. An overview of
various settings of Self-taught
Case 1
transfer learning Learning
No labeled data in a source domain
Inductive
Transfer Learning
Labeled data are available in a source domain
Labeled data are available in
Source and
a target domain
target tasks are
Multi-task
Case 2 learnt Learning
simultaneously
Transfer
Learning
Labeled data are
available only in a Assumption:
source domain Transductive different Domain
domains but
Transfer Learning single task Adaptation
No labeled data in
both source and
target domain Assumption: single domain
and single task
Unsupervised Sample Selection Bias /
Transfer Learning Covariance Shift
14. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
15. Approaches to Transfer Learning
Transfer learning approaches Description
Instance-transfer To re-weight some labeled data in a source
domain for use in the target domain
Feature-representation-transfer Find a “good” feature representation that reduces
difference between a source and a target domain
or minimizes error of models
Model-transfer Discover shared parameters or priors of models
between a source domain and a target domain
Relational-knowledge-transfer Build mapping of relational knowledge between
a source domain and a target domain.
16. Approaches to Transfer Learning
Inductive Transductive Unsupervised
Transfer Learning Transfer Learning Transfer Learning
Instance-transfer √ √
Feature-representation- √ √ √
transfer
Model-transfer √
Relational-knowledge- √
transfer
17. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Inductive Transfer Learning
Transductive Transfer Learning
Unsupervised Transfer Learning
18. Inductive Transfer Learning
Instance-transfer Approaches
• Assumption: the source domain and target domain data use
exactly the same features and labels.
• Motivation: Although the source domain data can not be
reused directly, there are some parts of the data that can still
be reused by re-weighting.
• Main Idea: Discriminatively adjust weighs of data in the
source domain for use in the target domain.
19. Inductive Transfer Learning
--- Instance-transfer Approaches
Non-standard SVMs
[Wu and Dietterich ICML-04]
Uniform weights Correct the decision boundary by re-weighting
Loss function on the Loss function on the
target domain data source domain data Regularization term
Differentiate the cost for misclassification of the target and source data
20. Inductive Transfer Learning
--- Instance-transfer Approaches
TrAdaBoost
[Dai et al. ICML-07]
Hedge ( β ) AdaBoost
[Freund et al. 1997] [Freund et al. 1997]
To decrease the weights To increase the weights of
of the misclassified data the misclassified data
The whole
Source domain training data set
target domain
labeled data labeled data
Classifiers trained on
re-weighted labeled data
Target domain
unlabeled data
21. Inductive Transfer Learning
Feature-representation-transfer Approaches
Supervised Feature Construction
[Argyriou et al. NIPS-06, NIPS-07]
Assumption: If t tasks are related to each other, then they may
share some common features which can benefit for all tasks.
Input: t tasks, each of them has its own training data.
Output: Common features learnt across t tasks and t models for t
tasks, respectively.
22. Supervised Feature Construction
[Argyriou et al. NIPS-06, NIPS-07]
Average of the empirical
error across t tasks Regularization to make the
representation sparse
Orthogonal Constraints
where
23. Inductive Transfer Learning
Feature-representation-transfer Approaches
Unsupervised Feature Construction
[Raina et al. ICML-07]
Three steps:
• Applying sparse coding [Lee et al. NIPS-07] algorithm to learn
higher-level representation from unlabeled data in the source
domain.
• Transforming the target data to new representations by new bases
learnt in the first step.
• Traditional discriminative models can be applied on new
representations of the target data with corresponding labels.
24. Unsupervised Feature Construction
[Raina et al. ICML-07]
Step1:
Input: Source domain data X S = {xS } and coefficient β
i
Output: New representations of the source domain data AS = {aS } i
and new bases B = {bi }
Step2:
Input: Target domain data X T = {xT } , coefficient β and bases B = {bi }
i
Output: New representations of the target domain data AT = {aT } i
25. Inductive Transfer Learning
Model-transfer Approaches
Regularization-based Method
[Evgeiou and Pontil, KDD-04]
Assumption: If t tasks are related to each other, then they may share some
parameters among individual models.
Assume f t = wt ⋅ x be a hyper-plane for task , where t ∈ {T , S } and
Common part Specific part for individual task
Regularization terms
for multiple tasks
Encode them into SVMs:
26. Inductive Transfer Learning
Relational-knowledge-transfer Approaches
TAMAR
[Mihalkova et al. AAAI-07]
Assumption: If the target domain and source domain are related, then there
may be some relationship between domains being similar, which can be used for
transfer learning
Input:
6. Relational data in the source domain and a statistical relational model,
Markov Logic Network (MLN), which has been learnt in the source
domain.
7. Relational data in the target domain.
Output: A new statistical relational model, MLN, in the target domain.
Goal: To learn a MLN in the target domain more efficiently and effectively.
27. TAMAR [Mihalkova et al. AAAI-07]
Two Stages:
2. Predicate Mapping
– Establish the mapping between predicates in the source
and target domain. Once a mapping is established, clauses
from the source domain can be translated into the target
domain.
3. Revising the Mapped Structure
– The clauses mapping from the source domain directly
may not be completely accurate and may need to be
revised, augmented , and re-weighted in order to properly
model the target data.
29. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Inductive Transfer Learning
Transductive Transfer Learning
Unsupervised Transfer Learning
30. Transductive Transfer Learning
Instance-transfer Approaches
Sample Selection Bias / Covariance Shift
[Zadrozny ICML-04, Schwaighofer JSPI-00]
Input: A lot of labeled data in the source domain and no labeled data in the
target domain.
Output: Models for use in the target domain data.
Assumption: The source domain and target domain are the same. In addition,
P (YS | X S ) P (YT | X T ) P( X S ) P( X T )
and are the same while and may be
different causing by different sampling process (training data and test data).
Main Idea: Re-weighting (important sampling) the source domain data.
31. Sample Selection Bias/Covariance Shift
To correct sample selection bias:
weights for source
domain data
How to estimate ?
One straightforward solution is to estimate P( X S ) and P( X T ) ,
respectively. However, estimating density function is a hard
problem.
32. Sample Selection Bias/Covariance Shift
Kernel Mean Match (KMM)
[Huang et al. NIPS 2006]
Main Idea: KMM tries to estimate directly instead of estimating
density function.
It can be proved that can be estimated by solving the following quadratic
programming (QP) optimization problem.
To match means between
training and test data in a RKHS
Theoretical Support: Maximum Mean Discrepancy (MMD) [Borgwardt et al.
BIOINFOMATICS-06]. The distance of distributions can be measured
by Euclid distance of their mean vectors in a RKHS.
33. Transductive Transfer Learning
Feature-representation-transfer Approaches
Domain Adaptation
[Blitzer et al. EMNL-06, Ben-David et al. NIPS-07, Daume III ACL-07]
Assumption: Single task across domains, which means P(YS | X S ) and P(YT | X T )
are the same while P ( X S ) and P ( X T ) may be different causing by feature
representations across domains.
Main Idea: Find a “good” feature representation that reduce the “distance”
between domains.
Input: A lot of labeled data in the source domain and only unlabeled data in the
target domain.
Output: A common representation between source domain data and target
domain data and a model on the new representation for use in the target
domain.
34. Domain Adaptation
Structural Correspondence Learning (SCL)
[Blitzer et al. EMNL-06, Blitzer et al. ACL-07, Ando and Zhang JMLR-05]
Motivation: If two domains are related to each other, then there may exist
some “pivot” features across both domain. Pivot features are features that
behave in the same way for discriminative learning in both domains.
Main Idea: To identify correspondences among features from different
domains by modeling their correlations with pivot features. Non-pivot features
form different domains that are correlated with many of the same pivot
features are assumed to correspond, and they are treated similarly in a
discriminative learner.
35. SCL
[Blitzer et al. EMNL-06, Blitzer et al. ACL-07, Ando and Zhang
JMLR-05]
a) Heuristically choose m pivot
features, which is task specific.
b) Transform each vector of pivot
feature to a vector of binary
values and then create
corresponding prediction
problem.
Learn parameters of each
prediction problem
Do Eigen Decomposition
on the matrix of
parameters and learn the
linear mapping function.
Use the learnt mapping function to
construct new features and train
classifiers onto the new representations.
36. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Inductive Transfer Learning
Transductive Transfer Learning
Unsupervised Transfer Learning
37. Unsupervised Transfer Learning
Feature-representation-transfer Approaches
Self-taught Clustering (STC)
[Dai et al. ICML-08]
Input: A lot of unlabeled data in a source domain and a few unlabeled data in a
target domain.
Goal: Clustering the target domain data.
Assumption: The source domain and target domain data share some common
features, which can help clustering in the target domain.
Main Idea: To extend the information theoretic co-clustering algorithm
[Dhillon et al. KDD-03] for transfer learning.
38. Self-taught Clustering (STC)
[Dai et al. ICML-08]
Common features
Target domain data
Source domain data
Co-clustering in the
source domain
Objective function that need to be minimized
Co-clustering in the target domain
where
Cluster functions
Output
39. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
40. Negative Transfer
Most approaches to transfer learning assume transferring knowledge across
domains be always positive.
However, in some cases, when two tasks are too dissimilar, brute-force
transfer may even hurt the performance of the target task, which is called
negative transfer [Rosenstein et al NIPS-05 Workshop].
Some researchers have studied how to measure relatedness among tasks
[Ben-David and Schuller NIPS-03, Bakker and Heskes JMLR-03].
How to design a mechanism to avoid negative transfer needs to be studied
theoretically.
41. Outline
Traditional Machine Learning vs. Transfer Learning
Why Transfer Learning?
Settings of Transfer Learning
Approaches to Transfer Learning
Negative Transfer
Conclusion
42. Conclusion
Inductive Transductive Unsupervised
Transfer Learning Transfer Learning Transfer Learning
Instance-transfer √ √
Feature-representation- √ √ √
transfer
Model-transfer √
Relational-knowledge- √
transfer
How to avoid negative transfer need to be attracted more attention!
Editor's Notes
Seems the changes is involved in moving to new tasks is more radical than moving to new domains.
Generality means human can learn to perform a variety of tasks.
Generality means human can learn to perform a variety of tasks.