Deep learning systems are susceptible to adversarial manipulation through techniques like generating adversarial samples and substitute models. By making small, targeted perturbations to inputs, an attacker can cause misclassifications or reduce a model's confidence without affecting human perception of the inputs. This is possible due to blind spots in how models learn representations that are different from human concepts. Defending against such attacks requires training models with adversarial techniques to make them more robust.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Huang Xiao presented on causative adversarial learning techniques. The presentation discussed:
1) Two types of adversarial attacks - causative attacks which poison training data to change a model's discriminant function, and exploratory attacks which manipulate test points to circumvent detection.
2) Examples of causative attacks including flipping labels on support vector machines and adding malicious points to maximize test error. Gradient ascent methods were used to optimize attack points.
3) Feature selection algorithms like Lasso were also shown to be vulnerable to poisoning attacks by contaminating training data to mislead feature selection. Gradient methods optimized attacks to maximize generalization error.
4) The talk concluded machine learning systems need to be more robust and
Deep learning is a class of machine learning algorithms that uses multiple layers of nonlinear processing units for feature extraction and transformation. It can be used for supervised learning tasks like classification and regression or unsupervised learning tasks like clustering. Deep learning models include deep neural networks, deep belief networks, and convolutional neural networks. Deep learning has been applied successfully in domains like computer vision, speech recognition, and natural language processing by companies like Google, Facebook, Microsoft, and others.
Generating Natural-Language Text with Neural NetworksJonathan Mugan
Automatic text generation enables computers to summarize text, to have conversations in customer-service and other settings, and to customize content based on the characteristics and goals of the human interlocutor. Using neural networks to automatically generate text is appealing because they can be trained through examples with no need to manually specify what should be said when. In this talk, we will provide an overview of the existing algorithms used in neural text generation, such as sequence2sequence models, reinforcement learning, variational methods, and generative adversarial networks. We will also discuss existing work that specifies how the content of generated text can be determined by manipulating a latent code. The talk will conclude with a discussion of current challenges and shortcomings of neural text generation.
Deep Learning: concepts and use cases (October 2018)Julien SIMON
An introduction to Deep Learning theory
Neurons & Neural Networks
The Training Process
Backpropagation
Optimizers
Common network architectures and use cases
Convolutional Neural Networks
Recurrent Neural Networks
Long Short Term Memory Networks
Generative Adversarial Networks
Getting started
Deep learning systems are susceptible to adversarial manipulation through techniques like generating adversarial samples and substitute models. By making small, targeted perturbations to inputs, an attacker can cause misclassifications or reduce a model's confidence without affecting human perception of the inputs. This is possible due to blind spots in how models learn representations that are different from human concepts. Defending against such attacks requires training models with adversarial techniques to make them more robust.
Artificial Intelligence, Machine Learning and Deep LearningSujit Pal
Slides for talk Abhishek Sharma and I gave at the Gennovation tech talks (https://gennovationtalks.com/) at Genesis. The talk was part of outreach for the Deep Learning Enthusiasts meetup group at San Francisco. My part of the talk is covered from slides 19-34.
Huang Xiao presented on causative adversarial learning techniques. The presentation discussed:
1) Two types of adversarial attacks - causative attacks which poison training data to change a model's discriminant function, and exploratory attacks which manipulate test points to circumvent detection.
2) Examples of causative attacks including flipping labels on support vector machines and adding malicious points to maximize test error. Gradient ascent methods were used to optimize attack points.
3) Feature selection algorithms like Lasso were also shown to be vulnerable to poisoning attacks by contaminating training data to mislead feature selection. Gradient methods optimized attacks to maximize generalization error.
4) The talk concluded machine learning systems need to be more robust and
Deep learning is a class of machine learning algorithms that uses multiple layers of nonlinear processing units for feature extraction and transformation. It can be used for supervised learning tasks like classification and regression or unsupervised learning tasks like clustering. Deep learning models include deep neural networks, deep belief networks, and convolutional neural networks. Deep learning has been applied successfully in domains like computer vision, speech recognition, and natural language processing by companies like Google, Facebook, Microsoft, and others.
Generating Natural-Language Text with Neural NetworksJonathan Mugan
Automatic text generation enables computers to summarize text, to have conversations in customer-service and other settings, and to customize content based on the characteristics and goals of the human interlocutor. Using neural networks to automatically generate text is appealing because they can be trained through examples with no need to manually specify what should be said when. In this talk, we will provide an overview of the existing algorithms used in neural text generation, such as sequence2sequence models, reinforcement learning, variational methods, and generative adversarial networks. We will also discuss existing work that specifies how the content of generated text can be determined by manipulating a latent code. The talk will conclude with a discussion of current challenges and shortcomings of neural text generation.
Deep Learning: concepts and use cases (October 2018)Julien SIMON
An introduction to Deep Learning theory
Neurons & Neural Networks
The Training Process
Backpropagation
Optimizers
Common network architectures and use cases
Convolutional Neural Networks
Recurrent Neural Networks
Long Short Term Memory Networks
Generative Adversarial Networks
Getting started
This document summarizes an adversarial examples presentation. It discusses how adversarial examples are samples modified to cause misclassification, gradient descent optimization techniques, neural network training methods, and black-box and white-box adversarial attack methods like Fast Gradient Sign Method. It also covers adversarial example defenses, uses of adversarial examples in research, and targeted perturbation algorithms.
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...GeekPwn Keen
Youtube: https://www.youtube.com/watch?v=ttmm2yo74z8
Moustafa Alzantot is a Ph.D. Candidate in Computer Science at UCLA. His research interests include machine learning, privacy, and mobile computing. He is an inventor of two US patents and the recipient of several awards including the COMESA 2014 innovation award. He worked as an intern at Google, Facebook, and Qualcom.
Yash Sharma is a visiting scientist at Cornell who recently graduated with a Bachelors and Masters in Electrical Engineering. His research has focused on adversarial examples, namely pushing the state-of-the-art in attacks in both limited access settings and challenging domains. He is interested in finding more principled solutions for resolving the robustness problem, as well as studying other practical issues which are inhibiting us from achieving AGI.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Deep Learning For Practitioners, lecture 2: Selecting the right applications...ananth
In this presentation we articulate when deep learning techniques yield best results from a practitioner's view point. Do we apply deep learning techniques for every machine learning problem? What characteristics of an application lends itself suitable for deep learning? Does more data automatically imply better results regardless of the algorithm or model? Does "automated feature learning" obviate the need for data preprocessing and feature design?
Research of adversarial example on a deep neural networkNAVER Engineering
최근 컴퓨터 성능이 발달되고 대량의 데이터 수집이 가능하게 되면서, 인공지능 기술 중에 딥뉴럴네트워크 (Deep Neural Network, DNN)을 이용한 인공지능 기술이 각광받고 있다.
특히, 딥뉴럴네트워크은 이미지 인식, 음성 인식, 패턴 분석 등 분야에 있어서 탁월한 성능을 보여주고 있다. 하지만 딥뉴럴네트워크의 보안문제 중 Adversarial example이 주목 받고 있다.
Adversarial example은 입력 데이터에 최소한의 데이터를 변조를 하여 딥뉴럴네트워크가 원래 class가 아닌 다른 class로 잘못 인식하게 만드는 공격이다.
따라서 Adversarial example은 딥뉴럴네트워크의 보안문제에 위협이 된다. 이번 발표에서는 Adversarial example에 대한 전체적인 내용과 발표자가 제안한 방법인 Friend-safe evasion attack 등에 대해서 소개하고자 한다.
An introduction to Machine Learning (and a little bit of Deep Learning)Thomas da Silva Paula
25-min talk about Machine Learning and a little bit of Deep Learning. Starts with some basic definitions (Supervised and Unsupervised Learning). Then, neural networks basic functionality is explained, ending up in Deep Learning and Convolutional Neural Networks.
Machine Learning Meetup that happened in Porto Alegre, Brazil.
“Automatically learning multiple levels of representations of the underlying distribution of the data to be modelled”
Deep learning algorithms have shown superior learning and classification performance.
In areas such as transfer learning, speech and handwritten character recognition, face recognition among others.
(I have referred many articles and experimental results provided by Stanford University)
This document provides an introduction to deep learning. It defines artificial intelligence, machine learning, data science, and deep learning. Machine learning is a subfield of AI that gives machines the ability to improve performance over time without explicit human intervention. Deep learning is a subfield of machine learning that builds artificial neural networks using multiple hidden layers, like the human brain. Popular deep learning techniques include convolutional neural networks, recurrent neural networks, and autoencoders. The document discusses key components and hyperparameters of deep learning models.
This document provides an introduction to deep learning. It begins by discussing modeling human intelligence with machines and the history of neural networks. It then covers concepts like supervised learning, loss functions, and gradient descent. Deep learning frameworks like Theano, Caffe, Keras, and Torch are also introduced. The document provides examples of deep learning applications and discusses challenges for the future of the field like understanding videos and text. Code snippets demonstrate basic network architecture.
This document provides an introduction to machine learning. It discusses how machine learning allows computers to learn from experience to improve their performance on tasks. Supervised learning is described, where the goal is to learn a function that maps inputs to outputs from a labeled dataset. Cross-validation techniques like the test set method, leave-one-out cross-validation, and k-fold cross-validation are introduced to evaluate model performance without overfitting. Applications of machine learning like medical diagnosis, recommendation systems, and autonomous driving are briefly outlined.
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)Márton Miháltz
A brief survey of current deep learning/neural network methods currently used in NLP: recurrent networks (LSTM, GRU), recursive networks, convolutional networks, hybrid architectures, attention models. We will look at specific papers in the literature, targeting sentiment analysis, text classification and other tasks.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Transfer learning aims to improve learning in a target domain by leveraging knowledge from a related source domain. It is useful when the target domain has limited labeled data. There are several approaches, including instance-based approaches that reweight or resample source instances, and feature-based approaches that learn a transformation to align features across domains. Spectral feature alignment is one technique that builds a graph of correlations between pivot features shared across domains and domain-specific features, then applies spectral clustering to derive new shared features.
Machine Learning: Generative and Discriminative Modelsbutest
The document discusses machine learning models, specifically generative and discriminative models. It provides examples of generative models like Naive Bayes classifiers and hidden Markov models. Discriminative models discussed include logistic regression and conditional random fields. The document contrasts how generative models estimate class-conditional probabilities while discriminative models directly estimate posterior probabilities. It also compares how hidden Markov models model sequential data generatively while conditional random fields model sequential data discriminatively.
A Multiscale Visualization of Attention in the Transformer Modeltaeseon ryu
안녕하세요 딥러닝 논문읽기 모임입니다 오늘 업로드된 논문 리뷰 영상은 2019 ACL 에서 발표된 A Multiscale Visualization of Attention in the Transformer Model 라는 제목의 논문입니다.
본 논문은 최고의 주가를 달리는 트랜스포머가 연산하는 과정을 다양한 관점에서 비주얼라이징을 할 수 있는 툴에 관련된 논문 입니다.
트랜스포머와 버트, GPT에 대한 간단한 소개와 더불어, 해당 비주얼라이징이 어떻게 활용 될 수 있는지에 대한 여러 유스케이스를 자연어 처리팀 백지윤님이 자세한 리뷰 도와 주셨습니다.
This is the first lecture on Applied Machine Learning. The course focuses on the emerging and modern aspects of this subject such as Deep Learning, Recurrent and Recursive Neural Networks (RNN), Long Short Term Memory (LSTM), Convolution Neural Networks (CNN), Hidden Markov Models (HMM). It deals with several application areas such as Natural Language Processing, Image Understanding etc. This presentation provides the landscape.
Transfer Learning and Fine-tuning Deep Neural NetworksPyData
This document outlines Anusua Trivedi's talk on transfer learning and fine-tuning deep neural networks. The talk covers traditional machine learning versus deep learning, using deep convolutional neural networks (DCNNs) for image analysis, transfer learning and fine-tuning DCNNs, recurrent neural networks (RNNs), and case studies applying these techniques to diabetic retinopathy prediction and fashion image caption generation.
This is a slide deck from a presentation, that my colleague Shirin Glander (https://www.slideshare.net/ShirinGlander/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, I just copied the two slide decks together. As I did the "surrounding" part, I added Shirin's part at the place when she took over and then added my concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
The document describes a scene understanding model that generates natural language descriptions of images. It discusses how humans understand scenes, then outlines the key components of the model: convolutional neural networks to extract image features, transfer learning from pre-trained models, and recurrent neural networks to generate captions. The presentation includes details on CNNs, LSTMs, training the model on Flickr 30k images and captions, and a demonstration of captions generated for sample images of varying complexity.
This is a slide deck from a presentation, that my colleague Uwe Friedrichsen (https://www.slideshare.net/ufried/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, Uwe copied the two slide decks together. As he did the "surrounding" part, he added my part at the place where I took over and then added concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
This document summarizes an adversarial examples presentation. It discusses how adversarial examples are samples modified to cause misclassification, gradient descent optimization techniques, neural network training methods, and black-box and white-box adversarial attack methods like Fast Gradient Sign Method. It also covers adversarial example defenses, uses of adversarial examples in research, and targeted perturbation algorithms.
PRACTICAL ADVERSARIAL ATTACKS AGAINST CHALLENGING MODELS ENVIRONMENTS - Moust...GeekPwn Keen
Youtube: https://www.youtube.com/watch?v=ttmm2yo74z8
Moustafa Alzantot is a Ph.D. Candidate in Computer Science at UCLA. His research interests include machine learning, privacy, and mobile computing. He is an inventor of two US patents and the recipient of several awards including the COMESA 2014 innovation award. He worked as an intern at Google, Facebook, and Qualcom.
Yash Sharma is a visiting scientist at Cornell who recently graduated with a Bachelors and Masters in Electrical Engineering. His research has focused on adversarial examples, namely pushing the state-of-the-art in attacks in both limited access settings and challenging domains. He is interested in finding more principled solutions for resolving the robustness problem, as well as studying other practical issues which are inhibiting us from achieving AGI.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Deep Learning For Practitioners, lecture 2: Selecting the right applications...ananth
In this presentation we articulate when deep learning techniques yield best results from a practitioner's view point. Do we apply deep learning techniques for every machine learning problem? What characteristics of an application lends itself suitable for deep learning? Does more data automatically imply better results regardless of the algorithm or model? Does "automated feature learning" obviate the need for data preprocessing and feature design?
Research of adversarial example on a deep neural networkNAVER Engineering
최근 컴퓨터 성능이 발달되고 대량의 데이터 수집이 가능하게 되면서, 인공지능 기술 중에 딥뉴럴네트워크 (Deep Neural Network, DNN)을 이용한 인공지능 기술이 각광받고 있다.
특히, 딥뉴럴네트워크은 이미지 인식, 음성 인식, 패턴 분석 등 분야에 있어서 탁월한 성능을 보여주고 있다. 하지만 딥뉴럴네트워크의 보안문제 중 Adversarial example이 주목 받고 있다.
Adversarial example은 입력 데이터에 최소한의 데이터를 변조를 하여 딥뉴럴네트워크가 원래 class가 아닌 다른 class로 잘못 인식하게 만드는 공격이다.
따라서 Adversarial example은 딥뉴럴네트워크의 보안문제에 위협이 된다. 이번 발표에서는 Adversarial example에 대한 전체적인 내용과 발표자가 제안한 방법인 Friend-safe evasion attack 등에 대해서 소개하고자 한다.
An introduction to Machine Learning (and a little bit of Deep Learning)Thomas da Silva Paula
25-min talk about Machine Learning and a little bit of Deep Learning. Starts with some basic definitions (Supervised and Unsupervised Learning). Then, neural networks basic functionality is explained, ending up in Deep Learning and Convolutional Neural Networks.
Machine Learning Meetup that happened in Porto Alegre, Brazil.
“Automatically learning multiple levels of representations of the underlying distribution of the data to be modelled”
Deep learning algorithms have shown superior learning and classification performance.
In areas such as transfer learning, speech and handwritten character recognition, face recognition among others.
(I have referred many articles and experimental results provided by Stanford University)
This document provides an introduction to deep learning. It defines artificial intelligence, machine learning, data science, and deep learning. Machine learning is a subfield of AI that gives machines the ability to improve performance over time without explicit human intervention. Deep learning is a subfield of machine learning that builds artificial neural networks using multiple hidden layers, like the human brain. Popular deep learning techniques include convolutional neural networks, recurrent neural networks, and autoencoders. The document discusses key components and hyperparameters of deep learning models.
This document provides an introduction to deep learning. It begins by discussing modeling human intelligence with machines and the history of neural networks. It then covers concepts like supervised learning, loss functions, and gradient descent. Deep learning frameworks like Theano, Caffe, Keras, and Torch are also introduced. The document provides examples of deep learning applications and discusses challenges for the future of the field like understanding videos and text. Code snippets demonstrate basic network architecture.
This document provides an introduction to machine learning. It discusses how machine learning allows computers to learn from experience to improve their performance on tasks. Supervised learning is described, where the goal is to learn a function that maps inputs to outputs from a labeled dataset. Cross-validation techniques like the test set method, leave-one-out cross-validation, and k-fold cross-validation are introduced to evaluate model performance without overfitting. Applications of machine learning like medical diagnosis, recommendation systems, and autonomous driving are briefly outlined.
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)Márton Miháltz
A brief survey of current deep learning/neural network methods currently used in NLP: recurrent networks (LSTM, GRU), recursive networks, convolutional networks, hybrid architectures, attention models. We will look at specific papers in the literature, targeting sentiment analysis, text classification and other tasks.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Transfer learning aims to improve learning in a target domain by leveraging knowledge from a related source domain. It is useful when the target domain has limited labeled data. There are several approaches, including instance-based approaches that reweight or resample source instances, and feature-based approaches that learn a transformation to align features across domains. Spectral feature alignment is one technique that builds a graph of correlations between pivot features shared across domains and domain-specific features, then applies spectral clustering to derive new shared features.
Machine Learning: Generative and Discriminative Modelsbutest
The document discusses machine learning models, specifically generative and discriminative models. It provides examples of generative models like Naive Bayes classifiers and hidden Markov models. Discriminative models discussed include logistic regression and conditional random fields. The document contrasts how generative models estimate class-conditional probabilities while discriminative models directly estimate posterior probabilities. It also compares how hidden Markov models model sequential data generatively while conditional random fields model sequential data discriminatively.
A Multiscale Visualization of Attention in the Transformer Modeltaeseon ryu
안녕하세요 딥러닝 논문읽기 모임입니다 오늘 업로드된 논문 리뷰 영상은 2019 ACL 에서 발표된 A Multiscale Visualization of Attention in the Transformer Model 라는 제목의 논문입니다.
본 논문은 최고의 주가를 달리는 트랜스포머가 연산하는 과정을 다양한 관점에서 비주얼라이징을 할 수 있는 툴에 관련된 논문 입니다.
트랜스포머와 버트, GPT에 대한 간단한 소개와 더불어, 해당 비주얼라이징이 어떻게 활용 될 수 있는지에 대한 여러 유스케이스를 자연어 처리팀 백지윤님이 자세한 리뷰 도와 주셨습니다.
This is the first lecture on Applied Machine Learning. The course focuses on the emerging and modern aspects of this subject such as Deep Learning, Recurrent and Recursive Neural Networks (RNN), Long Short Term Memory (LSTM), Convolution Neural Networks (CNN), Hidden Markov Models (HMM). It deals with several application areas such as Natural Language Processing, Image Understanding etc. This presentation provides the landscape.
Transfer Learning and Fine-tuning Deep Neural NetworksPyData
This document outlines Anusua Trivedi's talk on transfer learning and fine-tuning deep neural networks. The talk covers traditional machine learning versus deep learning, using deep convolutional neural networks (DCNNs) for image analysis, transfer learning and fine-tuning DCNNs, recurrent neural networks (RNNs), and case studies applying these techniques to diabetic retinopathy prediction and fashion image caption generation.
This is a slide deck from a presentation, that my colleague Shirin Glander (https://www.slideshare.net/ShirinGlander/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, I just copied the two slide decks together. As I did the "surrounding" part, I added Shirin's part at the place when she took over and then added my concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
The document describes a scene understanding model that generates natural language descriptions of images. It discusses how humans understand scenes, then outlines the key components of the model: convolutional neural networks to extract image features, transfer learning from pre-trained models, and recurrent neural networks to generate captions. The presentation includes details on CNNs, LSTMs, training the model on Flickr 30k images and captions, and a demonstration of captions generated for sample images of varying complexity.
This is a slide deck from a presentation, that my colleague Uwe Friedrichsen (https://www.slideshare.net/ufried/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, Uwe copied the two slide decks together. As he did the "surrounding" part, he added my part at the place where I took over and then added concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
This talk was presented in Startup Master Class 2017 - http://aaiitkblr.org/smc/ 2017 @ Christ College Bangalore. Hosted by IIT Kanpur Alumni Association and co-presented by IIT KGP Alumni Association, IITACB, PanIIT, IIMA and IIMB alumni.
My co-presenter was Biswa Gourav Singh. And contributor was Navin Manaswi.
http://dataconomy.com/2017/04/history-neural-networks/ - timeline for neural networks
This document provides an introduction to deep learning. It begins with an overview of artificial intelligence techniques like computer vision, speech processing, and natural language processing that benefit from deep learning. It then reviews the history of deep learning algorithms from perceptrons to modern deep neural networks. The core concepts of deep learning processes, neural network architectures, and training techniques like backpropagation are explained. Popular deep learning frameworks like TensorFlow, Keras, and PyTorch are also introduced. Finally, examples of convolutional neural networks, recurrent neural networks, and generative adversarial networks are briefly described along with tips for training deep neural networks and resources for further learning.
This document provides an introduction to deep learning. It discusses key concepts such as neural network layers that process input tensors, common layer types like convolutional and recurrent layers, and how networks are trained using stochastic gradient descent. Examples of deep learning applications that have achieved near-human level performance are also presented, such as image classification and speech recognition. The document then focuses on convolutional neural networks, covering concepts like convolution operations, spatial hierarchies, and max pooling. It concludes with a demonstration of digit and X-ray image classification using Keras and techniques for dealing with overfitting like dropout and data augmentation.
Deep learning is introduced along with its applications and key players in the field. The document discusses the problem space of inputs and outputs for deep learning systems. It describes what deep learning is, providing definitions and explaining the rise of neural networks. Key deep learning architectures like convolutional neural networks are overviewed along with a brief history and motivations for deep learning.
This document provides an overview of recurrent neural networks (RNNs) and long short-term memory (LSTM) networks. It discusses how RNNs can be used for sequence modeling tasks like sentiment analysis, machine translation, and speech recognition by incorporating context or memory from previous steps. LSTMs are presented as an improvement over basic RNNs that can learn long-term dependencies in sequences using forget gates, input gates, and output gates to control the flow of information through the network.
Deep Learning: Evolution of ML from Statistical to Brain-like Computing- Data...Impetus Technologies
Presentation on 'Deep Learning: Evolution of ML from Statistical to Brain-like Computing'
Speaker- Dr. Vijay Srinivas Agneeswaran,Director, Big Data Labs, Impetus
The main objective of the presentation is to give an overview of our cutting edge work on realizing distributed deep learning networks over GraphLab. The objectives can be summarized as below:
- First-hand experience and insights into implementation of distributed deep learning networks.
- Thorough view of GraphLab (including descriptions of code) and the extensions required to implement these networks.
- Details of how the extensions were realized/implemented in GraphLab source – they have been submitted to the community for evaluation.
- Arrhythmia detection use case as an application of the large scale distributed deep learning network.
Few shot learning/ one shot learning/ machine learningﺁﺻﻒ ﻋﻠﯽ ﻣﯿﺮ
The document discusses few-shot learning approaches. It begins with an introduction explaining that current deep learning models require large datasets but humans can learn from just a few examples. It then discusses the problem of few-shot learning, where models must perform classification, detection, or regression on novel categories represented by only a few samples. Popular approaches discussed include meta-learning methods like MAML and prototypical networks, metric learning methods like relation networks, and data augmentation methods. The document provides an overview of the goals and techniques of few-shot learning.
This document provides an introduction to deep learning, including definitions of key concepts. It discusses how deep learning is a subfield of machine learning focused on learning representations from data. The document outlines the main types of machine learning as supervised learning, unsupervised learning, self-supervised learning, and reinforcement learning. It also describes common deep learning components like layers, loss functions, optimizers, and popular frameworks like TensorFlow and PyTorch.
Learn to Build an App to Find Similar Images using Deep Learning- Piotr TeterwakPyData
This document discusses using deep learning and deep features to build an app that finds similar images. It begins with an overview of deep learning and how neural networks can learn complex patterns in data. The document then discusses how pre-trained neural networks can be used as feature extractors for other domains through transfer learning. This reduces data and tuning requirements compared to training new deep learning models. The rest of the document focuses on building an image similarity service using these techniques, including training a model with GraphLab Create and deploying it as a web service with Dato Predictive Services.
An Introduction to Deep Learning (May 2018)Julien SIMON
This document provides an introduction to deep learning, including common network architectures and use cases. It defines artificial intelligence, machine learning, and deep learning. It discusses how neural networks are trained using stochastic gradient descent and backpropagation to minimize loss and optimize weights. Common network types are described, such as convolutional neural networks for image recognition and LSTM networks for sequence prediction. Examples of deep learning applications include machine translation, object detection, segmentation, and generation of images, text, and video. Resources for learning more about deep learning are provided.
This document provides an overview of machine learning techniques for classification with imbalanced data. It discusses challenges with imbalanced datasets like most classifiers being biased towards the majority class. It then summarizes techniques for dealing with imbalanced data, including random over/under sampling, SMOTE, cost-sensitive classification, and collecting more data. [/SUMMARY]
Deep Learning Made Easy with Deep FeaturesTuri, Inc.
Deep learning models can learn hierarchical feature representations from raw input data. These learned features can then be used to build simple classifiers that achieve high accuracy, even when training data is limited. Transfer learning involves using features extracted from a model pre-trained on a large dataset to build classifiers for other related problems. This approach has been shown to outperform traditional feature engineering with hand-designed features. Deep features extracted from neural networks trained on large image or text datasets have proven to work well as general purpose features for other visual and language problems.
This document provides an outline for a presentation on machine learning and deep learning. It begins with an introduction to machine learning basics and types of learning. It then discusses what deep learning is and why it is useful. The main components and hyperparameters of deep learning models are explained, including activation functions, optimizers, cost functions, regularization methods, and tuning. Basic deep neural network architectures like convolutional and recurrent networks are described. An example application of relation extraction is provided. The document concludes by listing additional deep learning topics.
MDEC Data Matters Series: machine learning and Deep Learning, A PrimerPoo Kuan Hoong
The document provides an overview of machine learning and deep learning. It discusses the history and development of neural networks, including deep belief networks, convolutional neural networks, and recurrent neural networks. Applications of deep learning in areas like computer vision, natural language processing, and robotics are also covered. Finally, popular platforms, frameworks and libraries for developing deep learning models are presented, along with examples of pre-trained models that are available.
Separating Hype from Reality in Deep Learning with Sameer FarooquiDatabricks
Deep Learning is all the rage these days, but where does the reality of what Deep Learning can do end and the media hype begin? In this talk, I will dispel common myths about Deep Learning that are not necessarily true and help you decide whether you should practically use Deep Learning in your software stack.
I’ll begin with a technical overview of common neural network architectures like CNNs, RNNs, GANs and their common use cases like computer vision, language understanding or unsupervised machine learning. Then I’ll separate the hype from reality around questions like:
• When should you prefer traditional ML systems like scikit learn or Spark.ML instead of Deep Learning?
• Do you no longer need to do careful feature extraction and standardization if using Deep Learning?
• Do you really need terabytes of data when training neural networks or can you ‘steal’ pre-trained lower layers from public models by using transfer learning?
• How do you decide which activation function (like ReLU, leaky ReLU, ELU, etc) or optimizer (like Momentum, AdaGrad, RMSProp, Adam, etc) to use in your neural network?
• Should you randomly initialize the weights in your network or use more advanced strategies like Xavier or He initialization?
• How easy is it to overfit/overtrain a neural network and what are the common techniques to ovoid overfitting (like l1/l2 regularization, dropout and early stopping)?
Similar to Machine Duping 101: Pwning Deep Learning Systems (20)
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
4. Deep Learning
why would someone choose to use it?
(Semi)Automated
feature/representation
learning
One infrastructure for
multiple problems (sort of)
Hierarchical learning:
Divide task across
multiple layers
Efficient,
easily distributed &
parallelized
8. Training Deep Neural Networks
Step 1 of 2: Feed Forward
1. Each unit receives output of the neurons in the previous layer (+ bias
signal)
2. Computes weighted average of inputs
3. Apply weighted average through nonlinear activation function
4. For DNN classifier, send final layer output through softmax function
softmax
function
[0.34, 0.57, 0.09]logits
prediction
𝒇
𝑊1
activation
function
[17.0, 28.5, 4.50]
𝒇 𝒇
activation
function
activation
function
input 𝑊2
b1 b2
bias
unit
bias
unit
9. Training Deep Neural Networks
Step 2 of 2: Backpropagation
1. If the model made a wrong prediction, calculate the error
1. In this case, the correct class is 0, but the model predicted 1 with 57% confidence - error is thus 0.57
2. Assign blame: trace backwards to find the units that contributed to this wrong
prediction (and how much they contributed to the total error)
1. Partial differentiation of this error w.r.t. the unit’s activation value
3. Penalize those units by decreasing their weights and biases by an amount proportional
to their error contribution
4. Do the above efficiently with optimization algorithm e.g. Stochastic Gradient Descent
0.57
total error
𝒇
𝑊1
activation
function
28.5
𝒇 𝒇
activation
function
activation
function
𝑊2
b1 b2
bias
unit
bias
unit
12. Layered Learning
Lee, 2009, “Convolutional deep belief networks for scalable
unsupervised learning of hierarchical representations."
13. Beyond Multi Layer Perceptrons
Source: LeNet 5, LeCun et al.
Convolutional Neural Network
14. Beyond Multi Layer Perceptrons
Recurrent Neural Network
neural
network
input output
loop(s)
• Just a DNN with a feedback loop
• Previous time step feeds all
intermediate and final values into
next time step
• Introduces the concept of “memory”
to neural networks
15. Beyond Multi Layer Perceptrons
Recurrent Neural Network
input
output
time steps
network
depth
Y O L O !
16. Beyond Multi Layer Perceptrons
Disney, Finding Dory
Long Short-Term Memory (LSTM) RNN
• To make good predictions, we sometimes
need more context
• We need long-term memory capabilities
without extending the network’s recursion
indefinitely (unscalable)
Colah’s Blog, “Understanding LSTM Networks"
17. Deng et al. “Deep Learning: Methods and Applications”
19. Attack Taxonomy
ExploratoryCausative
(Manipulative test samples)
Applicable also to online-learning
models that continuously learn
from real-time test data
(Manipulative training samples)
Targeted
Indiscriminate
Training samples that move
classifier decision boundary in an
intentional direction
Training samples that increase FP/FN
→ renders classifier unusable
Adversarial input crafted to cause
an intentional misclassification
n/a
20. Why can we do this?
vs.
Statistical learning models don’t learn concepts the same way that we do.
BLINDSPOTS:
21. Adversarial Deep Learning
Intuitions
1. Run input x through the classifier model (or substitute/approximate model)
2. Based on model prediction, derive a perturbation tensor that
maximizes chances of misclassification:
1. Traverse the manifold to find blind spots in input space; or
2. Linear perturbation in direction of neural network’s cost function gradient; or
3. Select only input dimensions with high saliency* to perturb by the model’s Jacobian matrix
3. Scale the perturbation tensor by some magnitude, resulting in the
effective perturbation (δx) to x
1. Larger perturbation == higher probability for misclassification
2. Smaller perturbation == less likely for human detection
* saliency: amount of influence a selected dimension has on the entire model’s output
22. Adversarial Deep Learning
Intuitions
Szegedy, 2013: Traverse the manifold to find blind spots in the input space
• Adversarial samples == pockets in the manifold
• Difficult to efficiently find by brute force (high dimensional input)
• Optimize this search, take gradient of input w.r.t. target output class
23. Adversarial Deep Learning
Intuitions
Goodfellow, 2015: Linear adversarial perturbation
• Developed a linear view of adversarial examples
• Can just take the cost function gradient w.r.t. the sample (x) and original
predicted class (y)
• Easily found by backpropagation
24. Adversarial Deep Learning
Intuitions
Papernot, 2015: Saliency map + Jacobian matrix perturbation
• More complex derivations for why the Jacobian of the learned neural network
function is used
• Obtained with respect to input features rather than network parameters
• Forward propagation is used instead of backpropagation
• To reduce probability of human detection, only perturb the dimensions that
have the greatest impact on the output (salient dimensions)
27. Deep Neural Network Attacks
Adversary
knowledge
Attack
complexity
DIFFICULT
EASY
Architecture,
Training Tools,
Hyperparameters
Architecture
Training data
Oracle
Labeled Test
samples
Confidence
Reduction
Untargeted
M
isclassification
Targeted
M
isclassification
Source/Target
M
isclassification
Murphy, 2012 Szegedy, 2014
Papernot, 2016
Goodfellow, 2016
Xu, 2016
Nguyen, 2014
28. What can you do with limited knowledge?
• Quite a lot.
• Make good guesses: Infer the methodology from the task
• Image classification: ConvNet
• Speech recognition: LSTM-RNN
• Amazon ML, ML-as-a-service etc.: Shallow feed-forward network
• What if you can’t guess?
30. Black box attack methodology
1. Transferability
Adversarial samples that fool model A have a
good chance of fooling a previously unseen
model B
SVM, scikit-learn
Decision Tree
Matt's Webcorner, Stanford
Linear Classifier
(Logistic Regression)
Spectral Clustering, scikit-learn
Feed Forward Neural Network
31. Black box attack methodology
1. Transferability
Papernot et al. “Transferability in Machine Learning:
from Phenomena to Black-Box Attacks using Adversarial Samples”
32. Black box attack methodology
2. Substitute model
train a new model by treating the target model’s
output as a training labels
then, generate adversarial samples with
substitute model
input data
target
black box
model
training label
backpropagate
black box prediction
substitute
model
33. Why is this possible?
• Transferability?
• Still an open research problem
• Manifold learning problem
• Blind spots
• Model vs. Reality dimensionality mismatch
• IN GENERAL:
• Is the model not learning anything
at all?
34. What this means for us
• Deep learning algorithms (Machine Learning in general) are susceptible to
manipulative attacks
• Use with caution in critical deployments
• Don’t make false assumptions about what/how the model learns
• Evaluate a model’s adversarial resilience - not just accuracy/precision/recall
• Spend effort to make models more robust to tampering
35. Defending the machines
• Distillation
• Train model 2x, feed first DNN output logits into second DNN input layer
• Train model with adversarial samples
• i.e. ironing out imperfect knowledge learnt in the model
• Other miscellaneous tweaks
• Special regularization/loss-function methods (simulating adversarial content during training)
37. WHY DEEP-PWNING?
• lol why not
• “Penetration testing” of statistical/machine learning systems
• Train models with adversarial samples for increased robustness
39. Deep Learning and Privacy
• Deep learning also sees challenges in other areas relating to security &
privacy
• Adversary can reconstruct training samples from a trained black box
DNN model (Fredrikson, 2015)
• Can we precisely control the learning objective of a DNN model?
• Can we train a DNN model without the training agent having complete
access to all training data? (Shokri, 2015)
41. WHY DEEP-PWNING?
• MORE CRITICAL SYSTEMS RELY ON MACHINE LEARNING → MORE IMPORTANCE
ON ENSURING THEIR ROBUSTNESS
• WE NEED PEOPLE WITH BOTH SECURITY AND STATISTICAL SKILL SETS TO
DEVELOP ROBUST SYSTEMS AND EVALUATE NEW INFRASTRUCTURE