A simplified way of approaching machine learning and deep learning from the ground up. The case for deep learning and an attempt to develop intuition for how/why it works. Advantages, state-of-the-art, and trends.
Presented at NYU Center for Genomics for NY Deep Learning Meetup
Squeezing Deep Learning Into Mobile PhonesAnirudh Koul
A practical talk by Anirudh Koul aimed at how to run Deep Neural Networks to run on memory and energy constrained devices like smart phones. Highlights some frameworks and best practices.
Mastering Computer Vision Problems with State-of-the-art Deep LearningMiguel González-Fierro
Deep learning has been especially successful in computer vision tasks such as image classification because convolutional neural nets (CNNs) can create hierarchical
representations in an image. One of the most remarkable advances is ResNet, the CNN that surpassed human-level accuracy for the first time in history.
ImageNet competition has become the de facto benchmark for image classification in the research community. The “small” ImageNet data contains more than 1.2 million images distributed in 1,000 classes.
Miguel González-Fierro explains how to train a state of the art deep neural network, ResNet, using Microsoft RSever and MXNet with the ImageNet dataset. (While most of the deep learning libraries are programmed in C++ and Python, only MXNet offers an API for R programmers.) Miguel then demonstrates how to operationalize this training for real-world business problems related to image classification.
This talk was presented at Strata London 2017: https://conferences.oreilly.com/strata/strata-eu/public/schedule/detail/57428
A practical talk by Anirudh Koul aimed at how to run Deep Neural Networks to run on memory and energy constrained devices like smartphones. Highlights some frameworks and best practices.
Improving Hardware Efficiency for DNN ApplicationsChester Chen
Speaker: Dr. Hai (Helen) Li is the Clare Boothe Luce Associate Professor of Electrical and Computer Engineering and Co-director of the Duke Center for Evolutionary Intelligence at Duke University
In this talk, I will introduce a few recent research spotlights by the Duke Center for Evolutionary Intelligence. The talk will start with the structured sparsity learning (SSL) method which attempts to learn a compact structure from a bigger DNN to reduce computation cost. It generates a regularized structure with high execution efficiency. Our experiments on CPU, GPU, and FPGA platforms show on average 3~5 times speedup of convolutional layer computation of AlexNet. Then, the implementation and acceleration of DNN applications on mobile computing systems will be introduced. MoDNN is a local distributed system which partitions DNN models onto several mobile devices to accelerate computations. ApesNet is an efficient pixel-wise segmentation network, which understands road scenes in real-time, and has achieved promising accuracy. Our prospects on the adoption of emerging technology will also be given at the end of this talk, offering the audiences an alternative thinking about the future evolution and revolution of modern computing systems.
This presentation focuses on Deep Learning (DL) concepts, such as neural networks, backprop, activation functions, and Convolutional Neural Networks. You'll also learn how to incorporate Deep Learning in Android applications. Basic knowledge of matrices is helpful for this session, which is targeted primarily to beginners.
A simplified way of approaching machine learning and deep learning from the ground up. The case for deep learning and an attempt to develop intuition for how/why it works. Advantages, state-of-the-art, and trends.
Presented at NYU Center for Genomics for NY Deep Learning Meetup
Squeezing Deep Learning Into Mobile PhonesAnirudh Koul
A practical talk by Anirudh Koul aimed at how to run Deep Neural Networks to run on memory and energy constrained devices like smart phones. Highlights some frameworks and best practices.
Mastering Computer Vision Problems with State-of-the-art Deep LearningMiguel González-Fierro
Deep learning has been especially successful in computer vision tasks such as image classification because convolutional neural nets (CNNs) can create hierarchical
representations in an image. One of the most remarkable advances is ResNet, the CNN that surpassed human-level accuracy for the first time in history.
ImageNet competition has become the de facto benchmark for image classification in the research community. The “small” ImageNet data contains more than 1.2 million images distributed in 1,000 classes.
Miguel González-Fierro explains how to train a state of the art deep neural network, ResNet, using Microsoft RSever and MXNet with the ImageNet dataset. (While most of the deep learning libraries are programmed in C++ and Python, only MXNet offers an API for R programmers.) Miguel then demonstrates how to operationalize this training for real-world business problems related to image classification.
This talk was presented at Strata London 2017: https://conferences.oreilly.com/strata/strata-eu/public/schedule/detail/57428
A practical talk by Anirudh Koul aimed at how to run Deep Neural Networks to run on memory and energy constrained devices like smartphones. Highlights some frameworks and best practices.
Improving Hardware Efficiency for DNN ApplicationsChester Chen
Speaker: Dr. Hai (Helen) Li is the Clare Boothe Luce Associate Professor of Electrical and Computer Engineering and Co-director of the Duke Center for Evolutionary Intelligence at Duke University
In this talk, I will introduce a few recent research spotlights by the Duke Center for Evolutionary Intelligence. The talk will start with the structured sparsity learning (SSL) method which attempts to learn a compact structure from a bigger DNN to reduce computation cost. It generates a regularized structure with high execution efficiency. Our experiments on CPU, GPU, and FPGA platforms show on average 3~5 times speedup of convolutional layer computation of AlexNet. Then, the implementation and acceleration of DNN applications on mobile computing systems will be introduced. MoDNN is a local distributed system which partitions DNN models onto several mobile devices to accelerate computations. ApesNet is an efficient pixel-wise segmentation network, which understands road scenes in real-time, and has achieved promising accuracy. Our prospects on the adoption of emerging technology will also be given at the end of this talk, offering the audiences an alternative thinking about the future evolution and revolution of modern computing systems.
This presentation focuses on Deep Learning (DL) concepts, such as neural networks, backprop, activation functions, and Convolutional Neural Networks. You'll also learn how to incorporate Deep Learning in Android applications. Basic knowledge of matrices is helpful for this session, which is targeted primarily to beginners.
Presentation of few recent papers on Deep Learning ... in particular Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding, Song Han, Huizi Mao, William J. Dally International Conference on Learning Representations ICLR2016
From Conventional Machine Learning to Deep Learning and Beyond.pptxChun-Hao Chang
In this slide, Deep Learning are compared with Conventional Learning and the strength of DNN models will be explained.
The target audience are people who have the knowledge of Machine Learning or Data Mining but not familiar with Deep Learning.
Yinyin Liu presents at SD Robotics Meetup on November 8th, 2016. Deep learning has made great success in image understanding, speech, text recognition and natural language processing. Deep Learning also has tremendous potential to tackle the challenges in robotic vision, and sensorimotor learning in a robotic learning environment. In this talk, we will talk about how current and future deep learning technologies can be applied for robotic applications.
Deep learning on mobile - 2019 Practitioner's GuideAnirudh Koul
The 2019 Guide to Deep Learning on Mobile, from Inference to Training on iOS and Android smartphones. Featuring CoreML, Tensorflow Lite, MLKit, Fritz, AutoML Approaches (Hardware Aware Neural Architecture Search) to make models more efficient, and lots of videos. Presented by Anirudh Koul, Siddha Ganju and Meher Anand Kasam. More details at PracticalDL.ai in the upcoming O'Reilly Book 'Practical Deep Learning on Cloud & Mobile'
Talk given at PYCON Stockholm 2015
Intro to Deep Learning + taking pretrained imagenet network, extracting features, and RBM on top = 97 Accuracy after 1 hour (!) of training (in top 10% of kaggle cat vs dog competition)
This is a 2 hours overview on the deep learning status as for Q1 2017.
Starting with some basic concepts, continue to basic networks topologies , tools, HW/Accelerators and finally Intel's take on the the different fronts.
Faster deep learning solutions from training to inference - Michele Tameni - ...Codemotion
Intel Deep Learning SDK enables using of optimized open source deep-learning frameworks, including Caffe and TensorFlow through a step-by-step wizard or iPython interactive notebooks. It includes easy and fast installation of all depended libraries and advanced tools for easy data pre-processing and model training, optimization and deployment, providing an end-to-end solution to the problem. In addition, it supports scale-out on multiple computers for training, as well as using compression methods for deployment of the models on various platforms, addressing memory and speed constraints.
Deep learning goes beyond the traditional machine learning of big data and analytics. In this session, we will review the AWS offering, Amazon Machine Learning, and the AWS GPU-intensive family of servers that run native machine learning and deep-learning algorithms. We will also cover some basic deep-learning algorithms using open source software. Session sponsored by Day1 Solutions.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/wavecomp/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-nicol
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Chris Nicol, CTO at Wave Computing, presents the "New Dataflow Architecture for Machine Learning" tutorial at the May 2017 Embedded Vision Summit.
Data scientists have made tremendous advances in the use of deep neural networks (DNNs) to enhance business models and service offerings. But training DNNs can take a week or more using traditional hardware solutions that rely on legacy architectures that are limited in performance and scalability. New innovations that can reduce training time for both image-centric and text-centric deep neural networks will lead to an explosion of new applications. Dr. Chris Nicol, Wave Computing’s Chief Technology Officer, examines the performance challenge faced by data scientists today. Nicol outlines the technical factors underlying this bottleneck for systems relying on CPUs, GPUs, FPGAs and ASICs, and introduces a new dataflow-centric approach to DNN training.
On-device machine learning: TensorFlow on AndroidYufeng Guo
Machine learning has traditionally been the solely performed on servers and high performance machines. But there is great value is having on-device machine learning for mobile devices. Doing ML inference on mobile devices has huge potential and is still in its early stages. However, it's already more powerful than most realize.
In this demo-oriented talk, you will see some examples of deep learning models used for local prediction on mobile devices. Learn how to use TensorFlow to implement a machine learning model that is tailored to a custom dataset, and start making delightful experiences today!
Presentation of few recent papers on Deep Learning ... in particular Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding, Song Han, Huizi Mao, William J. Dally International Conference on Learning Representations ICLR2016
From Conventional Machine Learning to Deep Learning and Beyond.pptxChun-Hao Chang
In this slide, Deep Learning are compared with Conventional Learning and the strength of DNN models will be explained.
The target audience are people who have the knowledge of Machine Learning or Data Mining but not familiar with Deep Learning.
Yinyin Liu presents at SD Robotics Meetup on November 8th, 2016. Deep learning has made great success in image understanding, speech, text recognition and natural language processing. Deep Learning also has tremendous potential to tackle the challenges in robotic vision, and sensorimotor learning in a robotic learning environment. In this talk, we will talk about how current and future deep learning technologies can be applied for robotic applications.
Deep learning on mobile - 2019 Practitioner's GuideAnirudh Koul
The 2019 Guide to Deep Learning on Mobile, from Inference to Training on iOS and Android smartphones. Featuring CoreML, Tensorflow Lite, MLKit, Fritz, AutoML Approaches (Hardware Aware Neural Architecture Search) to make models more efficient, and lots of videos. Presented by Anirudh Koul, Siddha Ganju and Meher Anand Kasam. More details at PracticalDL.ai in the upcoming O'Reilly Book 'Practical Deep Learning on Cloud & Mobile'
Talk given at PYCON Stockholm 2015
Intro to Deep Learning + taking pretrained imagenet network, extracting features, and RBM on top = 97 Accuracy after 1 hour (!) of training (in top 10% of kaggle cat vs dog competition)
This is a 2 hours overview on the deep learning status as for Q1 2017.
Starting with some basic concepts, continue to basic networks topologies , tools, HW/Accelerators and finally Intel's take on the the different fronts.
Faster deep learning solutions from training to inference - Michele Tameni - ...Codemotion
Intel Deep Learning SDK enables using of optimized open source deep-learning frameworks, including Caffe and TensorFlow through a step-by-step wizard or iPython interactive notebooks. It includes easy and fast installation of all depended libraries and advanced tools for easy data pre-processing and model training, optimization and deployment, providing an end-to-end solution to the problem. In addition, it supports scale-out on multiple computers for training, as well as using compression methods for deployment of the models on various platforms, addressing memory and speed constraints.
Deep learning goes beyond the traditional machine learning of big data and analytics. In this session, we will review the AWS offering, Amazon Machine Learning, and the AWS GPU-intensive family of servers that run native machine learning and deep-learning algorithms. We will also cover some basic deep-learning algorithms using open source software. Session sponsored by Day1 Solutions.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/wavecomp/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-nicol
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Chris Nicol, CTO at Wave Computing, presents the "New Dataflow Architecture for Machine Learning" tutorial at the May 2017 Embedded Vision Summit.
Data scientists have made tremendous advances in the use of deep neural networks (DNNs) to enhance business models and service offerings. But training DNNs can take a week or more using traditional hardware solutions that rely on legacy architectures that are limited in performance and scalability. New innovations that can reduce training time for both image-centric and text-centric deep neural networks will lead to an explosion of new applications. Dr. Chris Nicol, Wave Computing’s Chief Technology Officer, examines the performance challenge faced by data scientists today. Nicol outlines the technical factors underlying this bottleneck for systems relying on CPUs, GPUs, FPGAs and ASICs, and introduces a new dataflow-centric approach to DNN training.
On-device machine learning: TensorFlow on AndroidYufeng Guo
Machine learning has traditionally been the solely performed on servers and high performance machines. But there is great value is having on-device machine learning for mobile devices. Doing ML inference on mobile devices has huge potential and is still in its early stages. However, it's already more powerful than most realize.
In this demo-oriented talk, you will see some examples of deep learning models used for local prediction on mobile devices. Learn how to use TensorFlow to implement a machine learning model that is tailored to a custom dataset, and start making delightful experiences today!
In recent months, Deep Learning has become the hottest topic in the IT industry. However, its arcane jargon and its intimidating equations often discourage software developers, who wrongly think that they’re “not smart enough”. Through code-level demos based on Apache MXNet, we’ll demonstrate how to build, train and use models based on different types of networks: multi-layer perceptrons, convolutional neural networks and long short-term memory networks. Finally, we’ll share some optimization tips which will help improve the training speed and the performance of your models.
Don't Start from Scratch: Transfer Learning for Novel Computer Vision Problem...StampedeCon
In this session, we’ll discuss approaches for applying convolutional neural networks to novel computer vision problems, even without having millions of images of your own. Pretrained models and generic image data sets from Google, Kaggle, universities, and other places can be leveraged and adapted to solve industry and business specific problems. We’ll discuss the approaches of transfer learning and fine tuning to help anyone get started on using deep learning to get cutting edge results on their computer vision problems.
Traditional Machine Learning had used handwritten features and modality-specific machine learning to classify images, text or recognize voices. Deep learning / Neural network identifies features and finds different patterns automatically. Time to build these complex tasks has been drastically reduced and accuracy has exponentially increased because of advancements in Deep learning. Neural networks have been partly inspired from how 86 billion neurons work in a human and become more of a mathematical and a computer problem. We will see by the end of the blog how neural networks can be intuitively understood and implemented as a set of matrix multiplications, cost function, and optimization algorithms.
Deep Learning: concepts and use cases (October 2018)Julien SIMON
An introduction to Deep Learning theory
Neurons & Neural Networks
The Training Process
Backpropagation
Optimizers
Common network architectures and use cases
Convolutional Neural Networks
Recurrent Neural Networks
Long Short Term Memory Networks
Generative Adversarial Networks
Getting started
The Frontier of Deep Learning in 2020 and BeyondNUS-ISS
This talk will be a summary of the recent advances in deep learning research, current trends in the industry, and the opportunities that lie ahead.
We will discuss topics in research such as:
Transformers, GPT-3, BERT
Neural Architecture Search, Evolutionary Search
Distillation, self-learning
NeRF
Self-Attention
Also shifting industry trends such as:
The move to free data
Rising importance of 3D vision
Using synthetic data (Sim2Real)
Mobile vision & Federated Learning
Artificial Intelligence, Machine Learning, Deep Learning
The 5 myths of AI
Deep Learning in action
Basics of Deep Learning
NVIDIA Volta V100 and AWS P3
Artificial Intelligence (AI) Interview Questions and Answers | EdurekaEdureka!
(** Machine Learning Engineer Masters Program: https://www.edureka.co/masters-progra... **)
This PPT on Artificial Intelligence Interview Questions covers all the important concepts involved in the field of AI. This PPT is ideal for both beginners as well as professionals who want to learn or brush up their knowledge on AI concepts. Below are the topics covered in this tutorial:
1. Artificial Intelligence Basic Level Interview Question
2. Artificial Intelligence Intermediate Level Interview Question
3. Artificial Intelligence Scenario based Interview Question
Check out the entire Machine Learning Playlist: https://bit.ly/2NG9tK4
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Neural Networks, Spark MLlib, Deep LearningAsim Jalis
What are neural networks? How to use the neural networks algorithm in Apache Spark MLlib? What is Deep Learning? Presented at Data Science Meetup at Galvanize on 2/17/2016.
For code see IPython/Jupyter/Toree notebook at http://nbviewer.jupyter.org/gist/asimjalis/4f911882a1ab963859ce
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Why is Deep learning hot right now? and How can we apply it on each day job?
1. Why is Deep Learning hot
right now, and How can we
apply it on each day job
ISSAM A. AL-ZINATI
OUTREACH & TECHNICAL ADVISOR
UCAS TECHNOLOGY INCUBATOR
ISSAM A. AL-ZINATI - UCASTI 1
2. What is Deep Learning
ISSAM A. AL-ZINATI - UCASTI 2
Is a Neural Network
3. What is Deep Learning
ISSAM A. AL-ZINATI - UCASTI 3
Is a Neural Network Neuron
Can run small specific
mathematical task
4. What is Deep Learning
ISSAM A. AL-ZINATI - UCASTI 4
Is a Neural Network Neuron
Can run small specific
mathematical task
Edge
Connects Neurons
Holds weights to adjust inputs
5. What is Deep Learning
ISSAM A. AL-ZINATI - UCASTI 5
Is a Neural Network
With More Layers
6. What is Deep Learning
ISSAM A. AL-ZINATI - UCASTI 6
Is a Neural Network
With More Layers
And More Neurons
9. Why Now- Scale
ISSAM A. AL-ZINATI - UCASTI 9
Data
Small Meduim Large
Performance Based on Data Size
Performance
The more data
you feed the
model, the
better results
you get
11. Why Now- Scale
ISSAM A. AL-ZINATI - UCASTI 11
Model Size & GPU
Small Meduim Large
Performance Based on Model Size
Performance
Bigger model could
achieve better
results.
GPUs help to train
those models in
much faster, 20X!!
12. Why Now– vs Others
ISSAM A. AL-ZINATI - UCASTI 12
What about other kind of machine learning algorithms, i.e. SVM, DT, Boosting, ….
Would they do better if they got more data and power?
13. Why Now– vs Others
Small Data Medium Data Large Data
Performance of NN VS Others
Based on Model Size and Data Amount
Others Small NN Medium NN Large NN
ISSAM A. AL-ZINATI - UCASTI 13
14. Why Now– End-To-End
ISSAM A. AL-ZINATI - UCASTI 14
Usual machine learning approach contains a pipeline of stages that are
responsible of feature extraction.
Each stage passes a set of engineered features which help model to better
understand the case it works on.
This approach is complex and prone to errors.
15. Why Now– End-To-End
ISSAM A. AL-ZINATI - UCASTI 15
Data (Audio)
Speech Recognition Pipeline
Audio
Features
Phonemes
Language
Model
Transcript
16. Why Now– End-To-End
ISSAM A. AL-ZINATI - UCASTI 16
Data (Audio)
Speech Recognition - DL
Audio
Features Phonemes Language
Model
Transcript
17. Why Now– End-To-End
ISSAM A. AL-ZINATI - UCASTI 17
Data (Audio)
Speech Recognition - DL
Audio
Features Phonemes Language
Model
Transcript
The Magic
18. How it wok – The Magic
ISSAM A. AL-ZINATI - UCASTI 18
19. How it work – No Magic
Deep Neural network is not magic. But it is very good at finding patterns.
“The hierarchy of concepts allows the computer to learn complicated concepts
by building them out of simpler ones. If we draw a graph showing how these
concepts are built on top of each other, the graph is deep, with many layers. For
this reason, we call this approach to AI deep learning”, Ian Goodfellow.
Deep Learning is Hierarchical Feature Learning.
ISSAM A. AL-ZINATI - UCASTI 19
20. Deep Learning Models
ISSAM A. AL-ZINATI - UCASTI 20
General
Model
FC
Sequence
Model
RNN
LSTM
Image
Model
CNN
Other
Models
Unsupervised
RL
21. Deep Learning Models
ISSAM A. AL-ZINATI - UCASTI 21
General
Model
FC
Sequence
Model
RNN
LSTM
Image
Model
CNN
Other
Models
Unsupervised
RL
Hot Research Topic
22. Advanced Deep Learning Models –
VGGNET - ResNet
Achieves 7.3% on ImageNet-2014 classification Challenge, come in the first
place.
It Used
120 million
parameters.
ISSAM A. AL-ZINATI - UCASTI 22
23. Advanced Deep Learning Models –
Google Inception V3
Achieves 5.64% on ImageNet-2015 classification Challenge, come in the second place.
ISSAM A. AL-ZINATI - UCASTI 23
24. Advanced Deep Learning Models –
Google Inception V3
Based on ConvNet concept with the addition
of inception module.
ISSAM A. AL-ZINATI - UCASTI 24
Using a network with a
computational cost of 5 billion
multiply-adds per inference and
with using less than 25 million
parameters.
25. Deep Learning Applications – Deep Voice
Baidu Research presents Deep Voice, a production-quality text-to-speech system
constructed entirely from deep neural networks.
Ground Truth
Generated Voice
ISSAM A. AL-ZINATI - UCASTI 25
26. Deep Learning Applications – Image
Captioning
Multimodal Recurrent Neural Architecture generates sentence descriptions from
images. Source.
ISSAM A. AL-ZINATI - UCASTI 26
"man in black shirt is playing guitar." "two young girls are playing with lego toy."
27. Deep Learning Applications – Generating
Videos
ISSAM A. AL-ZINATI - UCASTI 27
This approach was driven by using Adversarial Network to
1) Generate Videos
2) Conditional Video Generation based on Static Images
Source
30. Applying Deep Learning – Bias/Variance
The goal is to build a model that is close to human-level performance.
ISSAM A. AL-ZINATI - UCASTI 30
31. Applying Deep Learning – Bias/Variance
The goal is to build a model that is close to human-level performance.
ISSAM A. AL-ZINATI - UCASTI 31
Training Set – 70% Val Set – 15% Test Set – 15%
32. Applying Deep Learning – Bias/Variance
The goal is to build a model that is close to human-level performance.
ISSAM A. AL-ZINATI - UCASTI 32
Training Set – 70% Val Set – 15% Test Set – 15%
You need to know the following values:
1- Human-Level Error
2- Training Level Error
3- Validation Level Error
33. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 33
1%
5%
Human-Level
Training-Level
6%
Validation-Level
34. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 34
1%
5%
Human-Level
Training-Level
6%
Validation-Level
High bias/
underfitting
35. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 35
1%
5%
Human-Level
Training-Level
6%
Validation-Level
High bias/
underfitting
1- Bigger Model
2- Train Longer
3- New Model Arch
36. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 36
1%
2%
Human-Level
Training-Level
6%
Validation-Level
37. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 37
1%
2%
Human-Level
Training-Level
6%
Validation-Level
High variance/
overfitting
38. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 38
1%
2%
Human-Level
Training-Level
6%
Validation-Level
High variance/
overfitting
1- More Data
2- Early Stopping
3- Regularization
4- New Model Arch
39. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 39
1%
5%
Human-Level
Training-Level
10%
Validation-Level
40. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 40
1%
5%
Human-Level
Training-Level
10%
Validation-Level
High bias/
underfitting
High variance/
overfitting
41. Applying Deep Learning – Bias/Variance
ISSAM A. AL-ZINATI - UCASTI 41
1%
5%
Human-Level
Training-Level
10%
Validation-Level
High bias/
underfitting
1- Bigger Model
2- Train Longer
3- New Model Arch
High variance/
overfitting
1- More Data
2- Early Stopping
3- Regularization
4- New Model Arch
42. Applying Deep Learning – Data Synthesis
Usually, to overcome the problem of bias/variance we tend to create new set of
handy engineered features and try to retrain our model to see if we get more
accurate one.
In Deep Learning, Having more data can be a great solution to many scenarios.
But Its not always the case that we had this data ready to use.
So playing with data and try to create new handy engineered data set can be the
solution.
ISSAM A. AL-ZINATI - UCASTI 42
43. Applying Deep Learning – Data Synthesis
1) OCR
Getting more data for an OCR model is easy. We can follow these steps to get
those new data sets:
- Downloading images from the internet
- Generate text from MSWord in different font, size, color, blur, …
- Combine these two steps and you get millions of new images for training.
ISSAM A. AL-ZINATI - UCASTI 43
44. Applying Deep Learning – Data Synthesis
2) Speech Recognition
- Collect a set of clean audio files
- Collect random background sounds
- Combine these two steps and you get millions of new audio files for training.
ISSAM A. AL-ZINATI - UCASTI 44
45. Applying Deep Learning – Data Synthesis
3) NLP – Grammar correction
- Collect a set of correct sentences
- Randomly shuffle the word in this sentence
- Those new sentences are the new data set that we can fed to our model.
ISSAM A. AL-ZINATI - UCASTI 45
46. Applying Deep Learning – Data Synthesis
4) Image Recognition
- Having a set of labeled images
- Randomly add new effects to those images, i.e. rotate, blur, flip, luminosity, …
- Those new images are the new data set that we can fed to our model.
ISSAM A. AL-ZINATI - UCASTI 46
47. Applying Deep Learning – Data Synthesis
Data Syntheses has a limit, it can not always work but it good to start with.
ISSAM A. AL-ZINATI - UCASTI 47
48. Applying Deep Learning – Transfer Learning
Another approach to overcome the problem of bias/variance is to have:
1- Larger Model
2- New Model Architecture
But these two approaches needs more powerful machine to do the training.
Also, sometimes you don’t have enough resources to train it, i.e. Data and GPUs.
ISSAM A. AL-ZINATI - UCASTI 48
49. Applying Deep Learning – Transfer Learning
Imagine if we can run Google Inception V3 as our model for image classification,
would not be Great!!
Transfer learning allow us to use these popular model by replacing the last fully
connected layer (1000 class classifier) with our classifier. Here we are using the
other layers as a feature extractors.
ISSAM A. AL-ZINATI - UCASTI 49
50. Applying Deep Learning – Transfer Learning
1) Fixed feature extractor
◦ Import one of the famous model with its weights.
◦ Replace last layer with custom classifier. It could be another fully connected NN or
other ML models like SVM.
◦ Train the new classifier based on the features and weights that this network had
extracted already.
ISSAM A. AL-ZINATI - UCASTI 50
51. Applying Deep Learning – Transfer Learning
2) Fine-Tunning
◦ Import one of the famous model with its weights.
◦ Replace last layer with custom classifier. It could be another fully connected NN or
other ML models like SVM.
◦ Fine-tune the weights of the pretrained network by continuing the backpropagation.
Also, it will train your new classifier at the same time.
ISSAM A. AL-ZINATI - UCASTI 51
52. Applying Deep Learning – Transfer Learning
3) Retraining
◦ Import one of the famous model with its weights.
◦ Replace last layer with custom classifier. It could be another fully connected NN or
other ML models like SVM.
◦ Retrain the whole model.
ISSAM A. AL-ZINATI - UCASTI 52
53. Applying Deep Learning – Transfer Learning
Best practices
1- New dataset is small and similar to original dataset. – use the first approach.
2- New dataset is large and similar to the original dataset. – use the second
approach.
3- New dataset is small but very different from the original dataset. – use the
second approach but only on early activations in network.
4- New dataset is large and very different from the original dataset. – use the
third approach.
ISSAM A. AL-ZINATI - UCASTI 53
54. Applying Deep Learning – Use GPU/Cloud
The last point that we should consider is using GPUs on cloud.
There are two famous provider who give you the ability to configure a machine with GPU for
good prices
1- Amazon AWS – By using its P2 instances you can train and run your model under 1$ per hour.
Another advantage of using AWS is the ability to use preconfigure images that has every thing
installed and configured for you.
2- Microsoft Azure – By using its NC instances you can train and run your model for 1.05$ per
hour.
ISSAM A. AL-ZINATI - UCASTI 54