SlideShare a Scribd company logo
Autoencoders
Autoencoders
● Autoencoders are
○ Artificial neural networks
○ Capable of learning efficient representations of the input data, called
codings, without any supervision
○ The training set is unlabeled.
● These codings typically have a much lower dimensionality than the input
data, making autoencoders useful for dimensionality reduction
Autoencoders
Autoencoders
Why use Autoencoders?
● Useful for dimensionality reduction
● Autoencoders act as powerful feature detectors,
● And they can be used for unsupervised pre-training of deep neural
networks
● Lastly, they are capable of randomly generating new data that looks very
similar to the training data; this is called a generative model
Autoencoders
Autoencoders
Autoencoders
Why use Autoencoders?
For example, we could train an autoencoder on pictures of faces, and it
would then be able to generate new faces
Autoencoders
● Surprisingly, autoencoders work by simply learning to copy their inputs
to their outputs
● This may sound like a trivial task, but we will see that constraining the
network in various ways can make it rather difficult
Autoencoders
Autoencoders
Autoencoders
For example
● You can limit the size of the internal representation, or you can add noise
to the inputs and train the network to recover the original inputs.
● These constraints prevent the autoencoder from trivially copying the
inputs directly to the outputs, which forces it to learn efficient ways of
representing the data
● In short, the codings are byproducts of the autoencoder’s attempt to
learn the identity function under some constraints
Autoencoders
What we’ll learn ?
● We will explain in more depth how autoencoders work
● What types of constraints can be imposed
● And how to implement them using TensorFlow, whether it is for
○ Dimensionality reduction,
○ Feature extraction,
○ Insupervised pretraining,
○ Or as generative models
Autoencoders
Autoencoders
Let’s try to understand
“why constraining an
autoencoder during training pushes it to discover and exploit
patterns in the data”
With an example
Efficient Data Representations
Autoencoders
Which of the following number sequences do you find the easiest
to memorize?
● 40, 27, 25, 36, 81, 57, 10, 73, 19, 68
● 50, 25, 76, 38, 19, 58, 29, 88, 44, 22, 11, 34, 17, 52, 26, 13, 40, 20
Efficient Data Representations
Autoencoders
● At first glance, it would seem that the first sequence should be easier,
since it is much shorter
● However, if you look carefully at the second sequence, you may notice
that it follows two simple rules:
○ Even numbers are followed by their half,
○ And odd numbers are followed by their triple plus one
This is a famous sequence known as the hailstone sequence
Efficient Data Representations
Autoencoders
● Once you notice this pattern, the second sequence becomes much easier
to memorize than the first because you only need to memorize the two
rules,
○ The first number,
○ And the length of the sequence
Efficient Data Representations
Autoencoders
● Note that if we could quickly and easily memorize very long sequences,
you would not care much about the existence of a pattern in the
second sequence
● You would just learn every number by heart, and that would be that
● It is the fact that it is hard to memorize long sequences that makes
it useful to recognize patterns,
● And hopefully this clarifies why constraining an autoencoder during
training pushes it to discover and exploit patterns in the data
Efficient Data Representations
Autoencoders
The relationship between memory, perception, and pattern matching was
famously studied by William Chase and Herbert Simon in the early
1970s
Efficient Data Representations
William Chase Herbert Simon
Autoencoders
Let’s see how their study of chess players is similar to an
Autoencoder
● They observed that expert chess players were able to memorize the
positions of all the pieces in a game by looking at the board for just 5
seconds,
● A task that most people would find impossible
● However, this was only the case when the pieces were placed in realistic
positions from actual games, not when the pieces were placed randomly.
Efficient Data Representations
Autoencoders
Let’s see how their study of chess players is similar to an
Autoencoder
● Chess experts don’t have a much better memory than you and I,
● They just see chess patterns more easily thanks to their experience with
the game
● Noticing patterns helps them store information efficiently
Efficient Data Representations
Autoencoders
Let’s see how their study of chess players is similar to an
Autoencoder
● Just like the chess players in this memory experiment, an autoencoder
○ looks at the inputs,
○ converts them to an efficient internal representation,
○ and then spits out something that (hopefully) looks very close to the
inputs
Efficient Data Representations
Autoencoders
Composition of Autoencoder
An autoencoder is always composed of two parts:
● An encoder or recognition network that converts the inputs to an
internal representation,
● Followed by a decoder or generative network that converts the
internal representation to the outputs
Efficient Data Representations
Autoencoders
Composition of Autoencoder
Efficient Data Representations
Autoencoders
Composition of Autoencoder
Efficient Data Representations
An autoencoder
typically has the same
architecture as a
Multi-Layer
Perceptron, except that
the number of neurons
in the output layer must
be equal to the
number of inputs
Autoencoders
There is just one hidden
layer composed of two
neurons (the encoder),
and one output layer
composed of three
neurons (the decoder)
Composition of Autoencoder
Efficient Data Representations
Autoencoders
Composition of Autoencoder
Efficient Data Representations
The outputs are often called
the reconstructions
since the autoencoder tries
to reconstruct the inputs,
and the cost function
contains a reconstruction
loss that penalizes the model
when the reconstructions are
different from the inputs.
Autoencoders
Because the internal
representation has a lower
dimensionality than the input
data, it is 2D instead of
3D, the autoencoder is said
to be undercomplete
Composition of Autoencoder
Efficient Data Representations
Autoencoders
Composition of Autoencoder
Efficient Data Representations
● An undercomplete
autoencoder cannot
trivially copy its inputs to
the codings, yet it must
find a way to output a
copy of its inputs
● It is forced to learn the
most important features in
the input data and drop
the unimportant ones
Autoencoders
Let’s implement a very simple undercomplete autoencoder for
dimensionality reduction
Autoencoders
PCA with an Undercomplete Linear Autoencoder
If the autoencoder uses only linear activations and the cost function is
the Mean Squared Error (MSE), then it can be shown that it ends up
performing Principal Component Analysis
Now we will build a simple linear autoencoder to perform PCA on a 3D
dataset, projecting it to 2D
Autoencoders
>>> import tensorflow as tf
>>> from tensorflow.contrib.layers import fully_connected
>>> n_inputs = 3 # 3D inputs
>>> n_hidden = 2 # 2D codings
>>> n_outputs = n_inputs
>>> learning_rate = 0.01
>>> X = tf.placeholder(tf.float32, shape=[None, n_inputs])
>>> hidden = fully_connected(X, n_hidden, activation_fn=None)
>>> outputs = fully_connected(hidden, n_outputs, activation_fn=None)
>>> reconstruction_loss = tf.reduce_mean(tf.square(outputs - X)) # MSE
>>> optimizer = tf.train.AdamOptimizer(learning_rate)
>>> training_op = optimizer.minimize(reconstruction_loss)
>>> init = tf.global_variables_initializer()
PCA with an Undercomplete Linear Autoencoder
Run it on Notebook
Autoencoders
PCA with an Undercomplete Linear Autoencoder
The two things to note in the previous code are are:
● The number of outputs is equal to the number of inputs
● To perform simple PCA, we set activation_fn=None i.e., all neurons
are linear, and the cost function is the MSE.
Autoencoders
Now let’s load the dataset, train the model on the training set, and use it to
encode the test set i.e., project it to 2D
>>> X_train, X_test = [...] # load the dataset
>>> n_iterations = 1000
>>> codings = hidden # the output of the hidden layer provides the
codings
>>> with tf.Session() as sess:
init.run()
for iteration in range(n_iterations):
training_op.run(feed_dict={X: X_train}) # no labels
codings_val = codings.eval(feed_dict={X: X_test})
PCA with an Undercomplete Linear Autoencoder
Run it on Notebook
Autoencoders
Above figure shows the original 3D dataset (at the left) and the output of
the autoencoder’s hidden layer (i.e., the coding layer, at the right)
PCA with an Undercomplete Linear Autoencoder
Autoencoders
PCA with an Undercomplete Linear Autoencoder
As you can see, the autoencoder found the best 2D plane to project the
data onto, preserving as much variance in the data as it could just like PCA
Questions?
https://discuss.cloudxlab.com
reachus@cloudxlab.com

More Related Content

What's hot

Autoencoders
AutoencodersAutoencoders
Autoencoders
Ashok Govindarajan
 
Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
Akash Goel
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
Gaurav Mittal
 
Introduction to Deep Learning
Introduction to Deep LearningIntroduction to Deep Learning
Introduction to Deep Learning
Oswald Campesato
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
EdutechLearners
 
Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)
Prakhar Rastogi
 
Tutorial on Deep Generative Models
 Tutorial on Deep Generative Models Tutorial on Deep Generative Models
Tutorial on Deep Generative Models
MLReview
 
An introduction to Machine Learning
An introduction to Machine LearningAn introduction to Machine Learning
An introduction to Machine Learningbutest
 
Unsupervised learning
Unsupervised learningUnsupervised learning
Unsupervised learning
amalalhait
 
Neural Networks: Multilayer Perceptron
Neural Networks: Multilayer PerceptronNeural Networks: Multilayer Perceptron
Neural Networks: Multilayer Perceptron
Mostafa G. M. Mostafa
 
Deep Learning Tutorial
Deep Learning TutorialDeep Learning Tutorial
Deep Learning Tutorial
Amr Rashed
 
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Universitat Politècnica de Catalunya
 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural Networks
NAGUR SHAREEF SHAIK
 
Hyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine LearningHyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine Learning
Francesco Casalegno
 
GAN - Theory and Applications
GAN - Theory and ApplicationsGAN - Theory and Applications
GAN - Theory and Applications
Emanuele Ghelfi
 
Heuristic Search Techniques Unit -II.ppt
Heuristic Search Techniques Unit -II.pptHeuristic Search Techniques Unit -II.ppt
Heuristic Search Techniques Unit -II.ppt
karthikaparthasarath
 
Recurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: TheoryRecurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: Theory
Andrii Gakhov
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Uncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial IntelligenceUncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial Intelligence
Experfy
 

What's hot (20)

Autoencoders
AutoencodersAutoencoders
Autoencoders
 
Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
Introduction to Deep Learning
Introduction to Deep LearningIntroduction to Deep Learning
Introduction to Deep Learning
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
 
Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)
 
Tutorial on Deep Generative Models
 Tutorial on Deep Generative Models Tutorial on Deep Generative Models
Tutorial on Deep Generative Models
 
An introduction to Machine Learning
An introduction to Machine LearningAn introduction to Machine Learning
An introduction to Machine Learning
 
Unsupervised learning
Unsupervised learningUnsupervised learning
Unsupervised learning
 
Neural Networks: Multilayer Perceptron
Neural Networks: Multilayer PerceptronNeural Networks: Multilayer Perceptron
Neural Networks: Multilayer Perceptron
 
Deep Learning Tutorial
Deep Learning TutorialDeep Learning Tutorial
Deep Learning Tutorial
 
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural Networks
 
Hyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine LearningHyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine Learning
 
GAN - Theory and Applications
GAN - Theory and ApplicationsGAN - Theory and Applications
GAN - Theory and Applications
 
Heuristic Search Techniques Unit -II.ppt
Heuristic Search Techniques Unit -II.pptHeuristic Search Techniques Unit -II.ppt
Heuristic Search Techniques Unit -II.ppt
 
Recurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: TheoryRecurrent Neural Networks. Part 1: Theory
Recurrent Neural Networks. Part 1: Theory
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
 
Uncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial IntelligenceUncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial Intelligence
 

Similar to Autoencoders

Autoecoders.pptx
Autoecoders.pptxAutoecoders.pptx
Autoecoders.pptx
MirzaJahanzeb5
 
UNIT-4.pdf
UNIT-4.pdfUNIT-4.pdf
UNIT-4.pdf
NiharikaThakur32
 
UNIT-4.pdf
UNIT-4.pdfUNIT-4.pdf
UNIT-4.pdf
NiharikaThakur32
 
UNIT-4.pptx
UNIT-4.pptxUNIT-4.pptx
UNIT-4.pptx
NiharikaThakur32
 
ENNEoS Presentation - HackMiami
ENNEoS Presentation - HackMiamiENNEoS Presentation - HackMiami
ENNEoS Presentation - HackMiami
Drew Kirkpatrick
 
ENNEoS Presentation - CackalackyCon
ENNEoS Presentation - CackalackyConENNEoS Presentation - CackalackyCon
ENNEoS Presentation - CackalackyCon
Drew Kirkpatrick
 
Unsupervised Feature Learning
Unsupervised Feature LearningUnsupervised Feature Learning
Unsupervised Feature LearningAmgad Muhammad
 
MachinaFiesta: A Vision into Machine Learning 🚀
MachinaFiesta: A Vision into Machine Learning 🚀MachinaFiesta: A Vision into Machine Learning 🚀
MachinaFiesta: A Vision into Machine Learning 🚀
GDSCNiT
 
Introduction to Parallelization and performance optimization
Introduction to Parallelization and performance optimizationIntroduction to Parallelization and performance optimization
Introduction to Parallelization and performance optimization
CSUC - Consorci de Serveis Universitaris de Catalunya
 
Denoising autoencoder by Harish.R
Denoising autoencoder by Harish.RDenoising autoencoder by Harish.R
Denoising autoencoder by Harish.R
HARISH R
 
ML in Android
ML in AndroidML in Android
ML in Android
Jose Antonio Corbacho
 
Data Analysis with TensorFlow in PostgreSQL
Data Analysis with TensorFlow in PostgreSQLData Analysis with TensorFlow in PostgreSQL
Data Analysis with TensorFlow in PostgreSQL
EDB
 
MySQL Query Optimisation 101
MySQL Query Optimisation 101MySQL Query Optimisation 101
MySQL Query Optimisation 101
Federico Razzoli
 
CSSC ML Workshop
CSSC ML WorkshopCSSC ML Workshop
CSSC ML Workshop
GDSC UofT Mississauga
 
Computer Architecture and Organization
Computer Architecture and OrganizationComputer Architecture and Organization
Computer Architecture and Organization
ssuserdfc773
 
Autoencoder
AutoencoderAutoencoder
Autoencoder
Wataru Hirota
 
Module 1_ Introduction.pptx
Module 1_ Introduction.pptxModule 1_ Introduction.pptx
Module 1_ Introduction.pptx
nikshaikh786
 
Introduction to Tensor Flow for Optical Character Recognition (OCR)
Introduction to Tensor Flow for Optical Character Recognition (OCR)Introduction to Tensor Flow for Optical Character Recognition (OCR)
Introduction to Tensor Flow for Optical Character Recognition (OCR)
Vincenzo Santopietro
 
Meetup 29042015
Meetup 29042015Meetup 29042015
Meetup 29042015lbishal
 
Practical ML
Practical MLPractical ML
Practical ML
Antonio Pitasi
 

Similar to Autoencoders (20)

Autoecoders.pptx
Autoecoders.pptxAutoecoders.pptx
Autoecoders.pptx
 
UNIT-4.pdf
UNIT-4.pdfUNIT-4.pdf
UNIT-4.pdf
 
UNIT-4.pdf
UNIT-4.pdfUNIT-4.pdf
UNIT-4.pdf
 
UNIT-4.pptx
UNIT-4.pptxUNIT-4.pptx
UNIT-4.pptx
 
ENNEoS Presentation - HackMiami
ENNEoS Presentation - HackMiamiENNEoS Presentation - HackMiami
ENNEoS Presentation - HackMiami
 
ENNEoS Presentation - CackalackyCon
ENNEoS Presentation - CackalackyConENNEoS Presentation - CackalackyCon
ENNEoS Presentation - CackalackyCon
 
Unsupervised Feature Learning
Unsupervised Feature LearningUnsupervised Feature Learning
Unsupervised Feature Learning
 
MachinaFiesta: A Vision into Machine Learning 🚀
MachinaFiesta: A Vision into Machine Learning 🚀MachinaFiesta: A Vision into Machine Learning 🚀
MachinaFiesta: A Vision into Machine Learning 🚀
 
Introduction to Parallelization and performance optimization
Introduction to Parallelization and performance optimizationIntroduction to Parallelization and performance optimization
Introduction to Parallelization and performance optimization
 
Denoising autoencoder by Harish.R
Denoising autoencoder by Harish.RDenoising autoencoder by Harish.R
Denoising autoencoder by Harish.R
 
ML in Android
ML in AndroidML in Android
ML in Android
 
Data Analysis with TensorFlow in PostgreSQL
Data Analysis with TensorFlow in PostgreSQLData Analysis with TensorFlow in PostgreSQL
Data Analysis with TensorFlow in PostgreSQL
 
MySQL Query Optimisation 101
MySQL Query Optimisation 101MySQL Query Optimisation 101
MySQL Query Optimisation 101
 
CSSC ML Workshop
CSSC ML WorkshopCSSC ML Workshop
CSSC ML Workshop
 
Computer Architecture and Organization
Computer Architecture and OrganizationComputer Architecture and Organization
Computer Architecture and Organization
 
Autoencoder
AutoencoderAutoencoder
Autoencoder
 
Module 1_ Introduction.pptx
Module 1_ Introduction.pptxModule 1_ Introduction.pptx
Module 1_ Introduction.pptx
 
Introduction to Tensor Flow for Optical Character Recognition (OCR)
Introduction to Tensor Flow for Optical Character Recognition (OCR)Introduction to Tensor Flow for Optical Character Recognition (OCR)
Introduction to Tensor Flow for Optical Character Recognition (OCR)
 
Meetup 29042015
Meetup 29042015Meetup 29042015
Meetup 29042015
 
Practical ML
Practical MLPractical ML
Practical ML
 

More from CloudxLab

Understanding computer vision with Deep Learning
Understanding computer vision with Deep LearningUnderstanding computer vision with Deep Learning
Understanding computer vision with Deep Learning
CloudxLab
 
Deep Learning Overview
Deep Learning OverviewDeep Learning Overview
Deep Learning Overview
CloudxLab
 
Recurrent Neural Networks
Recurrent Neural NetworksRecurrent Neural Networks
Recurrent Neural Networks
CloudxLab
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language Processing
CloudxLab
 
Naive Bayes
Naive BayesNaive Bayes
Naive Bayes
CloudxLab
 
Training Deep Neural Nets
Training Deep Neural NetsTraining Deep Neural Nets
Training Deep Neural Nets
CloudxLab
 
Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
CloudxLab
 
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
CloudxLab
 
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLabAdvanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
CloudxLab
 
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
CloudxLab
 
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
CloudxLab
 
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLabApache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
CloudxLab
 
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLabIntroduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
CloudxLab
 
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLabIntroduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
CloudxLab
 
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
CloudxLab
 
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLabIntroduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
CloudxLab
 
Introduction to Deep Learning | CloudxLab
Introduction to Deep Learning | CloudxLabIntroduction to Deep Learning | CloudxLab
Introduction to Deep Learning | CloudxLab
CloudxLab
 
Dimensionality Reduction | Machine Learning | CloudxLab
Dimensionality Reduction | Machine Learning | CloudxLabDimensionality Reduction | Machine Learning | CloudxLab
Dimensionality Reduction | Machine Learning | CloudxLab
CloudxLab
 
Ensemble Learning and Random Forests
Ensemble Learning and Random ForestsEnsemble Learning and Random Forests
Ensemble Learning and Random Forests
CloudxLab
 
Decision Trees
Decision TreesDecision Trees
Decision Trees
CloudxLab
 

More from CloudxLab (20)

Understanding computer vision with Deep Learning
Understanding computer vision with Deep LearningUnderstanding computer vision with Deep Learning
Understanding computer vision with Deep Learning
 
Deep Learning Overview
Deep Learning OverviewDeep Learning Overview
Deep Learning Overview
 
Recurrent Neural Networks
Recurrent Neural NetworksRecurrent Neural Networks
Recurrent Neural Networks
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language Processing
 
Naive Bayes
Naive BayesNaive Bayes
Naive Bayes
 
Training Deep Neural Nets
Training Deep Neural NetsTraining Deep Neural Nets
Training Deep Neural Nets
 
Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
 
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
Apache Spark - Key Value RDD - Transformations | Big Data Hadoop Spark Tutori...
 
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLabAdvanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
Advanced Spark Programming - Part 2 | Big Data Hadoop Spark Tutorial | CloudxLab
 
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 2 | Big Data Hadoop Spark Tutori...
 
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
Apache Spark - Dataframes & Spark SQL - Part 1 | Big Data Hadoop Spark Tutori...
 
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLabApache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
Apache Spark - Running on a Cluster | Big Data Hadoop Spark Tutorial | CloudxLab
 
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLabIntroduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to SparkR | Big Data Hadoop Spark Tutorial | CloudxLab
 
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLabIntroduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
Introduction to NoSQL | Big Data Hadoop Spark Tutorial | CloudxLab
 
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
Introduction to MapReduce - Hadoop Streaming | Big Data Hadoop Spark Tutorial...
 
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLabIntroduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLab
 
Introduction to Deep Learning | CloudxLab
Introduction to Deep Learning | CloudxLabIntroduction to Deep Learning | CloudxLab
Introduction to Deep Learning | CloudxLab
 
Dimensionality Reduction | Machine Learning | CloudxLab
Dimensionality Reduction | Machine Learning | CloudxLabDimensionality Reduction | Machine Learning | CloudxLab
Dimensionality Reduction | Machine Learning | CloudxLab
 
Ensemble Learning and Random Forests
Ensemble Learning and Random ForestsEnsemble Learning and Random Forests
Ensemble Learning and Random Forests
 
Decision Trees
Decision TreesDecision Trees
Decision Trees
 

Recently uploaded

UiPath Test Automation using UiPath Test Suite series, part 4
UiPath Test Automation using UiPath Test Suite series, part 4UiPath Test Automation using UiPath Test Suite series, part 4
UiPath Test Automation using UiPath Test Suite series, part 4
DianaGray10
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
Paul Groth
 
Bits & Pixels using AI for Good.........
Bits & Pixels using AI for Good.........Bits & Pixels using AI for Good.........
Bits & Pixels using AI for Good.........
Alison B. Lowndes
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
James Anderson
 
Connector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a buttonConnector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a button
DianaGray10
 
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Product School
 
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
Product School
 
Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*
Frank van Harmelen
 
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdfFIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
Safe Software
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
Assuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyesAssuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyes
ThousandEyes
 
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
Product School
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
Product School
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
OnBoard
 
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Thierry Lestable
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
DevOps and Testing slides at DASA Connect
DevOps and Testing slides at DASA ConnectDevOps and Testing slides at DASA Connect
DevOps and Testing slides at DASA Connect
Kari Kakkonen
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
Jemma Hussein Allen
 

Recently uploaded (20)

UiPath Test Automation using UiPath Test Suite series, part 4
UiPath Test Automation using UiPath Test Suite series, part 4UiPath Test Automation using UiPath Test Suite series, part 4
UiPath Test Automation using UiPath Test Suite series, part 4
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
 
Bits & Pixels using AI for Good.........
Bits & Pixels using AI for Good.........Bits & Pixels using AI for Good.........
Bits & Pixels using AI for Good.........
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
 
Connector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a buttonConnector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a button
 
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...
 
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
From Daily Decisions to Bottom Line: Connecting Product Work to Revenue by VP...
 
Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*Neuro-symbolic is not enough, we need neuro-*semantic*
Neuro-symbolic is not enough, we need neuro-*semantic*
 
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdfFIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
Assuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyesAssuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyes
 
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
 
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
DevOps and Testing slides at DASA Connect
DevOps and Testing slides at DASA ConnectDevOps and Testing slides at DASA Connect
DevOps and Testing slides at DASA Connect
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
 

Autoencoders

  • 2. Autoencoders ● Autoencoders are ○ Artificial neural networks ○ Capable of learning efficient representations of the input data, called codings, without any supervision ○ The training set is unlabeled. ● These codings typically have a much lower dimensionality than the input data, making autoencoders useful for dimensionality reduction Autoencoders
  • 3. Autoencoders Why use Autoencoders? ● Useful for dimensionality reduction ● Autoencoders act as powerful feature detectors, ● And they can be used for unsupervised pre-training of deep neural networks ● Lastly, they are capable of randomly generating new data that looks very similar to the training data; this is called a generative model Autoencoders
  • 4. Autoencoders Autoencoders Why use Autoencoders? For example, we could train an autoencoder on pictures of faces, and it would then be able to generate new faces
  • 5. Autoencoders ● Surprisingly, autoencoders work by simply learning to copy their inputs to their outputs ● This may sound like a trivial task, but we will see that constraining the network in various ways can make it rather difficult Autoencoders
  • 6. Autoencoders Autoencoders For example ● You can limit the size of the internal representation, or you can add noise to the inputs and train the network to recover the original inputs. ● These constraints prevent the autoencoder from trivially copying the inputs directly to the outputs, which forces it to learn efficient ways of representing the data ● In short, the codings are byproducts of the autoencoder’s attempt to learn the identity function under some constraints
  • 7. Autoencoders What we’ll learn ? ● We will explain in more depth how autoencoders work ● What types of constraints can be imposed ● And how to implement them using TensorFlow, whether it is for ○ Dimensionality reduction, ○ Feature extraction, ○ Insupervised pretraining, ○ Or as generative models Autoencoders
  • 8. Autoencoders Let’s try to understand “why constraining an autoencoder during training pushes it to discover and exploit patterns in the data” With an example Efficient Data Representations
  • 9. Autoencoders Which of the following number sequences do you find the easiest to memorize? ● 40, 27, 25, 36, 81, 57, 10, 73, 19, 68 ● 50, 25, 76, 38, 19, 58, 29, 88, 44, 22, 11, 34, 17, 52, 26, 13, 40, 20 Efficient Data Representations
  • 10. Autoencoders ● At first glance, it would seem that the first sequence should be easier, since it is much shorter ● However, if you look carefully at the second sequence, you may notice that it follows two simple rules: ○ Even numbers are followed by their half, ○ And odd numbers are followed by their triple plus one This is a famous sequence known as the hailstone sequence Efficient Data Representations
  • 11. Autoencoders ● Once you notice this pattern, the second sequence becomes much easier to memorize than the first because you only need to memorize the two rules, ○ The first number, ○ And the length of the sequence Efficient Data Representations
  • 12. Autoencoders ● Note that if we could quickly and easily memorize very long sequences, you would not care much about the existence of a pattern in the second sequence ● You would just learn every number by heart, and that would be that ● It is the fact that it is hard to memorize long sequences that makes it useful to recognize patterns, ● And hopefully this clarifies why constraining an autoencoder during training pushes it to discover and exploit patterns in the data Efficient Data Representations
  • 13. Autoencoders The relationship between memory, perception, and pattern matching was famously studied by William Chase and Herbert Simon in the early 1970s Efficient Data Representations William Chase Herbert Simon
  • 14. Autoencoders Let’s see how their study of chess players is similar to an Autoencoder ● They observed that expert chess players were able to memorize the positions of all the pieces in a game by looking at the board for just 5 seconds, ● A task that most people would find impossible ● However, this was only the case when the pieces were placed in realistic positions from actual games, not when the pieces were placed randomly. Efficient Data Representations
  • 15. Autoencoders Let’s see how their study of chess players is similar to an Autoencoder ● Chess experts don’t have a much better memory than you and I, ● They just see chess patterns more easily thanks to their experience with the game ● Noticing patterns helps them store information efficiently Efficient Data Representations
  • 16. Autoencoders Let’s see how their study of chess players is similar to an Autoencoder ● Just like the chess players in this memory experiment, an autoencoder ○ looks at the inputs, ○ converts them to an efficient internal representation, ○ and then spits out something that (hopefully) looks very close to the inputs Efficient Data Representations
  • 17. Autoencoders Composition of Autoencoder An autoencoder is always composed of two parts: ● An encoder or recognition network that converts the inputs to an internal representation, ● Followed by a decoder or generative network that converts the internal representation to the outputs Efficient Data Representations
  • 19. Autoencoders Composition of Autoencoder Efficient Data Representations An autoencoder typically has the same architecture as a Multi-Layer Perceptron, except that the number of neurons in the output layer must be equal to the number of inputs
  • 20. Autoencoders There is just one hidden layer composed of two neurons (the encoder), and one output layer composed of three neurons (the decoder) Composition of Autoencoder Efficient Data Representations
  • 21. Autoencoders Composition of Autoencoder Efficient Data Representations The outputs are often called the reconstructions since the autoencoder tries to reconstruct the inputs, and the cost function contains a reconstruction loss that penalizes the model when the reconstructions are different from the inputs.
  • 22. Autoencoders Because the internal representation has a lower dimensionality than the input data, it is 2D instead of 3D, the autoencoder is said to be undercomplete Composition of Autoencoder Efficient Data Representations
  • 23. Autoencoders Composition of Autoencoder Efficient Data Representations ● An undercomplete autoencoder cannot trivially copy its inputs to the codings, yet it must find a way to output a copy of its inputs ● It is forced to learn the most important features in the input data and drop the unimportant ones
  • 24. Autoencoders Let’s implement a very simple undercomplete autoencoder for dimensionality reduction
  • 25. Autoencoders PCA with an Undercomplete Linear Autoencoder If the autoencoder uses only linear activations and the cost function is the Mean Squared Error (MSE), then it can be shown that it ends up performing Principal Component Analysis Now we will build a simple linear autoencoder to perform PCA on a 3D dataset, projecting it to 2D
  • 26. Autoencoders >>> import tensorflow as tf >>> from tensorflow.contrib.layers import fully_connected >>> n_inputs = 3 # 3D inputs >>> n_hidden = 2 # 2D codings >>> n_outputs = n_inputs >>> learning_rate = 0.01 >>> X = tf.placeholder(tf.float32, shape=[None, n_inputs]) >>> hidden = fully_connected(X, n_hidden, activation_fn=None) >>> outputs = fully_connected(hidden, n_outputs, activation_fn=None) >>> reconstruction_loss = tf.reduce_mean(tf.square(outputs - X)) # MSE >>> optimizer = tf.train.AdamOptimizer(learning_rate) >>> training_op = optimizer.minimize(reconstruction_loss) >>> init = tf.global_variables_initializer() PCA with an Undercomplete Linear Autoencoder Run it on Notebook
  • 27. Autoencoders PCA with an Undercomplete Linear Autoencoder The two things to note in the previous code are are: ● The number of outputs is equal to the number of inputs ● To perform simple PCA, we set activation_fn=None i.e., all neurons are linear, and the cost function is the MSE.
  • 28. Autoencoders Now let’s load the dataset, train the model on the training set, and use it to encode the test set i.e., project it to 2D >>> X_train, X_test = [...] # load the dataset >>> n_iterations = 1000 >>> codings = hidden # the output of the hidden layer provides the codings >>> with tf.Session() as sess: init.run() for iteration in range(n_iterations): training_op.run(feed_dict={X: X_train}) # no labels codings_val = codings.eval(feed_dict={X: X_test}) PCA with an Undercomplete Linear Autoencoder Run it on Notebook
  • 29. Autoencoders Above figure shows the original 3D dataset (at the left) and the output of the autoencoder’s hidden layer (i.e., the coding layer, at the right) PCA with an Undercomplete Linear Autoencoder
  • 30. Autoencoders PCA with an Undercomplete Linear Autoencoder As you can see, the autoencoder found the best 2D plane to project the data onto, preserving as much variance in the data as it could just like PCA