SlideShare a Scribd company logo
1 of 36
Plug & Play Generative Networks:
Conditional Iterative Generation of
Images in Latent Space
Anh Nguyen, Jason Yosinski, Yoshua Bengio, Alexey Dosovitskiy, Jeff Clune
2017
By Safaa Alnabulsi
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 2
Motivation
Challenges solved in generating Images:
 High quality images at higher resolutions (227 × 227):
Current image generative models often work well at low resolutions (e.g. 32 ×
32), but struggle to generate high-resolution (e.g. 128 × 128 or higher), due to many
challenges including difficulty in training and computationally expensive sampling
procedures.
Author | Safaa Alnabulsi 3
Motivation
Challenges solved in generating Images:
 High resolutions images
 Realistic and diverse samples within a class
Author | Safaa Alnabulsi 4
Motivation
Challenges solved in generating Images:
 High resolutions images
 Realistic and diverse samples within a class
 Does it for all 1000 ImageNet categories
Author | Safaa Alnabulsi 5
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 6
What is GAN?
 GAN consists of two models:
- Generative model (G): generates new data instances / models the distribution of
individual classes. p(x|c)
- Discriminative model (D): evaluates them for authenticity/ learns the boundary
between classes
Author | Safaa Alnabulsi 7
How does GAN work?
GANs are formulated as a game between two networks and it is important (and tricky!) to
keep them in balance!
Author | Safaa Alnabulsi
Cite:
Real / Fake
8
Plug and Play Generative Networks
Author | Safaa Alnabulsi 9
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 10
Probabilistic interpretation of iterative
image generation methods
MALA-approx: Metropolis-adjusted Langevin algorithm (MALA )+ Markov chain
Monte Carlo (MCMC) which uses the following transition operator:
Author | Safaa Alnabulsi 11
Probabilistic framework for Activation
Maximization
take a step from the current
image xt toward one that looks
more like a generic image (an
image from any class).
take a step from the current image
xt toward an image that causes the
classifier to output higher
confidence in the chosen class.
add a small amount of noise to
jump around the search space
to encourage a diversity of
images
Author | Safaa Alnabulsi
prior conditionjoint model
12
Probabilistic framework for Activation
Maximization
Author | Safaa Alnabulsi 13
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 14
Methods
Author | Safaa Alnabulsi 15
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 16
Method | DGN-AM: sampling without a
learned prior
 Deep Generator Network-based Activation Maximization
 Perform sampling in this lower-dimensional h-space.
 h in this case represents features extracted from the first fully connected layer
(called fc6) of a pre-trained AlexNet on 1000-class ImageNet classification
network
Author | Safaa Alnabulsi 17
Method | DGN-AM: sampling without a
learned prior
Once we trained the network G we find the equation for the MALA algorithm
No learned prior No noise
Author | Safaa Alnabulsi 18
Method | DGN-AM: sampling without a
learned prior
 proc:
• Sampling in input space h is faster
than image space x.
• produce realistic images at a high
resolution.
• It can also produce interesting new
types of images that G never saw
during training
Author | Safaa Alnabulsi
 cons:
• Low mixing speed
• Same image after many steps
• Lack of diversity
19
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 20
Method | PPGN-x: DAE model of p(x)
 What is DAE (Denoising Autoencoder) ?
DAE
Noisex
+
R(x)
Author | Safaa Alnabulsi 21
Method | PPGN-x: DAE model of p(x)
Author | Safaa Alnabulsi 22
Method | PPGN-x: DAE model of p(x)
 proc:
• sampling from the entire model.
Author | Safaa Alnabulsi
 cons:
• it models the data distribution
poorly.
• the chain mixes slowly (sampling in
the high-dimensional image space)
23
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h) <-- (This paper‘s model)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 24
Method | PPGN-h: Generator and DAE
model of p(h)
Author | Safaa Alnabulsi
To address the poor mixing speed of DGN-AM, they incorporate a proper p(h) prior
learned via a DAE into the sampling procedure
25
Method | PPGN-h: Generator and DAE
model of p(h)
Author | Safaa Alnabulsi
The update rule to sample h from this model:
26
Method | PPGN-h: Generator and DAE model
of p(h)
 proc:
• the chain mixes faster than PPGN-x
Author | Safaa Alnabulsi
 cons:
• Sample from PPGN-h are
qualitatively similar to those from
DGN-AM
• Samples still lack quality and
diversity. ( poor p(h) model
learned by the DAE).
27
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Resuls
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 28
Generating images with different condition
networks | Captioning
 PPGNs can be flexibly turned into a text-to-image model by combining the
prior with an image captioning network, and this process does not even
require additional training.
Author | Safaa Alnabulsi 29
Generating images with different condition
networks | Multifaceted Feature Visualization
 Instead of conditioning on a class output neuron, here we condition on a
hidden neuron, revealing many facets that a neuron has learned to detect
Author | Safaa Alnabulsi 30
Figure 6: Images synthesized to activate a hidden neuron (number 196) previously identified as a
“face detector neuron”
Index
 Motivation
 What is GAN?
 How does GAN work?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 31
Inpainting
 Because PPGNs can be interpreted
probabilistically, we can also sample
from them conditioned on part of an
image (in addition to the class
condition) to perform inpainting.
Author | Safaa Alnabulsi 32
Index
 Motivation
 What is GAN?
 Probabilistic interpretation of iterative image generation methods
 Methods and Experiments
o PDGN-AM: sampling without a learned prior
o PGN-x: DAE model of p(x)
o PPGN-h: Generator and DAE model of p(h)
 Additional Results
o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization )
o Inpainting
 Conclusion
Author | Safaa Alnabulsi 33
Conclusion
 P&P model generates images in 227x227 which is considered HD in Image
generation models.
 The most useful property of PPGN is the capability of “plug and play”—
allowing one to drop in a replaceable condition network and generate images
according to a condition specified ( Classs, caption or neuron ) at test time.
 Usage of PPGNs to synthesize images for videos or create arts with one or
even multiple condition networks at the same time.
 The approach is modality-agnostic and can be applied to many types of
data.
Author | Safaa Alnabulsi 34
Author | Safaa Alnabulsi 35
Author | Safaa Alnabulsi 36
Portrait of Edmond Belamy, 2018, created by GAN (Generative Adversarial Network).
Sold for $432,500 on 25 October at Christie’s in New York. Image © Obvious

More Related Content

Similar to Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space

Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...
Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...
Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...Lviv Data Science Summer School
 
Image colorization
Image colorizationImage colorization
Image colorizationYash Saraf
 
Image colorization
Image colorizationImage colorization
Image colorizationYash Saraf
 
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image TransformationDeep Learning JP
 
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)Matthias Trapp
 
Learning from Simulated and Unsupervised Images through Adversarial Training....
Learning from Simulated and Unsupervised Images through Adversarial Training....Learning from Simulated and Unsupervised Images through Adversarial Training....
Learning from Simulated and Unsupervised Images through Adversarial Training....eraser Juan José Calderón
 
Introduction to Generative Adversarial Networks (GANs)
Introduction to Generative Adversarial Networks (GANs)Introduction to Generative Adversarial Networks (GANs)
Introduction to Generative Adversarial Networks (GANs)Appsilon Data Science
 
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...Mirsaeid Abolghasemi
 
Evaluation of conditional images synthesis: generating a photorealistic image...
Evaluation of conditional images synthesis: generating a photorealistic image...Evaluation of conditional images synthesis: generating a photorealistic image...
Evaluation of conditional images synthesis: generating a photorealistic image...SamanthaGallone
 
Efficient aggregation for graph summarization
Efficient aggregation for graph summarizationEfficient aggregation for graph summarization
Efficient aggregation for graph summarizationaftab alam
 
Generative modeling with Convolutional Neural Networks
Generative modeling with Convolutional Neural NetworksGenerative modeling with Convolutional Neural Networks
Generative modeling with Convolutional Neural NetworksDenis Dus
 
Deep Learning for Computer Vision: Generative models and adversarial training...
Deep Learning for Computer Vision: Generative models and adversarial training...Deep Learning for Computer Vision: Generative models and adversarial training...
Deep Learning for Computer Vision: Generative models and adversarial training...Universitat Politècnica de Catalunya
 
Implementing Neural Style Transfer
Implementing Neural Style Transfer Implementing Neural Style Transfer
Implementing Neural Style Transfer Tahsin Mayeesha
 
The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)James McMurray
 
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...Universitat Politècnica de Catalunya
 
Using GANs to improve generalization in a semi-supervised setting - trying it...
Using GANs to improve generalization in a semi-supervised setting - trying it...Using GANs to improve generalization in a semi-supervised setting - trying it...
Using GANs to improve generalization in a semi-supervised setting - trying it...PyData
 
Semi-supervised learning with GANs
Semi-supervised learning with GANsSemi-supervised learning with GANs
Semi-supervised learning with GANsterek47
 

Similar to Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space (20)

Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...
Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...
Master defence 2020 - Vadym Korshunov - Region-Selected Image Generation with...
 
Image colorization
Image colorizationImage colorization
Image colorization
 
Image colorization
Image colorizationImage colorization
Image colorization
 
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation
[DLHacks 実装]Perceptual Adversarial Networks for Image-to-Image Transformation
 
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)
Interactive Stereoscopic Rendering for Non-Planar Projections (GRAPP 2009)
 
Learning from Simulated and Unsupervised Images through Adversarial Training....
Learning from Simulated and Unsupervised Images through Adversarial Training....Learning from Simulated and Unsupervised Images through Adversarial Training....
Learning from Simulated and Unsupervised Images through Adversarial Training....
 
Introduction to Generative Adversarial Networks (GANs)
Introduction to Generative Adversarial Networks (GANs)Introduction to Generative Adversarial Networks (GANs)
Introduction to Generative Adversarial Networks (GANs)
 
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
 
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...
A Survey of Generative Adversarial Neural Networks (GAN) for Text-to-Image Sy...
 
Evaluation of conditional images synthesis: generating a photorealistic image...
Evaluation of conditional images synthesis: generating a photorealistic image...Evaluation of conditional images synthesis: generating a photorealistic image...
Evaluation of conditional images synthesis: generating a photorealistic image...
 
Efficient aggregation for graph summarization
Efficient aggregation for graph summarizationEfficient aggregation for graph summarization
Efficient aggregation for graph summarization
 
Generative modeling with Convolutional Neural Networks
Generative modeling with Convolutional Neural NetworksGenerative modeling with Convolutional Neural Networks
Generative modeling with Convolutional Neural Networks
 
Deep Learning for Computer Vision: Generative models and adversarial training...
Deep Learning for Computer Vision: Generative models and adversarial training...Deep Learning for Computer Vision: Generative models and adversarial training...
Deep Learning for Computer Vision: Generative models and adversarial training...
 
Implementing Neural Style Transfer
Implementing Neural Style Transfer Implementing Neural Style Transfer
Implementing Neural Style Transfer
 
The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)The Gaussian Process Latent Variable Model (GPLVM)
The Gaussian Process Latent Variable Model (GPLVM)
 
M sc thesis proposal v4
M sc thesis proposal v4M sc thesis proposal v4
M sc thesis proposal v4
 
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
 
2021 04-01-dalle
2021 04-01-dalle2021 04-01-dalle
2021 04-01-dalle
 
Using GANs to improve generalization in a semi-supervised setting - trying it...
Using GANs to improve generalization in a semi-supervised setting - trying it...Using GANs to improve generalization in a semi-supervised setting - trying it...
Using GANs to improve generalization in a semi-supervised setting - trying it...
 
Semi-supervised learning with GANs
Semi-supervised learning with GANsSemi-supervised learning with GANs
Semi-supervised learning with GANs
 

Recently uploaded

B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxStephen266013
 
Ukraine War presentation: KNOW THE BASICS
Ukraine War presentation: KNOW THE BASICSUkraine War presentation: KNOW THE BASICS
Ukraine War presentation: KNOW THE BASICSAishani27
 
Midocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxMidocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxolyaivanovalion
 
04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationshipsccctableauusergroup
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfLars Albertsson
 
Introduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxIntroduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxfirstjob4
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxolyaivanovalion
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxJohnnyPlasten
 
BigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxBigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxolyaivanovalion
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...Suhani Kapoor
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Callshivangimorya083
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysismanisha194592
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfMarinCaroMartnezBerg
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptSonatrach
 
Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystSamantha Rae Coolbeth
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxEmmanuel Dauda
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptxthyngster
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxolyaivanovalion
 

Recently uploaded (20)

B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docx
 
Ukraine War presentation: KNOW THE BASICS
Ukraine War presentation: KNOW THE BASICSUkraine War presentation: KNOW THE BASICS
Ukraine War presentation: KNOW THE BASICS
 
Midocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxMidocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFx
 
04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdf
 
Introduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxIntroduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptx
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptx
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptx
 
BigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptxBigBuy dropshipping via API with DroFx.pptx
BigBuy dropshipping via API with DroFx.pptx
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysis
 
FESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdfFESE Capital Markets Fact Sheet 2024 Q1.pdf
FESE Capital Markets Fact Sheet 2024 Q1.pdf
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
 
Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data Analyst
 
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptx
 
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in  KishangarhDelhi 99530 vip 56974 Genuine Escort Service Call Girls in  Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptx
 

Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space

  • 1. Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space Anh Nguyen, Jason Yosinski, Yoshua Bengio, Alexey Dosovitskiy, Jeff Clune 2017 By Safaa Alnabulsi
  • 2. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 2
  • 3. Motivation Challenges solved in generating Images:  High quality images at higher resolutions (227 × 227): Current image generative models often work well at low resolutions (e.g. 32 × 32), but struggle to generate high-resolution (e.g. 128 × 128 or higher), due to many challenges including difficulty in training and computationally expensive sampling procedures. Author | Safaa Alnabulsi 3
  • 4. Motivation Challenges solved in generating Images:  High resolutions images  Realistic and diverse samples within a class Author | Safaa Alnabulsi 4
  • 5. Motivation Challenges solved in generating Images:  High resolutions images  Realistic and diverse samples within a class  Does it for all 1000 ImageNet categories Author | Safaa Alnabulsi 5
  • 6. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 6
  • 7. What is GAN?  GAN consists of two models: - Generative model (G): generates new data instances / models the distribution of individual classes. p(x|c) - Discriminative model (D): evaluates them for authenticity/ learns the boundary between classes Author | Safaa Alnabulsi 7
  • 8. How does GAN work? GANs are formulated as a game between two networks and it is important (and tricky!) to keep them in balance! Author | Safaa Alnabulsi Cite: Real / Fake 8
  • 9. Plug and Play Generative Networks Author | Safaa Alnabulsi 9
  • 10. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 10
  • 11. Probabilistic interpretation of iterative image generation methods MALA-approx: Metropolis-adjusted Langevin algorithm (MALA )+ Markov chain Monte Carlo (MCMC) which uses the following transition operator: Author | Safaa Alnabulsi 11
  • 12. Probabilistic framework for Activation Maximization take a step from the current image xt toward one that looks more like a generic image (an image from any class). take a step from the current image xt toward an image that causes the classifier to output higher confidence in the chosen class. add a small amount of noise to jump around the search space to encourage a diversity of images Author | Safaa Alnabulsi prior conditionjoint model 12
  • 13. Probabilistic framework for Activation Maximization Author | Safaa Alnabulsi 13
  • 14. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 14
  • 15. Methods Author | Safaa Alnabulsi 15
  • 16. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 16
  • 17. Method | DGN-AM: sampling without a learned prior  Deep Generator Network-based Activation Maximization  Perform sampling in this lower-dimensional h-space.  h in this case represents features extracted from the first fully connected layer (called fc6) of a pre-trained AlexNet on 1000-class ImageNet classification network Author | Safaa Alnabulsi 17
  • 18. Method | DGN-AM: sampling without a learned prior Once we trained the network G we find the equation for the MALA algorithm No learned prior No noise Author | Safaa Alnabulsi 18
  • 19. Method | DGN-AM: sampling without a learned prior  proc: • Sampling in input space h is faster than image space x. • produce realistic images at a high resolution. • It can also produce interesting new types of images that G never saw during training Author | Safaa Alnabulsi  cons: • Low mixing speed • Same image after many steps • Lack of diversity 19
  • 20. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 20
  • 21. Method | PPGN-x: DAE model of p(x)  What is DAE (Denoising Autoencoder) ? DAE Noisex + R(x) Author | Safaa Alnabulsi 21
  • 22. Method | PPGN-x: DAE model of p(x) Author | Safaa Alnabulsi 22
  • 23. Method | PPGN-x: DAE model of p(x)  proc: • sampling from the entire model. Author | Safaa Alnabulsi  cons: • it models the data distribution poorly. • the chain mixes slowly (sampling in the high-dimensional image space) 23
  • 24. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h) <-- (This paper‘s model)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 24
  • 25. Method | PPGN-h: Generator and DAE model of p(h) Author | Safaa Alnabulsi To address the poor mixing speed of DGN-AM, they incorporate a proper p(h) prior learned via a DAE into the sampling procedure 25
  • 26. Method | PPGN-h: Generator and DAE model of p(h) Author | Safaa Alnabulsi The update rule to sample h from this model: 26
  • 27. Method | PPGN-h: Generator and DAE model of p(h)  proc: • the chain mixes faster than PPGN-x Author | Safaa Alnabulsi  cons: • Sample from PPGN-h are qualitatively similar to those from DGN-AM • Samples still lack quality and diversity. ( poor p(h) model learned by the DAE). 27
  • 28. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Resuls o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 28
  • 29. Generating images with different condition networks | Captioning  PPGNs can be flexibly turned into a text-to-image model by combining the prior with an image captioning network, and this process does not even require additional training. Author | Safaa Alnabulsi 29
  • 30. Generating images with different condition networks | Multifaceted Feature Visualization  Instead of conditioning on a class output neuron, here we condition on a hidden neuron, revealing many facets that a neuron has learned to detect Author | Safaa Alnabulsi 30 Figure 6: Images synthesized to activate a hidden neuron (number 196) previously identified as a “face detector neuron”
  • 31. Index  Motivation  What is GAN?  How does GAN work?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 31
  • 32. Inpainting  Because PPGNs can be interpreted probabilistically, we can also sample from them conditioned on part of an image (in addition to the class condition) to perform inpainting. Author | Safaa Alnabulsi 32
  • 33. Index  Motivation  What is GAN?  Probabilistic interpretation of iterative image generation methods  Methods and Experiments o PDGN-AM: sampling without a learned prior o PGN-x: DAE model of p(x) o PPGN-h: Generator and DAE model of p(h)  Additional Results o Generating images with different condition networks (Captioning , Multifaceted Feature Visualization ) o Inpainting  Conclusion Author | Safaa Alnabulsi 33
  • 34. Conclusion  P&P model generates images in 227x227 which is considered HD in Image generation models.  The most useful property of PPGN is the capability of “plug and play”— allowing one to drop in a replaceable condition network and generate images according to a condition specified ( Classs, caption or neuron ) at test time.  Usage of PPGNs to synthesize images for videos or create arts with one or even multiple condition networks at the same time.  The approach is modality-agnostic and can be applied to many types of data. Author | Safaa Alnabulsi 34
  • 35. Author | Safaa Alnabulsi 35
  • 36. Author | Safaa Alnabulsi 36 Portrait of Edmond Belamy, 2018, created by GAN (Generative Adversarial Network). Sold for $432,500 on 25 October at Christie’s in New York. Image © Obvious

Editor's Notes

  1. Generating high-resolution, photo-realistic images has been a long-standing goal in machine learning
  2. Here are the steps a GAN takes: - The generator takes in random numbers and returns an image. - This generated image is fed into the discriminator alongside a stream of images taken from the actual dataset. - The discriminator takes in both real and fake images and returns probabilities, a number between 0 and 1, with 1 representing a prediction of authenticity and 0 representing fake. So you have a double feedback loop: - The discriminator is in a feedback loop with the ground truth of the images, which we know. - The generator is in a feedback loop with the discriminator. Ref: https://skymind.ai/wiki/generative-adversarial-network-gan
  3. In Plug and play, to generate images they combine two different types of networks: The Generator can be thought of as generic painter that draws wide variety of images and show them to the conditioner Which looks at the images and tell the generator what to draw next And this process is iterative replaceable conditioner
  4. for obtaining random samples, they use the Metropolis-adjusted Langevin algorithm (MALA) is a Markov chain Monte Carlo (MCMC) method probability distribution for which direct sampling is difficult. The furture state is based on the current step + gradient + noise
  5. joint model p(x, y), which can be decomposed into an image model and a clas- sification model:
  6. In every update they encourage the image to be realistic, class specific and diverse. It mixes fast because we sample in the latent space. I will cover that in the upcoming slides.
  7. In this paper, they propose a class of models called PPGNs that are composed of 1) a generator network G that is trained to draw a wide range of image types, and 2) a replaceable “condition” network C that tells G what to draw Starting from b
  8. Instead of sampling in the image space (i.e. in the space of individual pixels) they sample in the abstract, high-level feature space h of a generator G trained to reconstruct images x from compressed features h extracted from a pre-trained encoder E (f). So the input in here is no longer an image but rather a random noise vector .
  9. Because the generator network was trained to produce realistic images, it serves as a prior on p(x) since it ideally can only generate real images. However, this model has no learned prior on p(h) Explain the update equation: x is a deterministic variable, so we can simplify the model they define a Gaussian p(h) centered at 0 The final h is pushed through G to produce an image sample.
  10. They train a DAE for images and incorporate it to the sampling procedure as a p(x) prior to avoid fooling examples in here explain what the dae? A trained network which reconstructs an image after adding some random noise to it. We use it to calculate the score function p(x) by subtracting the output from input and divide them by the added noise (squared segma)
  11. We take here the MALA equation and put the formula from the DAE in first term, and this would represent how we generates images.
  12. I will be mentioning the mixing speed in the presnetion which means ( "mixing time" of the Markov chain ): The mixing time has a direct impact on sampling quality since, the smaller the mixing time, the faster the convergence of the Markov chain to the stationary distribution, and the smaller the correlation in the samples. In this approach we notice that and says the cons.
  13. In here they included all three ǫ terms.
  14. "mixing time" of the Markov chain: The mixing time has a direct impact on sampling quality since, the smaller the mixing time, the faster the convergence of the Markov chain to the stationary distribution, and the smaller the correlation in the samples.
  15. Inpainting is filling in missing pixels given the observed context regions. The model must understand the entire image to be able to reasonably fill in a large masked out region that is positioned randomly
  16. Fun Fact!