SlideShare a Scribd company logo
MisGAN
Learning from Incomplete Data with Generative Adversarial Networks
Steven Cheng-Xian Li
University of Massachusetts Amherst
Jihoo Kim
datartist@hanyang.ac.kr
Dept. of Computer and Software, Hanyang University
ICLR’19
Abstract
GANs provides an effective way to model complex distributions.
But, typical GANs require full-observed data during training.
In this paper, we present a GAN-based framework for learning from complex, high-
dimensional incomplete data
The proposed framework learns a complete data generator
along with a mask generator that models the missing data distribution.
We evaluate the proposed framework under the MCAR assumption.
1. Introduction
Unlike likelihood-based methods, GANs is an implicit probabilistic models
which represent a probability distribution through a generator
that learns to directly produce samples from the desired distribution.
GANs have been shown to be very successful in a range of applications
- Generating photorealistic images (2018)
- Image inpainting (2016, 2017)
Training GANs normally requires access to a large collection of fully-observed data.
However, it is not always possible to obtain a large amount of full-observed data.
1. Introduction
The generative process for incompletely observed data (2014, Little & Rubin)
the observed elements of x
the missing according to the mask m
the unknown parameters
of the mask distribution
the unknown parameters
of the data distribution
a binary mask that determines
which entries in x to reveal
a complete data vector
1. Introduction
The unknown parameters are estimated by maximizing the following marginal likelihood.
Little & Rubin (2014) characterize the missing data mechanism
in terms of independence between the complete data x, and the masks m.
①
②
③
1. Introduction
Most work on incomplete data assumes MCAR or MAR since under these assumptions
can be factorized into .
→ The missing data mechanism can be ignored when learning the data generating model
while yielding correct estimates for θ.
When does not admit efficient marginalization over , estimation of θ is usually
performed by maximizing a variational lower bound
1. Introduction
The primary contribution of this paper is the development of a GAN-based framework for
learning high-dimensional data distributions in the presence of incomplete observations.
Our framework introduces an auxiliary GAN for learning a mask distribution to model
the missingness.
The masks are used to “mask” generated complete data by filling the indicated missing
entries with a constant value.
The complete data generator is trained so that the resulting masked data are
indistinguishable from real incomplete data that are masked similarly.
1. Introduction
Our framework builds on the ideas of AmbientGAN (2018).
AmbientGAN modifies the discriminator of a GAN to distinguish corrupted real samples
from corrupted generated samples under a range of corruption processes.
Missing data can be seen as a special type of corruption.
AmbientGAN assumes the measurement process is known only by a few parameters,
which is not the case in general missing data problems.
We provide empirical evidence that the proposed framework is able to effectively learn
complex, high-dimensional data distributions from highly incomplete data.
We further show how the architecture can be used to generate high-quality imputations.
1. Introduction
1 , is observed.
2. MisGAN: A GAN for Missing Data
incomplete data
a partially-observed data vector
a corresponding mask
0 , is missing and contain arbitrary value that we should ignore.
It leads to a cleaner description of the proposed MisGAN.
It suggests how MisGAN can be implemented efficiently.
Instead of …
Two key ideas…
1. We explicitly model the missing data process using a mask generator.
Since the masks in the incomplete dataset are fully observed, we can estimate their distribution.
2. We train the complete data generator adversarially by masking its outputs using generated
masks and , and comparing to real incomplete data that are similarly masked by .
2. MisGAN: A GAN for Missing Data
Masking operator that fills in missing entries with a constant value .
2. MisGAN: A GAN for Missing Data
We use two generator-discriminator pairs
We focus on MCAR, where the two generators are independent of each other
and have their own noise distributions
Loss function for the masks
Loss function for the data
Fake MaskReal Mask
Fake DataReal Data
2. MisGAN: A GAN for Missing Data
We optimize the generators and the discriminators according to the following objectives
Loss function for the masks
Loss function for the data
The losses above follow the Wasserstein GAN formulation (Arjovsky, 2017)
coefficient
We find that choosing a small value
such as 𝜶 = 𝟎. 𝟐 improves performance
Wasserstein GAN (Arjovsky, 2017) Facebook AI Research
KL-Divergence and JS-Divergence
Wasserstein GAN (Arjovsky, 2017) Facebook AI Research
Wasserstein GAN (WGAN) proposes a new cost function
using Wasserstein distance that has a smoother gradient everywhere.
Arjovsky et al 2017 wrote a paper to illustrate the GAN problem mathematically.
Wasserstein GAN (Arjovsky, 2017) Facebook AI Research
2. MisGAN: A GAN for Missing Data
The data discriminator takes as input the masked samples as if the data are fully-observed.
This allows us to use any existing architecture designed for complete data.
The masks are binary. Discrete data generating processes have zero gradient almost everywhere.
To carry out gradient-based training for GANs, we relax the output of the mask generator .
The discriminator in MisGAN is unaware of which entries are missing in the masked input samples,
and does not even need to know which value is used for masking. (In next section, theoretical analysis)
Note that…
2. MisGAN: A GAN for Missing Data
3. Theoretical Results
Two important questions
Does the choice of the filled-in value
affect the ability to recover the data distribution?
Does information about the location of missing values
affect the ability to recover the data distribution?
Q1.
Q2.
3. Theoretical Results
3. Theoretical Results
=
3. Theoretical Results
3. Theoretical Results
4. Missing Data Imputation
We show how to impute missing data according to
by equipping MisGAN with an imputer accompanied by a corresponding discriminator .
Loss function for the masks
Loss function for the data
Loss function for the imputer
noise distribution
𝜶 = 𝟎. 𝟐
𝜷 = 𝟎. 𝟏
This encourages the generated masks
to match the distribution of the real masks
and the masked generated complete samples
to match masked real data.
This encourages the generated complete data
to match the distribution of the imputed real data
In addition to having the masked generated data
match the masked real data.
4. Missing Data Imputation
We can also train a stand-alone imputer using only
with a pre-trained data generator .
Moreover, it is also possible to train the imputer to target a different missing distribution
with a pre-trained data generator alone without access to the original (incomplete) training data
4. Missing Data Imputation
5. Experiments
Data
Missing data
distributions
Evaluation
metric
MNIST
CIFAR-10
CelebA
28x28 handwritten digits images
32x32 color images from 10 classes
64x64 face images (202,599)
The range of pixel values
is rescaled to
Square
observation
Dropout
Variable-size
rectangular
observation
All pixels are missing except for a square
occurring at a random location on the image
Each pixel is independently missing
according to a Bernoulli distribution
All pixels are missing except for a rectangular observed region
(width and height are drawn from 25% to 75% o the image length)
(Heusel, 2017)
5. Experiments
1. Architectures
2. Baseline
3. Results
MisGAN with convolutional networks – DCGAN (Radford, 2015)
MisGAN with fully connected networksFC-MisGAN
Conv-MisGAN
ConvAC
The generative convolutional arithmetic circuit (Sharir, 2016)
→ capable of learning from large-scale incomplete data
Figure 3
Figure 4
Figure 5
Figure 6
5.1 Empirical Study of MisGAN on MNIST
Next slides...
5. Experiments
Training Samples
Generated data Samples
Generated mask Samples
Generated data Samples
Conv-MisGAN
FC-MisGAN
5.1 Empirical Study of MisGAN on MNIST
5. Experiments 5.1 Empirical Study of MisGAN on MNIST
MisGAN outperforms ConvAC
Data samples generated by Conv-MisGAN
Mask samples generated by Conv-MisGAN
Data samples generated by MisGAN
Variable-size
Square
5. Experiments
4. Ablation study
5.1 Empirical Study of MisGAN on MNIST
We point out that the mask discriminator in MisGAN is important for learning the correct distribution.
Two failure cases of AmbientGAN, which is essentially equivalent to a MisGAN without the mask discriminator.
Generated data samplesGenerated mask samples Generated data samples Generated mask samples
rescale
5. Experiments 5.1 Empirical Study of MisGAN on MNIST
5. Missing data imputation
Inside of box → observed pixels
Outside of box → generated pixels Each row → same incomplete input
The imputer can produce a variety of different imputed results
5. Experiments 5.2 Quantitative Evaluation
1. Baselines
3. Architecture
2. Evaluation of
imputation
4. Results
We focus on evaluating MisGAN on the missing data imputation task
zero/mean imputation
matrix factorization
GAIN (Generative Adversarial Imputation Network)
FID between the imputed data and the original fully-observed data
For MNIST → Fully-connected imputer network
For CIFAR-10 and CelebA → Five-layer U-Net architecture (Ronneberger, 2015)
Next slides...
5. Experiments 5.2 Quantitative Evaluation
MisGAN consistently outperforms other methods in all cases, especially under high missing rates.
Training MisGAN is more stable than training GAIN.
6. Discussion and Future Work
This work presents and evaluates a high flexible framework for learning
standard GAN data generators in the presence of missing data.
We only focus on the MCAR case in this work.
MisGAN can be easily extended to cases both MAR and NMAR.
We have tried the modified architecture and it showed similar results.
This suggests that the extra dependencies may not adversely affect learnability.
We leave the formal evaluation of this modified framework for future work.

More Related Content

What's hot

【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
Deep Learning JP
 
DeepLearningとWord2Vecを用いた画像レコメンドの考察
DeepLearningとWord2Vecを用いた画像レコメンドの考察DeepLearningとWord2Vecを用いた画像レコメンドの考察
DeepLearningとWord2Vecを用いた画像レコメンドの考察
Tadaichiro Nakano
 
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
Taehoon Kim
 
[Ridge-i 論文よみかい] Wasserstein auto encoder
[Ridge-i 論文よみかい] Wasserstein auto encoder[Ridge-i 論文よみかい] Wasserstein auto encoder
[Ridge-i 論文よみかい] Wasserstein auto encoder
Masanari Kimura
 
辺彩色
辺彩色辺彩色
辺彩色
Ken Ogura
 
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
Deep Learning JP
 
Superpixel Sampling Networks
Superpixel Sampling NetworksSuperpixel Sampling Networks
Superpixel Sampling Networks
yukihiro domae
 
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
Deep Learning JP
 
MINI PROJECT 2023 deepfake detection.pptx
MINI PROJECT 2023 deepfake detection.pptxMINI PROJECT 2023 deepfake detection.pptx
MINI PROJECT 2023 deepfake detection.pptx
swathiravishankar3
 
差分プライバシーとは何か? (定義 & 解釈編)
差分プライバシーとは何か? (定義 & 解釈編)差分プライバシーとは何か? (定義 & 解釈編)
差分プライバシーとは何か? (定義 & 解釈編)
Kentaro Minami
 
RTBにおける機械学習の活用事例
RTBにおける機械学習の活用事例RTBにおける機械学習の活用事例
RTBにおける機械学習の活用事例
MicroAd, Inc.(Engineer)
 
십분딥러닝_16_WGAN (Wasserstein GANs)
십분딥러닝_16_WGAN (Wasserstein GANs)십분딥러닝_16_WGAN (Wasserstein GANs)
십분딥러닝_16_WGAN (Wasserstein GANs)
HyunKyu Jeon
 
Improved Trainings of Wasserstein GANs (WGAN-GP)
Improved Trainings of Wasserstein GANs (WGAN-GP)Improved Trainings of Wasserstein GANs (WGAN-GP)
Improved Trainings of Wasserstein GANs (WGAN-GP)
Sangwoo Mo
 
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
Deep Learning JP
 
Batch normalization effectiveness_20190206
Batch normalization effectiveness_20190206Batch normalization effectiveness_20190206
Batch normalization effectiveness_20190206
Masakazu Shinoda
 
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
Haezoom Inc.
 
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
Deep Learning JP
 
ベイジアンディープニューラルネット
ベイジアンディープニューラルネットベイジアンディープニューラルネット
ベイジアンディープニューラルネット
Yuta Kashino
 
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
Morpho, Inc.
 
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
Deep Learning JP
 

What's hot (20)

【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
【DL輪読会】StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-I...
 
DeepLearningとWord2Vecを用いた画像レコメンドの考察
DeepLearningとWord2Vecを用いた画像レコメンドの考察DeepLearningとWord2Vecを用いた画像レコメンドの考察
DeepLearningとWord2Vecを用いた画像レコメンドの考察
 
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
텐서플로우 설치도 했고 튜토리얼도 봤고 기초 예제도 짜봤다면 TensorFlow KR Meetup 2016
 
[Ridge-i 論文よみかい] Wasserstein auto encoder
[Ridge-i 論文よみかい] Wasserstein auto encoder[Ridge-i 論文よみかい] Wasserstein auto encoder
[Ridge-i 論文よみかい] Wasserstein auto encoder
 
辺彩色
辺彩色辺彩色
辺彩色
 
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
[DL輪読会]Live-Streaming Fraud Detection: A Heterogeneous Graph Neural Network A...
 
Superpixel Sampling Networks
Superpixel Sampling NetworksSuperpixel Sampling Networks
Superpixel Sampling Networks
 
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
 
MINI PROJECT 2023 deepfake detection.pptx
MINI PROJECT 2023 deepfake detection.pptxMINI PROJECT 2023 deepfake detection.pptx
MINI PROJECT 2023 deepfake detection.pptx
 
差分プライバシーとは何か? (定義 & 解釈編)
差分プライバシーとは何か? (定義 & 解釈編)差分プライバシーとは何か? (定義 & 解釈編)
差分プライバシーとは何か? (定義 & 解釈編)
 
RTBにおける機械学習の活用事例
RTBにおける機械学習の活用事例RTBにおける機械学習の活用事例
RTBにおける機械学習の活用事例
 
십분딥러닝_16_WGAN (Wasserstein GANs)
십분딥러닝_16_WGAN (Wasserstein GANs)십분딥러닝_16_WGAN (Wasserstein GANs)
십분딥러닝_16_WGAN (Wasserstein GANs)
 
Improved Trainings of Wasserstein GANs (WGAN-GP)
Improved Trainings of Wasserstein GANs (WGAN-GP)Improved Trainings of Wasserstein GANs (WGAN-GP)
Improved Trainings of Wasserstein GANs (WGAN-GP)
 
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
【DL輪読会】Which Shortcut Cues Will DNNs Choose? A Study from the Parameter-Space...
 
Batch normalization effectiveness_20190206
Batch normalization effectiveness_20190206Batch normalization effectiveness_20190206
Batch normalization effectiveness_20190206
 
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
Variational Autoencoder를 여러 가지 각도에서 이해하기 (Understanding Variational Autoencod...
 
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
 
ベイジアンディープニューラルネット
ベイジアンディープニューラルネットベイジアンディープニューラルネット
ベイジアンディープニューラルネット
 
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
(文献紹介)エッジ保存フィルタ:Side Window Filter, Curvature Filter
 
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
[DL輪読会]AutoAugment: LearningAugmentation Strategies from Data & Learning Data...
 

Similar to [Paper Review] MisGAN: Learning from Incomplete Data with Generative Adversarial Networks (ICLR'19)

An Extensive Review on Generative Adversarial Networks GAN’s
An Extensive Review on Generative Adversarial Networks GAN’sAn Extensive Review on Generative Adversarial Networks GAN’s
An Extensive Review on Generative Adversarial Networks GAN’s
ijtsrd
 
Anomaly Detection Using Generative Adversarial Network(GAN)
Anomaly Detection Using Generative Adversarial Network(GAN)Anomaly Detection Using Generative Adversarial Network(GAN)
Anomaly Detection Using Generative Adversarial Network(GAN)
Asha Aher
 
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdfTop Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
Artificial Intelligence Board of America
 
Gans - Generative Adversarial Nets
Gans - Generative Adversarial NetsGans - Generative Adversarial Nets
Gans - Generative Adversarial Nets
SajalRastogi8
 
An ann approach for network
An ann approach for networkAn ann approach for network
An ann approach for network
IJNSA Journal
 
2. visualization in data mining
2. visualization in data mining2. visualization in data mining
2. visualization in data mining
Azad public school
 
Time series anomaly detection using cnn coupled with data augmentation using ...
Time series anomaly detection using cnn coupled with data augmentation using ...Time series anomaly detection using cnn coupled with data augmentation using ...
Time series anomaly detection using cnn coupled with data augmentation using ...
Prasenjeet Acharjee
 
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
IJNSA Journal
 
Nips 2016 tutorial generative adversarial networks review
Nips 2016 tutorial  generative adversarial networks reviewNips 2016 tutorial  generative adversarial networks review
Nips 2016 tutorial generative adversarial networks review
Minho Heo
 
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORETEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
IJCI JOURNAL
 
Progress of Machine Learning in the Field of Intrusion Detection Systems
Progress of Machine Learning in the Field of Intrusion Detection SystemsProgress of Machine Learning in the Field of Intrusion Detection Systems
Progress of Machine Learning in the Field of Intrusion Detection Systems
ijcisjournal
 
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
ijcisjournal
 
Botnet detection using Wgans for security
Botnet detection using Wgans for securityBotnet detection using Wgans for security
Botnet detection using Wgans for security
ssuser3f5a831
 
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
IRJET- Generating 3D Models Using 3D Generative Adversarial NetworkIRJET- Generating 3D Models Using 3D Generative Adversarial Network
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
IRJET Journal
 
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
IRJET Journal
 
Image Masking.pdf
Image Masking.pdfImage Masking.pdf
Image Masking.pdf
farin11
 
IRJET - Hand Gesture Recognition to Perform System Operations
IRJET -  	  Hand Gesture Recognition to Perform System OperationsIRJET -  	  Hand Gesture Recognition to Perform System Operations
IRJET - Hand Gesture Recognition to Perform System Operations
IRJET Journal
 
Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)
Prakhar Rastogi
 
Fake News Detection using Deep Learning
Fake News Detection using Deep LearningFake News Detection using Deep Learning
Fake News Detection using Deep Learning
NIET Journal of Engineering & Technology (NIETJET)
 
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILUREAN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
IJCI JOURNAL
 

Similar to [Paper Review] MisGAN: Learning from Incomplete Data with Generative Adversarial Networks (ICLR'19) (20)

An Extensive Review on Generative Adversarial Networks GAN’s
An Extensive Review on Generative Adversarial Networks GAN’sAn Extensive Review on Generative Adversarial Networks GAN’s
An Extensive Review on Generative Adversarial Networks GAN’s
 
Anomaly Detection Using Generative Adversarial Network(GAN)
Anomaly Detection Using Generative Adversarial Network(GAN)Anomaly Detection Using Generative Adversarial Network(GAN)
Anomaly Detection Using Generative Adversarial Network(GAN)
 
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdfTop Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
Top Machine Learning Algorithms Used By AI Professionals ARTiBA.pdf
 
Gans - Generative Adversarial Nets
Gans - Generative Adversarial NetsGans - Generative Adversarial Nets
Gans - Generative Adversarial Nets
 
An ann approach for network
An ann approach for networkAn ann approach for network
An ann approach for network
 
2. visualization in data mining
2. visualization in data mining2. visualization in data mining
2. visualization in data mining
 
Time series anomaly detection using cnn coupled with data augmentation using ...
Time series anomaly detection using cnn coupled with data augmentation using ...Time series anomaly detection using cnn coupled with data augmentation using ...
Time series anomaly detection using cnn coupled with data augmentation using ...
 
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
AN ANN APPROACH FOR NETWORK INTRUSION DETECTION USING ENTROPY BASED FEATURE S...
 
Nips 2016 tutorial generative adversarial networks review
Nips 2016 tutorial  generative adversarial networks reviewNips 2016 tutorial  generative adversarial networks review
Nips 2016 tutorial generative adversarial networks review
 
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORETEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
TEXT GENERATION WITH GAN NETWORKS USING FEEDBACK SCORE
 
Progress of Machine Learning in the Field of Intrusion Detection Systems
Progress of Machine Learning in the Field of Intrusion Detection SystemsProgress of Machine Learning in the Field of Intrusion Detection Systems
Progress of Machine Learning in the Field of Intrusion Detection Systems
 
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
11421ijcPROGRESS OF MACHINE LEARNING IN THE FIELD OF INTRUSION DETECTION SYST...
 
Botnet detection using Wgans for security
Botnet detection using Wgans for securityBotnet detection using Wgans for security
Botnet detection using Wgans for security
 
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
IRJET- Generating 3D Models Using 3D Generative Adversarial NetworkIRJET- Generating 3D Models Using 3D Generative Adversarial Network
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
 
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
Locate, Size and Count: Accurately Resolving People in Dense Crowds via Detec...
 
Image Masking.pdf
Image Masking.pdfImage Masking.pdf
Image Masking.pdf
 
IRJET - Hand Gesture Recognition to Perform System Operations
IRJET -  	  Hand Gesture Recognition to Perform System OperationsIRJET -  	  Hand Gesture Recognition to Perform System Operations
IRJET - Hand Gesture Recognition to Perform System Operations
 
Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)Generative Adversarial Network (GAN)
Generative Adversarial Network (GAN)
 
Fake News Detection using Deep Learning
Fake News Detection using Deep LearningFake News Detection using Deep Learning
Fake News Detection using Deep Learning
 
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILUREAN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
AN IMPROVED CTGAN FOR DATA PROCESSING METHOD OF IMBALANCED DISK FAILURE
 

Recently uploaded

一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
slg6lamcq
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
rwarrenll
 
一比一原版(BU毕业证)波士顿大学毕业证成绩单
一比一原版(BU毕业证)波士顿大学毕业证成绩单一比一原版(BU毕业证)波士顿大学毕业证成绩单
一比一原版(BU毕业证)波士顿大学毕业证成绩单
ewymefz
 
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
ukgaet
 
Empowering Data Analytics Ecosystem.pptx
Empowering Data Analytics Ecosystem.pptxEmpowering Data Analytics Ecosystem.pptx
Empowering Data Analytics Ecosystem.pptx
benishzehra469
 
The affect of service quality and online reviews on customer loyalty in the E...
The affect of service quality and online reviews on customer loyalty in the E...The affect of service quality and online reviews on customer loyalty in the E...
The affect of service quality and online reviews on customer loyalty in the E...
jerlynmaetalle
 
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
Subhajit Sahu
 
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
mbawufebxi
 
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
pchutichetpong
 
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptxData_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
AnirbanRoy608946
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
74nqk8xf
 
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
vcaxypu
 
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
slg6lamcq
 
一比一原版(TWU毕业证)西三一大学毕业证成绩单
一比一原版(TWU毕业证)西三一大学毕业证成绩单一比一原版(TWU毕业证)西三一大学毕业证成绩单
一比一原版(TWU毕业证)西三一大学毕业证成绩单
ocavb
 
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
nscud
 
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
axoqas
 
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
NABLAS株式会社
 
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
2023240532
 
一比一原版(NYU毕业证)纽约大学毕业证成绩单
一比一原版(NYU毕业证)纽约大学毕业证成绩单一比一原版(NYU毕业证)纽约大学毕业证成绩单
一比一原版(NYU毕业证)纽约大学毕业证成绩单
ewymefz
 
Predicting Product Ad Campaign Performance: A Data Analysis Project Presentation
Predicting Product Ad Campaign Performance: A Data Analysis Project PresentationPredicting Product Ad Campaign Performance: A Data Analysis Project Presentation
Predicting Product Ad Campaign Performance: A Data Analysis Project Presentation
Boston Institute of Analytics
 

Recently uploaded (20)

一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
 
一比一原版(BU毕业证)波士顿大学毕业证成绩单
一比一原版(BU毕业证)波士顿大学毕业证成绩单一比一原版(BU毕业证)波士顿大学毕业证成绩单
一比一原版(BU毕业证)波士顿大学毕业证成绩单
 
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
一比一原版(UVic毕业证)维多利亚大学毕业证成绩单
 
Empowering Data Analytics Ecosystem.pptx
Empowering Data Analytics Ecosystem.pptxEmpowering Data Analytics Ecosystem.pptx
Empowering Data Analytics Ecosystem.pptx
 
The affect of service quality and online reviews on customer loyalty in the E...
The affect of service quality and online reviews on customer loyalty in the E...The affect of service quality and online reviews on customer loyalty in the E...
The affect of service quality and online reviews on customer loyalty in the E...
 
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...
 
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
 
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...
 
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptxData_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
 
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
一比一原版(RUG毕业证)格罗宁根大学毕业证成绩单
 
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
一比一原版(Adelaide毕业证书)阿德莱德大学毕业证如何办理
 
一比一原版(TWU毕业证)西三一大学毕业证成绩单
一比一原版(TWU毕业证)西三一大学毕业证成绩单一比一原版(TWU毕业证)西三一大学毕业证成绩单
一比一原版(TWU毕业证)西三一大学毕业证成绩单
 
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
一比一原版(CBU毕业证)卡普顿大学毕业证成绩单
 
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
哪里卖(usq毕业证书)南昆士兰大学毕业证研究生文凭证书托福证书原版一模一样
 
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
【社内勉強会資料_Octo: An Open-Source Generalist Robot Policy】
 
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...
 
一比一原版(NYU毕业证)纽约大学毕业证成绩单
一比一原版(NYU毕业证)纽约大学毕业证成绩单一比一原版(NYU毕业证)纽约大学毕业证成绩单
一比一原版(NYU毕业证)纽约大学毕业证成绩单
 
Predicting Product Ad Campaign Performance: A Data Analysis Project Presentation
Predicting Product Ad Campaign Performance: A Data Analysis Project PresentationPredicting Product Ad Campaign Performance: A Data Analysis Project Presentation
Predicting Product Ad Campaign Performance: A Data Analysis Project Presentation
 

[Paper Review] MisGAN: Learning from Incomplete Data with Generative Adversarial Networks (ICLR'19)

  • 1. MisGAN Learning from Incomplete Data with Generative Adversarial Networks Steven Cheng-Xian Li University of Massachusetts Amherst Jihoo Kim datartist@hanyang.ac.kr Dept. of Computer and Software, Hanyang University ICLR’19
  • 2. Abstract GANs provides an effective way to model complex distributions. But, typical GANs require full-observed data during training. In this paper, we present a GAN-based framework for learning from complex, high- dimensional incomplete data The proposed framework learns a complete data generator along with a mask generator that models the missing data distribution. We evaluate the proposed framework under the MCAR assumption.
  • 3. 1. Introduction Unlike likelihood-based methods, GANs is an implicit probabilistic models which represent a probability distribution through a generator that learns to directly produce samples from the desired distribution. GANs have been shown to be very successful in a range of applications - Generating photorealistic images (2018) - Image inpainting (2016, 2017) Training GANs normally requires access to a large collection of fully-observed data. However, it is not always possible to obtain a large amount of full-observed data.
  • 4. 1. Introduction The generative process for incompletely observed data (2014, Little & Rubin) the observed elements of x the missing according to the mask m the unknown parameters of the mask distribution the unknown parameters of the data distribution a binary mask that determines which entries in x to reveal a complete data vector
  • 5. 1. Introduction The unknown parameters are estimated by maximizing the following marginal likelihood. Little & Rubin (2014) characterize the missing data mechanism in terms of independence between the complete data x, and the masks m. ① ② ③
  • 6. 1. Introduction Most work on incomplete data assumes MCAR or MAR since under these assumptions can be factorized into . → The missing data mechanism can be ignored when learning the data generating model while yielding correct estimates for θ. When does not admit efficient marginalization over , estimation of θ is usually performed by maximizing a variational lower bound
  • 7. 1. Introduction The primary contribution of this paper is the development of a GAN-based framework for learning high-dimensional data distributions in the presence of incomplete observations. Our framework introduces an auxiliary GAN for learning a mask distribution to model the missingness. The masks are used to “mask” generated complete data by filling the indicated missing entries with a constant value. The complete data generator is trained so that the resulting masked data are indistinguishable from real incomplete data that are masked similarly.
  • 8. 1. Introduction Our framework builds on the ideas of AmbientGAN (2018). AmbientGAN modifies the discriminator of a GAN to distinguish corrupted real samples from corrupted generated samples under a range of corruption processes. Missing data can be seen as a special type of corruption. AmbientGAN assumes the measurement process is known only by a few parameters, which is not the case in general missing data problems.
  • 9. We provide empirical evidence that the proposed framework is able to effectively learn complex, high-dimensional data distributions from highly incomplete data. We further show how the architecture can be used to generate high-quality imputations. 1. Introduction
  • 10. 1 , is observed. 2. MisGAN: A GAN for Missing Data incomplete data a partially-observed data vector a corresponding mask 0 , is missing and contain arbitrary value that we should ignore. It leads to a cleaner description of the proposed MisGAN. It suggests how MisGAN can be implemented efficiently. Instead of …
  • 11. Two key ideas… 1. We explicitly model the missing data process using a mask generator. Since the masks in the incomplete dataset are fully observed, we can estimate their distribution. 2. We train the complete data generator adversarially by masking its outputs using generated masks and , and comparing to real incomplete data that are similarly masked by . 2. MisGAN: A GAN for Missing Data Masking operator that fills in missing entries with a constant value .
  • 12. 2. MisGAN: A GAN for Missing Data We use two generator-discriminator pairs We focus on MCAR, where the two generators are independent of each other and have their own noise distributions Loss function for the masks Loss function for the data Fake MaskReal Mask Fake DataReal Data
  • 13. 2. MisGAN: A GAN for Missing Data We optimize the generators and the discriminators according to the following objectives Loss function for the masks Loss function for the data The losses above follow the Wasserstein GAN formulation (Arjovsky, 2017) coefficient We find that choosing a small value such as 𝜶 = 𝟎. 𝟐 improves performance
  • 14. Wasserstein GAN (Arjovsky, 2017) Facebook AI Research KL-Divergence and JS-Divergence
  • 15. Wasserstein GAN (Arjovsky, 2017) Facebook AI Research Wasserstein GAN (WGAN) proposes a new cost function using Wasserstein distance that has a smoother gradient everywhere. Arjovsky et al 2017 wrote a paper to illustrate the GAN problem mathematically.
  • 16. Wasserstein GAN (Arjovsky, 2017) Facebook AI Research
  • 17. 2. MisGAN: A GAN for Missing Data The data discriminator takes as input the masked samples as if the data are fully-observed. This allows us to use any existing architecture designed for complete data. The masks are binary. Discrete data generating processes have zero gradient almost everywhere. To carry out gradient-based training for GANs, we relax the output of the mask generator . The discriminator in MisGAN is unaware of which entries are missing in the masked input samples, and does not even need to know which value is used for masking. (In next section, theoretical analysis) Note that…
  • 18. 2. MisGAN: A GAN for Missing Data
  • 19. 3. Theoretical Results Two important questions Does the choice of the filled-in value affect the ability to recover the data distribution? Does information about the location of missing values affect the ability to recover the data distribution? Q1. Q2.
  • 24. 4. Missing Data Imputation We show how to impute missing data according to by equipping MisGAN with an imputer accompanied by a corresponding discriminator . Loss function for the masks Loss function for the data Loss function for the imputer noise distribution 𝜶 = 𝟎. 𝟐 𝜷 = 𝟎. 𝟏 This encourages the generated masks to match the distribution of the real masks and the masked generated complete samples to match masked real data. This encourages the generated complete data to match the distribution of the imputed real data In addition to having the masked generated data match the masked real data.
  • 25. 4. Missing Data Imputation We can also train a stand-alone imputer using only with a pre-trained data generator . Moreover, it is also possible to train the imputer to target a different missing distribution with a pre-trained data generator alone without access to the original (incomplete) training data
  • 26. 4. Missing Data Imputation
  • 27. 5. Experiments Data Missing data distributions Evaluation metric MNIST CIFAR-10 CelebA 28x28 handwritten digits images 32x32 color images from 10 classes 64x64 face images (202,599) The range of pixel values is rescaled to Square observation Dropout Variable-size rectangular observation All pixels are missing except for a square occurring at a random location on the image Each pixel is independently missing according to a Bernoulli distribution All pixels are missing except for a rectangular observed region (width and height are drawn from 25% to 75% o the image length) (Heusel, 2017)
  • 28. 5. Experiments 1. Architectures 2. Baseline 3. Results MisGAN with convolutional networks – DCGAN (Radford, 2015) MisGAN with fully connected networksFC-MisGAN Conv-MisGAN ConvAC The generative convolutional arithmetic circuit (Sharir, 2016) → capable of learning from large-scale incomplete data Figure 3 Figure 4 Figure 5 Figure 6 5.1 Empirical Study of MisGAN on MNIST Next slides...
  • 29. 5. Experiments Training Samples Generated data Samples Generated mask Samples Generated data Samples Conv-MisGAN FC-MisGAN 5.1 Empirical Study of MisGAN on MNIST
  • 30. 5. Experiments 5.1 Empirical Study of MisGAN on MNIST MisGAN outperforms ConvAC Data samples generated by Conv-MisGAN Mask samples generated by Conv-MisGAN Data samples generated by MisGAN Variable-size Square
  • 31. 5. Experiments 4. Ablation study 5.1 Empirical Study of MisGAN on MNIST We point out that the mask discriminator in MisGAN is important for learning the correct distribution. Two failure cases of AmbientGAN, which is essentially equivalent to a MisGAN without the mask discriminator. Generated data samplesGenerated mask samples Generated data samples Generated mask samples rescale
  • 32. 5. Experiments 5.1 Empirical Study of MisGAN on MNIST 5. Missing data imputation Inside of box → observed pixels Outside of box → generated pixels Each row → same incomplete input The imputer can produce a variety of different imputed results
  • 33. 5. Experiments 5.2 Quantitative Evaluation 1. Baselines 3. Architecture 2. Evaluation of imputation 4. Results We focus on evaluating MisGAN on the missing data imputation task zero/mean imputation matrix factorization GAIN (Generative Adversarial Imputation Network) FID between the imputed data and the original fully-observed data For MNIST → Fully-connected imputer network For CIFAR-10 and CelebA → Five-layer U-Net architecture (Ronneberger, 2015) Next slides...
  • 34. 5. Experiments 5.2 Quantitative Evaluation MisGAN consistently outperforms other methods in all cases, especially under high missing rates. Training MisGAN is more stable than training GAIN.
  • 35. 6. Discussion and Future Work This work presents and evaluates a high flexible framework for learning standard GAN data generators in the presence of missing data. We only focus on the MCAR case in this work. MisGAN can be easily extended to cases both MAR and NMAR. We have tried the modified architecture and it showed similar results. This suggests that the extra dependencies may not adversely affect learnability. We leave the formal evaluation of this modified framework for future work.