Subclass deep neural networks

retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Title of presentation
Subtitle
Name of presenter
Date
Subclass deep neural networks: re-enabling neglected classes
in deep network training for multimedia classification
N. Gkalelis,V. Mezaris
CERTH-ITI,Thermi -Thessaloniki,Greece
26th Int. Conf. Multimedia Modeling
Daejeon, Korea, January 2020
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Outline
2
• Problem statement
• Related work
• Identification of neglected classes
• Subclass DNNs
• Experiments
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
3
• Deep neural networks (DNNs) have shown remarkable performance in many
machine learning problems
• They are being commercially deployed in various domains
Problem statement
• Multimedia
understanding
• Self-driving cars • Edge computing
Image Credits:V2Gov
ImageCredits: [1]
[1]Chen, J., Ran, X.: Deep LearningWith Edge Computing:A Review, Proc. of the IEEE, vol. 107, no. 8, (Aug. 2019)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
4
One weakness of DNNs is that some classes are getting less attention during
training, for instance, due to:
• Overfitting: error in the training set becomes negligible although generalization
error is still high
• Simply, the contributions of the misclassified observations to the updating of the
weights associated with these classes cancel out
 How to identify these neglected classes?
 How to boost the classification performance of neglected classes?
Problem statement
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
5
Related work
• The problem of neglected classes sometimes is related with imbalanced class
learning, where the distribution of training data across different classes is skewed
• There are relatively few works studying the class imbalance problem in DNNs
(e.g. see [1, 2])
 As observed in our experiments neglected classes may also appear with balanced
training datasets (e.g. CIFAR datasets)
 The identification of classes getting less attention from the DNN during the
training procedure is a relatively unexplored topic
[1] Sarafianos, N., Xu, X., Kakadiaris, I.A.: Deep imbalanced attribute classification using visual attention aggregation. ECCV. Munich, Germany
(Sep 2018)
[2] Lin,T., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal Loss for DenseObject Detection, 2017 IEEE ICCV,Venice, 2017.
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
6
Identification of neglected classes: motivation
• During backpropagation, a strong gradient update gi for the weight vectors wi at
the output layer is typically produced for classes with high training error ei
• In this way, the overall network is guided to extract nonlinear features at different
layers in order to provide a linear separable subspace at the output layer for these
classes
wi
wi’ = wi + gi
gi
length of the gradient update ||gi|| is large as
expected due to a high error in the output node i
weight vector is updated (as well as
weight vectors in the layers below) in
order to reduce the output node error
weight vector of node i
(associated with class i)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
7
Identification of neglected classes: motivation
• However, it is observed that for certain output nodes the gradient update is close
to zero although the classes associated with these nodes still have a large error
• The result is that the DNN neglects to extract the necessary discriminant features
for reducing the error for these classes
• We investigate this case in detail and based on the analysis we propose a new
measure to identify the so-called neglected classes
The length of the gradient
update ||gi|| is small despite a
large error in the output node i! the updated weight vector of output node i
remains relatively unchanged because the length
of the gradient update ||gi|| is close to zero
wi’  wi
gi
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
8
Identification of neglected classes: background
• Suppose a DNN with sigmoid (SG) output layer and cross entropy (CE) loss
xk
xf,k
xj,k
x1,k
wi
w1,i
wj,i
bi
weight vector and bias term of i-th node
input vector to the i-th node associated with the k-th training observation
output of the i-th node associated with k-th training observation
i-th output node of the SG layer (associated with i-th class)
wf,i
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
9
Identification of neglected classes: background
gradient update of wi
label of k-th observation w.r.t. class i
Learning rate
input vector to the i-th output
node associated with the k-th
training observation in the batch
output of the i-th node associated
with k-th training observation
Number of batch observations
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
10
Identification of neglected classes Unnormalized gradient update
(w.r.t i-th node)
Unnormalized
gradient update
of positive class
observations
Unnormalized gradient update
of negative class observations
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
11
Identification of neglected classes
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
12
Identification of neglected classes
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
13
Identification of neglected classes
Length of
unnormalized
gradient update of
positive and
negative class
observations
Length of unnormalized gradient
update (w.r.t i-th node)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
14
Training of neglected classes: Subclass DNNs
• A subclass-based learning framework is applied to improve the classification
performance of neglected classes
• Subclass-based classification techniques have been successfully used in the
shallow learning paradigm [1-5].
• To this end, neglected classes are augmented and partitioned to subclasses and
the created labeling information is used to create and train a subclass DNN
[1] Hastie,T.,Tibshirani, R.: Discriminant analysis by Gaussian mixtures. Journal of the Royal Statistical Society. Series B 58(1), (Jul 1996)
[2] Manli, Z., Martinez, A.M.: Subclass discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 28(8), (Aug 2006)
[3] Escalera, S. et al.: Subclass problem-dependent design for error-correcting output codes. IEEETrans. PatternAnal. Mach. Intell. 30(6), (Jun
2008)
[4]You, D., Hamsici, O.C., Martinez, A.M.: Kernel optimization in discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 33(3), (Mar 2011)
[5] Gkalelis, N., Mezaris,V., Kompatsiaris, I., Stathaki,T.: Mixture subclass discriminant analysis link to restricted Gaussian model and other
generalizations. IEEETrans. Neural Netw. Learn. Syst. 24(1), (Jan 2013)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
15
Training of neglected classes: Subclass DNNs
Number of training observations
Number of classes
Number of subclasses of class i
label of k-th observation
w.r.t subclass (i,j)
output of the node (i.j) associated
with k-th training observation
label of k-th observation
w.r.t class i
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
16
Multiclass classification
• CIFAR10 [1]: 10 classes, 32 x 32 color images, 50000 training and 10000 testing
observations
• CIFAR100 [1]: as CIFAR10 but with 100 classes
• SVHN [2]: 10 classes, 32 x 32 color images, 73257 training, 26032 testing and
531131 extra observations
[1] Krizhevsky, A.: Learning multiple layers of features from tiny images.Tech. rep. (2009)
[2] Netzer,Y.,Wang,T., Coates, A., Bissacco, A.,Wu, B., Ng, A.Y.: Reading digits in natural images with unsupervised feature learning. NIPS
Workshops.Venice, Italy, (Oct 2017)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
17
Experimental setup (similar setup with [1])
• Images are normalized to zero mean and unit variance
• Data augmentation is applied as in [1] (cropping, mirroring, cutout, etc.)
• DNNs [2,3]: VGG16, WRN-28-10 (CIFAR-10, -100), WRN-16-8 (SVHN)
• CE loss, Minibatch SGD, Nesterov momentum 0.9, batch size 128, weight decay
0.0005, 200 epochs, initial learning rate 0.1, exponential schedule (learning rate
multiplied by 0.1 for CIFAR or 0.2 for SVHN, at epochs 60, 120 and 160)
[1] Devries,T.,Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 (2017)
[2]Zagoruyko, S., Komodakis, N.:Wide residual networks. BMVC.York, UK, (Sep 2016)
[3] Simonyan, K., Zisserman, A.:Very deep convolutional networks for large-scale image recognition. ICLR. San Diego, CA, USA (May 2015)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
18
SDNN training
• DNN is trained for 30 epochs in order to compute a reliable neglection score for
each class
• The classes with highest neglection score are selected: 2 classes for CIFAR-10 and
SVHN, 10 classes for CIFAR-100
• Observations of the selected classes are doubled using the augmentation
approach presented in [1]
• Selected classes are partitioned to 2 subclasses using k-means
• SDNN is trained using the annotations at class and subclass level
[1] Inoue, H.: Data augmentation by pairing samples for images classification. arXiv:1801.02929 (2018)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
19
Results
• The proposed subclass DNNs (SVGG16, SWRN) provide improved performance over their
conventional variants (VGG16 [1], WRN [2])
• Considering that [2] is among the state-of-the-art approaches, the attained CCR performance
improvement is significant
• Training time overhead is negligible for the medium-size datasets (CIFAR-10, -100) and relatively
small for the larger SVHN; testing time is only a few seconds for all DNNs
VGG16 [1] SVGG16 WRN [2] SWRN
CIFAR-10 93.5% (2.6 h) 94.8% (2.7 h) 96.92% (7.1 h) 97.14% (8.1 h)
CIFAR-100 71.24% (2.6 h) 73.67% (2.6 h) 81.59% (7.1 h) 82.17% (7.5 h)
SVHN 98.16% (29.1 h) 98.35 (34.1 h) 98.7% (33.1 h) 98.81% (42.7 h)
[1] Simonyan, K., Zisserman, A.:Very deep convolutional networks for large-scale image recognition. In: ICLR. San Diego, CA, USA (May 2015)
[2] Devries,T.,Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 (2017)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
20
Multilabel classification
• YouTube-8M (YT8M) [1]: largest public dataset for multilabel video classification
• 3862 classes, 3888919 training, 1112356 evaluation and 1133323 testing videos
• 1024- and 128-dimensional visual and audio feature vectors are provided for each
video
[1]Abu-El-Haija, S., Kothari, N., Lee, J., Natsev, A.P.,Toderici, G.,Varadarajan, B.,Vijayanarasimhan, S.:YouTube-8M:A large-scale video
classification benchmark. In: arXiv:1609.08675 (2016)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
21
Experimental setup
• Two experiments: using the visual feature vectors directly; concatenating the
visual and audio feature vectors
• L2-normalization is applied to feature vectors in all cases
• DNN: 1 conv layer with 64 1d-filters, max-pooling, dropout, SG output layer
• CE loss, Minibatch SGD, batch size 512, weight decay 0.0005, 5 epochs, initial
learning rate 0.001, exponential schedule (learning rate multiplied by 0.95 at
every epoch)
• The evaluation metrics of YT8M challenge [1] are utilized for performance
evaluation; GAP@20 is used as the primary evaluation metric
[1]Abu-El-Haija, S., Kothari, N., Lee, J., Natsev, A.P.,Toderici, G.,Varadarajan, B.,Vijayanarasimhan, S.:YouTube-8M:A large-scale video
classification benchmark. In: arXiv:1609.08675 (2016)
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
22
SDNN training
• DNN is trained for 1/3 of an epoch; a neglection score for each class is computed
• 386 classes with highest neglection score are selected
• Observations of the selected classes are doubled by applying extrapolation in
feature space [1]
• A efficient variant of the nearest neighbor-based clustering algorithm [2,3] is
used to partition the neglected classes to subclasses
• SDNN is trained using the annotations at class and subclass level
[1] DeVries,T.,Taylor, G.W.: Dataset augmentation in feature space. ICLRWorkshops.Toulon, France (Apr 2017)
[2] Manli, Z., Martinez, A.M.: Subclass discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 28(8), 1274-1286 (Aug 2006)
[3]You, D., Hamsici, O.C., Martinez, A.M.: Kernel optimization in discriminant analysis. IEEE PAMI 33(3), 631-638 (Mar 2011)
Experiments
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
23
Results
• The proposed SDNN outperforms (in terms of GAP@20) the DNN by 1% and 1.5% using visual
and audiovisual features, respectively; both networks outperform Logistic Regression (LR)
• A performance gain of 3% is observed by exploiting the audio information for both networks
Visual Visual + Audio
LR DNN SDNN LR DNN SDNN
Hit@1 82.4% 82.5% 83.2% 82.3% 85.2% 85.7%
PERR 71.9% 72.2% 72.9% 71.8% 75.4% 75.9%
mAP 41.2% 42.3% 45.2% 40.1% 45.6% 47.9%
GAP@20 77.1% 77.6% 78.6% 77% 80.7% 82.2%
Ttr (min) 18.7 59.2 66.2 18.9 60.3 67.1
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Experiments
24
Results
• Comparison with best single-model results in YT8M
• These methods use frame-level features and build upon stronger and computationally
demanding feature descriptors (e.g. Fisher Vectors, VLAD, FVNet, DBoF and other)
• The proposed SDNN performs in par with top-performers in the literature despite the fact that we
utilize only the video-level features provided from YT8M
[1] Huang, P.,Yuan,Y., Lan, Z., Jiang, L., Hauptmann,A.G.:Video representation learning and latent concept mining for large-scale multi-label
video classification. arXiv:1707.01408 (2017)
[2] Na, S.,Yu,Y., Lee, S., Kim, J., Kim, G.: Encoding video and label priors for multilabel video classification onYouTube-8M dataset. In: CVPR
Workshops (2017)
[3] Bober-Irizar, M., Husain, S., Ong, E.J., Bober, M.: Cultivating DNN diversity for large scale video labelling. In: CVPRWorkshops (2017)
[4] Skalic, M., Pekalski, M., Pan, X.E.: Deep learning methods for ecient large scale video labeling. In: CVPRWorkshops (2017)
[1] [2] [3] [4] SDNN
GAP@20 82.15% 80.9% 82.5% 82.25% 82.2%
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
Summary and next steps
25
Summary
• A new criterion was suggested to identify the neglected classes, i.e. classes that getting less
attention from the DNN during the training procedure
• Subclass DNNs were proposed, where the identified classes are augmented and partitioned to
subclasses, and, a new CE Loss is applied to exploit effectively the subclass labelling information
• The proposed approach was evaluated successfully in two different problem domains, multiclass
classification (CIFAR-10, CIFAR-100, SVHN) and multilabel classification (YT8M)
Next steps
• Further improve the classification performance of the proposed approach using a mixture of
experts model to combine the subclass scores at the output layer
• Extend the proposed approach to the semi-supervised paradigm in order to exploit more
effectively partially labelled datasets
retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project
26
Thank you for your attention!
Questions?
Vasileios Mezaris, bmezaris@iti.gr
Code publicly available at:
https://github.com/bmezaris/subclass_deep_neural_networks
This work was supported by the EUs Horizon 2020 research and innovation
programme under grant agreement H2020-780656 ReTV
1 of 26

Recommended

LSTM Structured Pruning by
LSTM Structured PruningLSTM Structured Pruning
LSTM Structured PruningVasileiosMezaris
510 views18 slides
Video summarization using clustering by
Video summarization using clusteringVideo summarization using clustering
Video summarization using clusteringSahil Biswas
2.9K views15 slides
Cloud, Fog, or Edge: Where and When to Compute? by
Cloud, Fog, or Edge: Where and When to Compute?Cloud, Fog, or Edge: Where and When to Compute?
Cloud, Fog, or Edge: Where and When to Compute?Förderverein Technische Fakultät
2.7K views43 slides
Media4Math's Fall 2012 Catalog by
Media4Math's Fall 2012 CatalogMedia4Math's Fall 2012 Catalog
Media4Math's Fall 2012 CatalogMedia4math
572 views8 slides
Icml2018 naver review by
Icml2018 naver reviewIcml2018 naver review
Icml2018 naver reviewNAVER Engineering
1.7K views110 slides
Multi-View Video Coding Algorithms/Techniques: A Comprehensive Study by
Multi-View Video Coding Algorithms/Techniques: A Comprehensive StudyMulti-View Video Coding Algorithms/Techniques: A Comprehensive Study
Multi-View Video Coding Algorithms/Techniques: A Comprehensive StudyIJERA Editor
71 views6 slides

More Related Content

What's hot

"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio... by
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio..."Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...Edge AI and Vision Alliance
1.7K views35 slides
Understanding user interactivity for immersive communications and its impact ... by
Understanding user interactivity for immersive communications and its impact ...Understanding user interactivity for immersive communications and its impact ...
Understanding user interactivity for immersive communications and its impact ...lauratoni4
363 views24 slides
An Intelligent Approach for Effective Retrieval of Content from Large Data Se... by
An Intelligent Approach for Effective Retrieval of Content from Large Data Se...An Intelligent Approach for Effective Retrieval of Content from Large Data Se...
An Intelligent Approach for Effective Retrieval of Content from Large Data Se...IJCSIS Research Publications
67 views8 slides
Network recasting by
Network recastingNetwork recasting
Network recastingNAVER Engineering
530 views66 slides
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Models by
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep ModelsMediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Models
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Modelsmultimediaeval
177 views16 slides
Devil in the Details: Analysing the Performance of ConvNet Features by
Devil in the Details: Analysing the Performance of ConvNet FeaturesDevil in the Details: Analysing the Performance of ConvNet Features
Devil in the Details: Analysing the Performance of ConvNet FeaturesKen Chatfield
1.7K views27 slides

What's hot(10)

"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio... by Edge AI and Vision Alliance
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio..."Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
"Quantizing Deep Networks for Efficient Inference at the Edge," a Presentatio...
Understanding user interactivity for immersive communications and its impact ... by lauratoni4
Understanding user interactivity for immersive communications and its impact ...Understanding user interactivity for immersive communications and its impact ...
Understanding user interactivity for immersive communications and its impact ...
lauratoni4363 views
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Models by multimediaeval
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep ModelsMediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Models
MediaEval 2016 - HUCVL Predicting Interesting Key Frames with Deep Models
multimediaeval177 views
Devil in the Details: Analysing the Performance of ConvNet Features by Ken Chatfield
Devil in the Details: Analysing the Performance of ConvNet FeaturesDevil in the Details: Analysing the Performance of ConvNet Features
Devil in the Details: Analysing the Performance of ConvNet Features
Ken Chatfield1.7K views

Similar to Subclass deep neural networks

Fractional step discriminant pruning by
Fractional step discriminant pruningFractional step discriminant pruning
Fractional step discriminant pruningVasileiosMezaris
90 views20 slides
Neural Architectures for Still Images - Xavier Giro- UPC Barcelona 2019 by
Neural Architectures for Still Images - Xavier Giro- UPC Barcelona 2019Neural Architectures for Still Images - Xavier Giro- UPC Barcelona 2019
Neural Architectures for Still Images - Xavier Giro- UPC Barcelona 2019Universitat Politècnica de Catalunya
399 views58 slides
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020 by
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Universitat Politècnica de Catalunya
333 views64 slides
Multimodal Residual Networks for Visual QA by
Multimodal Residual Networks for Visual QAMultimodal Residual Networks for Visual QA
Multimodal Residual Networks for Visual QAJin-Hwa Kim
261 views25 slides
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks) by
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)Fellowship at Vodafone FutureLab
2K views22 slides
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks) by
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)Fellowship at Vodafone FutureLab
389 views22 slides

Similar to Subclass deep neural networks(20)

Multimodal Residual Networks for Visual QA by Jin-Hwa Kim
Multimodal Residual Networks for Visual QAMultimodal Residual Networks for Visual QA
Multimodal Residual Networks for Visual QA
Jin-Hwa Kim261 views
RICE INSECTS CLASSIFICATION USIING TRANSFER LEARNING AND CNN by IRJET Journal
RICE INSECTS CLASSIFICATION USIING TRANSFER LEARNING AND CNNRICE INSECTS CLASSIFICATION USIING TRANSFER LEARNING AND CNN
RICE INSECTS CLASSIFICATION USIING TRANSFER LEARNING AND CNN
IRJET Journal4 views
A Pointing Gesture-based Signal to Text Communication System Using OpenCV in ... by IRJET Journal
A Pointing Gesture-based Signal to Text Communication System Using OpenCV in ...A Pointing Gesture-based Signal to Text Communication System Using OpenCV in ...
A Pointing Gesture-based Signal to Text Communication System Using OpenCV in ...
IRJET Journal5 views
Reservoir computing fast deep learning for sequences by Claudio Gallicchio
Reservoir computing   fast deep learning for sequencesReservoir computing   fast deep learning for sequences
Reservoir computing fast deep learning for sequences
Claudio Gallicchio395 views
imageclassification-160206090009.pdf by KammetaJoshna
imageclassification-160206090009.pdfimageclassification-160206090009.pdf
imageclassification-160206090009.pdf
KammetaJoshna7 views
Image classification with Deep Neural Networks by Yogendra Tamang
Image classification with Deep Neural NetworksImage classification with Deep Neural Networks
Image classification with Deep Neural Networks
Yogendra Tamang7.6K views
Garbage Classification Using Deep Learning Techniques by IRJET Journal
Garbage Classification Using Deep Learning TechniquesGarbage Classification Using Deep Learning Techniques
Garbage Classification Using Deep Learning Techniques
IRJET Journal278 views
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis by Olga Scrivner
Workshop nwav 47 - LVS - Tool for Quantitative Data AnalysisWorkshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Workshop nwav 47 - LVS - Tool for Quantitative Data Analysis
Olga Scrivner278 views
Machine Learning Algorithms for Image Classification of Hand Digits and Face ... by IRJET Journal
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
IRJET Journal43 views
Masking preprocessing in transfer learning for damage building detection by IAESIJAI
Masking preprocessing in transfer learning for damage building detectionMasking preprocessing in transfer learning for damage building detection
Masking preprocessing in transfer learning for damage building detection
IAESIJAI21 views

More from VasileiosMezaris

TAME: Trainable Attention Mechanism for Explanations by
TAME: Trainable Attention Mechanism for ExplanationsTAME: Trainable Attention Mechanism for Explanations
TAME: Trainable Attention Mechanism for ExplanationsVasileiosMezaris
110 views22 slides
Gated-ViGAT by
Gated-ViGATGated-ViGAT
Gated-ViGATVasileiosMezaris
28 views18 slides
Explaining video summarization based on the focus of attention by
Explaining video summarization based on the focus of attentionExplaining video summarization based on the focus of attention
Explaining video summarization based on the focus of attentionVasileiosMezaris
69 views35 slides
Combining textual and visual features for Ad-hoc Video Search by
Combining textual and visual features for Ad-hoc Video SearchCombining textual and visual features for Ad-hoc Video Search
Combining textual and visual features for Ad-hoc Video SearchVasileiosMezaris
9 views18 slides
Explaining the decisions of image/video classifiers by
Explaining the decisions of image/video classifiersExplaining the decisions of image/video classifiers
Explaining the decisions of image/video classifiersVasileiosMezaris
62 views67 slides
Learning visual explanations for DCNN-based image classifiers using an attent... by
Learning visual explanations for DCNN-based image classifiers using an attent...Learning visual explanations for DCNN-based image classifiers using an attent...
Learning visual explanations for DCNN-based image classifiers using an attent...VasileiosMezaris
58 views16 slides

More from VasileiosMezaris(20)

TAME: Trainable Attention Mechanism for Explanations by VasileiosMezaris
TAME: Trainable Attention Mechanism for ExplanationsTAME: Trainable Attention Mechanism for Explanations
TAME: Trainable Attention Mechanism for Explanations
VasileiosMezaris110 views
Explaining video summarization based on the focus of attention by VasileiosMezaris
Explaining video summarization based on the focus of attentionExplaining video summarization based on the focus of attention
Explaining video summarization based on the focus of attention
VasileiosMezaris69 views
Combining textual and visual features for Ad-hoc Video Search by VasileiosMezaris
Combining textual and visual features for Ad-hoc Video SearchCombining textual and visual features for Ad-hoc Video Search
Combining textual and visual features for Ad-hoc Video Search
Explaining the decisions of image/video classifiers by VasileiosMezaris
Explaining the decisions of image/video classifiersExplaining the decisions of image/video classifiers
Explaining the decisions of image/video classifiers
VasileiosMezaris62 views
Learning visual explanations for DCNN-based image classifiers using an attent... by VasileiosMezaris
Learning visual explanations for DCNN-based image classifiers using an attent...Learning visual explanations for DCNN-based image classifiers using an attent...
Learning visual explanations for DCNN-based image classifiers using an attent...
VasileiosMezaris58 views
Are all combinations equal? Combining textual and visual features with multi... by VasileiosMezaris
Are all combinations equal?  Combining textual and visual features with multi...Are all combinations equal?  Combining textual and visual features with multi...
Are all combinations equal? Combining textual and visual features with multi...
VasileiosMezaris12 views
Hard-Negatives Selection Strategy for Cross-Modal Retrieval by VasileiosMezaris
Hard-Negatives Selection Strategy for Cross-Modal RetrievalHard-Negatives Selection Strategy for Cross-Modal Retrieval
Hard-Negatives Selection Strategy for Cross-Modal Retrieval
VasileiosMezaris77 views
Misinformation on the internet: Video and AI by VasileiosMezaris
Misinformation on the internet: Video and AIMisinformation on the internet: Video and AI
Misinformation on the internet: Video and AI
VasileiosMezaris176 views
Icme2020 tutorial video_summarization_part1 by VasileiosMezaris
Icme2020 tutorial video_summarization_part1Icme2020 tutorial video_summarization_part1
Icme2020 tutorial video_summarization_part1
VasileiosMezaris1K views
Video, AI and News: video analysis and verification technologies for supporti... by VasileiosMezaris
Video, AI and News: video analysis and verification technologies for supporti...Video, AI and News: video analysis and verification technologies for supporti...
Video, AI and News: video analysis and verification technologies for supporti...
VasileiosMezaris349 views
Unsupervised Video Summarization via Attention-Driven Adversarial Learning by VasileiosMezaris
Unsupervised Video Summarization via Attention-Driven Adversarial LearningUnsupervised Video Summarization via Attention-Driven Adversarial Learning
Unsupervised Video Summarization via Attention-Driven Adversarial Learning
VasileiosMezaris1.3K views
Video & AI: capabilities and limitations of AI in detecting video manipulations by VasileiosMezaris
Video & AI: capabilities and limitations of AI in detecting video manipulationsVideo & AI: capabilities and limitations of AI in detecting video manipulations
Video & AI: capabilities and limitations of AI in detecting video manipulations
VasileiosMezaris283 views

Recently uploaded

1978 NASA News Release Log by
1978 NASA News Release Log1978 NASA News Release Log
1978 NASA News Release Logpurrterminator
9 views146 slides
Guinea Pig as a Model for Translation Research by
Guinea Pig as a Model for Translation ResearchGuinea Pig as a Model for Translation Research
Guinea Pig as a Model for Translation ResearchPervaizDar1
11 views21 slides
TF-FAIR.pdf by
TF-FAIR.pdfTF-FAIR.pdf
TF-FAIR.pdfDirk Roorda
6 views120 slides
Batrachospermum.pptx by
Batrachospermum.pptxBatrachospermum.pptx
Batrachospermum.pptxnisarahmad632316
43 views37 slides
Light Pollution for LVIS students by
Light Pollution for LVIS studentsLight Pollution for LVIS students
Light Pollution for LVIS studentsCWBarthlmew
6 views12 slides
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ... by
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...ILRI
5 views1 slide

Recently uploaded(20)

Guinea Pig as a Model for Translation Research by PervaizDar1
Guinea Pig as a Model for Translation ResearchGuinea Pig as a Model for Translation Research
Guinea Pig as a Model for Translation Research
PervaizDar111 views
Light Pollution for LVIS students by CWBarthlmew
Light Pollution for LVIS studentsLight Pollution for LVIS students
Light Pollution for LVIS students
CWBarthlmew6 views
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ... by ILRI
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...
ILRI5 views
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ... by ILRI
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...
Small ruminant keepers’ knowledge, attitudes and practices towards peste des ...
ILRI5 views
Pollination By Nagapradheesh.M.pptx by MNAGAPRADHEESH
Pollination By Nagapradheesh.M.pptxPollination By Nagapradheesh.M.pptx
Pollination By Nagapradheesh.M.pptx
MNAGAPRADHEESH16 views
Metatheoretical Panda-Samaneh Borji.pdf by samanehborji
Metatheoretical Panda-Samaneh Borji.pdfMetatheoretical Panda-Samaneh Borji.pdf
Metatheoretical Panda-Samaneh Borji.pdf
samanehborji16 views
MODULE-9-Biotechnology, Genetically Modified Organisms, and Gene Therapy.pdf by KerryNuez1
MODULE-9-Biotechnology, Genetically Modified Organisms, and Gene Therapy.pdfMODULE-9-Biotechnology, Genetically Modified Organisms, and Gene Therapy.pdf
MODULE-9-Biotechnology, Genetically Modified Organisms, and Gene Therapy.pdf
KerryNuez124 views
Open Access Publishing in Astrophysics by Peter Coles
Open Access Publishing in AstrophysicsOpen Access Publishing in Astrophysics
Open Access Publishing in Astrophysics
Peter Coles808 views
ENTOMOLOGY PPT ON BOMBYCIDAE AND SATURNIIDAE.pptx by MN
ENTOMOLOGY PPT ON BOMBYCIDAE AND SATURNIIDAE.pptxENTOMOLOGY PPT ON BOMBYCIDAE AND SATURNIIDAE.pptx
ENTOMOLOGY PPT ON BOMBYCIDAE AND SATURNIIDAE.pptx
MN7 views
RemeOs science and clinical evidence by PetrusViitanen1
RemeOs science and clinical evidenceRemeOs science and clinical evidence
RemeOs science and clinical evidence
PetrusViitanen136 views
Conventional and non-conventional methods for improvement of cucurbits.pptx by gandhi976
Conventional and non-conventional methods for improvement of cucurbits.pptxConventional and non-conventional methods for improvement of cucurbits.pptx
Conventional and non-conventional methods for improvement of cucurbits.pptx
gandhi97618 views

Subclass deep neural networks

  • 1. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Title of presentation Subtitle Name of presenter Date Subclass deep neural networks: re-enabling neglected classes in deep network training for multimedia classification N. Gkalelis,V. Mezaris CERTH-ITI,Thermi -Thessaloniki,Greece 26th Int. Conf. Multimedia Modeling Daejeon, Korea, January 2020
  • 2. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Outline 2 • Problem statement • Related work • Identification of neglected classes • Subclass DNNs • Experiments
  • 3. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 3 • Deep neural networks (DNNs) have shown remarkable performance in many machine learning problems • They are being commercially deployed in various domains Problem statement • Multimedia understanding • Self-driving cars • Edge computing Image Credits:V2Gov ImageCredits: [1] [1]Chen, J., Ran, X.: Deep LearningWith Edge Computing:A Review, Proc. of the IEEE, vol. 107, no. 8, (Aug. 2019)
  • 4. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 4 One weakness of DNNs is that some classes are getting less attention during training, for instance, due to: • Overfitting: error in the training set becomes negligible although generalization error is still high • Simply, the contributions of the misclassified observations to the updating of the weights associated with these classes cancel out  How to identify these neglected classes?  How to boost the classification performance of neglected classes? Problem statement
  • 5. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 5 Related work • The problem of neglected classes sometimes is related with imbalanced class learning, where the distribution of training data across different classes is skewed • There are relatively few works studying the class imbalance problem in DNNs (e.g. see [1, 2])  As observed in our experiments neglected classes may also appear with balanced training datasets (e.g. CIFAR datasets)  The identification of classes getting less attention from the DNN during the training procedure is a relatively unexplored topic [1] Sarafianos, N., Xu, X., Kakadiaris, I.A.: Deep imbalanced attribute classification using visual attention aggregation. ECCV. Munich, Germany (Sep 2018) [2] Lin,T., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal Loss for DenseObject Detection, 2017 IEEE ICCV,Venice, 2017.
  • 6. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 6 Identification of neglected classes: motivation • During backpropagation, a strong gradient update gi for the weight vectors wi at the output layer is typically produced for classes with high training error ei • In this way, the overall network is guided to extract nonlinear features at different layers in order to provide a linear separable subspace at the output layer for these classes wi wi’ = wi + gi gi length of the gradient update ||gi|| is large as expected due to a high error in the output node i weight vector is updated (as well as weight vectors in the layers below) in order to reduce the output node error weight vector of node i (associated with class i)
  • 7. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 7 Identification of neglected classes: motivation • However, it is observed that for certain output nodes the gradient update is close to zero although the classes associated with these nodes still have a large error • The result is that the DNN neglects to extract the necessary discriminant features for reducing the error for these classes • We investigate this case in detail and based on the analysis we propose a new measure to identify the so-called neglected classes The length of the gradient update ||gi|| is small despite a large error in the output node i! the updated weight vector of output node i remains relatively unchanged because the length of the gradient update ||gi|| is close to zero wi’  wi gi
  • 8. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 8 Identification of neglected classes: background • Suppose a DNN with sigmoid (SG) output layer and cross entropy (CE) loss xk xf,k xj,k x1,k wi w1,i wj,i bi weight vector and bias term of i-th node input vector to the i-th node associated with the k-th training observation output of the i-th node associated with k-th training observation i-th output node of the SG layer (associated with i-th class) wf,i
  • 9. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 9 Identification of neglected classes: background gradient update of wi label of k-th observation w.r.t. class i Learning rate input vector to the i-th output node associated with the k-th training observation in the batch output of the i-th node associated with k-th training observation Number of batch observations
  • 10. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 10 Identification of neglected classes Unnormalized gradient update (w.r.t i-th node) Unnormalized gradient update of positive class observations Unnormalized gradient update of negative class observations
  • 11. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 11 Identification of neglected classes
  • 12. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 12 Identification of neglected classes
  • 13. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 13 Identification of neglected classes Length of unnormalized gradient update of positive and negative class observations Length of unnormalized gradient update (w.r.t i-th node)
  • 14. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 14 Training of neglected classes: Subclass DNNs • A subclass-based learning framework is applied to improve the classification performance of neglected classes • Subclass-based classification techniques have been successfully used in the shallow learning paradigm [1-5]. • To this end, neglected classes are augmented and partitioned to subclasses and the created labeling information is used to create and train a subclass DNN [1] Hastie,T.,Tibshirani, R.: Discriminant analysis by Gaussian mixtures. Journal of the Royal Statistical Society. Series B 58(1), (Jul 1996) [2] Manli, Z., Martinez, A.M.: Subclass discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 28(8), (Aug 2006) [3] Escalera, S. et al.: Subclass problem-dependent design for error-correcting output codes. IEEETrans. PatternAnal. Mach. Intell. 30(6), (Jun 2008) [4]You, D., Hamsici, O.C., Martinez, A.M.: Kernel optimization in discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 33(3), (Mar 2011) [5] Gkalelis, N., Mezaris,V., Kompatsiaris, I., Stathaki,T.: Mixture subclass discriminant analysis link to restricted Gaussian model and other generalizations. IEEETrans. Neural Netw. Learn. Syst. 24(1), (Jan 2013)
  • 15. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 15 Training of neglected classes: Subclass DNNs Number of training observations Number of classes Number of subclasses of class i label of k-th observation w.r.t subclass (i,j) output of the node (i.j) associated with k-th training observation label of k-th observation w.r.t class i
  • 16. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 16 Multiclass classification • CIFAR10 [1]: 10 classes, 32 x 32 color images, 50000 training and 10000 testing observations • CIFAR100 [1]: as CIFAR10 but with 100 classes • SVHN [2]: 10 classes, 32 x 32 color images, 73257 training, 26032 testing and 531131 extra observations [1] Krizhevsky, A.: Learning multiple layers of features from tiny images.Tech. rep. (2009) [2] Netzer,Y.,Wang,T., Coates, A., Bissacco, A.,Wu, B., Ng, A.Y.: Reading digits in natural images with unsupervised feature learning. NIPS Workshops.Venice, Italy, (Oct 2017)
  • 17. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 17 Experimental setup (similar setup with [1]) • Images are normalized to zero mean and unit variance • Data augmentation is applied as in [1] (cropping, mirroring, cutout, etc.) • DNNs [2,3]: VGG16, WRN-28-10 (CIFAR-10, -100), WRN-16-8 (SVHN) • CE loss, Minibatch SGD, Nesterov momentum 0.9, batch size 128, weight decay 0.0005, 200 epochs, initial learning rate 0.1, exponential schedule (learning rate multiplied by 0.1 for CIFAR or 0.2 for SVHN, at epochs 60, 120 and 160) [1] Devries,T.,Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 (2017) [2]Zagoruyko, S., Komodakis, N.:Wide residual networks. BMVC.York, UK, (Sep 2016) [3] Simonyan, K., Zisserman, A.:Very deep convolutional networks for large-scale image recognition. ICLR. San Diego, CA, USA (May 2015)
  • 18. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 18 SDNN training • DNN is trained for 30 epochs in order to compute a reliable neglection score for each class • The classes with highest neglection score are selected: 2 classes for CIFAR-10 and SVHN, 10 classes for CIFAR-100 • Observations of the selected classes are doubled using the augmentation approach presented in [1] • Selected classes are partitioned to 2 subclasses using k-means • SDNN is trained using the annotations at class and subclass level [1] Inoue, H.: Data augmentation by pairing samples for images classification. arXiv:1801.02929 (2018)
  • 19. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 19 Results • The proposed subclass DNNs (SVGG16, SWRN) provide improved performance over their conventional variants (VGG16 [1], WRN [2]) • Considering that [2] is among the state-of-the-art approaches, the attained CCR performance improvement is significant • Training time overhead is negligible for the medium-size datasets (CIFAR-10, -100) and relatively small for the larger SVHN; testing time is only a few seconds for all DNNs VGG16 [1] SVGG16 WRN [2] SWRN CIFAR-10 93.5% (2.6 h) 94.8% (2.7 h) 96.92% (7.1 h) 97.14% (8.1 h) CIFAR-100 71.24% (2.6 h) 73.67% (2.6 h) 81.59% (7.1 h) 82.17% (7.5 h) SVHN 98.16% (29.1 h) 98.35 (34.1 h) 98.7% (33.1 h) 98.81% (42.7 h) [1] Simonyan, K., Zisserman, A.:Very deep convolutional networks for large-scale image recognition. In: ICLR. San Diego, CA, USA (May 2015) [2] Devries,T.,Taylor, G.W.: Improved regularization of convolutional neural networks with cutout. arXiv:1708.04552 (2017)
  • 20. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 20 Multilabel classification • YouTube-8M (YT8M) [1]: largest public dataset for multilabel video classification • 3862 classes, 3888919 training, 1112356 evaluation and 1133323 testing videos • 1024- and 128-dimensional visual and audio feature vectors are provided for each video [1]Abu-El-Haija, S., Kothari, N., Lee, J., Natsev, A.P.,Toderici, G.,Varadarajan, B.,Vijayanarasimhan, S.:YouTube-8M:A large-scale video classification benchmark. In: arXiv:1609.08675 (2016)
  • 21. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 21 Experimental setup • Two experiments: using the visual feature vectors directly; concatenating the visual and audio feature vectors • L2-normalization is applied to feature vectors in all cases • DNN: 1 conv layer with 64 1d-filters, max-pooling, dropout, SG output layer • CE loss, Minibatch SGD, batch size 512, weight decay 0.0005, 5 epochs, initial learning rate 0.001, exponential schedule (learning rate multiplied by 0.95 at every epoch) • The evaluation metrics of YT8M challenge [1] are utilized for performance evaluation; GAP@20 is used as the primary evaluation metric [1]Abu-El-Haija, S., Kothari, N., Lee, J., Natsev, A.P.,Toderici, G.,Varadarajan, B.,Vijayanarasimhan, S.:YouTube-8M:A large-scale video classification benchmark. In: arXiv:1609.08675 (2016)
  • 22. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 22 SDNN training • DNN is trained for 1/3 of an epoch; a neglection score for each class is computed • 386 classes with highest neglection score are selected • Observations of the selected classes are doubled by applying extrapolation in feature space [1] • A efficient variant of the nearest neighbor-based clustering algorithm [2,3] is used to partition the neglected classes to subclasses • SDNN is trained using the annotations at class and subclass level [1] DeVries,T.,Taylor, G.W.: Dataset augmentation in feature space. ICLRWorkshops.Toulon, France (Apr 2017) [2] Manli, Z., Martinez, A.M.: Subclass discriminant analysis. IEEETrans. PatternAnal. Mach. Intell. 28(8), 1274-1286 (Aug 2006) [3]You, D., Hamsici, O.C., Martinez, A.M.: Kernel optimization in discriminant analysis. IEEE PAMI 33(3), 631-638 (Mar 2011) Experiments
  • 23. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 23 Results • The proposed SDNN outperforms (in terms of GAP@20) the DNN by 1% and 1.5% using visual and audiovisual features, respectively; both networks outperform Logistic Regression (LR) • A performance gain of 3% is observed by exploiting the audio information for both networks Visual Visual + Audio LR DNN SDNN LR DNN SDNN Hit@1 82.4% 82.5% 83.2% 82.3% 85.2% 85.7% PERR 71.9% 72.2% 72.9% 71.8% 75.4% 75.9% mAP 41.2% 42.3% 45.2% 40.1% 45.6% 47.9% GAP@20 77.1% 77.6% 78.6% 77% 80.7% 82.2% Ttr (min) 18.7 59.2 66.2 18.9 60.3 67.1
  • 24. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Experiments 24 Results • Comparison with best single-model results in YT8M • These methods use frame-level features and build upon stronger and computationally demanding feature descriptors (e.g. Fisher Vectors, VLAD, FVNet, DBoF and other) • The proposed SDNN performs in par with top-performers in the literature despite the fact that we utilize only the video-level features provided from YT8M [1] Huang, P.,Yuan,Y., Lan, Z., Jiang, L., Hauptmann,A.G.:Video representation learning and latent concept mining for large-scale multi-label video classification. arXiv:1707.01408 (2017) [2] Na, S.,Yu,Y., Lee, S., Kim, J., Kim, G.: Encoding video and label priors for multilabel video classification onYouTube-8M dataset. In: CVPR Workshops (2017) [3] Bober-Irizar, M., Husain, S., Ong, E.J., Bober, M.: Cultivating DNN diversity for large scale video labelling. In: CVPRWorkshops (2017) [4] Skalic, M., Pekalski, M., Pan, X.E.: Deep learning methods for ecient large scale video labeling. In: CVPRWorkshops (2017) [1] [2] [3] [4] SDNN GAP@20 82.15% 80.9% 82.5% 82.25% 82.2%
  • 25. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project Summary and next steps 25 Summary • A new criterion was suggested to identify the neglected classes, i.e. classes that getting less attention from the DNN during the training procedure • Subclass DNNs were proposed, where the identified classes are augmented and partitioned to subclasses, and, a new CE Loss is applied to exploit effectively the subclass labelling information • The proposed approach was evaluated successfully in two different problem domains, multiclass classification (CIFAR-10, CIFAR-100, SVHN) and multilabel classification (YT8M) Next steps • Further improve the classification performance of the proposed approach using a mixture of experts model to combine the subclass scores at the output layer • Extend the proposed approach to the semi-supervised paradigm in order to exploit more effectively partially labelled datasets
  • 26. retv-project.eu @ReTV_EU @ReTVproject retv-project retv_project 26 Thank you for your attention! Questions? Vasileios Mezaris, bmezaris@iti.gr Code publicly available at: https://github.com/bmezaris/subclass_deep_neural_networks This work was supported by the EUs Horizon 2020 research and innovation programme under grant agreement H2020-780656 ReTV