SlideShare a Scribd company logo
Task-Adaptive Neural Network Search
with Meta-Contrastive Learning
Wonyong Jeong∗,#,$, Hayeon Lee∗,%,$, Geon Park∗,#,$,
Eunyoung Hyung#, Jinheon Baek#, and Sung Ju Hwang#,%,$
Graduate Shool of AI!, KAIST, Seoul, South Korea
School of Computing"
, KAIST, Daejeon, South Korea
AITRICS#, Seoul, South Korea
∗: 𝐸𝑞𝑢𝑎𝑙 𝐶𝑜𝑛𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛
1
Motivation
In most cases, the exhaustive trial-and-error and brute force efforts have been often
required to design and tune the neural networks to get good models on given datasets.
Neural Architecture Search (NAS) alleviates such costs by automatically building neural
architectures performing even higher than hand-crafted networks.
Search Strategy
Performance Estimation Strategy
Architecture Search Space
Feedback
Trials
Human Model
Manual Design Process
Network
Architecture
Estimated
Performance
Optimal
Architecture
Neural Architecture Search (NAS)
2
Motivation: The Limitations
Most conventional NAS approaches search for only optimal architectures without
generating parameters, which requires additional training steps on a given dataset.
While some recent NAS methods* depend on a supernet pretrained on ImageNet, they
may be suboptimal if the target tasks are highly dissimilar from ImageNet.
Pretraining Supernet on Large-Scale Dataset
Additional Training Phases on Target Dataset
*[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020.
3
Neural Network Search (NNS)
What if we can search not only optimal architectures but also relevant parameters on a
given dataset and conditions to reduce the additional training costs?
We newly introduce a novel problem of Neural Network Search (NNS), whose goal is to
search for the optimal pretrained networks for a given dataset and conditions.
Neural Network Search
Target
Dataset
Desired
Conditions
Optimal
Network
Relevant
Knowledge
Latency
Accuracy
…
# Params
4
Challenges
To do this, several critical and essential challenges should be properly tackled, such as
where to search and how to find the relevant pretrained models.
While tackling such challenges, we plan to construct our own model-zoo and learn the
cross-modal retrieval space to perform successful neural network search.
How to construct the model-zoo?
Neural Network Search
How to learn the cross modal space?
How to encode parameters? How to encode datasets?
…
…
…
…
5
TANS: Task-Adaptive Neural Network Search
To address such challenges, we newly propose our novel method, namely Task-Adaptive
Neural Network Search with Meta-Contrastive Learning (TANS).
TANS consists of several components, efficient model-zoo construction, model and
query encoders, performance predictor, and meta-contrastive learning framework.
6
Methodology: Model Encoder & Functional Embeddings
To learn the cross-modal retrieval space, we should properly encode both models and
datasets. For embedding pretrained models, how can we encode model parameters?
Our idea is to utilize individual model outputs from the single criteria input which is
unbiasedly generated from the Gaussian distribution, namely functional embeddings.
Unbiased Criteria Input
Generated from Gaussian dist.
Feed Forward
Across All Models
Models’ Individual
Interpretations
on the Criteria Input
7
Methodology: Model Encoder & Functional Embeddings
For architectural topology information, we adopt OFA*’s topological encodings which
contains number of layers, kernel sizes, and channel expansion ratios.
We then merge functional embeddings 𝑣M and topology information 𝑣N to learn model
embeddings 𝑚 such that model encoder 𝐸O 𝑣N, 𝑣M; 𝜙 ∶ ℳ → ℝP
Model
Encoder
⨁
Network Architecture
Functional Embedding
Model Embedding
*[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020.
8
Methodology: Query Encoder & Performance Predictor
We design simple pooling-based set encoder for our query encoder 𝐸Q 𝐷; 𝜃 : 𝒬 →
ℝP so that it can produce permutation-invariant query representation 𝑞.
Also, our performance predictor S 𝑚, 𝑞; 𝜓 takes both model embeddings 𝑚 and
query representations 𝑞 to estimate the performance with the given pair.
Model
Encoder
⨁
Network Architecture
Functional Embedding
Model
Embedding
Query
Encoder
Query
Embedding
Query Dataset
Performance
Predictor
Estimated
Performance
9
Methodology: Meta-Contrastive Learning
Putting model and query encoders and performance predictor altogether, we perform
amortized meta-contrastive learning to learn the cross-modal retrieval space.
Our algorithm maximizes distances of irrelevant model and query embeddings while
minimizing the matched pairs, being guided by our performance predictors.
Model
Encoder
Query
Encoder
Query
Embedding
Performance
Predictor
Model
Embedding
𝒒
𝒎$
𝒎$
𝒎$
𝒎%
Cross-Modal Latent Space for Model-Query Pairs
𝒒
𝒎%
𝒎$
𝒎$
𝒎$
Maximize Distance of
Negative Pairs
Minimize Distance
of Positive Pair
Guide Learning based on
Performance of Given Pairs
10
Methodology: Learning Objective
We design contrastive loss ℒT for model embeddings and ℒQ for query embeddings on
our cross-modal retrieval space, optimizing the parameters 𝜃 and 𝜙.
Further we optimize our performance predictor while learning the cross modal space
for accurately estimating the performance on given dataset and model pairs via MSE.
𝒒!
, 𝒎
𝒒"
, 𝒎
Mean Square Error:
11
We use an uncertainty-guided approach to iteratively select the dataset-model pairs
that are expected to expand the pareto frontier the most from the current state.
We can significantly reduce the size of the model zoo, while also having higher
performance compared to the randomly constructed model zoo.
Top-1 accuracy on dataset D
# params
Architecture B
Architecture A
Architecture C
Expected improvement of the
pareto front by training
Architecture B on D
Expected improvement of
the pareto front by training
Architecture C on D
Current pareto front
Methodology: Model-Zoo Construction
12
Experimental Setup: Datasets
We collect 96 real-world image datasets from Kaggle. We split them into 86 meta-
training and 10 meta-test datasets with no class-wise, instance-wise overlapping.
We further partition the meta-training datasets into 140 sub-datasets, so that each has
maximum 20 classes when the number of classes are extremely large. 13
Experimental Setup: Model-Zoo Construction
We train 100 neural network architectures sampled from OFA* space on 140 meta-
training datasets to construct the Model-Zoo consisting of 100*140 trained models.
In order to make this process more efficient, we can employ the efficient model zoo
construction algorithm to reduce the number of training rounds.
Model-Zoo Construction
from Real-world Datasets
N M
*[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020.
14
Experimental Setup: Baseline Models
We use six baselines in four categories, such as base architecture, conventional NAS,
weight-sharing approaches, and data-driven Meta-NAS.
MobileNet-V3 [1]
Conventional NAS
Weight-sharing NAS
Data-driven Meta-NAS
Base Architecture
PC-DARTS [2]
DrNAS [3]
FBNet-A [4]
Once-for-All [5]
MetaD2A [6]
[1] Howard, A et al. Searching for mobilenetv3, ICCV 2019.
[2] Xu, Y et al. Pc-darts: Partial channel connections for memory-efficient architecture search, ICLR 2020.
[3] Chen, X et al. Dr{nas}: Dirichlet neural architecture search, ICLR 2021.
[4] Wu, B et al. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. CVPR 2019.
[5] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020.
[6] Lee, H et al. Rapid neural architecture search by learn- ing to generate graphs from datasets. ICLR 2021.
15
Experimental Results: Meta-test Performance
TANS outperforms all baselines with almost zero search time and also greatly reduces
the training time as TANS can utilize a relevant pretrained knowledge .
Method
Pre-trained
Resource
Training
Epoch
Search
Time
(GPU sec)
Training
Time
(GPU sec)
Speed
Up
Accura
cy
(%)
MobileNetV3 ImageNet 1k 50 - 257 1.00× 94.20
PC-DARTS Scratch 500 1100.37 5721 0.04× 79.22
DrNAS Scratch 500 1501.75 5659 0.04× 84.06
FBNet-A ImageNet 1K 50 - 293 0.88× 93.00
OFA ImageNet 1K 50 121.90 226 0.74× 93.89
MetaD2A ImageNet 1K 50 2.59 345 0.74× 95.24
TANS (Ours)
Retrieved
task
50 0.002 200 1.28× 96.28
Averaged Performance of Searched (Retrieved) Networks on 10 unseen real-world datasets
5 unseen real-world datasets
16
Experimental Results: Semantic Similarity
We show example images from the unseen meta-test query dataset (Query) and meta-
train model-zoo datasets (Retrieval) that the retrieved models are pretrained on.
In most cases, our method matches semantically similar datasets to the query datasets.
Even for the semantically-dissimilar cases, our models still outperform other baselines.
Similar Cases Dissimilar Cases
Query Retrieval Query Retrieval
17
Experimental Results: Analysis & Ablation Study
We examine how accurately our model retrieves the paired network when the meta-
training dataset is given (we used unseen validation examples).
The meta-contrastive learning allows the model to accurately retrieve the same paired
models when the correspondent meta-train datasets are given.
Model
Recall
@Top 1
Recall
@Top 5
Mean
Random 2.14 2.86 69.04
Largest Parameter 3.57 7.14 51.85
TANS + Cosine Sim. Loss 9.29 12.86 46.02
TANS + Hard Neg. Loss 72.14 84.29 4.86
TANS + Meta-Contrastive Loss 80.71 96.43 1.9
TANS w/o Predictor 80.00 96.43 2.23
The Cross-Modal Retrieval Performance Visualization of The Cross-Modal Space
18
Experimental Results: Analysis & Ablation Study
With our performance predictor, we obtain 1.5 %p - 8%p performance gains on 10
meta-test datasets compared to the top 3 retrieved candidates.
Our efficient model-zoo construction algorithm selects Pareto-optimal network and
dataset pairs, creating the higher performing model-zoo over the naïve construction.
Performance Gain (%)
Effectiveness of Performance Predictor Effectiveness of our Model-zoo Construction Algorithm
19
Conclusion
• We newly introduced a novel problem of Neural Network Search (NNS), whose goal is to
search for the optimal pretrained networks for a given dataset and conditions.
• We propose a novel cross-modal retrieval framework to retrieve a pretrained network from
the model zoo for a given task via amortized meta-learning with contrastive objective.
• We propose an efficient model-zoo construction method to construct an effective database
of dataset-architecture pairs considering the model performance.
• We train and validate TANS on a newly collected large-scale database, on which our method
outperforms all NAS & AutoML baselines with almost no architecture search cost and
significantly fewer fine-tuning steps.
20
21
Thank You !

More Related Content

What's hot

Neural network
Neural networkNeural network
Neural network
Saddam Hussain
 
Pattern Recognition using Artificial Neural Network
Pattern Recognition using Artificial Neural NetworkPattern Recognition using Artificial Neural Network
Pattern Recognition using Artificial Neural Network
Editor IJCATR
 
ANN load forecasting
ANN load forecastingANN load forecasting
ANN load forecasting
Dr Ashok Tiwari
 
Kernel, RKHS, and Gaussian Processes
Kernel, RKHS, and Gaussian ProcessesKernel, RKHS, and Gaussian Processes
Kernel, RKHS, and Gaussian Processes
Sungjoon Choi
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognitionHưng Đặng
 
IROS 2017 Slides
IROS 2017 SlidesIROS 2017 Slides
IROS 2017 Slides
Sungjoon Choi
 
Introduction to Neural Network
Introduction to Neural NetworkIntroduction to Neural Network
Introduction to Neural Network
Yan Xu
 
Deep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter TuningDeep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter Tuning
Shubhmay Potdar
 
Advance deep learning
Advance deep learningAdvance deep learning
Advance deep learning
aliaKhan71
 
Artificial Neural Network Paper Presentation
Artificial Neural Network Paper PresentationArtificial Neural Network Paper Presentation
Artificial Neural Network Paper Presentation
guestac67362
 
Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)
spartacus131211
 
Forecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniquesForecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniquesHitesh Dua
 
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
Myungyon Kim
 
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksPR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
Taesu Kim
 
Learning to learn unlearned feature for segmentation
Learning to learn unlearned feature for segmentationLearning to learn unlearned feature for segmentation
Learning to learn unlearned feature for segmentation
NAVER Engineering
 
Neural Network Classification and its Applications in Insurance Industry
Neural Network Classification and its Applications in Insurance IndustryNeural Network Classification and its Applications in Insurance Industry
Neural Network Classification and its Applications in Insurance IndustryInderjeet Singh
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
Faria Priya
 
201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search
DaeJin Kim
 
Neural networks
Neural networksNeural networks
Neural networks
Rizwan Rizzu
 
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate HelpdeskDeep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Saurabh Saxena
 

What's hot (20)

Neural network
Neural networkNeural network
Neural network
 
Pattern Recognition using Artificial Neural Network
Pattern Recognition using Artificial Neural NetworkPattern Recognition using Artificial Neural Network
Pattern Recognition using Artificial Neural Network
 
ANN load forecasting
ANN load forecastingANN load forecasting
ANN load forecasting
 
Kernel, RKHS, and Gaussian Processes
Kernel, RKHS, and Gaussian ProcessesKernel, RKHS, and Gaussian Processes
Kernel, RKHS, and Gaussian Processes
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
IROS 2017 Slides
IROS 2017 SlidesIROS 2017 Slides
IROS 2017 Slides
 
Introduction to Neural Network
Introduction to Neural NetworkIntroduction to Neural Network
Introduction to Neural Network
 
Deep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter TuningDeep Dive into Hyperparameter Tuning
Deep Dive into Hyperparameter Tuning
 
Advance deep learning
Advance deep learningAdvance deep learning
Advance deep learning
 
Artificial Neural Network Paper Presentation
Artificial Neural Network Paper PresentationArtificial Neural Network Paper Presentation
Artificial Neural Network Paper Presentation
 
Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)Artificial Neural Network(Artificial intelligence)
Artificial Neural Network(Artificial intelligence)
 
Forecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniquesForecasting of Sales using Neural network techniques
Forecasting of Sales using Neural network techniques
 
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
Deep Learning and Tensorflow Implementation(딥러닝, 텐서플로우, 파이썬, CNN)_Myungyon Ki...
 
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksPR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
 
Learning to learn unlearned feature for segmentation
Learning to learn unlearned feature for segmentationLearning to learn unlearned feature for segmentation
Learning to learn unlearned feature for segmentation
 
Neural Network Classification and its Applications in Insurance Industry
Neural Network Classification and its Applications in Insurance IndustryNeural Network Classification and its Applications in Insurance Industry
Neural Network Classification and its Applications in Insurance Industry
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search
 
Neural networks
Neural networksNeural networks
Neural networks
 
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate HelpdeskDeep Learning Enabled Question Answering System to Automate Corporate Helpdesk
Deep Learning Enabled Question Answering System to Automate Corporate Helpdesk
 

Similar to Task Adaptive Neural Network Search with Meta-Contrastive Learning

Introduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-LearnIntroduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-Learn
Benjamin Bengfort
 
StackNet Meta-Modelling framework
StackNet Meta-Modelling frameworkStackNet Meta-Modelling framework
StackNet Meta-Modelling framework
Sri Ambati
 
Few shot learning/ one shot learning/ machine learning
Few shot learning/ one shot learning/ machine learningFew shot learning/ one shot learning/ machine learning
Few shot learning/ one shot learning/ machine learning
ﺁﺻﻒ ﻋﻠﯽ ﻣﯿﺮ
 
Machine Learning_Unit 2_Full.ppt.pdf
Machine Learning_Unit 2_Full.ppt.pdfMachine Learning_Unit 2_Full.ppt.pdf
Machine Learning_Unit 2_Full.ppt.pdf
Dr.DHANALAKSHMI SENTHILKUMAR
 
A Survey of Machine Learning Methods Applied to Computer ...
A Survey of Machine Learning Methods Applied to Computer ...A Survey of Machine Learning Methods Applied to Computer ...
A Survey of Machine Learning Methods Applied to Computer ...butest
 
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
IAEME Publication
 
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
IEEEFINALYEARSTUDENTPROJECTS
 
Guiding through a typical Machine Learning Pipeline
Guiding through a typical Machine Learning PipelineGuiding through a typical Machine Learning Pipeline
Guiding through a typical Machine Learning Pipeline
Michael Gerke
 
OpenML 2019
OpenML 2019OpenML 2019
OpenML 2019
Joaquin Vanschoren
 
How to Build a Neural Network and Make Predictions
How to Build a Neural Network and Make PredictionsHow to Build a Neural Network and Make Predictions
How to Build a Neural Network and Make Predictions
Developer Helps
 
Getting started with Machine Learning
Getting started with Machine LearningGetting started with Machine Learning
Getting started with Machine Learning
Gaurav Bhalotia
 
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
IEEEFINALYEARSTUDENTPROJECTS
 
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
IEEEMEMTECHSTUDENTSPROJECTS
 
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
IEEEFINALYEARSTUDENTPROJECT
 
IEEE Datamining 2016 Title and Abstract
IEEE  Datamining 2016 Title and AbstractIEEE  Datamining 2016 Title and Abstract
IEEE Datamining 2016 Title and Abstract
tsysglobalsolutions
 
Tuning the Untunable - Insights on Deep Learning Optimization
Tuning the Untunable - Insights on Deep Learning OptimizationTuning the Untunable - Insights on Deep Learning Optimization
Tuning the Untunable - Insights on Deep Learning Optimization
SigOpt
 
Improving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docImproving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docbutest
 
Everything you need to know about AutoML
Everything you need to know about AutoMLEverything you need to know about AutoML
Everything you need to know about AutoML
Arpitha Gurumurthy
 
A Stacked Generalization Ensemble Approach for Improved Intrusion Detection
A Stacked Generalization Ensemble Approach for Improved Intrusion DetectionA Stacked Generalization Ensemble Approach for Improved Intrusion Detection
A Stacked Generalization Ensemble Approach for Improved Intrusion Detection
IJCSIS Research Publications
 
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET Journal
 

Similar to Task Adaptive Neural Network Search with Meta-Contrastive Learning (20)

Introduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-LearnIntroduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-Learn
 
StackNet Meta-Modelling framework
StackNet Meta-Modelling frameworkStackNet Meta-Modelling framework
StackNet Meta-Modelling framework
 
Few shot learning/ one shot learning/ machine learning
Few shot learning/ one shot learning/ machine learningFew shot learning/ one shot learning/ machine learning
Few shot learning/ one shot learning/ machine learning
 
Machine Learning_Unit 2_Full.ppt.pdf
Machine Learning_Unit 2_Full.ppt.pdfMachine Learning_Unit 2_Full.ppt.pdf
Machine Learning_Unit 2_Full.ppt.pdf
 
A Survey of Machine Learning Methods Applied to Computer ...
A Survey of Machine Learning Methods Applied to Computer ...A Survey of Machine Learning Methods Applied to Computer ...
A Survey of Machine Learning Methods Applied to Computer ...
 
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
 
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
IEEE 2014 JAVA DATA MINING PROJECTS Active learning of constraints for semi s...
 
Guiding through a typical Machine Learning Pipeline
Guiding through a typical Machine Learning PipelineGuiding through a typical Machine Learning Pipeline
Guiding through a typical Machine Learning Pipeline
 
OpenML 2019
OpenML 2019OpenML 2019
OpenML 2019
 
How to Build a Neural Network and Make Predictions
How to Build a Neural Network and Make PredictionsHow to Build a Neural Network and Make Predictions
How to Build a Neural Network and Make Predictions
 
Getting started with Machine Learning
Getting started with Machine LearningGetting started with Machine Learning
Getting started with Machine Learning
 
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
IEEE 2014 JAVA DATA MINING PROJECTS Mining weakly labeled web facial images f...
 
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
 
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
2014 IEEE JAVA DATA MINING PROJECT Mining weakly labeled web facial images fo...
 
IEEE Datamining 2016 Title and Abstract
IEEE  Datamining 2016 Title and AbstractIEEE  Datamining 2016 Title and Abstract
IEEE Datamining 2016 Title and Abstract
 
Tuning the Untunable - Insights on Deep Learning Optimization
Tuning the Untunable - Insights on Deep Learning OptimizationTuning the Untunable - Insights on Deep Learning Optimization
Tuning the Untunable - Insights on Deep Learning Optimization
 
Improving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docImproving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..doc
 
Everything you need to know about AutoML
Everything you need to know about AutoMLEverything you need to know about AutoML
Everything you need to know about AutoML
 
A Stacked Generalization Ensemble Approach for Improved Intrusion Detection
A Stacked Generalization Ensemble Approach for Improved Intrusion DetectionA Stacked Generalization Ensemble Approach for Improved Intrusion Detection
A Stacked Generalization Ensemble Approach for Improved Intrusion Detection
 
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
 

More from MLAI2

Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
MLAI2
 
Online Hyperparameter Meta-Learning with Hypergradient Distillation
Online Hyperparameter Meta-Learning with Hypergradient DistillationOnline Hyperparameter Meta-Learning with Hypergradient Distillation
Online Hyperparameter Meta-Learning with Hypergradient Distillation
MLAI2
 
Online Coreset Selection for Rehearsal-based Continual Learning
Online Coreset Selection for Rehearsal-based Continual LearningOnline Coreset Selection for Rehearsal-based Continual Learning
Online Coreset Selection for Rehearsal-based Continual Learning
MLAI2
 
Representational Continuity for Unsupervised Continual Learning
Representational Continuity for Unsupervised Continual LearningRepresentational Continuity for Unsupervised Continual Learning
Representational Continuity for Unsupervised Continual Learning
MLAI2
 
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual LearningSequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
MLAI2
 
Skill-Based Meta-Reinforcement Learning
Skill-Based Meta-Reinforcement LearningSkill-Based Meta-Reinforcement Learning
Skill-Based Meta-Reinforcement Learning
MLAI2
 
Edge Representation Learning with Hypergraphs
Edge Representation Learning with HypergraphsEdge Representation Learning with Hypergraphs
Edge Representation Learning with Hypergraphs
MLAI2
 
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
MLAI2
 
Mini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
Mini-Batch Consistent Slot Set Encoder For Scalable Set EncodingMini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
Mini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
MLAI2
 
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
MLAI2
 
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-LearningMeta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
MLAI2
 
Accurate Learning of Graph Representations with Graph Multiset Pooling
Accurate Learning of Graph Representations with Graph Multiset PoolingAccurate Learning of Graph Representations with Graph Multiset Pooling
Accurate Learning of Graph Representations with Graph Multiset Pooling
MLAI2
 
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
MLAI2
 
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
MLAI2
 
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MLAI2
 
Adversarial Self-Supervised Contrastive Learning
Adversarial Self-Supervised Contrastive LearningAdversarial Self-Supervised Contrastive Learning
Adversarial Self-Supervised Contrastive Learning
MLAI2
 
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
MLAI2
 
Neural Mask Generator : Learning to Generate Adaptive Word Maskings for Langu...
Neural Mask Generator : Learning to Generate Adaptive WordMaskings for Langu...Neural Mask Generator : Learning to Generate Adaptive WordMaskings for Langu...
Neural Mask Generator : Learning to Generate Adaptive Word Maskings for Langu...
MLAI2
 
Adversarial Neural Pruning with Latent Vulnerability Suppression
Adversarial Neural Pruning with Latent Vulnerability SuppressionAdversarial Neural Pruning with Latent Vulnerability Suppression
Adversarial Neural Pruning with Latent Vulnerability Suppression
MLAI2
 
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
MLAI2
 

More from MLAI2 (20)

Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic Unce...
 
Online Hyperparameter Meta-Learning with Hypergradient Distillation
Online Hyperparameter Meta-Learning with Hypergradient DistillationOnline Hyperparameter Meta-Learning with Hypergradient Distillation
Online Hyperparameter Meta-Learning with Hypergradient Distillation
 
Online Coreset Selection for Rehearsal-based Continual Learning
Online Coreset Selection for Rehearsal-based Continual LearningOnline Coreset Selection for Rehearsal-based Continual Learning
Online Coreset Selection for Rehearsal-based Continual Learning
 
Representational Continuity for Unsupervised Continual Learning
Representational Continuity for Unsupervised Continual LearningRepresentational Continuity for Unsupervised Continual Learning
Representational Continuity for Unsupervised Continual Learning
 
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual LearningSequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
Sequential Reptile_Inter-Task Gradient Alignment for Multilingual Learning
 
Skill-Based Meta-Reinforcement Learning
Skill-Based Meta-Reinforcement LearningSkill-Based Meta-Reinforcement Learning
Skill-Based Meta-Reinforcement Learning
 
Edge Representation Learning with Hypergraphs
Edge Representation Learning with HypergraphsEdge Representation Learning with Hypergraphs
Edge Representation Learning with Hypergraphs
 
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Genera...
 
Mini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
Mini-Batch Consistent Slot Set Encoder For Scalable Set EncodingMini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
Mini-Batch Consistent Slot Set Encoder For Scalable Set Encoding
 
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint L...
 
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-LearningMeta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
 
Accurate Learning of Graph Representations with Graph Multiset Pooling
Accurate Learning of Graph Representations with Graph Multiset PoolingAccurate Learning of Graph Representations with Graph Multiset Pooling
Accurate Learning of Graph Representations with Graph Multiset Pooling
 
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
Contrastive Learning with Adversarial Perturbations for Conditional Text Gene...
 
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
Clinical Risk Prediction with Temporal Probabilistic Asymmetric Multi-Task Le...
 
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
 
Adversarial Self-Supervised Contrastive Learning
Adversarial Self-Supervised Contrastive LearningAdversarial Self-Supervised Contrastive Learning
Adversarial Self-Supervised Contrastive Learning
 
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
Learning to Extrapolate Knowledge: Transductive Few-shot Out-of-Graph Link Pr...
 
Neural Mask Generator : Learning to Generate Adaptive Word Maskings for Langu...
Neural Mask Generator : Learning to Generate Adaptive WordMaskings for Langu...Neural Mask Generator : Learning to Generate Adaptive WordMaskings for Langu...
Neural Mask Generator : Learning to Generate Adaptive Word Maskings for Langu...
 
Adversarial Neural Pruning with Latent Vulnerability Suppression
Adversarial Neural Pruning with Latent Vulnerability SuppressionAdversarial Neural Pruning with Latent Vulnerability Suppression
Adversarial Neural Pruning with Latent Vulnerability Suppression
 
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
Generating Diverse and Consistent QA pairs from Contexts with Information-Max...
 

Recently uploaded

Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
Alpen-Adria-Universität
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
Safe Software
 
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfSAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
Peter Spielvogel
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
Neo4j
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
Adtran
 
By Design, not by Accident - Agile Venture Bolzano 2024
By Design, not by Accident - Agile Venture Bolzano 2024By Design, not by Accident - Agile Venture Bolzano 2024
By Design, not by Accident - Agile Venture Bolzano 2024
Pierluigi Pugliese
 
National Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practicesNational Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practices
Quotidiano Piemontese
 
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptxSecstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
nkrafacyberclub
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
Neo4j
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Nexer Digital
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems S.M.S.A.
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
KatiaHIMEUR1
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
Matthew Sinclair
 
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdfFIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
Ana-Maria Mihalceanu
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
Neo4j
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance
 

Recently uploaded (20)

Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
 
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfSAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdf
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
 
By Design, not by Accident - Agile Venture Bolzano 2024
By Design, not by Accident - Agile Venture Bolzano 2024By Design, not by Accident - Agile Venture Bolzano 2024
By Design, not by Accident - Agile Venture Bolzano 2024
 
National Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practicesNational Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practices
 
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptxSecstrike : Reverse Engineering & Pwnable tools for CTF.pptx
Secstrike : Reverse Engineering & Pwnable tools for CTF.pptx
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
 
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdfFIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
FIDO Alliance Osaka Seminar: FIDO Security Aspects.pdf
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
 

Task Adaptive Neural Network Search with Meta-Contrastive Learning

  • 1. Task-Adaptive Neural Network Search with Meta-Contrastive Learning Wonyong Jeong∗,#,$, Hayeon Lee∗,%,$, Geon Park∗,#,$, Eunyoung Hyung#, Jinheon Baek#, and Sung Ju Hwang#,%,$ Graduate Shool of AI!, KAIST, Seoul, South Korea School of Computing" , KAIST, Daejeon, South Korea AITRICS#, Seoul, South Korea ∗: 𝐸𝑞𝑢𝑎𝑙 𝐶𝑜𝑛𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 1
  • 2. Motivation In most cases, the exhaustive trial-and-error and brute force efforts have been often required to design and tune the neural networks to get good models on given datasets. Neural Architecture Search (NAS) alleviates such costs by automatically building neural architectures performing even higher than hand-crafted networks. Search Strategy Performance Estimation Strategy Architecture Search Space Feedback Trials Human Model Manual Design Process Network Architecture Estimated Performance Optimal Architecture Neural Architecture Search (NAS) 2
  • 3. Motivation: The Limitations Most conventional NAS approaches search for only optimal architectures without generating parameters, which requires additional training steps on a given dataset. While some recent NAS methods* depend on a supernet pretrained on ImageNet, they may be suboptimal if the target tasks are highly dissimilar from ImageNet. Pretraining Supernet on Large-Scale Dataset Additional Training Phases on Target Dataset *[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020. 3
  • 4. Neural Network Search (NNS) What if we can search not only optimal architectures but also relevant parameters on a given dataset and conditions to reduce the additional training costs? We newly introduce a novel problem of Neural Network Search (NNS), whose goal is to search for the optimal pretrained networks for a given dataset and conditions. Neural Network Search Target Dataset Desired Conditions Optimal Network Relevant Knowledge Latency Accuracy … # Params 4
  • 5. Challenges To do this, several critical and essential challenges should be properly tackled, such as where to search and how to find the relevant pretrained models. While tackling such challenges, we plan to construct our own model-zoo and learn the cross-modal retrieval space to perform successful neural network search. How to construct the model-zoo? Neural Network Search How to learn the cross modal space? How to encode parameters? How to encode datasets? … … … … 5
  • 6. TANS: Task-Adaptive Neural Network Search To address such challenges, we newly propose our novel method, namely Task-Adaptive Neural Network Search with Meta-Contrastive Learning (TANS). TANS consists of several components, efficient model-zoo construction, model and query encoders, performance predictor, and meta-contrastive learning framework. 6
  • 7. Methodology: Model Encoder & Functional Embeddings To learn the cross-modal retrieval space, we should properly encode both models and datasets. For embedding pretrained models, how can we encode model parameters? Our idea is to utilize individual model outputs from the single criteria input which is unbiasedly generated from the Gaussian distribution, namely functional embeddings. Unbiased Criteria Input Generated from Gaussian dist. Feed Forward Across All Models Models’ Individual Interpretations on the Criteria Input 7
  • 8. Methodology: Model Encoder & Functional Embeddings For architectural topology information, we adopt OFA*’s topological encodings which contains number of layers, kernel sizes, and channel expansion ratios. We then merge functional embeddings 𝑣M and topology information 𝑣N to learn model embeddings 𝑚 such that model encoder 𝐸O 𝑣N, 𝑣M; 𝜙 ∶ ℳ → ℝP Model Encoder ⨁ Network Architecture Functional Embedding Model Embedding *[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020. 8
  • 9. Methodology: Query Encoder & Performance Predictor We design simple pooling-based set encoder for our query encoder 𝐸Q 𝐷; 𝜃 : 𝒬 → ℝP so that it can produce permutation-invariant query representation 𝑞. Also, our performance predictor S 𝑚, 𝑞; 𝜓 takes both model embeddings 𝑚 and query representations 𝑞 to estimate the performance with the given pair. Model Encoder ⨁ Network Architecture Functional Embedding Model Embedding Query Encoder Query Embedding Query Dataset Performance Predictor Estimated Performance 9
  • 10. Methodology: Meta-Contrastive Learning Putting model and query encoders and performance predictor altogether, we perform amortized meta-contrastive learning to learn the cross-modal retrieval space. Our algorithm maximizes distances of irrelevant model and query embeddings while minimizing the matched pairs, being guided by our performance predictors. Model Encoder Query Encoder Query Embedding Performance Predictor Model Embedding 𝒒 𝒎$ 𝒎$ 𝒎$ 𝒎% Cross-Modal Latent Space for Model-Query Pairs 𝒒 𝒎% 𝒎$ 𝒎$ 𝒎$ Maximize Distance of Negative Pairs Minimize Distance of Positive Pair Guide Learning based on Performance of Given Pairs 10
  • 11. Methodology: Learning Objective We design contrastive loss ℒT for model embeddings and ℒQ for query embeddings on our cross-modal retrieval space, optimizing the parameters 𝜃 and 𝜙. Further we optimize our performance predictor while learning the cross modal space for accurately estimating the performance on given dataset and model pairs via MSE. 𝒒! , 𝒎 𝒒" , 𝒎 Mean Square Error: 11
  • 12. We use an uncertainty-guided approach to iteratively select the dataset-model pairs that are expected to expand the pareto frontier the most from the current state. We can significantly reduce the size of the model zoo, while also having higher performance compared to the randomly constructed model zoo. Top-1 accuracy on dataset D # params Architecture B Architecture A Architecture C Expected improvement of the pareto front by training Architecture B on D Expected improvement of the pareto front by training Architecture C on D Current pareto front Methodology: Model-Zoo Construction 12
  • 13. Experimental Setup: Datasets We collect 96 real-world image datasets from Kaggle. We split them into 86 meta- training and 10 meta-test datasets with no class-wise, instance-wise overlapping. We further partition the meta-training datasets into 140 sub-datasets, so that each has maximum 20 classes when the number of classes are extremely large. 13
  • 14. Experimental Setup: Model-Zoo Construction We train 100 neural network architectures sampled from OFA* space on 140 meta- training datasets to construct the Model-Zoo consisting of 100*140 trained models. In order to make this process more efficient, we can employ the efficient model zoo construction algorithm to reduce the number of training rounds. Model-Zoo Construction from Real-world Datasets N M *[Once-for-All] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020. 14
  • 15. Experimental Setup: Baseline Models We use six baselines in four categories, such as base architecture, conventional NAS, weight-sharing approaches, and data-driven Meta-NAS. MobileNet-V3 [1] Conventional NAS Weight-sharing NAS Data-driven Meta-NAS Base Architecture PC-DARTS [2] DrNAS [3] FBNet-A [4] Once-for-All [5] MetaD2A [6] [1] Howard, A et al. Searching for mobilenetv3, ICCV 2019. [2] Xu, Y et al. Pc-darts: Partial channel connections for memory-efficient architecture search, ICLR 2020. [3] Chen, X et al. Dr{nas}: Dirichlet neural architecture search, ICLR 2021. [4] Wu, B et al. Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search. CVPR 2019. [5] Cai, H et al. Once-for-all: Train one network and specialize it for efficient deployment. ICLR 2020. [6] Lee, H et al. Rapid neural architecture search by learn- ing to generate graphs from datasets. ICLR 2021. 15
  • 16. Experimental Results: Meta-test Performance TANS outperforms all baselines with almost zero search time and also greatly reduces the training time as TANS can utilize a relevant pretrained knowledge . Method Pre-trained Resource Training Epoch Search Time (GPU sec) Training Time (GPU sec) Speed Up Accura cy (%) MobileNetV3 ImageNet 1k 50 - 257 1.00× 94.20 PC-DARTS Scratch 500 1100.37 5721 0.04× 79.22 DrNAS Scratch 500 1501.75 5659 0.04× 84.06 FBNet-A ImageNet 1K 50 - 293 0.88× 93.00 OFA ImageNet 1K 50 121.90 226 0.74× 93.89 MetaD2A ImageNet 1K 50 2.59 345 0.74× 95.24 TANS (Ours) Retrieved task 50 0.002 200 1.28× 96.28 Averaged Performance of Searched (Retrieved) Networks on 10 unseen real-world datasets 5 unseen real-world datasets 16
  • 17. Experimental Results: Semantic Similarity We show example images from the unseen meta-test query dataset (Query) and meta- train model-zoo datasets (Retrieval) that the retrieved models are pretrained on. In most cases, our method matches semantically similar datasets to the query datasets. Even for the semantically-dissimilar cases, our models still outperform other baselines. Similar Cases Dissimilar Cases Query Retrieval Query Retrieval 17
  • 18. Experimental Results: Analysis & Ablation Study We examine how accurately our model retrieves the paired network when the meta- training dataset is given (we used unseen validation examples). The meta-contrastive learning allows the model to accurately retrieve the same paired models when the correspondent meta-train datasets are given. Model Recall @Top 1 Recall @Top 5 Mean Random 2.14 2.86 69.04 Largest Parameter 3.57 7.14 51.85 TANS + Cosine Sim. Loss 9.29 12.86 46.02 TANS + Hard Neg. Loss 72.14 84.29 4.86 TANS + Meta-Contrastive Loss 80.71 96.43 1.9 TANS w/o Predictor 80.00 96.43 2.23 The Cross-Modal Retrieval Performance Visualization of The Cross-Modal Space 18
  • 19. Experimental Results: Analysis & Ablation Study With our performance predictor, we obtain 1.5 %p - 8%p performance gains on 10 meta-test datasets compared to the top 3 retrieved candidates. Our efficient model-zoo construction algorithm selects Pareto-optimal network and dataset pairs, creating the higher performing model-zoo over the naïve construction. Performance Gain (%) Effectiveness of Performance Predictor Effectiveness of our Model-zoo Construction Algorithm 19
  • 20. Conclusion • We newly introduced a novel problem of Neural Network Search (NNS), whose goal is to search for the optimal pretrained networks for a given dataset and conditions. • We propose a novel cross-modal retrieval framework to retrieve a pretrained network from the model zoo for a given task via amortized meta-learning with contrastive objective. • We propose an efficient model-zoo construction method to construct an effective database of dataset-architecture pairs considering the model performance. • We train and validate TANS on a newly collected large-scale database, on which our method outperforms all NAS & AutoML baselines with almost no architecture search cost and significantly fewer fine-tuning steps. 20