SlideShare a Scribd company logo
Semi-supervised Learning Survey
- 20 years of evaluation
Subarna Chowdhury Soma
SJSU ID: 014549587
Paper
 A survey on semi-supervised learning
◦ Jesper E. van Engelen &Holger H. Hoos
◦ Received: 3 December 2018 / Revised: 20 September 2019 / Accepted: 29
September 2019 /
◦ Published online: 15 November 2019
Overview
 Semi-supervised learning
 Basic Concepts and Assumptions
 Taxonomy of Semi-supervised learning methods
 Semi-supervised learning methods
 Conclusions and future scope
Semi-supervised Learning
◦ In supervised learning, one is presented with a set of data points consisting of some
input x and a corresponding output value y. The goal is, then, to construct a classifier
or regressor that can estimate the output value for previously unseen inputs
◦ In unsupervised learning, no specific output value is provided; goal is to infer some
underlying structure from the inputs
◦ Semi-supervised learning is about making use of unlabeled data. It is useful when the
available labeled data is small.Unlabeled data is cheap Although not every type of
unlabeled data is useful
Basic Concepts and Assumptions
 Assumptions of semi-supervised learning
◦ An important pre-condition of semi-supervised learning is that the underlying data distribution, p(x), contains
information about the posterior distribution p(y|x). Considering this condition, we can use unlabeled data to
gain information about p(x), and thereby about p(y|x).
◦ The smoothness assumption : if two samples x and x1 are close in the input space, their labels y and
y1 should be the same
◦ The low-density assumption : the decision boundary should not pass through high-density areas in
the input space
◦ The manifold assumption : same low-dimensional manifold datapoints should have the same label
Illustrations of the
semi-supervised
learning
assumptions.
Basic Concepts and Assumptions
 Connection to Clustering
◦ According to Chapelle et al. 2006b, data points that belong to the same cluster come
from the same class. This assumption is called the semi-supervised learning assumption
of clusters.
Taxonomy of Semi-
supervised
learning methods
Inductive methods
◦ Inductive methods build a classifier to generate predictions for any point in the
input space
◦ To train the classifier, unlabeled data can be used.
◦ Once, a model is built from the training phase and later can be used to predict the
new data point labels.
Inductive methods
1. Wrapper methods
2. Unsupervised pre-processing
3. Intrinsically semi-supervised
1. Wrapper methods
o In the training step, one or more supervised classifiers are trained using the
labeled data and pseudo-labeled data from earlier iteration steps.
o Next, to infer labels from the previously trained unlabeled data, the resulting
classifiers are used in the pseudo-labeled step.
o Most confident predictions are used as the pseudo-label at the next iteration.
1. Wrapper methods
1. 1 Self-training:
o This approach uses one supervised classifier that is re-trained iteratively based on its most confident predictions
1.2 Co-training:
o Here classifiers are iteratively re-trained on the most confident predictions of each other
o For co-training to succeed, the classifiers are supposed to be adequately different
1.3 Boosting:
o Like conventional boosting methods, they build ensemble classifiers by sequentially building individual classifiers
o Moreover, each individual classifier is trained using both the labeled data and the most confident predictions of
previously trained unlabeled data like before
2. Unsupervised pre-processing
◦ Unsupervised preprocessing, either extract useful features from unlabeled data,
pre-cluster the data or use an unsupervised manner to compute the initial
parameters of a supervised learning procedure
2. Unsupervised pre-processing
2. 1 Feature Extraction:
o These methods try to find a transformation of the input data that helps to improve classification and makes it
computationally efficient
2.2 Cluster-then-label:
o Cluster-then-label procedures form a group of methods that explicitly join the clustering and classification
processes
o At first, it uses an unsupervised or semi-supervised clustering algorithm to all the input data, and then use the
resulting clusters lead the classification process
2.3 Pre-training:
o This approach naturally fits in deep learning areas where each layer of the hierarchical model is a latent
representation of the input data.
o In pre-training methods, a decision boundary is conducted by unlabeled data before applying supervised training
3. Intrinsically semi-supervised
◦ Here unlabeled data is directly incorporated into the objective function or optimization
procedure of the learning part.
◦ Most of these methods are direct extensions of supervised learning methods by
transforming objective functions to include unlabeled data.
3. Intrinsically semi-supervised
3. 1 Maximum Margin:
o These methods aim to maximize the distance between the given data points and the decision
boundary
o The semi-supervised low-density assumption is the basis of these approaches
3.2 Perturbation-based:
o The perturbation-based methods directly incorporate the smoothness assumption.
o This concept implies that when a data point is perturbed with a small amount of noise, the
predictions for noisy and clean inputs should not be different
3. Intrinsically semi-supervised
3.3 Manifolds:
o According to the manifold assumption: (a) the input space is composed of multiple lower-
dimensional manifolds on which all data points lie and
o (b) data points sharing the same lower-dimensional manifold have the same label
3.4 Generative Models:
o Generative approaches model the process that generates data.
o For classification problems, a generative model is conditioned on a given y label.
Transductive methods
◦ Transductive algorithms do not produce a predictor that can operate over the entire
input space
◦ Actually, transductive methods yield a set of predictions for the provided unlabeled
data.
◦ Therefore, a training phase and a testing phase are not distinguishable
◦ Some well-established transductive graph-based methods are:
 General framework for graph-based methods,
 Inference in graphs,
 Probabilistic label assignments: Markov random fields
Future scope
1. One important issue to address in near future is, possible performance degradation caused by the introduction of
unlabeled data. Limited supervised techniques only perform better than their supervised counterparts or base learners
in specific cases (Li and Zhou 2015; Singh et al. 2009).
2. Recent studies have shown that perturbation-based methods with neural networks consistently outperform their
supervised counterparts. This flexibility of using considerable advantage of using neural networks should be explored
more and should gain more popularity in the field of semi-supervised neural networks.
3. Recently, automated machine learning (AutoML) has been used widely to achieve the robustness of the models.
These approaches include meta-learning and neural architecture for automatic algorithm selection and hyperparameter
optimization. While AutoML techniques are successfully being used to supervised learning, there is a lack of
improvement of research in the semi-supervised field. This field should be studied more to bring striking results to
semi-supervised approaches.
Future scope
4. Another important step towards the advancement of semi-supervised approaches are having
standardized software packages or library dedicated to this domain. Currently, some generic
packages like the KEEL software package do exist that include a semi-supervised learning module
(Triguero et al. 2017).
5. One vital sector that needs serious attention from researchers is building a strong connection
between clustering and classification. Essentially, both approaches are a special branch of semi-
supervised, where either only labeled or unlabeled data is considered. The recent hype in generative
models can be seen as a good sign to address this field.
Key References:
◦ J. E. van Engelen and H. H. Hoos, “A survey on semi-supervised learning,” Machine Learning, vol. 109, no. 2, pp. 373–440,
feb 2020.
◦ Bengio, Y., Delalleau, O., & Le Roux, N. (2006). Chapter 11. Label propagation and quadratic criterion. In O. Chapelle, B.
Schölkopf, & A. Zien (Eds.), Semi-supervised learning (pp. 193–216). Cambridge: The
MIT Press.
◦ Bengio, Y., Delalleau, O., & Le Roux, N. (2006). Chapter 11. Label propagation and quadratic criterion. In O. Chapelle, B.
Schölkopf, & A. Zien (Eds.), Semi-supervised learning (pp. 193–216). Cambridge: The MIT Press.
◦ Goldberg, A. B., Zhu, X., Singh, A., Xu, Z.,&Nowak, R. D. (2009).Multi-manifold semi-supervised learning. In Proceedings
of the 12th international conference on artificial intelligence and statistics (pp. 169–176).
◦ Haffari, G. R., & Sarkar, A. (2007). Analysis of semi-supervised learning with the Yarowsky algorithm. In Proceedings of
the 23rd conference on uncertainty in artificial intelligence (pp. 159–166).
Thanks

More Related Content

What's hot

Technological persuasive pedagogy a new way to persuade students in the compu...
Technological persuasive pedagogy a new way to persuade students in the compu...Technological persuasive pedagogy a new way to persuade students in the compu...
Technological persuasive pedagogy a new way to persuade students in the compu...
Alexander Decker
 
FYP ppt
FYP pptFYP ppt
FYP ppt
Aniqa Bano
 
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
Dila Nurlaila
 
Analysing/Interpreting Quantitative Research
Analysing/Interpreting  Quantitative Research Analysing/Interpreting  Quantitative Research
Analysing/Interpreting Quantitative Research
HariBolKafle
 
Mb0050
Mb0050Mb0050
Mb0050 research methodology
Mb0050   research methodologyMb0050   research methodology
Mb0050 research methodology
smumbahelp
 
Student Performance Data Mining Project Report
Student Performance Data Mining Project ReportStudent Performance Data Mining Project Report
Student Performance Data Mining Project Report
Jinnah University for Women
 
Data analysis and Presentation
Data analysis and PresentationData analysis and Presentation
Data analysis and Presentation
Jignesh Kariya
 
activelearning.ppt
activelearning.pptactivelearning.ppt
activelearning.ppt
butest
 
A cluster-based analysis to diagnose students’ learning achievements
A cluster-based analysis to diagnose students’ learning achievementsA cluster-based analysis to diagnose students’ learning achievements
A cluster-based analysis to diagnose students’ learning achievements
Miguel R. Artacho
 
Test equating using irt. final
Test equating using irt. finalTest equating using irt. final
Test equating using irt. final
munsif123
 
B05110409
B05110409B05110409
B05110409
IOSR-JEN
 
Independent Study Guide
Independent Study GuideIndependent Study Guide
Independent Study Guide
Mahendar Kumar Papu
 
Quantitative data analysis - Attitudes Towards Research
Quantitative data analysis - Attitudes Towards ResearchQuantitative data analysis - Attitudes Towards Research
Quantitative data analysis - Attitudes Towards Research
Lee Cox
 
De carlo rizk 2010 icelw
De carlo rizk 2010 icelwDe carlo rizk 2010 icelw
De carlo rizk 2010 icelw
Ting Yuan, Ed.D.
 
Test construction
Test constructionTest construction
Test construction
munsif123
 
Lec1-Into
Lec1-IntoLec1-Into
Lec1-Into
butest
 
Galambos_SlidesNEAIR2015
Galambos_SlidesNEAIR2015Galambos_SlidesNEAIR2015
Galambos_SlidesNEAIR2015
Nora Galambos, PhD
 
Big shadow test
Big shadow testBig shadow test
Big shadow test
munsif123
 

What's hot (19)

Technological persuasive pedagogy a new way to persuade students in the compu...
Technological persuasive pedagogy a new way to persuade students in the compu...Technological persuasive pedagogy a new way to persuade students in the compu...
Technological persuasive pedagogy a new way to persuade students in the compu...
 
FYP ppt
FYP pptFYP ppt
FYP ppt
 
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
Seminar Proposal Tugas Akhir - SPK Pemilihan tema Tugas Akhir menggunakan Ana...
 
Analysing/Interpreting Quantitative Research
Analysing/Interpreting  Quantitative Research Analysing/Interpreting  Quantitative Research
Analysing/Interpreting Quantitative Research
 
Mb0050
Mb0050Mb0050
Mb0050
 
Mb0050 research methodology
Mb0050   research methodologyMb0050   research methodology
Mb0050 research methodology
 
Student Performance Data Mining Project Report
Student Performance Data Mining Project ReportStudent Performance Data Mining Project Report
Student Performance Data Mining Project Report
 
Data analysis and Presentation
Data analysis and PresentationData analysis and Presentation
Data analysis and Presentation
 
activelearning.ppt
activelearning.pptactivelearning.ppt
activelearning.ppt
 
A cluster-based analysis to diagnose students’ learning achievements
A cluster-based analysis to diagnose students’ learning achievementsA cluster-based analysis to diagnose students’ learning achievements
A cluster-based analysis to diagnose students’ learning achievements
 
Test equating using irt. final
Test equating using irt. finalTest equating using irt. final
Test equating using irt. final
 
B05110409
B05110409B05110409
B05110409
 
Independent Study Guide
Independent Study GuideIndependent Study Guide
Independent Study Guide
 
Quantitative data analysis - Attitudes Towards Research
Quantitative data analysis - Attitudes Towards ResearchQuantitative data analysis - Attitudes Towards Research
Quantitative data analysis - Attitudes Towards Research
 
De carlo rizk 2010 icelw
De carlo rizk 2010 icelwDe carlo rizk 2010 icelw
De carlo rizk 2010 icelw
 
Test construction
Test constructionTest construction
Test construction
 
Lec1-Into
Lec1-IntoLec1-Into
Lec1-Into
 
Galambos_SlidesNEAIR2015
Galambos_SlidesNEAIR2015Galambos_SlidesNEAIR2015
Galambos_SlidesNEAIR2015
 
Big shadow test
Big shadow testBig shadow test
Big shadow test
 

Similar to Semi-supervised Learning Survey - 20 years of evaluation

Student Performance Evaluation in Education Sector Using Prediction and Clust...
Student Performance Evaluation in Education Sector Using Prediction and Clust...Student Performance Evaluation in Education Sector Using Prediction and Clust...
Student Performance Evaluation in Education Sector Using Prediction and Clust...
IJSRD
 
Knowledge Management in the Cloud: Benefits and Risks
Knowledge Management in the Cloud: Benefits and RisksKnowledge Management in the Cloud: Benefits and Risks
Knowledge Management in the Cloud: Benefits and Risks
Editor IJCATR
 
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
Editor IJCATR
 
When deep learners change their mind learning dynamics for active learning
When deep learners change their mind  learning dynamics for active learningWhen deep learners change their mind  learning dynamics for active learning
When deep learners change their mind learning dynamics for active learning
Devansh16
 
A Survey on the Classification Techniques In Educational Data Mining
A Survey on the Classification Techniques In Educational Data MiningA Survey on the Classification Techniques In Educational Data Mining
A Survey on the Classification Techniques In Educational Data Mining
Editor IJCATR
 
L016136369
L016136369L016136369
L016136369
IOSR Journals
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
Adetimehin Oluwasegun Matthew
 
De carlo rizk 2010 icelw
De carlo rizk 2010 icelwDe carlo rizk 2010 icelw
De carlo rizk 2010 icelw
Ting Yuan, Ed.D.
 
Endsem AI merged.pdf
Endsem AI merged.pdfEndsem AI merged.pdf
Endsem AI merged.pdf
ShivamMishra603376
 
Week 4 advanced labeling, augmentation and data preprocessing
Week 4   advanced labeling, augmentation and data preprocessingWeek 4   advanced labeling, augmentation and data preprocessing
Week 4 advanced labeling, augmentation and data preprocessing
Ajay Taneja
 
BTech Pattern Recognition Notes
BTech Pattern Recognition NotesBTech Pattern Recognition Notes
BTech Pattern Recognition Notes
Ashutosh Agrahari
 
Detection of Online Learning Activity Scopes
Detection of Online Learning Activity ScopesDetection of Online Learning Activity Scopes
Detection of Online Learning Activity Scopes
Syeda Sana
 
Machine Learning - Deep Learning
Machine Learning - Deep LearningMachine Learning - Deep Learning
Machine Learning - Deep Learning
Adetimehin Oluwasegun Matthew
 
Educational Data Mining to Analyze Students Performance – Concept Plan
Educational Data Mining to Analyze Students Performance – Concept PlanEducational Data Mining to Analyze Students Performance – Concept Plan
Educational Data Mining to Analyze Students Performance – Concept Plan
IRJET Journal
 
IRJET- Academic Performance Analysis System
IRJET- Academic Performance Analysis SystemIRJET- Academic Performance Analysis System
IRJET- Academic Performance Analysis System
IRJET Journal
 
Review: Semi-Supervised Learning Methods for Word Sense Disambiguation
Review: Semi-Supervised Learning Methods for Word Sense DisambiguationReview: Semi-Supervised Learning Methods for Word Sense Disambiguation
Review: Semi-Supervised Learning Methods for Word Sense Disambiguation
IOSR Journals
 
Prognostication of the placement of students applying machine learning algori...
Prognostication of the placement of students applying machine learning algori...Prognostication of the placement of students applying machine learning algori...
Prognostication of the placement of students applying machine learning algori...
BIJIAM Journal
 
Effective Feature Selection for Feature Possessing Group Structure
Effective Feature Selection for Feature Possessing Group StructureEffective Feature Selection for Feature Possessing Group Structure
Effective Feature Selection for Feature Possessing Group Structure
rahulmonikasharma
 
For iiii year students of cse ML-UNIT-V.pptx
For iiii year students of cse ML-UNIT-V.pptxFor iiii year students of cse ML-UNIT-V.pptx
For iiii year students of cse ML-UNIT-V.pptx
SureshPolisetty2
 
Ijetr042132
Ijetr042132Ijetr042132

Similar to Semi-supervised Learning Survey - 20 years of evaluation (20)

Student Performance Evaluation in Education Sector Using Prediction and Clust...
Student Performance Evaluation in Education Sector Using Prediction and Clust...Student Performance Evaluation in Education Sector Using Prediction and Clust...
Student Performance Evaluation in Education Sector Using Prediction and Clust...
 
Knowledge Management in the Cloud: Benefits and Risks
Knowledge Management in the Cloud: Benefits and RisksKnowledge Management in the Cloud: Benefits and Risks
Knowledge Management in the Cloud: Benefits and Risks
 
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
An Evaluation of Feature Selection Methods for Positive- Unlabeled Learning i...
 
When deep learners change their mind learning dynamics for active learning
When deep learners change their mind  learning dynamics for active learningWhen deep learners change their mind  learning dynamics for active learning
When deep learners change their mind learning dynamics for active learning
 
A Survey on the Classification Techniques In Educational Data Mining
A Survey on the Classification Techniques In Educational Data MiningA Survey on the Classification Techniques In Educational Data Mining
A Survey on the Classification Techniques In Educational Data Mining
 
L016136369
L016136369L016136369
L016136369
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
 
De carlo rizk 2010 icelw
De carlo rizk 2010 icelwDe carlo rizk 2010 icelw
De carlo rizk 2010 icelw
 
Endsem AI merged.pdf
Endsem AI merged.pdfEndsem AI merged.pdf
Endsem AI merged.pdf
 
Week 4 advanced labeling, augmentation and data preprocessing
Week 4   advanced labeling, augmentation and data preprocessingWeek 4   advanced labeling, augmentation and data preprocessing
Week 4 advanced labeling, augmentation and data preprocessing
 
BTech Pattern Recognition Notes
BTech Pattern Recognition NotesBTech Pattern Recognition Notes
BTech Pattern Recognition Notes
 
Detection of Online Learning Activity Scopes
Detection of Online Learning Activity ScopesDetection of Online Learning Activity Scopes
Detection of Online Learning Activity Scopes
 
Machine Learning - Deep Learning
Machine Learning - Deep LearningMachine Learning - Deep Learning
Machine Learning - Deep Learning
 
Educational Data Mining to Analyze Students Performance – Concept Plan
Educational Data Mining to Analyze Students Performance – Concept PlanEducational Data Mining to Analyze Students Performance – Concept Plan
Educational Data Mining to Analyze Students Performance – Concept Plan
 
IRJET- Academic Performance Analysis System
IRJET- Academic Performance Analysis SystemIRJET- Academic Performance Analysis System
IRJET- Academic Performance Analysis System
 
Review: Semi-Supervised Learning Methods for Word Sense Disambiguation
Review: Semi-Supervised Learning Methods for Word Sense DisambiguationReview: Semi-Supervised Learning Methods for Word Sense Disambiguation
Review: Semi-Supervised Learning Methods for Word Sense Disambiguation
 
Prognostication of the placement of students applying machine learning algori...
Prognostication of the placement of students applying machine learning algori...Prognostication of the placement of students applying machine learning algori...
Prognostication of the placement of students applying machine learning algori...
 
Effective Feature Selection for Feature Possessing Group Structure
Effective Feature Selection for Feature Possessing Group StructureEffective Feature Selection for Feature Possessing Group Structure
Effective Feature Selection for Feature Possessing Group Structure
 
For iiii year students of cse ML-UNIT-V.pptx
For iiii year students of cse ML-UNIT-V.pptxFor iiii year students of cse ML-UNIT-V.pptx
For iiii year students of cse ML-UNIT-V.pptx
 
Ijetr042132
Ijetr042132Ijetr042132
Ijetr042132
 

Recently uploaded

How to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For FlutterHow to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For Flutter
Daiki Mogmet Ito
 
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUHCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
panagenda
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
Kari Kakkonen
 
TrustArc Webinar - 2024 Global Privacy Survey
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc Webinar - 2024 Global Privacy Survey
TrustArc Webinar - 2024 Global Privacy Survey
TrustArc
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems S.M.S.A.
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
Neo4j
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
Neo4j
 
How to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptxHow to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptx
danishmna97
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
Matthew Sinclair
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
DianaGray10
 
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
Neo4j
 
Serial Arm Control in Real Time Presentation
Serial Arm Control in Real Time PresentationSerial Arm Control in Real Time Presentation
Serial Arm Control in Real Time Presentation
tolgahangng
 
Mind map of terminologies used in context of Generative AI
Mind map of terminologies used in context of Generative AIMind map of terminologies used in context of Generative AI
Mind map of terminologies used in context of Generative AI
Kumud Singh
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
Matthew Sinclair
 
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with SlackLet's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
shyamraj55
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
名前 です男
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
SOFTTECHHUB
 
AI 101: An Introduction to the Basics and Impact of Artificial Intelligence
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceAI 101: An Introduction to the Basics and Impact of Artificial Intelligence
AI 101: An Introduction to the Basics and Impact of Artificial Intelligence
IndexBug
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
mikeeftimakis1
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
DianaGray10
 

Recently uploaded (20)

How to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For FlutterHow to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For Flutter
 
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUHCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAU
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
 
TrustArc Webinar - 2024 Global Privacy Survey
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc Webinar - 2024 Global Privacy Survey
TrustArc Webinar - 2024 Global Privacy Survey
 
Uni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdfUni Systems Copilot event_05062024_C.Vlachos.pdf
Uni Systems Copilot event_05062024_C.Vlachos.pdf
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
 
How to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptxHow to Get CNIC Information System with Paksim Ga.pptx
How to Get CNIC Information System with Paksim Ga.pptx
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
 
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
 
Serial Arm Control in Real Time Presentation
Serial Arm Control in Real Time PresentationSerial Arm Control in Real Time Presentation
Serial Arm Control in Real Time Presentation
 
Mind map of terminologies used in context of Generative AI
Mind map of terminologies used in context of Generative AIMind map of terminologies used in context of Generative AI
Mind map of terminologies used in context of Generative AI
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
 
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with SlackLet's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slack
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
 
AI 101: An Introduction to the Basics and Impact of Artificial Intelligence
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceAI 101: An Introduction to the Basics and Impact of Artificial Intelligence
AI 101: An Introduction to the Basics and Impact of Artificial Intelligence
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
 

Semi-supervised Learning Survey - 20 years of evaluation

  • 1. Semi-supervised Learning Survey - 20 years of evaluation Subarna Chowdhury Soma SJSU ID: 014549587
  • 2. Paper  A survey on semi-supervised learning ◦ Jesper E. van Engelen &Holger H. Hoos ◦ Received: 3 December 2018 / Revised: 20 September 2019 / Accepted: 29 September 2019 / ◦ Published online: 15 November 2019
  • 3. Overview  Semi-supervised learning  Basic Concepts and Assumptions  Taxonomy of Semi-supervised learning methods  Semi-supervised learning methods  Conclusions and future scope
  • 4. Semi-supervised Learning ◦ In supervised learning, one is presented with a set of data points consisting of some input x and a corresponding output value y. The goal is, then, to construct a classifier or regressor that can estimate the output value for previously unseen inputs ◦ In unsupervised learning, no specific output value is provided; goal is to infer some underlying structure from the inputs ◦ Semi-supervised learning is about making use of unlabeled data. It is useful when the available labeled data is small.Unlabeled data is cheap Although not every type of unlabeled data is useful
  • 5. Basic Concepts and Assumptions  Assumptions of semi-supervised learning ◦ An important pre-condition of semi-supervised learning is that the underlying data distribution, p(x), contains information about the posterior distribution p(y|x). Considering this condition, we can use unlabeled data to gain information about p(x), and thereby about p(y|x). ◦ The smoothness assumption : if two samples x and x1 are close in the input space, their labels y and y1 should be the same ◦ The low-density assumption : the decision boundary should not pass through high-density areas in the input space ◦ The manifold assumption : same low-dimensional manifold datapoints should have the same label
  • 7. Basic Concepts and Assumptions  Connection to Clustering ◦ According to Chapelle et al. 2006b, data points that belong to the same cluster come from the same class. This assumption is called the semi-supervised learning assumption of clusters.
  • 9. Inductive methods ◦ Inductive methods build a classifier to generate predictions for any point in the input space ◦ To train the classifier, unlabeled data can be used. ◦ Once, a model is built from the training phase and later can be used to predict the new data point labels.
  • 10. Inductive methods 1. Wrapper methods 2. Unsupervised pre-processing 3. Intrinsically semi-supervised
  • 11. 1. Wrapper methods o In the training step, one or more supervised classifiers are trained using the labeled data and pseudo-labeled data from earlier iteration steps. o Next, to infer labels from the previously trained unlabeled data, the resulting classifiers are used in the pseudo-labeled step. o Most confident predictions are used as the pseudo-label at the next iteration.
  • 12. 1. Wrapper methods 1. 1 Self-training: o This approach uses one supervised classifier that is re-trained iteratively based on its most confident predictions 1.2 Co-training: o Here classifiers are iteratively re-trained on the most confident predictions of each other o For co-training to succeed, the classifiers are supposed to be adequately different 1.3 Boosting: o Like conventional boosting methods, they build ensemble classifiers by sequentially building individual classifiers o Moreover, each individual classifier is trained using both the labeled data and the most confident predictions of previously trained unlabeled data like before
  • 13. 2. Unsupervised pre-processing ◦ Unsupervised preprocessing, either extract useful features from unlabeled data, pre-cluster the data or use an unsupervised manner to compute the initial parameters of a supervised learning procedure
  • 14. 2. Unsupervised pre-processing 2. 1 Feature Extraction: o These methods try to find a transformation of the input data that helps to improve classification and makes it computationally efficient 2.2 Cluster-then-label: o Cluster-then-label procedures form a group of methods that explicitly join the clustering and classification processes o At first, it uses an unsupervised or semi-supervised clustering algorithm to all the input data, and then use the resulting clusters lead the classification process 2.3 Pre-training: o This approach naturally fits in deep learning areas where each layer of the hierarchical model is a latent representation of the input data. o In pre-training methods, a decision boundary is conducted by unlabeled data before applying supervised training
  • 15. 3. Intrinsically semi-supervised ◦ Here unlabeled data is directly incorporated into the objective function or optimization procedure of the learning part. ◦ Most of these methods are direct extensions of supervised learning methods by transforming objective functions to include unlabeled data.
  • 16. 3. Intrinsically semi-supervised 3. 1 Maximum Margin: o These methods aim to maximize the distance between the given data points and the decision boundary o The semi-supervised low-density assumption is the basis of these approaches 3.2 Perturbation-based: o The perturbation-based methods directly incorporate the smoothness assumption. o This concept implies that when a data point is perturbed with a small amount of noise, the predictions for noisy and clean inputs should not be different
  • 17. 3. Intrinsically semi-supervised 3.3 Manifolds: o According to the manifold assumption: (a) the input space is composed of multiple lower- dimensional manifolds on which all data points lie and o (b) data points sharing the same lower-dimensional manifold have the same label 3.4 Generative Models: o Generative approaches model the process that generates data. o For classification problems, a generative model is conditioned on a given y label.
  • 18. Transductive methods ◦ Transductive algorithms do not produce a predictor that can operate over the entire input space ◦ Actually, transductive methods yield a set of predictions for the provided unlabeled data. ◦ Therefore, a training phase and a testing phase are not distinguishable ◦ Some well-established transductive graph-based methods are:  General framework for graph-based methods,  Inference in graphs,  Probabilistic label assignments: Markov random fields
  • 19. Future scope 1. One important issue to address in near future is, possible performance degradation caused by the introduction of unlabeled data. Limited supervised techniques only perform better than their supervised counterparts or base learners in specific cases (Li and Zhou 2015; Singh et al. 2009). 2. Recent studies have shown that perturbation-based methods with neural networks consistently outperform their supervised counterparts. This flexibility of using considerable advantage of using neural networks should be explored more and should gain more popularity in the field of semi-supervised neural networks. 3. Recently, automated machine learning (AutoML) has been used widely to achieve the robustness of the models. These approaches include meta-learning and neural architecture for automatic algorithm selection and hyperparameter optimization. While AutoML techniques are successfully being used to supervised learning, there is a lack of improvement of research in the semi-supervised field. This field should be studied more to bring striking results to semi-supervised approaches.
  • 20. Future scope 4. Another important step towards the advancement of semi-supervised approaches are having standardized software packages or library dedicated to this domain. Currently, some generic packages like the KEEL software package do exist that include a semi-supervised learning module (Triguero et al. 2017). 5. One vital sector that needs serious attention from researchers is building a strong connection between clustering and classification. Essentially, both approaches are a special branch of semi- supervised, where either only labeled or unlabeled data is considered. The recent hype in generative models can be seen as a good sign to address this field.
  • 21. Key References: ◦ J. E. van Engelen and H. H. Hoos, “A survey on semi-supervised learning,” Machine Learning, vol. 109, no. 2, pp. 373–440, feb 2020. ◦ Bengio, Y., Delalleau, O., & Le Roux, N. (2006). Chapter 11. Label propagation and quadratic criterion. In O. Chapelle, B. Schölkopf, & A. Zien (Eds.), Semi-supervised learning (pp. 193–216). Cambridge: The MIT Press. ◦ Bengio, Y., Delalleau, O., & Le Roux, N. (2006). Chapter 11. Label propagation and quadratic criterion. In O. Chapelle, B. Schölkopf, & A. Zien (Eds.), Semi-supervised learning (pp. 193–216). Cambridge: The MIT Press. ◦ Goldberg, A. B., Zhu, X., Singh, A., Xu, Z.,&Nowak, R. D. (2009).Multi-manifold semi-supervised learning. In Proceedings of the 12th international conference on artificial intelligence and statistics (pp. 169–176). ◦ Haffari, G. R., & Sarkar, A. (2007). Analysis of semi-supervised learning with the Yarowsky algorithm. In Proceedings of the 23rd conference on uncertainty in artificial intelligence (pp. 159–166).