SlideShare a Scribd company logo
1 of 16
Download to read offline
Semi-supervised learning with
adversarially missing label information

           by U. Syad and B. Taskar @ Univ. Penn



                            Akisato Kimura, Ph.D
        [E-mail] akisato@ieee.org   [Twitter] @_akisato
Abstract
   Semi-supervised learning in an adversarial setting
       Consider a general framework, including
           Covariate shift : Missing labels are not necessarily selected at
            random.
           Multiple instance learning : Correspondences between labels
            and instances are not provided explicitly.
           Graph-based regularization : Only information about which
            instances are likely to have the same label.
       but not including
           Malicious label noise setting : Labelers are allowed to mislabel a
            part of the training set.
       Labelers can remove labels, but cannot change them.


    2                                   NIPS2010 reading @ Cybozu   December 26, 2010
Framework
                   : m labeled training examples
                                : examples
                                : labels
                      : unknown PDF


                     : label regularization function
                      : soft labeling of the training examples
               : probability so that     has label
       The algorithm can access to only the examples and
        the label regularization function .
       Correct labeling might be near the minimum of ,
        but not necessarily (thus it is called “adversarial”).

    3                                   NIPS2010 reading @ Cybozu   December 26, 2010
Label regularization functions (1)
   Fully supervised learning             A set of indexes of unlabeled examples

                                                              otherwise
       [Note] Labels   also denotes the correct soft labeling.
   Semi-supervised learning                               Don’t mind how q labels
                                                           examples not in I
    
    
                   otherwise         A set of indexes of non-zero components

   Ambiguous labeling       (   multiple instance learning)
                          a set of possible labels of the example
                                                                          otherwise



    4                              NIPS2010 reading @ Cybozu         December 26, 2010
Label regularization functions (2)
   Laplacian regularization
                                                       positive semi-definite
               is small       and        are believed to have the
        same label
   Any combinations are also in this framework.
       Ex.                SSL with Laplacian regularization


   We don’t specify how the function is selected.
       Ex. Semi-supervised learning :
        Usually, we can’t know how the unlabeled samples were
        selected (“covariate shift” setting).


    5                                NIPS2010 reading @ Cybozu      December 26, 2010
Loss functions for model learning
   Any types of loss functions can be applicable, but
   Particularly we focus on
       Negative log likelihood of a log-linear PDF



       Binary loss function of a linear classifier



               a model/classifier parameter
                  a feature function (“basis function” in PRML)



    6                               NIPS2010 reading @ Cybozu   December 26, 2010
Algorithm
   Objective: Find a parameter                  achieving


        Its validity will be disclosed later.
        For brevity, we abbreviate the objective as              .
                                  The order is opposite !!
   Basic algorithm:
    Game for Adversarially Missing Evidence (GAME)
    1.    Constants          are given in advance.
    2.    Find s.t.
    3.    Find s.t.
    4.    Output the parameter estimate

    7                               NIPS2010 reading @ Cybozu   December 26, 2010
Algorithm in detail
   Consider the case of the negative log likelihood
   Step 3
       The label regularization function can be ignored.
       This step is equivalent to maximization of the likelihood of
        a log-linear model.       Many solvers we have


   Step 2
       Take the dual of the inner minimization
                                                                     Easy to
                                                                     optimize
                                                               concave
                                                               Shannon entropy

    8                              NIPS2010 reading @ Cybozu     December 26, 2010
Convergence properties (1)
   Theorem 1       (Converse theorem)
       If           , then for all parameters                      and label
        regularization functions


        satisfies with probability            .          Similar to the objective
                                                         function of the algorithm




    9                                NIPS2010 reading @ Cybozu        December 26, 2010
Convergence properties (2)
   Theorem 2          (Direct theorem: lower bound)
        Consider the binary loss function, and a finite set .
         For all learning algorithms and example distributions                    ,
         there exists a labeling function   that satisfies




         with probability               under some assumptions.

            The assumptions can be seen in the paper (1,2 and 3).
            Some detailed conditions are omitted for the simplicity.
            The parameter    is obtained by an learning algorithm.

    10                                   NIPS2010 reading @ Cybozu   December 26, 2010
Convergence properties (3)
   Theorem 3          (Direct theorem: upper bound)
        Consider the binary loss function. If           ,
         there exists a family of label regularization function
         and a learning algorithm that satisfy



         with probability           under some assumptions.

            Assumption 3 can be removed.
            Some detailed conditions are omitted for the simplicity.
            The parameter    is obtained by the learning algorithm.



    11                                   NIPS2010 reading @ Cybozu   December 26, 2010
Back to the algorithm part …
   Objective: Find a parameter                 achieving


   Minimum upper bound in the converse theorem



        Do not minimize over            .
        Instead, add a regularization part to leave unconstrained.




    12                             NIPS2010 reading @ Cybozu   December 26, 2010
Experiments
   Exp. 1: Binary classification
        Datasets: COIL (object recognition), BCI (EEG scans)



   Exp. 2: Multi-class categorization
        Dataset: Labeled Faces in the Wild




    13                            NIPS2010 reading @ Cybozu   December 26, 2010
Results
     BCI


                                                         COIL




           Faces




14                 NIPS2010 reading @ Cybozu   December 26, 2010
この論文から何を思うか?
   「一般化」論文は得てして面白くない。
        その特殊ケースの結果を知っているので、
         「おお、そうだったのか」という驚きを得られにくい。


   が、一般化が新しい見方を提供することは見逃せない。
        知られていない特殊ケースの発見
        特殊ケースの組み合わせによる新しい知見/手法
   さらに、一般化で技術を俯瞰できるようになることも重要。


   なので、ここからもう一歩先を考えてみましょう!

    15                NIPS2010 reading @ Cybozu   December 26, 2010
Fin.
   Thank you for your kind attention.




    16                       NIPS2010 reading @ Cybozu   December 26, 2010

More Related Content

Similar to NIPS2010 reading: Semi-supervised learning with adversarially missing label information

Random Features Strengthen Graph Neural Networks
Random Features Strengthen Graph Neural NetworksRandom Features Strengthen Graph Neural Networks
Random Features Strengthen Graph Neural Networksjoisino
 
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAEFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAIJCI JOURNAL
 
論文サーベイ(Sasaki)
論文サーベイ(Sasaki)論文サーベイ(Sasaki)
論文サーベイ(Sasaki)Hajime Sasaki
 
Polikar10missing
Polikar10missingPolikar10missing
Polikar10missingkagupta
 
Creating Documentation Your Users Will Love
Creating Documentation Your Users Will LoveCreating Documentation Your Users Will Love
Creating Documentation Your Users Will LoveEna Arel
 
Adhyann – a hybrid part of-speech tagger
Adhyann – a hybrid part of-speech taggerAdhyann – a hybrid part of-speech tagger
Adhyann – a hybrid part of-speech taggerijitjournal
 
Chapter 6 OOPS Concept
Chapter 6 OOPS ConceptChapter 6 OOPS Concept
Chapter 6 OOPS ConceptAmrit Kaur
 
SBML and related resources 
and standardization efforts
SBML and related resources 
and standardization effortsSBML and related resources 
and standardization efforts
SBML and related resources 
and standardization effortsMike Hucka
 
Event driven, mobile artificial intelligence algorithms
Event driven, mobile artificial intelligence algorithmsEvent driven, mobile artificial intelligence algorithms
Event driven, mobile artificial intelligence algorithmsDinesh More
 
Algorithm
AlgorithmAlgorithm
Algorithmseobear
 
ML DL AI DS BD - An Introduction
ML DL AI DS BD - An IntroductionML DL AI DS BD - An Introduction
ML DL AI DS BD - An IntroductionDony Riyanto
 
Generics Tutorial
Generics TutorialGenerics Tutorial
Generics Tutorialwasntgosu
 
EuroAD 2021: ChainRules.jl
EuroAD 2021: ChainRules.jl EuroAD 2021: ChainRules.jl
EuroAD 2021: ChainRules.jl Lyndon White
 
Substructrual surrogates for learning decomposable classification problems: i...
Substructrual surrogates for learning decomposable classification problems: i...Substructrual surrogates for learning decomposable classification problems: i...
Substructrual surrogates for learning decomposable classification problems: i...kknsastry
 
Cs6502 ooad-cse-vst-au-unit-v dce
Cs6502 ooad-cse-vst-au-unit-v dceCs6502 ooad-cse-vst-au-unit-v dce
Cs6502 ooad-cse-vst-au-unit-v dcetagoreengineering
 
A Survey on Fuzzy Association Rule Mining Methodologies
A Survey on Fuzzy Association Rule Mining MethodologiesA Survey on Fuzzy Association Rule Mining Methodologies
A Survey on Fuzzy Association Rule Mining MethodologiesIOSR Journals
 

Similar to NIPS2010 reading: Semi-supervised learning with adversarially missing label information (20)

SDOW (ISWC2011)
SDOW (ISWC2011)SDOW (ISWC2011)
SDOW (ISWC2011)
 
Test Doubles
Test DoublesTest Doubles
Test Doubles
 
Random Features Strengthen Graph Neural Networks
Random Features Strengthen Graph Neural NetworksRandom Features Strengthen Graph Neural Networks
Random Features Strengthen Graph Neural Networks
 
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATAEFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
EFFICIENT FEATURE SUBSET SELECTION MODEL FOR HIGH DIMENSIONAL DATA
 
UML
UMLUML
UML
 
論文サーベイ(Sasaki)
論文サーベイ(Sasaki)論文サーベイ(Sasaki)
論文サーベイ(Sasaki)
 
Polikar10missing
Polikar10missingPolikar10missing
Polikar10missing
 
Creating Documentation Your Users Will Love
Creating Documentation Your Users Will LoveCreating Documentation Your Users Will Love
Creating Documentation Your Users Will Love
 
Adhyann – a hybrid part of-speech tagger
Adhyann – a hybrid part of-speech taggerAdhyann – a hybrid part of-speech tagger
Adhyann – a hybrid part of-speech tagger
 
Chapter 6 OOPS Concept
Chapter 6 OOPS ConceptChapter 6 OOPS Concept
Chapter 6 OOPS Concept
 
Lecture19
Lecture19Lecture19
Lecture19
 
SBML and related resources 
and standardization efforts
SBML and related resources 
and standardization effortsSBML and related resources 
and standardization efforts
SBML and related resources 
and standardization efforts
 
Event driven, mobile artificial intelligence algorithms
Event driven, mobile artificial intelligence algorithmsEvent driven, mobile artificial intelligence algorithms
Event driven, mobile artificial intelligence algorithms
 
Algorithm
AlgorithmAlgorithm
Algorithm
 
ML DL AI DS BD - An Introduction
ML DL AI DS BD - An IntroductionML DL AI DS BD - An Introduction
ML DL AI DS BD - An Introduction
 
Generics Tutorial
Generics TutorialGenerics Tutorial
Generics Tutorial
 
EuroAD 2021: ChainRules.jl
EuroAD 2021: ChainRules.jl EuroAD 2021: ChainRules.jl
EuroAD 2021: ChainRules.jl
 
Substructrual surrogates for learning decomposable classification problems: i...
Substructrual surrogates for learning decomposable classification problems: i...Substructrual surrogates for learning decomposable classification problems: i...
Substructrual surrogates for learning decomposable classification problems: i...
 
Cs6502 ooad-cse-vst-au-unit-v dce
Cs6502 ooad-cse-vst-au-unit-v dceCs6502 ooad-cse-vst-au-unit-v dce
Cs6502 ooad-cse-vst-au-unit-v dce
 
A Survey on Fuzzy Association Rule Mining Methodologies
A Survey on Fuzzy Association Rule Mining MethodologiesA Survey on Fuzzy Association Rule Mining Methodologies
A Survey on Fuzzy Association Rule Mining Methodologies
 

More from Akisato Kimura

Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Akisato Kimura
 
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Akisato Kimura
 
多変量解析の一般化
多変量解析の一般化多変量解析の一般化
多変量解析の一般化Akisato Kimura
 
CVPR2016 reading - 特徴量学習とクロスモーダル転移について
CVPR2016 reading - 特徴量学習とクロスモーダル転移についてCVPR2016 reading - 特徴量学習とクロスモーダル転移について
CVPR2016 reading - 特徴量学習とクロスモーダル転移についてAkisato Kimura
 
NIPS2015 reading - Learning visual biases from human imagination
NIPS2015 reading - Learning visual biases from human imaginationNIPS2015 reading - Learning visual biases from human imagination
NIPS2015 reading - Learning visual biases from human imaginationAkisato Kimura
 
CVPR2015 reading "Global refinement of random forest"
CVPR2015 reading "Global refinement of random forest"CVPR2015 reading "Global refinement of random forest"
CVPR2015 reading "Global refinement of random forest"Akisato Kimura
 
CVPR2015 reading "Understainding image virality" (in Japanese)
CVPR2015 reading "Understainding image virality" (in Japanese)CVPR2015 reading "Understainding image virality" (in Japanese)
CVPR2015 reading "Understainding image virality" (in Japanese)Akisato Kimura
 
Computational models of human visual attention driven by auditory cues
Computational models of human visual attention driven by auditory cuesComputational models of human visual attention driven by auditory cues
Computational models of human visual attention driven by auditory cuesAkisato Kimura
 
NIPS2014 reading - Top rank optimization in linear time
NIPS2014 reading - Top rank optimization in linear timeNIPS2014 reading - Top rank optimization in linear time
NIPS2014 reading - Top rank optimization in linear timeAkisato Kimura
 
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...Akisato Kimura
 
ICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged informationICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged informationAkisato Kimura
 
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...Akisato Kimura
 
IJCAI13 Paper review: Large-scale spectral clustering on graphs
IJCAI13 Paper review: Large-scale spectral clustering on graphsIJCAI13 Paper review: Large-scale spectral clustering on graphs
IJCAI13 Paper review: Large-scale spectral clustering on graphsAkisato Kimura
 
関西CVPR勉強会 2012.10.28
関西CVPR勉強会 2012.10.28関西CVPR勉強会 2012.10.28
関西CVPR勉強会 2012.10.28Akisato Kimura
 
関西CVPR勉強会 2012.7.29
関西CVPR勉強会 2012.7.29関西CVPR勉強会 2012.7.29
関西CVPR勉強会 2012.7.29Akisato Kimura
 
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)Akisato Kimura
 
関西CVPRML勉強会(特定物体認識) 2012.1.14
関西CVPRML勉強会(特定物体認識) 2012.1.14関西CVPRML勉強会(特定物体認識) 2012.1.14
関西CVPRML勉強会(特定物体認識) 2012.1.14Akisato Kimura
 
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -Akisato Kimura
 
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明Akisato Kimura
 

More from Akisato Kimura (20)

Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
 
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
Paper reading - Dropout as a Bayesian Approximation: Representing Model Uncer...
 
多変量解析の一般化
多変量解析の一般化多変量解析の一般化
多変量解析の一般化
 
CVPR2016 reading - 特徴量学習とクロスモーダル転移について
CVPR2016 reading - 特徴量学習とクロスモーダル転移についてCVPR2016 reading - 特徴量学習とクロスモーダル転移について
CVPR2016 reading - 特徴量学習とクロスモーダル転移について
 
NIPS2015 reading - Learning visual biases from human imagination
NIPS2015 reading - Learning visual biases from human imaginationNIPS2015 reading - Learning visual biases from human imagination
NIPS2015 reading - Learning visual biases from human imagination
 
CVPR2015 reading "Global refinement of random forest"
CVPR2015 reading "Global refinement of random forest"CVPR2015 reading "Global refinement of random forest"
CVPR2015 reading "Global refinement of random forest"
 
CVPR2015 reading "Understainding image virality" (in Japanese)
CVPR2015 reading "Understainding image virality" (in Japanese)CVPR2015 reading "Understainding image virality" (in Japanese)
CVPR2015 reading "Understainding image virality" (in Japanese)
 
Computational models of human visual attention driven by auditory cues
Computational models of human visual attention driven by auditory cuesComputational models of human visual attention driven by auditory cues
Computational models of human visual attention driven by auditory cues
 
NIPS2014 reading - Top rank optimization in linear time
NIPS2014 reading - Top rank optimization in linear timeNIPS2014 reading - Top rank optimization in linear time
NIPS2014 reading - Top rank optimization in linear time
 
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...
CVPR2014 reading "Reconstructing storyline graphs for image recommendation fr...
 
ICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged informationICCV2013 reading: Learning to rank using privileged information
ICCV2013 reading: Learning to rank using privileged information
 
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...
ACMMM 2013 reading: Large-scale visual sentiment ontology and detectors using...
 
IJCAI13 Paper review: Large-scale spectral clustering on graphs
IJCAI13 Paper review: Large-scale spectral clustering on graphsIJCAI13 Paper review: Large-scale spectral clustering on graphs
IJCAI13 Paper review: Large-scale spectral clustering on graphs
 
関西CVPR勉強会 2012.10.28
関西CVPR勉強会 2012.10.28関西CVPR勉強会 2012.10.28
関西CVPR勉強会 2012.10.28
 
関西CVPR勉強会 2012.7.29
関西CVPR勉強会 2012.7.29関西CVPR勉強会 2012.7.29
関西CVPR勉強会 2012.7.29
 
ICWSM12 Brief Review
ICWSM12 Brief ReviewICWSM12 Brief Review
ICWSM12 Brief Review
 
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)
関西CVPRML勉強会 2012.2.18 (一般物体認識 - データセット)
 
関西CVPRML勉強会(特定物体認識) 2012.1.14
関西CVPRML勉強会(特定物体認識) 2012.1.14関西CVPRML勉強会(特定物体認識) 2012.1.14
関西CVPRML勉強会(特定物体認識) 2012.1.14
 
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -
人間の視覚的注意を予測するモデル - 動的ベイジアンネットワークに基づく 最新のアプローチ -
 
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明
IBIS2011 企画セッション「CV/PRで独自の進化を遂げる学習・最適化技術」 趣旨説明
 

Recently uploaded

04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
HTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation StrategiesHTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation StrategiesBoston Institute of Analytics
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilV3cube
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProduct Anonymous
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobeapidays
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc
 

Recently uploaded (20)

04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
HTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation StrategiesHTML Injection Attacks: Impact and Mitigation Strategies
HTML Injection Attacks: Impact and Mitigation Strategies
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Developing An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of BrazilDeveloping An App To Navigate The Roads of Brazil
Developing An App To Navigate The Roads of Brazil
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 

NIPS2010 reading: Semi-supervised learning with adversarially missing label information

  • 1. Semi-supervised learning with adversarially missing label information by U. Syad and B. Taskar @ Univ. Penn Akisato Kimura, Ph.D [E-mail] akisato@ieee.org [Twitter] @_akisato
  • 2. Abstract  Semi-supervised learning in an adversarial setting  Consider a general framework, including  Covariate shift : Missing labels are not necessarily selected at random.  Multiple instance learning : Correspondences between labels and instances are not provided explicitly.  Graph-based regularization : Only information about which instances are likely to have the same label.  but not including  Malicious label noise setting : Labelers are allowed to mislabel a part of the training set.  Labelers can remove labels, but cannot change them. 2 NIPS2010 reading @ Cybozu December 26, 2010
  • 3. Framework  : m labeled training examples  : examples  : labels  : unknown PDF  : label regularization function  : soft labeling of the training examples  : probability so that has label  The algorithm can access to only the examples and the label regularization function .  Correct labeling might be near the minimum of , but not necessarily (thus it is called “adversarial”). 3 NIPS2010 reading @ Cybozu December 26, 2010
  • 4. Label regularization functions (1)  Fully supervised learning A set of indexes of unlabeled examples  otherwise  [Note] Labels also denotes the correct soft labeling.  Semi-supervised learning Don’t mind how q labels examples not in I    otherwise A set of indexes of non-zero components  Ambiguous labeling ( multiple instance learning)  a set of possible labels of the example  otherwise 4 NIPS2010 reading @ Cybozu December 26, 2010
  • 5. Label regularization functions (2)  Laplacian regularization  positive semi-definite  is small and are believed to have the same label  Any combinations are also in this framework.  Ex. SSL with Laplacian regularization  We don’t specify how the function is selected.  Ex. Semi-supervised learning : Usually, we can’t know how the unlabeled samples were selected (“covariate shift” setting). 5 NIPS2010 reading @ Cybozu December 26, 2010
  • 6. Loss functions for model learning  Any types of loss functions can be applicable, but  Particularly we focus on  Negative log likelihood of a log-linear PDF  Binary loss function of a linear classifier  a model/classifier parameter  a feature function (“basis function” in PRML) 6 NIPS2010 reading @ Cybozu December 26, 2010
  • 7. Algorithm  Objective: Find a parameter achieving  Its validity will be disclosed later.  For brevity, we abbreviate the objective as . The order is opposite !!  Basic algorithm: Game for Adversarially Missing Evidence (GAME) 1. Constants are given in advance. 2. Find s.t. 3. Find s.t. 4. Output the parameter estimate 7 NIPS2010 reading @ Cybozu December 26, 2010
  • 8. Algorithm in detail  Consider the case of the negative log likelihood  Step 3  The label regularization function can be ignored.  This step is equivalent to maximization of the likelihood of a log-linear model. Many solvers we have  Step 2  Take the dual of the inner minimization Easy to optimize concave Shannon entropy 8 NIPS2010 reading @ Cybozu December 26, 2010
  • 9. Convergence properties (1)  Theorem 1 (Converse theorem)  If , then for all parameters and label regularization functions satisfies with probability . Similar to the objective function of the algorithm 9 NIPS2010 reading @ Cybozu December 26, 2010
  • 10. Convergence properties (2)  Theorem 2 (Direct theorem: lower bound)  Consider the binary loss function, and a finite set . For all learning algorithms and example distributions , there exists a labeling function that satisfies with probability under some assumptions.  The assumptions can be seen in the paper (1,2 and 3).  Some detailed conditions are omitted for the simplicity.  The parameter is obtained by an learning algorithm. 10 NIPS2010 reading @ Cybozu December 26, 2010
  • 11. Convergence properties (3)  Theorem 3 (Direct theorem: upper bound)  Consider the binary loss function. If , there exists a family of label regularization function and a learning algorithm that satisfy with probability under some assumptions.  Assumption 3 can be removed.  Some detailed conditions are omitted for the simplicity.  The parameter is obtained by the learning algorithm. 11 NIPS2010 reading @ Cybozu December 26, 2010
  • 12. Back to the algorithm part …  Objective: Find a parameter achieving  Minimum upper bound in the converse theorem  Do not minimize over .  Instead, add a regularization part to leave unconstrained. 12 NIPS2010 reading @ Cybozu December 26, 2010
  • 13. Experiments  Exp. 1: Binary classification  Datasets: COIL (object recognition), BCI (EEG scans)  Exp. 2: Multi-class categorization  Dataset: Labeled Faces in the Wild 13 NIPS2010 reading @ Cybozu December 26, 2010
  • 14. Results BCI COIL Faces 14 NIPS2010 reading @ Cybozu December 26, 2010
  • 15. この論文から何を思うか?  「一般化」論文は得てして面白くない。  その特殊ケースの結果を知っているので、 「おお、そうだったのか」という驚きを得られにくい。  が、一般化が新しい見方を提供することは見逃せない。  知られていない特殊ケースの発見  特殊ケースの組み合わせによる新しい知見/手法  さらに、一般化で技術を俯瞰できるようになることも重要。  なので、ここからもう一歩先を考えてみましょう! 15 NIPS2010 reading @ Cybozu December 26, 2010
  • 16. Fin.  Thank you for your kind attention. 16 NIPS2010 reading @ Cybozu December 26, 2010