SlideShare a Scribd company logo
1 of 22
- Presented By
Anshu Dipit
Likitha Seeram
UTA : CSE6363 : Machine Learning 1

 A new algorithm for solving L1 – regularized least squares
problem which is more efficient for learning sparse coding
bases
 A new approach for L2 – constrained least squares
problem which result in significant speed up for sparse
coding
Goal
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 What is Sparse Coding?
 Sparse Coding applications in Computer Vision
Image Denoising Image Restoration
Introduction
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Sparse coding is a method for discovering good basis
vectors automatically using only unlabeled data
 It learns the basis functions that capture high-level
features in the data
Input Features selected
Sparse Coding Problem
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Sparse coding is a method for discovering good basis
vectors automatically using only unlabeled data
 It is similar to PCA
 Given a training set of m vectors
where
we attempt to find a succinct representation for each xi
using basis vectors and a sparse vector
such that
 Note that the basis can be overcomplete, i.e., n>k
Sparse Coding Problem
1 2, ,[ ], mx xX x L
k
ix R
1 2, , , k
nbb b L R n
sR
1 2
1
[ , , , ]
n
i j j n
j
x s b b b b s

  L
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 The goal of sparse coding is to present input vectors as
weighted linear combinations of ‘basis vectors’, which capture
high level patterns in input data
 The optimization problem in sparse coding
where
and ᶲ is a sparse penalty function (we consider L1 penalty
function).
Sparse Coding Problem
1 1, , , , ,[ ], [ ], [ 1, ]m n mx bB b S sX x s  L L L
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 The formulation of LASSO
where 𝑥, 𝑦 are vectors, 𝐴 is a matrix and 𝛾 is a constant.
 Basic idea of the algorithm is
 To get the most useful attributes in a vector (data record)
 To guess the sign of each component (attribute) of 𝑥 (data
record), thus, guessing the impact of any changes in the
attribute to the classification of the data record.
A new algorithm to solve
LASSO
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram

UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
First Sign Search Algorithm
L1 Regularized Least
Square
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Consider optimization problem (given in the LASSO slide) augmented
with the additional constraint that 𝑥 is consistent with a given active
set and sign vector. Then, if the current coefficients 𝑥 𝑐 are consistent
with the active set and sign vector, but are not optimal for the
augmented problem at the start of Step 3, the feature-sign step is
guaranteed to strictly reduce the objective.
 Consider optimization problem (LASSO equation) augmented with the
additional constraint that 𝑥 is consistent with a given active set and
sign vector. If the coefficients 𝑥 𝑐 at the start of Step 2 are optimal for
the augmented problem, but are not optimal for problem (LASSO
equation), the feature-sign step is guaranteed to strictly reduce the
objective.
 The feature-sign search algorithm converges to a global optimum of
the optimization problem in a finite number of steps.
Proofs of the Algorithm
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram

UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
Learning Bases using
Lagrange Dual
L2 Constrained Least
Square
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Solving optimization problem over bases B and given fixed
coefficients S.
 This is least squares problem with quadratic constraints,
which can be efficiently solved using Lagrange dual.
 After the calculations, we find the optimal bases 𝐵 as
follows :

UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
Experiment
 Performance of the algorithms was evaluated on four natural
stimulus datasets:
 Natural Images
 Speech
 Stereo Images
 Natural Image Videos
 All experiments were conducted on a Linux machine with AMD
Opteron 2GHz CPU and 2GB RAM
 All the algorithms were implemented in MATLAB
Experiment
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Evaluating Feature sign search algorithm for learning coefficients with L1
sparsity function
 Running time and error are compared with other
coefficient learning algorithms
 For each dataset, a test set of 100 input vectors and a training set of 1000
input vectors was used
Values: Running Time (Relative Error)
Evaluating Feature Sign
Search Algorithm
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Running time (in seconds) for different algorithm combinations
of coefficient learning and basis learning algorithms using
different sparsity functions is shown below:
Time Taken for learning
Bases
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Using these efficient algorithms they were able to learn
overcomplete bases of natural images
1024 bases 2000 bases
(14 X 14 pixels each) (20 X 20 pixels each)
Learning overcomplete
natural images
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 Sparse coding can model the interaction (inhibition)
between the bases (neurons) by sparsifying their
coefficients (activations), and our algorithms enable these
phenomena to be tested with highly overcomplete bases.
 They evaluated whether end-stopping behavior could be
observed in sparse coding framework. The results seemed
consistent with the end stopping behavior of the neurons.
 Using the learned overcomplete bases, they tested for
center-surround non classical receptive field (nCRF)
effects.
Replicating Complex
Neuroscience phenomena
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 They applied sparse coding approaches to self-taught learning,
a new machine learning formalism.
 A supervised learning problem along with additional unlabeled
instances that may not have same class labels as labeled
instances.
 Sparse coding algorithms are applied to unlabeled data to learn
bases which gives a higher level representation of images, thus
making supervised learning task easier.
 This approach proved 11-36% reductions in test error.
 Related Work: R. Raina, A. Battle, H. Lee, B. Packer, and A. Y.
Ng. Self-taught learning. In NIPS Workshop on Learning when
test and training inputs have different distributions, 2006
Application to self-taught
learning
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 In this paper, sparse coding is formulated as a
combination of two convex optimization problems
 Efficient algorithms for these problems were presented:
the feature-sign search for solving the L1-least squares
problem to learn coefficients, and a Lagrange dual
method for the L2-constrained least squares problem to
learn the bases for any sparsity penalty function.
 Partially explain the phenomena of end-stopping and
nCRF surround suppression in V1 neurons.
Conclusion
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
 http://ai.stanford.edu/~hllee/softwares/nips06-
sparsecoding.htm
 https://www.cs.ubc.ca/~schmidtm/MLRG/sparseCoding.pdf
 http://slideplayer.com/slide/4953202/
 https://www.youtube.com/watch?v=7a0_iEruGoM
 http://www.inb.uni-
luebeck.de/publications/pdfs/HoLaBaMa12.pdf
References
UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram

Thank You
UTA : CSE6363 : Machine Learning 22

More Related Content

What's hot

Me 443 1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443   1 what is mathematica Erdi Karaçal Mechanical Engineer University of...Me 443   1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443 1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
Erdi Karaçal
 
HW2-1_05.doc
HW2-1_05.docHW2-1_05.doc
HW2-1_05.doc
butest
 
Programming in python
Programming in pythonProgramming in python
Programming in python
Ivan Rojas
 

What's hot (18)

Slide1
Slide1Slide1
Slide1
 
Beginners Guide to Non-Negative Matrix Factorization
Beginners Guide to Non-Negative Matrix FactorizationBeginners Guide to Non-Negative Matrix Factorization
Beginners Guide to Non-Negative Matrix Factorization
 
Me 443 2 tour of mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443   2 tour of mathematica Erdi Karaçal Mechanical Engineer University of...Me 443   2 tour of mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443 2 tour of mathematica Erdi Karaçal Mechanical Engineer University of...
 
Me 443 1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443   1 what is mathematica Erdi Karaçal Mechanical Engineer University of...Me 443   1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
Me 443 1 what is mathematica Erdi Karaçal Mechanical Engineer University of...
 
Algorithms with-java-advanced-1.0
Algorithms with-java-advanced-1.0Algorithms with-java-advanced-1.0
Algorithms with-java-advanced-1.0
 
A045030105
A045030105A045030105
A045030105
 
HW2-1_05.doc
HW2-1_05.docHW2-1_05.doc
HW2-1_05.doc
 
K Nearest Neighbor Algorithm
K Nearest Neighbor AlgorithmK Nearest Neighbor Algorithm
K Nearest Neighbor Algorithm
 
Datastructures using c++
Datastructures using c++Datastructures using c++
Datastructures using c++
 
Independent Component Analysis
Independent Component Analysis Independent Component Analysis
Independent Component Analysis
 
DATA STRUCTURE AND ALGORITHM FULL NOTES
DATA STRUCTURE AND ALGORITHM FULL NOTESDATA STRUCTURE AND ALGORITHM FULL NOTES
DATA STRUCTURE AND ALGORITHM FULL NOTES
 
Programming in python
Programming in pythonProgramming in python
Programming in python
 
Chapitre 08
Chapitre 08Chapitre 08
Chapitre 08
 
Me 443 4 plotting curves Erdi Karaçal Mechanical Engineer University of Gaz...
Me 443   4 plotting curves Erdi Karaçal Mechanical Engineer University of Gaz...Me 443   4 plotting curves Erdi Karaçal Mechanical Engineer University of Gaz...
Me 443 4 plotting curves Erdi Karaçal Mechanical Engineer University of Gaz...
 
A NEW PARALLEL ALGORITHM FOR COMPUTING MINIMUM SPANNING TREE
A NEW PARALLEL ALGORITHM FOR COMPUTING MINIMUM SPANNING TREEA NEW PARALLEL ALGORITHM FOR COMPUTING MINIMUM SPANNING TREE
A NEW PARALLEL ALGORITHM FOR COMPUTING MINIMUM SPANNING TREE
 
Parallel processing technique for high speed image segmentation using color
Parallel processing technique for high speed image segmentation using colorParallel processing technique for high speed image segmentation using color
Parallel processing technique for high speed image segmentation using color
 
Big o notation
Big o notationBig o notation
Big o notation
 
Introduction to Data Structures & Algorithms
Introduction to Data Structures & AlgorithmsIntroduction to Data Structures & Algorithms
Introduction to Data Structures & Algorithms
 

Similar to Efficient Sparse Coding Algorithms

Kernal based speaker specific feature extraction and its applications in iTau...
Kernal based speaker specific feature extraction and its applications in iTau...Kernal based speaker specific feature extraction and its applications in iTau...
Kernal based speaker specific feature extraction and its applications in iTau...
TELKOMNIKA JOURNAL
 
Machine Learning Notes for beginners ,Step by step
Machine Learning Notes for beginners ,Step by stepMachine Learning Notes for beginners ,Step by step
Machine Learning Notes for beginners ,Step by step
SanjanaSaxena17
 
The Most Important Algorithms
The Most Important AlgorithmsThe Most Important Algorithms
The Most Important Algorithms
wensheng wei
 

Similar to Efficient Sparse Coding Algorithms (20)

Kernal based speaker specific feature extraction and its applications in iTau...
Kernal based speaker specific feature extraction and its applications in iTau...Kernal based speaker specific feature extraction and its applications in iTau...
Kernal based speaker specific feature extraction and its applications in iTau...
 
A Comparative study of K-SVD and WSQ Algorithms in Fingerprint Compression Te...
A Comparative study of K-SVD and WSQ Algorithms in Fingerprint Compression Te...A Comparative study of K-SVD and WSQ Algorithms in Fingerprint Compression Te...
A Comparative study of K-SVD and WSQ Algorithms in Fingerprint Compression Te...
 
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
 
chapter 1
chapter 1chapter 1
chapter 1
 
House price prediction
House price predictionHouse price prediction
House price prediction
 
fINAL ML PPT.pptx
fINAL ML PPT.pptxfINAL ML PPT.pptx
fINAL ML PPT.pptx
 
Occurrence Prediction_NLP
Occurrence Prediction_NLPOccurrence Prediction_NLP
Occurrence Prediction_NLP
 
Survey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsSurvey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique Algorithms
 
Y34147151
Y34147151Y34147151
Y34147151
 
20 bayes learning
20 bayes learning20 bayes learning
20 bayes learning
 
VCE Unit 01 (1).pptx
VCE Unit 01 (1).pptxVCE Unit 01 (1).pptx
VCE Unit 01 (1).pptx
 
Machine Learning Notes for beginners ,Step by step
Machine Learning Notes for beginners ,Step by stepMachine Learning Notes for beginners ,Step by step
Machine Learning Notes for beginners ,Step by step
 
Vol 16 No 2 - July-December 2016
Vol 16 No 2 - July-December 2016Vol 16 No 2 - July-December 2016
Vol 16 No 2 - July-December 2016
 
A New Method Based on MDA to Enhance the Face Recognition Performance
A New Method Based on MDA to Enhance the Face Recognition PerformanceA New Method Based on MDA to Enhance the Face Recognition Performance
A New Method Based on MDA to Enhance the Face Recognition Performance
 
Gk3611601162
Gk3611601162Gk3611601162
Gk3611601162
 
IMPROVEMENT IN IMAGE DENOISING OF HANDWRITTEN DIGITS USING AUTOENCODERS IN DE...
IMPROVEMENT IN IMAGE DENOISING OF HANDWRITTEN DIGITS USING AUTOENCODERS IN DE...IMPROVEMENT IN IMAGE DENOISING OF HANDWRITTEN DIGITS USING AUTOENCODERS IN DE...
IMPROVEMENT IN IMAGE DENOISING OF HANDWRITTEN DIGITS USING AUTOENCODERS IN DE...
 
The Most Important Algorithms
The Most Important AlgorithmsThe Most Important Algorithms
The Most Important Algorithms
 
Discrete structure ch 3 short question's
Discrete structure ch 3 short question'sDiscrete structure ch 3 short question's
Discrete structure ch 3 short question's
 
DAA UNIT 3
DAA UNIT 3DAA UNIT 3
DAA UNIT 3
 
Neural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabNeural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlab
 

Recently uploaded

Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
VictoriaMetrics
 
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
Medical / Health Care (+971588192166) Mifepristone and Misoprostol tablets 200mg
 

Recently uploaded (20)

WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
WSO2CON 2024 - Cloud Native Middleware: Domain-Driven Design, Cell-Based Arch...
 
WSO2CON 2024 - IoT Needs CIAM: The Importance of Centralized IAM in a Growing...
WSO2CON 2024 - IoT Needs CIAM: The Importance of Centralized IAM in a Growing...WSO2CON 2024 - IoT Needs CIAM: The Importance of Centralized IAM in a Growing...
WSO2CON 2024 - IoT Needs CIAM: The Importance of Centralized IAM in a Growing...
 
WSO2CON 2024 - Architecting AI in the Enterprise: APIs and Applications
WSO2CON 2024 - Architecting AI in the Enterprise: APIs and ApplicationsWSO2CON 2024 - Architecting AI in the Enterprise: APIs and Applications
WSO2CON 2024 - Architecting AI in the Enterprise: APIs and Applications
 
WSO2Con2024 - WSO2's IAM Vision: Identity-Led Digital Transformation
WSO2Con2024 - WSO2's IAM Vision: Identity-Led Digital TransformationWSO2Con2024 - WSO2's IAM Vision: Identity-Led Digital Transformation
WSO2Con2024 - WSO2's IAM Vision: Identity-Led Digital Transformation
 
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
Large-scale Logging Made Easy: Meetup at Deutsche Bank 2024
 
WSO2Con2024 - Software Delivery in Hybrid Environments
WSO2Con2024 - Software Delivery in Hybrid EnvironmentsWSO2Con2024 - Software Delivery in Hybrid Environments
WSO2Con2024 - Software Delivery in Hybrid Environments
 
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
%in Stilfontein+277-882-255-28 abortion pills for sale in Stilfontein
 
WSO2Con2024 - Low-Code Integration Tooling
WSO2Con2024 - Low-Code Integration ToolingWSO2Con2024 - Low-Code Integration Tooling
WSO2Con2024 - Low-Code Integration Tooling
 
AI & Machine Learning Presentation Template
AI & Machine Learning Presentation TemplateAI & Machine Learning Presentation Template
AI & Machine Learning Presentation Template
 
WSO2CON 2024 - Designing Event-Driven Enterprises: Stories of Transformation
WSO2CON 2024 - Designing Event-Driven Enterprises: Stories of TransformationWSO2CON 2024 - Designing Event-Driven Enterprises: Stories of Transformation
WSO2CON 2024 - Designing Event-Driven Enterprises: Stories of Transformation
 
%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in soweto%in Soweto+277-882-255-28 abortion pills for sale in soweto
%in Soweto+277-882-255-28 abortion pills for sale in soweto
 
WSO2CON 2024 - Freedom First—Unleashing Developer Potential with Open Source
WSO2CON 2024 - Freedom First—Unleashing Developer Potential with Open SourceWSO2CON 2024 - Freedom First—Unleashing Developer Potential with Open Source
WSO2CON 2024 - Freedom First—Unleashing Developer Potential with Open Source
 
Architecture decision records - How not to get lost in the past
Architecture decision records - How not to get lost in the pastArchitecture decision records - How not to get lost in the past
Architecture decision records - How not to get lost in the past
 
WSO2Con2024 - Navigating the Digital Landscape: Transforming Healthcare with ...
WSO2Con2024 - Navigating the Digital Landscape: Transforming Healthcare with ...WSO2Con2024 - Navigating the Digital Landscape: Transforming Healthcare with ...
WSO2Con2024 - Navigating the Digital Landscape: Transforming Healthcare with ...
 
Crypto Cloud Review - How To Earn Up To $500 Per DAY Of Bitcoin 100% On AutoP...
Crypto Cloud Review - How To Earn Up To $500 Per DAY Of Bitcoin 100% On AutoP...Crypto Cloud Review - How To Earn Up To $500 Per DAY Of Bitcoin 100% On AutoP...
Crypto Cloud Review - How To Earn Up To $500 Per DAY Of Bitcoin 100% On AutoP...
 
WSO2Con2024 - Enabling Transactional System's Exponential Growth With Simplicity
WSO2Con2024 - Enabling Transactional System's Exponential Growth With SimplicityWSO2Con2024 - Enabling Transactional System's Exponential Growth With Simplicity
WSO2Con2024 - Enabling Transactional System's Exponential Growth With Simplicity
 
WSO2CON 2024 - Not Just Microservices: Rightsize Your Services!
WSO2CON 2024 - Not Just Microservices: Rightsize Your Services!WSO2CON 2024 - Not Just Microservices: Rightsize Your Services!
WSO2CON 2024 - Not Just Microservices: Rightsize Your Services!
 
WSO2Con2024 - From Blueprint to Brilliance: WSO2's Guide to API-First Enginee...
WSO2Con2024 - From Blueprint to Brilliance: WSO2's Guide to API-First Enginee...WSO2Con2024 - From Blueprint to Brilliance: WSO2's Guide to API-First Enginee...
WSO2Con2024 - From Blueprint to Brilliance: WSO2's Guide to API-First Enginee...
 
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
Abortion Pill Prices Tembisa [(+27832195400*)] 🏥 Women's Abortion Clinic in T...
 
WSO2CON 2024 - Does Open Source Still Matter?
WSO2CON 2024 - Does Open Source Still Matter?WSO2CON 2024 - Does Open Source Still Matter?
WSO2CON 2024 - Does Open Source Still Matter?
 

Efficient Sparse Coding Algorithms

  • 1. - Presented By Anshu Dipit Likitha Seeram UTA : CSE6363 : Machine Learning 1
  • 2.   A new algorithm for solving L1 – regularized least squares problem which is more efficient for learning sparse coding bases  A new approach for L2 – constrained least squares problem which result in significant speed up for sparse coding Goal UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 3.  What is Sparse Coding?  Sparse Coding applications in Computer Vision Image Denoising Image Restoration Introduction UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 4.  Sparse coding is a method for discovering good basis vectors automatically using only unlabeled data  It learns the basis functions that capture high-level features in the data Input Features selected Sparse Coding Problem UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 5.  Sparse coding is a method for discovering good basis vectors automatically using only unlabeled data  It is similar to PCA  Given a training set of m vectors where we attempt to find a succinct representation for each xi using basis vectors and a sparse vector such that  Note that the basis can be overcomplete, i.e., n>k Sparse Coding Problem 1 2, ,[ ], mx xX x L k ix R 1 2, , , k nbb b L R n sR 1 2 1 [ , , , ] n i j j n j x s b b b b s    L UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 6.  The goal of sparse coding is to present input vectors as weighted linear combinations of ‘basis vectors’, which capture high level patterns in input data  The optimization problem in sparse coding where and ᶲ is a sparse penalty function (we consider L1 penalty function). Sparse Coding Problem 1 1, , , , ,[ ], [ ], [ 1, ]m n mx bB b S sX x s  L L L UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 7.  The formulation of LASSO where 𝑥, 𝑦 are vectors, 𝐴 is a matrix and 𝛾 is a constant.  Basic idea of the algorithm is  To get the most useful attributes in a vector (data record)  To guess the sign of each component (attribute) of 𝑥 (data record), thus, guessing the impact of any changes in the attribute to the classification of the data record. A new algorithm to solve LASSO UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 8.  UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram First Sign Search Algorithm L1 Regularized Least Square
  • 9. UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 10.  Consider optimization problem (given in the LASSO slide) augmented with the additional constraint that 𝑥 is consistent with a given active set and sign vector. Then, if the current coefficients 𝑥 𝑐 are consistent with the active set and sign vector, but are not optimal for the augmented problem at the start of Step 3, the feature-sign step is guaranteed to strictly reduce the objective.  Consider optimization problem (LASSO equation) augmented with the additional constraint that 𝑥 is consistent with a given active set and sign vector. If the coefficients 𝑥 𝑐 at the start of Step 2 are optimal for the augmented problem, but are not optimal for problem (LASSO equation), the feature-sign step is guaranteed to strictly reduce the objective.  The feature-sign search algorithm converges to a global optimum of the optimization problem in a finite number of steps. Proofs of the Algorithm UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 11.  UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram Learning Bases using Lagrange Dual L2 Constrained Least Square
  • 12. UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram  Solving optimization problem over bases B and given fixed coefficients S.  This is least squares problem with quadratic constraints, which can be efficiently solved using Lagrange dual.  After the calculations, we find the optimal bases 𝐵 as follows :
  • 13.  UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram Experiment
  • 14.  Performance of the algorithms was evaluated on four natural stimulus datasets:  Natural Images  Speech  Stereo Images  Natural Image Videos  All experiments were conducted on a Linux machine with AMD Opteron 2GHz CPU and 2GB RAM  All the algorithms were implemented in MATLAB Experiment UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 15.  Evaluating Feature sign search algorithm for learning coefficients with L1 sparsity function  Running time and error are compared with other coefficient learning algorithms  For each dataset, a test set of 100 input vectors and a training set of 1000 input vectors was used Values: Running Time (Relative Error) Evaluating Feature Sign Search Algorithm UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 16.  Running time (in seconds) for different algorithm combinations of coefficient learning and basis learning algorithms using different sparsity functions is shown below: Time Taken for learning Bases UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 17.  Using these efficient algorithms they were able to learn overcomplete bases of natural images 1024 bases 2000 bases (14 X 14 pixels each) (20 X 20 pixels each) Learning overcomplete natural images UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 18.  Sparse coding can model the interaction (inhibition) between the bases (neurons) by sparsifying their coefficients (activations), and our algorithms enable these phenomena to be tested with highly overcomplete bases.  They evaluated whether end-stopping behavior could be observed in sparse coding framework. The results seemed consistent with the end stopping behavior of the neurons.  Using the learned overcomplete bases, they tested for center-surround non classical receptive field (nCRF) effects. Replicating Complex Neuroscience phenomena UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 19.  They applied sparse coding approaches to self-taught learning, a new machine learning formalism.  A supervised learning problem along with additional unlabeled instances that may not have same class labels as labeled instances.  Sparse coding algorithms are applied to unlabeled data to learn bases which gives a higher level representation of images, thus making supervised learning task easier.  This approach proved 11-36% reductions in test error.  Related Work: R. Raina, A. Battle, H. Lee, B. Packer, and A. Y. Ng. Self-taught learning. In NIPS Workshop on Learning when test and training inputs have different distributions, 2006 Application to self-taught learning UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 20.  In this paper, sparse coding is formulated as a combination of two convex optimization problems  Efficient algorithms for these problems were presented: the feature-sign search for solving the L1-least squares problem to learn coefficients, and a Lagrange dual method for the L2-constrained least squares problem to learn the bases for any sparsity penalty function.  Partially explain the phenomena of end-stopping and nCRF surround suppression in V1 neurons. Conclusion UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 21.  http://ai.stanford.edu/~hllee/softwares/nips06- sparsecoding.htm  https://www.cs.ubc.ca/~schmidtm/MLRG/sparseCoding.pdf  http://slideplayer.com/slide/4953202/  https://www.youtube.com/watch?v=7a0_iEruGoM  http://www.inb.uni- luebeck.de/publications/pdfs/HoLaBaMa12.pdf References UTA : CSE6363 : Machine Learning Anshu Dipit / Likitha Seeram
  • 22.  Thank You UTA : CSE6363 : Machine Learning 22

Editor's Notes

  1. 2 optimization problems over 2 subset of variables.
  2. Sparse coding provides a class of algorithms for finding succinct representations of stimuli; given only unlabeled input data, it discovers basis functions that capture higher-level features in the data
  3. Digit Recognition. Features capture significant properties of the digits
  4. Sparse coding can be applied for learning over complete basis, in which the number of bases is greater than the input dimension.
  5. Beta is constant. Assuming uniform prior on basis. This objective is to iteratively optimize by alternatingly optimizing w.r.t B and S while holding the other constant.
  6. LARS – Least Angle Regression, Chen et al’s interior point method. Relative error calculation – f obj is final objective value attained by the algorithm and f* is best objective value attained from among all the algorithms
  7. As a result, we can see that Lagrange dual was much faster than gradient descent with projections
  8. 1024 bases in 2 hours. 2000 bases in 10 hours. This is not possible using gradient descent method of basis learning.
  9. V1 neurons – primary visual cortex
  10. Paper link - http://ai.stanford.edu/~hllee/nips06-sparsecoding.pdf