Sparse & Redundant Representation Modeling of Images: Theory and Applications  Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa 32000, IsraelSeventh International Conference on     Curves and SurfacesAvignon - FRANCE June 24-30, 2010This research was supported by the European Community's FP7-FET program SMALL under grant agreement no. 225913
2   This Talk Gives and Overview On … A decade of tremendous progress in the field of Sparse and Redundant RepresentationsNumerical ProblemsTheoryApplicationsSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
3AgendaPart II – Theoretical & Numerical FoundationsPart I – Denoising by Sparse & Redundant RepresentationsPart III – Dictionary Learning         & The K-SVD Algorithm Part IV 	– Back to Denoising … and Beyond – handling stills and video denoising & inpainting, demosaicing, super-res., and compressionPart V	–Summary & ConclusionsSparsity and Redundancy are valuable and well-founded tools for modeling data.
When used in image processing, they lead to state-of-the-art results. Today we will show that Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
4Part IDenoising by                              Sparse & Redundant                 RepresentationsSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
5?Remove Additive Noise   Noise Removal?Our story begins with image denoising …Important: (i) Practical application; (ii) A convenient platform                       (being the simplest inverse problem) for testing basic ideas in image processing, and then generalizing to more complex problems.
Many Considered Directions: Partial differential equations, Statistical estimators, Adaptive filters, Inverse problems & regularization,          Wavelets, Example-based techniques, Sparse representations, …Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
6 Relation to measurementsPrior or regularizationThomasBayes                                    1702 - 1761   Denoising By Energy Minimization Many of the proposed image denoising algorithms are related to the minimization of an energy function of the formy : Given measurements  x : Unknown to be recoveredThis is in-fact a Bayesian point of view, adopting the Maximum-A-posteriori Probability (MAP) estimation.
Clearly, the wisdom in such an approach is within the choice of the prior – modeling the images of interest. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
7EnergySmoothnessAdapt+ SmoothRobust Statistics Hidden Markov Models,
 Compression algorithms as priors,
 …Total-VariationWavelet SparsitySparse & Redundant   The Evolution of G(x)During the past several decades we have made all sort of guesses about the prior G(x) for images:   Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
8Every column in    D (dictionary) is    a prototype signal (atom).
The vector is generated randomly with few (say L) non-zeros at random locations and with random values. NNA sparse & random vectorKA fixed Dictionary   Sparse Modeling of Signals MWe shall refer to this model as SparselandSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
9MMultiply by DSparseland  Signals are SpecialInteresting Model:Simple:Every generated   signal is built as a linear combination of fewatoms   from our dictionaryD
Rich:A general model: the obtained signals are a union     of many low-dimensional Gaussians.
Familiar: We have been  using this model in other context for a while now (wavelet, JPEG, …).Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
10As p  0 we  get a count         of the non-zeros in the vector1-1+1   Sparse & Redundant Rep. Modeling?Our signal  model is thus: Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
11D-y=            -   Back to Our MAP Energy Function We L0 norm is effectively                                                                  counting the number of                                                                  non-zeros in .
The vector  is the                                                            representation (sparse/redundant)                                                     of the desired                                                                                               signal x.
The core idea: while few (L out of K) atoms can be merged        to form the true signal, the noise cannot be fitted well. Thus, we obtain an effective projection of the noise onto a very         low-dimensional space, thus getting denoising effect. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
12Wait! There are Some Issues Numerical Problems: How should we solve or approximate the solution of the problem					      or                                                                           or                                    ?Theoretical Problems: Is there a unique sparse representation? If we are to approximate the solution somehow, how close will we get?
Practical Problems: What dictionary D should we use, such that all this leads to effective denoising? Will all this work in applications?Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
Image Denoising & Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad13To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsWhat do    we do?  There are some issues: TheoreticalHow to approximate?What about D?Great!      No?
14Part IITheoretical &                   Numerical Foundations Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
Sparse and Redundant                    Signal Representation, and Its Role in Image Processing15   Lets Start with the Noiseless ProblemSuppose we build a signal by the relationWe aim to find the signal’s representation: Known Why should we necessarily get            ?It might happen that eventually                    .Uniqueness
16*Definition:Given a matrix D, =Spark{D} is the smallestnumber of columns that are linearly dependent.Donoho & E. (‘02) Example:Spark = 3*	In tensor decomposition, Kruskal defined something similar already in 1989.   Matrix “Spark”Rank  = 4Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
17Suppose this problem has been solved somehowUniquenessIf we found a representation that satisfy Then necessarily it is unique (the sparsest).Donoho & E. (‘02) MThis result implies that if       generates signals using “sparse enough” , the solution of the above will find it exactly.   Uniqueness RuleSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
18This is a combinatorial problem, proven to be NP-Hard! Solve the LS problem for each support                                  There are (K) such supportsL   Our Goal  Here is a recipe for solving this problem:Gather all the supports {Si}i        of cardinality L   LS error ≤ ε2?Set L=1 YesNoSet L=L+1 Assume: K=1000, L=10 (known!), 1 nano-sec per each LS        We shall need ~8e+6 years to solve this problem !!!!!DoneSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
19   Lets Approximate   Greedy methodsBuild the solution one non-zero element at a timeRelaxation methodsSmooth the L0 and use continuous optimization techniquesSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
20   Relaxation – The Basis Pursuit (BP)Instead of solvingSolve InsteadThis is known as the Basis-Pursuit (BP) [Chen, Donoho & Saunders (’95)].
The newly defined problem is convex (quad. programming).
Very efficient solvers can be deployed:
Interior point methods [Chen, Donoho, & Saunders (‘95)] [Kim, Koh, Lustig, Boyd, & D. Gorinevsky (`07)].
Sequential shrinkage for union of ortho-bases [Bruce et.al. (‘98)].
Iterative shrinkage [Figuerido & Nowak (‘03)] [Daubechies, Defrise, & De-Mole (‘04)]                     [E. (‘05)] [E., Matalon, & Zibulevsky (‘06)] [Beck & Teboulle (`09)] … Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
21   Go Greedy: Matching Pursuit (MP)The MPis one of the greedy algorithms that finds one atom at a time [Mallat & Zhang (’93)].
Step 1: find the one atom that  best matches the signal.
Next steps: given the previously found atoms, find the next one to best fit the rsidual.
The algorithm stops when the error            is below the destination threshold.
The Orthogonal MP (OMP) is an improved version that re-evaluates the coefficients by Least-Squares after each round.Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
22Pursuit Algorithms?Why should they workThere are various algorithms designed for approximating the solution of this problem: Greedy Algorithms: Matching Pursuit, Orthogonal Matching Pursuit (OMP), Least-Squares-OMP, Weak Matching Pursuit, Block Matching Pursuit [1993-today].
Relaxation Algorithms: Basis Pursuit (a.k.a. LASSO), Dnatzig Selector & numerical ways to handle them [1995-today].
Hybrid Algorithms: StOMP, CoSaMP, Subspace Pursuit, Iterative Hard-Thresholding [2007-today].
…Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
23D=DTDTDThe Mutual Coherence is a property of the dictionary (just like the “Spark”). In fact, the following relation             can be shown: The Mutual CoherenceComputeAssume normalized columnsThe Mutual Coherence M is the largest off-diagonal           entry in absolute value.Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
24   BP and MP Equivalence (No Noise)EquivalenceGiven a signal x with a representation            ,assuming that                         , BP and MP   are guaranteed to find the sparsest solution. Donoho & E. (‘02) Gribonval & Nielsen (‘03)Tropp (‘03) Temlyakov (‘03)MP and BP are different in general (hard to say which is better).
The above result corresponds to the worst-case, and as such, it is too pessimistic.
Average performance results are available too, showing much better bounds [Donoho (`04)] [Candes et.al. (‘04)] [Tanner et.al. (‘05)]             [E. (‘06)] [Tropp et.al. (‘06)] … [Candes et. al. (‘09)]. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
25   BP Stability for the Noisy Case Given a signal                with a representationsatisfying                    and a white Gaussian noise                   , BP will show  stability, i.e., Stability*Ben-Haim, Eldar & E. (‘09)* With very high   probability (as    K goes to ∞)For =0 we get a weaker version of the previous result.
This result is the oracle’s error, multuiplied by C·logK.
Similar results exist for other pursuit algorithms (Dantzig Selector, Orthogonal Matching Pursuit, CoSaMP, Subspace Pursuit, …)Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
Image Denoising & Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad26To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsProblems?What do    we do?  We have seen that there are approximation methods to find the sparsest solution, and there are theoretical results that guarantee their success.The Dictionary D should be found somehow !!!What           next?
27Part IIIDictionary Learning:                         The K-SVD AlgorithmSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
28D should be chosen such that it sparsifies the representationsThe approach we will take for building D is training it,   based on Learning from          Image ExamplesOne approach to choose D is from a known set of transforms (Steerable wavelet, Curvelet, Contourlets, Bandlets, Shearlets…)   What Should D Be? Our Assumption: Good-behaved Images                                      have a sparse representationSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
29DXAEach example has a sparse representation with no more than L atomsEach example is                    a linear combination                   of atoms from DMeasure of Quality for D[Field & Olshausen (‘96)][Engan et. al. (‘99)][Lewicki & Sejnowski (‘00)][Cotter et. al. (‘03)][Gribonval et. al. (‘04)][Aharon, E. & Bruckstein (‘04)] [Aharon, E. & Bruckstein (‘05)]Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
30DInitialize         DSparse CodingNearest NeighborXTDictionary UpdateColumn-by-Column by  Mean computation over the relevant examples   K–Means For Clustering Clustering: An extreme sparse representation  Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
31DInitialize         DSparse CodingUse Matching PursuitXTDictionary UpdateColumn-by-Column by  SVD computation over the relevant examples   The K–SVD Algorithm – General [Aharon, E. & Bruckstein (‘04,‘05)]Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
32DD is known!  For the jth item           we solve XT   K–SVD: Sparse Coding StageSolved by                            A Pursuit AlgorithmSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
33Fixing all A and D apart from the kth column, and seek both dk and the kth column in A to better fit the residual!We should solve:   K–SVD: Dictionary Update StageWe refer only to the examples that use the column dkDSVDSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
Image Denoising & Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad34To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsProblems?What do    we do?  We have seen approximation methods that find the sparsest solution, and theoretical results that guarantee their success. We also saw a way to learn DWill it all work in applications? What           next?
35Part IVBack to Denoising …                 and Beyond –                     Combining it AllSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
36kDNExtracts a patch in the ij locationOur prior   From Local to Global TreatmentThe K-SVD algorithm is reasonable for low-dimension signals (N in the range 10-400). As N grows, the complexity and the memory requirements of the K-SVD become prohibitive.
So, how should large images be handled?

C&s sparse june_2010

  • 1.
    Sparse & RedundantRepresentation Modeling of Images: Theory and Applications Michael Elad The Computer Science Department The Technion – Israel Institute of technology Haifa 32000, IsraelSeventh International Conference on Curves and SurfacesAvignon - FRANCE June 24-30, 2010This research was supported by the European Community's FP7-FET program SMALL under grant agreement no. 225913
  • 2.
    2 This Talk Gives and Overview On … A decade of tremendous progress in the field of Sparse and Redundant RepresentationsNumerical ProblemsTheoryApplicationsSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 3.
    3AgendaPart II –Theoretical & Numerical FoundationsPart I – Denoising by Sparse & Redundant RepresentationsPart III – Dictionary Learning & The K-SVD Algorithm Part IV – Back to Denoising … and Beyond – handling stills and video denoising & inpainting, demosaicing, super-res., and compressionPart V –Summary & ConclusionsSparsity and Redundancy are valuable and well-founded tools for modeling data.
  • 4.
    When used inimage processing, they lead to state-of-the-art results. Today we will show that Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 5.
    4Part IDenoising by Sparse & Redundant RepresentationsSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 6.
    5?Remove Additive Noise Noise Removal?Our story begins with image denoising …Important: (i) Practical application; (ii) A convenient platform (being the simplest inverse problem) for testing basic ideas in image processing, and then generalizing to more complex problems.
  • 7.
    Many Considered Directions:Partial differential equations, Statistical estimators, Adaptive filters, Inverse problems & regularization, Wavelets, Example-based techniques, Sparse representations, …Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 8.
    6 Relation tomeasurementsPrior or regularizationThomasBayes 1702 - 1761 Denoising By Energy Minimization Many of the proposed image denoising algorithms are related to the minimization of an energy function of the formy : Given measurements x : Unknown to be recoveredThis is in-fact a Bayesian point of view, adopting the Maximum-A-posteriori Probability (MAP) estimation.
  • 9.
    Clearly, the wisdomin such an approach is within the choice of the prior – modeling the images of interest. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 10.
  • 11.
  • 12.
    …Total-VariationWavelet SparsitySparse& Redundant The Evolution of G(x)During the past several decades we have made all sort of guesses about the prior G(x) for images: Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 13.
    8Every column in D (dictionary) is a prototype signal (atom).
  • 14.
    The vector isgenerated randomly with few (say L) non-zeros at random locations and with random values. NNA sparse & random vectorKA fixed Dictionary Sparse Modeling of Signals MWe shall refer to this model as SparselandSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 15.
    9MMultiply by DSparseland Signals are SpecialInteresting Model:Simple:Every generated signal is built as a linear combination of fewatoms from our dictionaryD
  • 16.
    Rich:A general model:the obtained signals are a union of many low-dimensional Gaussians.
  • 17.
    Familiar: We havebeen using this model in other context for a while now (wavelet, JPEG, …).Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 18.
    10As p 0 we get a count of the non-zeros in the vector1-1+1 Sparse & Redundant Rep. Modeling?Our signal model is thus: Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 19.
    11D-y= - Back to Our MAP Energy Function We L0 norm is effectively counting the number of non-zeros in .
  • 20.
    The vector is the representation (sparse/redundant) of the desired signal x.
  • 21.
    The core idea:while few (L out of K) atoms can be merged to form the true signal, the noise cannot be fitted well. Thus, we obtain an effective projection of the noise onto a very low-dimensional space, thus getting denoising effect. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 22.
    12Wait! There areSome Issues Numerical Problems: How should we solve or approximate the solution of the problem or or ?Theoretical Problems: Is there a unique sparse representation? If we are to approximate the solution somehow, how close will we get?
  • 23.
    Practical Problems: Whatdictionary D should we use, such that all this leads to effective denoising? Will all this work in applications?Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 24.
    Image Denoising &Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad13To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsWhat do we do? There are some issues: TheoreticalHow to approximate?What about D?Great! No?
  • 25.
    14Part IITheoretical & Numerical Foundations Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 26.
    Sparse and Redundant Signal Representation, and Its Role in Image Processing15 Lets Start with the Noiseless ProblemSuppose we build a signal by the relationWe aim to find the signal’s representation: Known Why should we necessarily get ?It might happen that eventually .Uniqueness
  • 27.
    16*Definition:Given a matrixD, =Spark{D} is the smallestnumber of columns that are linearly dependent.Donoho & E. (‘02) Example:Spark = 3* In tensor decomposition, Kruskal defined something similar already in 1989. Matrix “Spark”Rank = 4Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 28.
    17Suppose this problemhas been solved somehowUniquenessIf we found a representation that satisfy Then necessarily it is unique (the sparsest).Donoho & E. (‘02) MThis result implies that if generates signals using “sparse enough” , the solution of the above will find it exactly. Uniqueness RuleSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 29.
    18This is acombinatorial problem, proven to be NP-Hard! Solve the LS problem for each support There are (K) such supportsL Our Goal Here is a recipe for solving this problem:Gather all the supports {Si}i of cardinality L LS error ≤ ε2?Set L=1 YesNoSet L=L+1 Assume: K=1000, L=10 (known!), 1 nano-sec per each LS We shall need ~8e+6 years to solve this problem !!!!!DoneSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 30.
    19 Lets Approximate Greedy methodsBuild the solution one non-zero element at a timeRelaxation methodsSmooth the L0 and use continuous optimization techniquesSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 31.
    20 Relaxation – The Basis Pursuit (BP)Instead of solvingSolve InsteadThis is known as the Basis-Pursuit (BP) [Chen, Donoho & Saunders (’95)].
  • 32.
    The newly definedproblem is convex (quad. programming).
  • 33.
    Very efficient solverscan be deployed:
  • 34.
    Interior point methods[Chen, Donoho, & Saunders (‘95)] [Kim, Koh, Lustig, Boyd, & D. Gorinevsky (`07)].
  • 35.
    Sequential shrinkage forunion of ortho-bases [Bruce et.al. (‘98)].
  • 36.
    Iterative shrinkage [Figuerido& Nowak (‘03)] [Daubechies, Defrise, & De-Mole (‘04)] [E. (‘05)] [E., Matalon, & Zibulevsky (‘06)] [Beck & Teboulle (`09)] … Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 37.
    21 Go Greedy: Matching Pursuit (MP)The MPis one of the greedy algorithms that finds one atom at a time [Mallat & Zhang (’93)].
  • 38.
    Step 1: findthe one atom that best matches the signal.
  • 39.
    Next steps: giventhe previously found atoms, find the next one to best fit the rsidual.
  • 40.
    The algorithm stopswhen the error is below the destination threshold.
  • 41.
    The Orthogonal MP(OMP) is an improved version that re-evaluates the coefficients by Least-Squares after each round.Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 42.
    22Pursuit Algorithms?Why shouldthey workThere are various algorithms designed for approximating the solution of this problem: Greedy Algorithms: Matching Pursuit, Orthogonal Matching Pursuit (OMP), Least-Squares-OMP, Weak Matching Pursuit, Block Matching Pursuit [1993-today].
  • 43.
    Relaxation Algorithms: BasisPursuit (a.k.a. LASSO), Dnatzig Selector & numerical ways to handle them [1995-today].
  • 44.
    Hybrid Algorithms: StOMP,CoSaMP, Subspace Pursuit, Iterative Hard-Thresholding [2007-today].
  • 45.
    …Sparse and RedundantRepresentation Modeling of Signals – Theory and Applications By: Michael Elad
  • 46.
    23D=DTDTDThe Mutual Coherenceis a property of the dictionary (just like the “Spark”). In fact, the following relation can be shown: The Mutual CoherenceComputeAssume normalized columnsThe Mutual Coherence M is the largest off-diagonal entry in absolute value.Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 47.
    24 BP and MP Equivalence (No Noise)EquivalenceGiven a signal x with a representation ,assuming that , BP and MP are guaranteed to find the sparsest solution. Donoho & E. (‘02) Gribonval & Nielsen (‘03)Tropp (‘03) Temlyakov (‘03)MP and BP are different in general (hard to say which is better).
  • 48.
    The above resultcorresponds to the worst-case, and as such, it is too pessimistic.
  • 49.
    Average performance resultsare available too, showing much better bounds [Donoho (`04)] [Candes et.al. (‘04)] [Tanner et.al. (‘05)] [E. (‘06)] [Tropp et.al. (‘06)] … [Candes et. al. (‘09)]. Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 50.
    25 BP Stability for the Noisy Case Given a signal with a representationsatisfying and a white Gaussian noise , BP will show stability, i.e., Stability*Ben-Haim, Eldar & E. (‘09)* With very high probability (as K goes to ∞)For =0 we get a weaker version of the previous result.
  • 51.
    This result isthe oracle’s error, multuiplied by C·logK.
  • 52.
    Similar results existfor other pursuit algorithms (Dantzig Selector, Orthogonal Matching Pursuit, CoSaMP, Subspace Pursuit, …)Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 53.
    Image Denoising &Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad26To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsProblems?What do we do? We have seen that there are approximation methods to find the sparsest solution, and there are theoretical results that guarantee their success.The Dictionary D should be found somehow !!!What next?
  • 54.
    27Part IIIDictionary Learning: The K-SVD AlgorithmSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 55.
    28D should bechosen such that it sparsifies the representationsThe approach we will take for building D is training it, based on Learning from Image ExamplesOne approach to choose D is from a known set of transforms (Steerable wavelet, Curvelet, Contourlets, Bandlets, Shearlets…) What Should D Be? Our Assumption: Good-behaved Images have a sparse representationSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 56.
    29DXAEach example hasa sparse representation with no more than L atomsEach example is a linear combination of atoms from DMeasure of Quality for D[Field & Olshausen (‘96)][Engan et. al. (‘99)][Lewicki & Sejnowski (‘00)][Cotter et. al. (‘03)][Gribonval et. al. (‘04)][Aharon, E. & Bruckstein (‘04)] [Aharon, E. & Bruckstein (‘05)]Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 57.
    30DInitialize DSparse CodingNearest NeighborXTDictionary UpdateColumn-by-Column by Mean computation over the relevant examples K–Means For Clustering Clustering: An extreme sparse representation Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 58.
    31DInitialize DSparse CodingUse Matching PursuitXTDictionary UpdateColumn-by-Column by SVD computation over the relevant examples The K–SVD Algorithm – General [Aharon, E. & Bruckstein (‘04,‘05)]Sparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 59.
    32DD is known! For the jth item we solve XT K–SVD: Sparse Coding StageSolved by A Pursuit AlgorithmSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 60.
    33Fixing all Aand D apart from the kth column, and seek both dk and the kth column in A to better fit the residual!We should solve: K–SVD: Dictionary Update StageWe refer only to the examples that use the column dkDSVDSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 61.
    Image Denoising &Beyond Via Learned Dictionaries and Sparse representationsBy: Michael Elad34To Summarize So Far …Image denoising (and many other problems in image processing) requires a model for the desired imageWe proposed a model for signals/images based on sparse and redundant representationsProblems?What do we do? We have seen approximation methods that find the sparsest solution, and theoretical results that guarantee their success. We also saw a way to learn DWill it all work in applications? What next?
  • 62.
    35Part IVBack toDenoising … and Beyond – Combining it AllSparse and Redundant Representation Modeling of Signals – Theory and Applications By: Michael Elad
  • 63.
    36kDNExtracts a patchin the ij locationOur prior From Local to Global TreatmentThe K-SVD algorithm is reasonable for low-dimension signals (N in the range 10-400). As N grows, the complexity and the memory requirements of the K-SVD become prohibitive.
  • 64.
    So, how shouldlarge images be handled?