SlideShare a Scribd company logo
1 of 57
Download to read offline
Fundamental Limits of Recovering Tree Sparse Vectors
from Noisy Linear Measurements
a`(1)
a`(2)
a`(5)
a`(3)
a`(4) a`(6) a`(7)EE-8500 Seminar
Akshay Soni
University of Minnesota
sonix022@umn.edu
(joint work with J. Haupt)
aupt
Minnesota
Computer Engineering
essive Imaging
al Learned Dictionaries
Supported by
Data Everywhere
Integral to Science, Engineering, Discovery
Inevitable Data Deluge!
The Economist, February 2010
Novel Sensing Architectures
20 %
sample
40 %
sample
original
Single	
  Pixel	
  Images	
  -­‐-­‐	
  h0p://dsp.rice.edu/cscamera	
  
	
  
9me	
  	
  
Key Idea – Sparsity
frequency	
  
Many signals exhibit sparsity in
the canonical or ‘pixel basis’
Communication signals often
have sparse frequency content
Natural images often have sparse
wavelet representationDWT
DFT
-- Background --
Sparsity and Structured Sparsity
A Model for Sparse Signals
Union of Subspace Model
signal support
numbe
signal c
arse Signal Model
signal support set
number of nonzero
signal components
signal support set
number of no
signal compo
Signals of interest are vectors x 2 Rn
Structured Sparsity
Tree Sparsity in Wavelets Grid Sparsity in Networks Graph Sparsity – background
subtraction
9
(a) Wavelet Tree Sparsity (b) Background Subtracted Image: Graph Sparsity
Figure 1.3: Structured sparsity. (a) The brain image has tree sparsity after wavelet transfor-
mation; (b) The background subtracted image has graph sparsity.
From above introductions, we know that there exists literature on structured sparsity, with
empirical evidence showing that one can achieve better performance by imposing additional
structures. However, none of the previous work was able to establish a general theoretical
framework for structured sparsity that can quantify its effectiveness. The goal of this thesis
is to develop such a general theory that addresses the following issues, where we pay special
attention to the benefit of structured sparsity over the standard non-structured sparsity:
• Quantifying structured sparsity;
• The minimal number of measurements required in compressive sensing;
• locations of nonzeros are inter-dependent
• structure knowledge can be used during sensing, inference or both
Structured Sparsity
Our focus – Tree Structured Sparsity!
Tree Sparsity in Wavelets Grid Sparsity in Networks Graph Sparsity – background
subtraction
9
(a) Wavelet Tree Sparsity (b) Background Subtracted Image: Graph Sparsity
Figure 1.3: Structured sparsity. (a) The brain image has tree sparsity after wavelet transfor-
mation; (b) The background subtracted image has graph sparsity.
From above introductions, we know that there exists literature on structured sparsity, with
empirical evidence showing that one can achieve better performance by imposing additional
structures. However, none of the previous work was able to establish a general theoretical
framework for structured sparsity that can quantify its effectiveness. The goal of this thesis
is to develop such a general theory that addresses the following issues, where we pay special
attention to the benefit of structured sparsity over the standard non-structured sparsity:
• Quantifying structured sparsity;
• The minimal number of measurements required in compressive sensing;
• locations of nonzeros are inter-dependent
• structure knowledge can be used during sensing, inference or both
Tree Structured Sparsity
1
52
3 4 6 7
Characteristics of tree structure
1 2 3 4 5 6 7
Tree Structured Sparsity – Why?
Wavelets!
•  Tree sparsity naturally arises in the wavelet
coefficients of many signals
•  for e.g. natural images
•  Several prior efforts that examined wavelet tree
structure specialized sensing techniques
•  for e.g. in dynamic MRI [*] and compressive
imaging [**]
•  Previous work was either experimental or
analyzed only in noise-free settings
[*]	
  L.	
  P.	
  Panych	
  and	
  F.	
  A.	
  Jolesz,	
  “A	
  dynamically	
  adap9ve	
  imaging	
  algorithm	
  for	
  wavelet-­‐encoded	
  MRI,”	
  Magne9c	
  Resonance	
  in	
  Medicine,	
  vol.	
  32,	
  
no.	
  6,	
  pp.	
  738–748,	
  1994.	
  
	
  
[**]	
  M.	
  W.	
  Seeger	
  and	
  H.	
  Nickisch,	
  “Compressed	
  sensing	
  and	
  Bayesian	
  experimental	
  design,”	
  in	
  Proc.	
  ICML,	
  2008,	
  pp.	
  912–919.	
  
[**]	
  S.	
  Deutsch,	
  A.	
  Averbuch,	
  and	
  S.	
  Dekel,	
  “Adap9ve	
  compressed	
  image	
  sensing	
  based	
  on	
  wavelet	
  modeling	
  and	
  direct	
  sampling,”	
  in	
  Proc.	
  Intl.	
  Conf	
  
on	
  Sampling	
  Theory	
  and	
  Applica9ons,	
  2009.	
  
-- Sensing Sparse Signals --
Noisy Linear Measurement Model
Sensing Strategies
Sensing Strategies
Non-Adaptive Sensing Adaptive Sensing
• j-th measurement vector aj is a function of {al, yl}j 1
l=1
for each j = 2, 3, . . . , m.
Measurement
vectors
y
y1
y2
yj
ym
Exact Support Recovery (ESR)
1 2 3 4 5 6 7
so that |xi| µ > 0, i 2 S,
Task of Interest:
Primary questions:
Exact Support Recovery (ESR)
1 2 3 4 5 6 7
so that |xi| µ > 0, i 2 S,
Task of Interest:
-- Adaptive Sensing of Tree-Sparse Signals --
A Simple Algorithm with Guarantees
Few Tree Specifics
•  Signal components are coefficients in an
orthonormal representation (canonical basis
without loss of generality)
•  We consider binary trees (all results may be
extended to trees with any degree)
1
52
3 4 6 7
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
(can also measure each location r 1 times
and average to reduce e↵ective noise)
Theorem (2011 & 2013): AS & J. Haupt
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
Choose any 2 (0, 1) and set ⌧ =
p
2 2 log(4k/ ). If the signal x being acquired
by our procedure is k-tree sparse, and the nonzero components of x satisfy
|xi|
s
24

1 + log
✓
4
◆ s
2
✓
k
m
◆
log k,
for every i 2 S(x), then with probability at least 1 , a “repeated measure-
ment” variant of algorithm to the left that acquires r measurements at each
observed location terminates after collecting m  r(2k + 1) measurements, and
produces support estimate ˆS satisfying ˆS = S(x)
Question:
Can any other “smart” scheme recover support of a tree-
sparse signal having “significantly” smaller magnitude?
i.e., is this the best one can hope for?
Theorem (2011 & 2013): AS & J. Haupt
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
Choose any 2 (0, 1) and set ⌧ =
p
2 2 log(4k/ ). If the signal x being acquired
by our procedure is k-tree sparse, and the nonzero components of x satisfy
|xi|
s
24

1 + log
✓
4
◆ s
2
✓
k
m
◆
log k,
for every i 2 S(x), then with probability at least 1 , a “repeated measure-
ment” variant of algorithm to the left that acquires r measurements at each
observed location terminates after collecting m  r(2k + 1) measurements, and
produces support estimate ˆS satisfying ˆS = S(x)
-- Our Investigation in Context --
Fundamental Limits for ESR
The Big Picture: Minimum Signal Amplitudes for ESR
Let’s identify necessary conditions for ESR in
each case…
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
[*]	
  S.	
  Aeron,	
  V.	
  Saligrama,	
  and	
  M.	
  Zhao,	
  "Informa9on	
  Theore9c	
  Bounds	
  for	
  Compressed	
  Sensing,"	
  IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  
vol.56,	
  no.10,	
  pp.5111-­‐5130,	
  2010	
  
[*]	
  M.	
  J.	
  Wainwright,	
  ”Sharp	
  thresholds	
  for	
  high-­‐dimensional	
  and	
  noisy	
  sparsity	
  recovery	
  using	
  l1-­‐constrained	
  quadra9c	
  programming	
  (lasso),	
  "	
  
IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  vol.55,	
  no.5,	
  pp.2183-­‐2202,	
  2009	
  	
  
[*]	
  M.	
  J.	
  Wainwright,	
  ”Informa9on-­‐theore9c	
  limita9ons	
  on	
  sparsity	
  recovery	
  in	
  the	
  high-­‐dimensional	
  and	
  noisy	
  sehng,	
  "	
  IEEE	
  Transac9ons	
  on	
  
Informa9on	
  Theory,	
  vol.55,	
  no.12,	
  	
  2009	
  	
  
[*]	
   W.	
   Wang,	
   M.	
   J.	
   Wainwright	
   and	
   K.	
   Ramchandran,	
   ”Informa9on-­‐theore9c	
   limits	
   on	
   sparse	
   signal	
   recovery:	
   Dense	
   versus	
   sparse	
  
measurement	
  matrices,	
  "	
  IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  vol.56,	
  no.6,	
  pp.2967-­‐2979,	
  2010	
  	
  
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
[*]	
  S.	
  Aeron,	
  V.	
  Saligrama,	
  and	
  M.	
  Zhao,	
  "Informa9on	
  Theore9c	
  Bounds	
  for	
  Compressed	
  Sensing,"	
  IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  
vol.56,	
  no.10,	
  pp.5111-­‐5130,	
  2010	
  
[*]	
  M.	
  J.	
  Wainwright,	
  ”Sharp	
  thresholds	
  for	
  high-­‐dimensional	
  and	
  noisy	
  sparsity	
  recovery	
  using	
  l1-­‐constrained	
  quadra9c	
  programming	
  (lasso),	
  "	
  
IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  vol.55,	
  no.5,	
  pp.2183-­‐2202,	
  2009	
  	
  
[*]	
  M.	
  J.	
  Wainwright,	
  ”Informa9on-­‐theore9c	
  limita9ons	
  on	
  sparsity	
  recovery	
  in	
  the	
  high-­‐dimensional	
  and	
  noisy	
  sehng,	
  "	
  IEEE	
  Transac9ons	
  on	
  
Informa9on	
  Theory,	
  vol.55,	
  no.12,	
  	
  2009	
  	
  
[*]	
   W.	
   Wang,	
   M.	
   J.	
   Wainwright	
   and	
   K.	
   Ramchandran,	
   ”Informa9on-­‐theore9c	
   limits	
   on	
   sparse	
   signal	
   recovery:	
   Dense	
   versus	
   sparse	
  
measurement	
  matrices,	
  "	
  IEEE	
  Transac9ons	
  on	
  Informa9on	
  Theory,	
  vol.56,	
  no.6,	
  pp.2967-­‐2979,	
  2010	
  	
  
uncompressed or
compressed
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
[*]	
   M.	
   Malloy	
   and	
   R.	
   Nowak,	
   “Sequen9al	
   analysis	
   in	
   high-­‐dimensional	
   mul9ple	
   tes9ng	
   and	
   sparse	
   recovery,”	
   in	
   Proc.	
   IEEE	
   Intl.	
   Symp.	
   on	
  
Informa9on	
  Theory,	
  2011,	
  pp.	
  2661-­‐2665.	
  
Adaptivity may at best improve log(n) to log(k)!
-- Problem Formulation --
Tree-Sparse Model
Signal Model:
Sensing Strategies:
Observations:
1
52
3 4 6 7
{Am, ym} : short hand for {aj, yj}m
j=1
Notations:
Adaptive : aj depends on {al, yl}j 1
l=1 , subject to constraint kajk2
2 = 1 8 j
Support estimate:
amplitude
parameter (>=0) Set of all k-node
rooted sub-trees
(in underlying tree)
Non Adaptive : here Gaussian; row aj of A is independent and aj ⇠ N(0, I/n)
Mm : class of all adaptive (or non-adaptive) sensing strategies based on m measurements
a mapping from observations ! subset of {1, 2, . . . , n}
(Maximum) Risk of a support estimator:
Element whose support
is most difficult to estimate
Minimax Risk:
Our aim – quantify errors corresponding to these
hard cases!
Preliminaries:
for estimators and sensing strategies M 2 M
In words, error of the best estimator when estimating the support of the “most di cult”
If R⇤
Xµ,k,M > 0 =) regardless of and M 2 M, we have at least one signal x 2 Xµ,k for
Note
In words, worst-case performance of when estimating the “most di cult”
-- Non-Adaptive Tree-Structured Sensing --
Fundamental Limits
Theorem (2013): AS & J. Haupt
Non-Adaptive Tree-Structured Sensing – fundamental limits
Implications: no uniform guarantees can be made for any estimation
procedure for recovering the support of tree-sparse signals when
signal amplitude is “too small”.
For ESR with non-adaptive sensing a necessary condition is:
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
[*]	
  AS	
  and	
  J.	
  Haupt,	
  “On	
  the	
  Fundamental	
  Limits	
  of	
  Recovering	
  Tree	
  Sparse	
  Vectors	
  from	
  Noisy	
  Linear	
  Measurement,”	
  IEEE	
  Transac9ons	
  on	
  
Informa9on	
  Theory,	
  	
  	
  2013	
  (accepted	
  for	
  publica9on).	
  
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Same necessary conditions as for adaptive + unstructured!
Structure or Adaptivity in isolation may at best improve
log(n) to log(k)
[*]	
  AS	
  and	
  J.	
  Haupt,	
  “On	
  the	
  Fundamental	
  Limits	
  of	
  Recovering	
  Tree	
  Sparse	
  Vectors	
  from	
  Noisy	
  Linear	
  Measurement,”	
  IEEE	
  Transac9ons	
  on	
  
Informa9on	
  Theory,	
  	
  	
  2013	
  (accepted	
  for	
  publica9on).	
  
Proof Idea – Non-Adaptive + Tree-Sparse
Restrict to a “Smaller Set”:
Convert to a Multiple-Hypothesis testing problem:
We can get a lower bound on minimax risk over a smaller subset of signals!
minimax prob. of error for
multiple hypothesis testing problem
Introduc9on	
  to	
  Nonparametric	
  Es9ma9on	
  –	
  A.B.	
  Tsybokov	
  
sup
x2Xµ,k
Prx ( (Am, ym; M) 6= S(x)) sup
x2X0
µ,k
Prx ( (Am, ym; M) 6= S(x))
For any X0
µ,k ✓ Xµ,k,
=)
• get lower bound on pe,L using Fano’s inequality (or similar ideas)
-- Adaptive Tree-Structured Sensing --
Fundamental Limits
Theorem (2013): AS & J. Haupt
Adaptive Tree-Structured Sensing – fundamental limits
For ESR with non-adaptive sensing a necessary condition is:
Proof Idea: this problem is as hard as recovering the location of one
nonzero given all other k-1 nonzero locations.
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
[*]	
  AS	
  and	
  J.	
  Haupt,	
  “On	
  the	
  Fundamental	
  Limits	
  of	
  Recovering	
  Tree	
  Sparse	
  Vectors	
  from	
  Noisy	
  Linear	
  Measurement,”	
  IEEE	
  Transac9ons	
  on	
  
Informa9on	
  Theory,	
  	
  2013	
  (accepted	
  for	
  publica9on).	
  
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Recall, for our simple tree-structured adaptive algorithm the
sufficient condition for ESR was
which is only log(k) factor away from the lower bound.
We cannot do much better than the simple proposed algorithm!
µ
q
2 k
m log k,
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
(when m > n)
Note: for adaptive + unstructured, our proof
ideas can show in case of m < n, a
necessary condition for ESR is
µ
q
2 n k+1
m
The Big Picture:
The Big Picture:
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Related Works:
[*]	
  A.	
  Krishnamurthy,	
  J.	
  Sharpnack,	
  and	
  A.	
  Singh,	
  “Recovering	
  block-­‐structured	
  ac9va9ons	
  using	
  compressive	
  measurements,”	
  Submi0ed	
  2012.	
  	
  
Question:
Can any other “smart” scheme recover support of a tree-
sparse signal having “significantly” smaller magnitude?
Theorem (2011 & 2013): AS & J. Haupt
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
Choose any 2 (0, 1) and set ⌧ =
p
2 2 log(4k/ ). If the signal x being acquired
by our procedure is k-tree sparse, and the nonzero components of x satisfy
|xi|
s
24

1 + log
✓
4
◆ s
2
✓
k
m
◆
log k,
for every i 2 S(x), then with probability at least 1 , a “repeated measure-
ment” variant of algorithm to the left that acquires r measurements at each
observed location terminates after collecting m  r(2k + 1) measurements, and
produces support estimate ˆS satisfying ˆS = S(x)
Answer:
No! We’re within log(k) of minimax optimal
Question:
Can any other “smart” scheme recover support of a tree-
sparse signal having “significantly” smaller magnitude?
Theorem (2011 & 2013): AS & J. Haupt
Tree Structured Adaptive Support Recovery
1	
  
5	
  2	
  
3	
   4	
   6	
   7	
  
Q[1] = {5}
y5 = eT
5 x + w
ˆS ˆS [ {5}
Q {6, 7} [ Q{5}
Q[1] = {6} ˆS = {1, 5}
y6 = eT
6 x + w
suppose |y5| > ⌧
suppose |y6| < ⌧
Q Q{6}
Q[1] = {7}
ˆS = {1, 5}
y7 = eT
7 x + w
suppose |y7| < ⌧
Q Q{7}
Q[1] = {;}
ˆS = {1, 5}
Choose any 2 (0, 1) and set ⌧ =
p
2 2 log(4k/ ). If the signal x being acquired
by our procedure is k-tree sparse, and the nonzero components of x satisfy
|xi|
s
24

1 + log
✓
4
◆ s
2
✓
k
m
◆
log k,
for every i 2 S(x), then with probability at least 1 , a “repeated measure-
ment” variant of algorithm to the left that acquires r measurements at each
observed location terminates after collecting m  r(2k + 1) measurements, and
produces support estimate ˆS satisfying ˆS = S(x)
-- Experimental Evaluation --
Simulation Setup
Non-adaptive + unstructured:
Non-adaptive + tree sparsity:
Adaptive + unstructured:
Adaptive + tree sparsity:
10
−1
10
0
10
1
10
2
0
0.2
0.4
0.6
0.8
1
Amplitude parameter µ
Prob.Error
n=28
−1
10
−1
10
0
10
1
10
2
0
0.2
0.4
0.6
0.8
1
Amplitude parameter µ
Prob.Error
n=2
10
−1
10
−1
10
0
10
1
10
2
0
0.2
0.4
0.6
0.8
1
Amplitude parameter µProb.Error
n=212
−1
4 orders of magnitude
[*]	
   M.	
   Malloy	
   and	
   R.	
   Nowak,	
   “Near-­‐op9mal	
   adap9ve	
   compressive	
   sensing,”	
   in	
   Proc.	
   Asilomar	
   Conf.	
   on	
   Signals,	
  
Systems,	
  and	
  Computers,	
  2012.	
  
-- Next Step --
1) MSE estimation implications?
MSE estimation implications
Unstructured + Non-Adaptive:
If the measurement matrix Am satisfies the norm constraint kAmk2
F  m, then
we have minimax MSE bound
Unstructured + Adaptive:
infbx,M2Mna
supx:|S(x)|=k E
⇥
kbx(Am, ym; M) xk2
2
⇤
c 2 n
m k log n,
[	
   *	
   ]	
   E.	
   J.	
   Cand`es	
   and	
   M.	
   A.	
   Davenport,	
   “How	
   well	
   can	
   we	
   es9mate	
   a	
   sparse	
   vector?,”	
   Applied	
   and	
  
Computa9onal	
  Harmonic	
  Analysis,	
  vol.	
  34,	
  no.	
  2,	
  pp.	
  317–323,	
  2013	
  
	
  
[	
  **	
  ]	
  E.	
  Arias-­‐Castro,	
  E.	
  J.	
  Candes,	
  and	
  M.	
  Davenport,	
  “On	
  the	
  fundamental	
  limits	
  of	
  adap9ve	
  sensing,”	
  
Submi0ed,	
  2011,	
  online	
  at	
  arxiv.org/abs/1111.4646.	
  
c > 0 is a constant. [ * ]
c0
> 0 is another constant. [ ** ]
MSE estimation implications
Unstructured + Non-Adaptive:
If the measurement matrix Am satisfies the norm constraint kAmk2
F  m, then
we have minimax MSE bound
Unstructured + Adaptive:
c > 0 is a constant.
infbx,M2Mna
supx:|S(x)|=k E
⇥
kbx(Am, ym; M) xk2
2
⇤
c 2 n
m k log n,
c0
> 0 is another constant.
Tree Structured + Non-Adaptive:
Tree Structured + Adaptive:
MSE estimation implications
Unstructured + Non-Adaptive:
If the measurement matrix Am satisfies the norm constraint kAmk2
F  m, then
we have minimax MSE bound
Unstructured + Adaptive:
c > 0 is a constant.
infbx,M2Mna
supx:|S(x)|=k E
⇥
kbx(Am, ym; M) xk2
2
⇤
c 2 n
m k log n,
c0
> 0 is another constant.
Tree-sparse + our adaptive procedure:
There exists a two-stage (support recovery followed by direct measurements)
adaptive compressive sensing procedure for k-tree sparse signals that produces,
from O(k) measurements, an estimate ˆx satisfying
kˆx xk2
2 = O
✓
2
✓
k
m
◆
k
◆
,
with high probability, provided the nonzero signal component amplitudes exceed
a constant times
q
2 k
m log k.
-- Next Step --
2) Learning Adaptive Sensing Representations (LASeR)
LASeR
Use Dictionary Learning and training data to learn tree-sparse representationsLearning Adaptive Sensing
!"#$%$%&'
(#)#'
*)"+,)+"-.'
*/#"0$)1'
2.#/34-'
*-%0$%&'
52*-6'
52*-67'5-#"%$%&'2.#/34-'*-%0$%&'6-/"-0-%)#38%0'
Example images (128 ⇥ 128)
52*-6'
52*-67'5-#"%$%&'2.#/34-'*-%0$%&'6-/"-0-%)#38%0'
(PICS) http://pics.psych.stir.ac.uk/
Example images (128 ⇥ 128)
Learn representation for 163 images from
Psychological Image Collection at Stirling
Wavelet Tree
Sensing
PCA
CS LASSO
CS Tree LASSO
LASeR
m = 50 m = 80m = 20
R = 128⇥128
32
“Sensing Energy”
Qualitative Results
Details	
   &	
   examples	
   of	
   LASeR	
   in	
   ac9on:	
   AS	
   and	
   J.	
   Haupt,	
   “Efficient	
   adap9ve	
   compressive	
   sensing	
   using	
   sparse	
   hierarchical	
   learned	
  
dic9onaries,”	
  in	
  Proc.	
  Asilomar	
  Conf.	
  on	
  Signals,	
  Systems	
  and	
  Computers,	
  2011,	
  pp.	
  1250-­‐1254.	
  
Tree Elements
Present in Sparse
Representation
Original Image
Overall Taxonomy
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Sufficient condition for ESR for our algorithm:
=)nearly optimal!!	
  
µ
q
2 k
m log k
Overall Taxonomy
Non-Adaptive	
   Adaptive	
  
Non-Adaptive	
   Adaptive	
  
Unstructured	
  
Unstructured	
  
TreeSparse	
  
TreeSparse	
  
Thank You!
Akshay Soni
University of Minnesota
sonix022@umn.edu
Sufficient condition for ESR for our algorithm:
=)nearly optimal!!	
  
µ
q
2 k
m log k

More Related Content

What's hot

Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...
Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...
Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...SimranKetha
 
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network DynamicsJAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamicshirokazutanaka
 
. An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic .... An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic ...butest
 
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENT
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENTAPPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENT
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENTsipij
 
Background Estimation Using Principal Component Analysis Based on Limited Mem...
Background Estimation Using Principal Component Analysis Based on Limited Mem...Background Estimation Using Principal Component Analysis Based on Limited Mem...
Background Estimation Using Principal Component Analysis Based on Limited Mem...IJECEIAES
 
Data-Driven Motion Estimation With Spatial Adaptation
Data-Driven Motion Estimation With Spatial AdaptationData-Driven Motion Estimation With Spatial Adaptation
Data-Driven Motion Estimation With Spatial AdaptationCSCJournals
 
Pre-computation for ABC in image analysis
Pre-computation for ABC in image analysisPre-computation for ABC in image analysis
Pre-computation for ABC in image analysisMatt Moores
 
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-Mesh
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-MeshImproved Parallel Algorithm for Time Series Based Forecasting Using OTIS-Mesh
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-MeshIDES Editor
 
PROMISE 2011: "Handling missing data in software effort prediction with naive...
PROMISE 2011: "Handling missing data in software effort prediction with naive...PROMISE 2011: "Handling missing data in software effort prediction with naive...
PROMISE 2011: "Handling missing data in software effort prediction with naive...CS, NcState
 
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rulesJAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning ruleshirokazutanaka
 
Principal Component Analysis For Novelty Detection
Principal Component Analysis For Novelty DetectionPrincipal Component Analysis For Novelty Detection
Principal Component Analysis For Novelty DetectionJordan McBain
 
20191018 reservoir computing
20191018 reservoir computing20191018 reservoir computing
20191018 reservoir computingCHIA-HSIANG KAO
 
Fuzzy entropy based optimal
Fuzzy entropy based optimalFuzzy entropy based optimal
Fuzzy entropy based optimalijsc
 
Machine learning in science and industry — day 4
Machine learning in science and industry — day 4Machine learning in science and industry — day 4
Machine learning in science and industry — day 4arogozhnikov
 
A comprehensive study of different image super resolution reconstruction algo...
A comprehensive study of different image super resolution reconstruction algo...A comprehensive study of different image super resolution reconstruction algo...
A comprehensive study of different image super resolution reconstruction algo...IAEME Publication
 

What's hot (19)

Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...
Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...
Deep Learning based Frameworks for Handling Imbalance in DGA, Email, and URL ...
 
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network DynamicsJAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
 
. An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic .... An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic ...
 
Ijciet 10 01_153-2
Ijciet 10 01_153-2Ijciet 10 01_153-2
Ijciet 10 01_153-2
 
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENT
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENTAPPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENT
APPLICATION OF CONVOLUTIONAL NEURAL NETWORK IN LAWN MEASUREMENT
 
Background Estimation Using Principal Component Analysis Based on Limited Mem...
Background Estimation Using Principal Component Analysis Based on Limited Mem...Background Estimation Using Principal Component Analysis Based on Limited Mem...
Background Estimation Using Principal Component Analysis Based on Limited Mem...
 
Dx25751756
Dx25751756Dx25751756
Dx25751756
 
Data-Driven Motion Estimation With Spatial Adaptation
Data-Driven Motion Estimation With Spatial AdaptationData-Driven Motion Estimation With Spatial Adaptation
Data-Driven Motion Estimation With Spatial Adaptation
 
50120140504015
5012014050401550120140504015
50120140504015
 
Pre-computation for ABC in image analysis
Pre-computation for ABC in image analysisPre-computation for ABC in image analysis
Pre-computation for ABC in image analysis
 
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-Mesh
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-MeshImproved Parallel Algorithm for Time Series Based Forecasting Using OTIS-Mesh
Improved Parallel Algorithm for Time Series Based Forecasting Using OTIS-Mesh
 
PROMISE 2011: "Handling missing data in software effort prediction with naive...
PROMISE 2011: "Handling missing data in software effort prediction with naive...PROMISE 2011: "Handling missing data in software effort prediction with naive...
PROMISE 2011: "Handling missing data in software effort prediction with naive...
 
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rulesJAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
 
Principal Component Analysis For Novelty Detection
Principal Component Analysis For Novelty DetectionPrincipal Component Analysis For Novelty Detection
Principal Component Analysis For Novelty Detection
 
20191018 reservoir computing
20191018 reservoir computing20191018 reservoir computing
20191018 reservoir computing
 
CLIM: Transition Workshop - Modeling Weather-related House Insurance Claims w...
CLIM: Transition Workshop - Modeling Weather-related House Insurance Claims w...CLIM: Transition Workshop - Modeling Weather-related House Insurance Claims w...
CLIM: Transition Workshop - Modeling Weather-related House Insurance Claims w...
 
Fuzzy entropy based optimal
Fuzzy entropy based optimalFuzzy entropy based optimal
Fuzzy entropy based optimal
 
Machine learning in science and industry — day 4
Machine learning in science and industry — day 4Machine learning in science and industry — day 4
Machine learning in science and industry — day 4
 
A comprehensive study of different image super resolution reconstruction algo...
A comprehensive study of different image super resolution reconstruction algo...A comprehensive study of different image super resolution reconstruction algo...
A comprehensive study of different image super resolution reconstruction algo...
 

Viewers also liked

Robust Face Recognition under Varying Illumination and Occlusion Considering ...
Robust Face Recognition under Varying Illumination and Occlusion Considering ...Robust Face Recognition under Varying Illumination and Occlusion Considering ...
Robust Face Recognition under Varying Illumination and Occlusion Considering ...singing_wei
 
Advanced lock in amplifier for detection of phase transitions in liquid crystals
Advanced lock in amplifier for detection of phase transitions in liquid crystalsAdvanced lock in amplifier for detection of phase transitions in liquid crystals
Advanced lock in amplifier for detection of phase transitions in liquid crystalsIAEME Publication
 
Time Series Analysis:Basic Stochastic Signal Recovery
Time Series Analysis:Basic Stochastic Signal RecoveryTime Series Analysis:Basic Stochastic Signal Recovery
Time Series Analysis:Basic Stochastic Signal RecoveryDaniel Cuneo
 
Algorithms for Sparse Signal Recovery in Compressed Sensing
Algorithms for Sparse Signal Recovery in Compressed SensingAlgorithms for Sparse Signal Recovery in Compressed Sensing
Algorithms for Sparse Signal Recovery in Compressed SensingAqib Ejaz
 
Brandt - Superconductors and Vortices at Radio Frequency Magnetic Fields
Brandt - Superconductors and Vortices at Radio Frequency Magnetic FieldsBrandt - Superconductors and Vortices at Radio Frequency Magnetic Fields
Brandt - Superconductors and Vortices at Radio Frequency Magnetic Fieldsthinfilmsworkshop
 
Signals and noise
Signals and noiseSignals and noise
Signals and noiseRavi Kant
 
Superconductivity of materials
Superconductivity of materialsSuperconductivity of materials
Superconductivity of materialsSmit Parikh
 
Signals and noise
Signals and noiseSignals and noise
Signals and noiseRavi Kant
 
Superconductivity
SuperconductivitySuperconductivity
Superconductivityilmyong
 
Introduction to compressive sensing
Introduction to compressive sensingIntroduction to compressive sensing
Introduction to compressive sensingAhmed Nasser Agag
 
Superconductors presentation
Superconductors presentationSuperconductors presentation
Superconductors presentationIslam Mohamed
 
Ppt compressed sensing a tutorial
Ppt compressed sensing a tutorialPpt compressed sensing a tutorial
Ppt compressed sensing a tutorialTerence Gao
 

Viewers also liked (20)

Robust Face Recognition under Varying Illumination and Occlusion Considering ...
Robust Face Recognition under Varying Illumination and Occlusion Considering ...Robust Face Recognition under Varying Illumination and Occlusion Considering ...
Robust Face Recognition under Varying Illumination and Occlusion Considering ...
 
Advanced lock in amplifier for detection of phase transitions in liquid crystals
Advanced lock in amplifier for detection of phase transitions in liquid crystalsAdvanced lock in amplifier for detection of phase transitions in liquid crystals
Advanced lock in amplifier for detection of phase transitions in liquid crystals
 
Super Conductivity
Super ConductivitySuper Conductivity
Super Conductivity
 
Time Series Analysis:Basic Stochastic Signal Recovery
Time Series Analysis:Basic Stochastic Signal RecoveryTime Series Analysis:Basic Stochastic Signal Recovery
Time Series Analysis:Basic Stochastic Signal Recovery
 
Algorithms for Sparse Signal Recovery in Compressed Sensing
Algorithms for Sparse Signal Recovery in Compressed SensingAlgorithms for Sparse Signal Recovery in Compressed Sensing
Algorithms for Sparse Signal Recovery in Compressed Sensing
 
Brandt - Superconductors and Vortices at Radio Frequency Magnetic Fields
Brandt - Superconductors and Vortices at Radio Frequency Magnetic FieldsBrandt - Superconductors and Vortices at Radio Frequency Magnetic Fields
Brandt - Superconductors and Vortices at Radio Frequency Magnetic Fields
 
Signals and noise
Signals and noiseSignals and noise
Signals and noise
 
Superconductivity of materials
Superconductivity of materialsSuperconductivity of materials
Superconductivity of materials
 
Signals and noise
Signals and noiseSignals and noise
Signals and noise
 
Superconductivity
SuperconductivitySuperconductivity
Superconductivity
 
Superconductivity
SuperconductivitySuperconductivity
Superconductivity
 
Introduction to compressive sensing
Introduction to compressive sensingIntroduction to compressive sensing
Introduction to compressive sensing
 
Lec17 sparse signal processing & applications
Lec17 sparse signal processing & applicationsLec17 sparse signal processing & applications
Lec17 sparse signal processing & applications
 
Superconductors presentation
Superconductors presentationSuperconductors presentation
Superconductors presentation
 
Ppt compressed sensing a tutorial
Ppt compressed sensing a tutorialPpt compressed sensing a tutorial
Ppt compressed sensing a tutorial
 
Sparse representation and compressive sensing
Sparse representation and compressive sensingSparse representation and compressive sensing
Sparse representation and compressive sensing
 
Line coding
Line coding Line coding
Line coding
 
Noise pollution
Noise pollutionNoise pollution
Noise pollution
 
Superconductors
SuperconductorsSuperconductors
Superconductors
 
Superconductivity
SuperconductivitySuperconductivity
Superconductivity
 

Similar to Fundamental Limits of Recovering Tree Sparse Vectors from Noisy Linear Measurements

block-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfblock-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfJunghyun Lee
 
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...Anirbit Mukherjee
 
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.Anirbit Mukherjee
 
A Compressed Sensing Approach to Image Reconstruction
A Compressed Sensing Approach to Image ReconstructionA Compressed Sensing Approach to Image Reconstruction
A Compressed Sensing Approach to Image Reconstructionijsrd.com
 
An efficient approach to wavelet image Denoising
An efficient approach to wavelet image DenoisingAn efficient approach to wavelet image Denoising
An efficient approach to wavelet image Denoisingijcsit
 
International Journal of Computer Science and Security Volume (2) Issue (5)
International Journal of Computer Science and Security Volume (2) Issue (5)International Journal of Computer Science and Security Volume (2) Issue (5)
International Journal of Computer Science and Security Volume (2) Issue (5)CSCJournals
 
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...EL-Hachemi Guerrout
 
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...Chris Rackauckas
 
Manish Kurse PhD research slides
Manish Kurse PhD research slidesManish Kurse PhD research slides
Manish Kurse PhD research slidesmanishkurse
 
20141003.journal club
20141003.journal club20141003.journal club
20141003.journal clubHayaru SHOUNO
 
T. Lucas Makinen x Imperial SBI Workshop
T. Lucas Makinen x Imperial SBI WorkshopT. Lucas Makinen x Imperial SBI Workshop
T. Lucas Makinen x Imperial SBI WorkshopLucasMakinen1
 
Exploring temporal graph data with Python: 
a study on tensor decomposition o...
Exploring temporal graph data with Python: 
a study on tensor decomposition o...Exploring temporal graph data with Python: 
a study on tensor decomposition o...
Exploring temporal graph data with Python: 
a study on tensor decomposition o...André Panisson
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chaptersChristian Robert
 
Image Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkImage Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkCSCJournals
 

Similar to Fundamental Limits of Recovering Tree Sparse Vectors from Noisy Linear Measurements (20)

block-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfblock-mdp-masters-defense.pdf
block-mdp-masters-defense.pdf
 
Sparse codes for natural images
Sparse codes for natural imagesSparse codes for natural images
Sparse codes for natural images
 
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...
My invited talk at the 2018 Annual Meeting of SIAM (Society of Industrial and...
 
Lausanne 2019 #2
Lausanne 2019 #2Lausanne 2019 #2
Lausanne 2019 #2
 
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
 
A Compressed Sensing Approach to Image Reconstruction
A Compressed Sensing Approach to Image ReconstructionA Compressed Sensing Approach to Image Reconstruction
A Compressed Sensing Approach to Image Reconstruction
 
Estimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample SetsEstimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample Sets
 
An efficient approach to wavelet image Denoising
An efficient approach to wavelet image DenoisingAn efficient approach to wavelet image Denoising
An efficient approach to wavelet image Denoising
 
International Journal of Computer Science and Security Volume (2) Issue (5)
International Journal of Computer Science and Security Volume (2) Issue (5)International Journal of Computer Science and Security Volume (2) Issue (5)
International Journal of Computer Science and Security Volume (2) Issue (5)
 
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
 
Data mining
Data mining Data mining
Data mining
 
Statistical Physics Studies of Machine Learning Problems by Lenka Zdeborova, ...
Statistical Physics Studies of Machine Learning Problems by Lenka Zdeborova, ...Statistical Physics Studies of Machine Learning Problems by Lenka Zdeborova, ...
Statistical Physics Studies of Machine Learning Problems by Lenka Zdeborova, ...
 
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...
 
Manish Kurse PhD research slides
Manish Kurse PhD research slidesManish Kurse PhD research slides
Manish Kurse PhD research slides
 
20141003.journal club
20141003.journal club20141003.journal club
20141003.journal club
 
T. Lucas Makinen x Imperial SBI Workshop
T. Lucas Makinen x Imperial SBI WorkshopT. Lucas Makinen x Imperial SBI Workshop
T. Lucas Makinen x Imperial SBI Workshop
 
Exploring temporal graph data with Python: 
a study on tensor decomposition o...
Exploring temporal graph data with Python: 
a study on tensor decomposition o...Exploring temporal graph data with Python: 
a study on tensor decomposition o...
Exploring temporal graph data with Python: 
a study on tensor decomposition o...
 
kape_science
kape_sciencekape_science
kape_science
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chapters
 
Image Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic FrameworkImage Denoising Based On Sparse Representation In A Probabilistic Framework
Image Denoising Based On Sparse Representation In A Probabilistic Framework
 

Recently uploaded

CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionDr.Costas Sachpazis
 
Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.eptoze12
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)Dr SOUNDIRARAJ N
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidNikhilNagaraju
 
computer application and construction management
computer application and construction managementcomputer application and construction management
computer application and construction managementMariconPadriquez1
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxPoojaBan
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile servicerehmti665
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...srsj9000
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptSAURABHKUMAR892774
 
Introduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHIntroduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHC Sai Kiran
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx959SahilShah
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort servicejennyeacort
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfme23b1001
 

Recently uploaded (20)

CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
 
POWER SYSTEMS-1 Complete notes examples
POWER SYSTEMS-1 Complete notes  examplesPOWER SYSTEMS-1 Complete notes  examples
POWER SYSTEMS-1 Complete notes examples
 
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCRCall Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
 
Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.Oxy acetylene welding presentation note.
Oxy acetylene welding presentation note.
 
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Serviceyoung call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
 
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
UNIT III ANALOG ELECTRONICS (BASIC ELECTRONICS)
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfid
 
computer application and construction management
computer application and construction managementcomputer application and construction management
computer application and construction management
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptx
 
Call Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile serviceCall Girls Delhi {Jodhpur} 9711199012 high profile service
Call Girls Delhi {Jodhpur} 9711199012 high profile service
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
Arduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.pptArduino_CSE ece ppt for working and principal of arduino.ppt
Arduino_CSE ece ppt for working and principal of arduino.ppt
 
Introduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECHIntroduction to Machine Learning Unit-3 for II MECH
Introduction to Machine Learning Unit-3 for II MECH
 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdf
 

Fundamental Limits of Recovering Tree Sparse Vectors from Noisy Linear Measurements

  • 1. Fundamental Limits of Recovering Tree Sparse Vectors from Noisy Linear Measurements a`(1) a`(2) a`(5) a`(3) a`(4) a`(6) a`(7)EE-8500 Seminar Akshay Soni University of Minnesota sonix022@umn.edu (joint work with J. Haupt) aupt Minnesota Computer Engineering essive Imaging al Learned Dictionaries Supported by
  • 3. Integral to Science, Engineering, Discovery
  • 4. Inevitable Data Deluge! The Economist, February 2010
  • 5. Novel Sensing Architectures 20 % sample 40 % sample original Single  Pixel  Images  -­‐-­‐  h0p://dsp.rice.edu/cscamera    
  • 6. 9me     Key Idea – Sparsity frequency   Many signals exhibit sparsity in the canonical or ‘pixel basis’ Communication signals often have sparse frequency content Natural images often have sparse wavelet representationDWT DFT
  • 7. -- Background -- Sparsity and Structured Sparsity
  • 8. A Model for Sparse Signals Union of Subspace Model signal support numbe signal c arse Signal Model signal support set number of nonzero signal components signal support set number of no signal compo Signals of interest are vectors x 2 Rn
  • 9. Structured Sparsity Tree Sparsity in Wavelets Grid Sparsity in Networks Graph Sparsity – background subtraction 9 (a) Wavelet Tree Sparsity (b) Background Subtracted Image: Graph Sparsity Figure 1.3: Structured sparsity. (a) The brain image has tree sparsity after wavelet transfor- mation; (b) The background subtracted image has graph sparsity. From above introductions, we know that there exists literature on structured sparsity, with empirical evidence showing that one can achieve better performance by imposing additional structures. However, none of the previous work was able to establish a general theoretical framework for structured sparsity that can quantify its effectiveness. The goal of this thesis is to develop such a general theory that addresses the following issues, where we pay special attention to the benefit of structured sparsity over the standard non-structured sparsity: • Quantifying structured sparsity; • The minimal number of measurements required in compressive sensing; • locations of nonzeros are inter-dependent • structure knowledge can be used during sensing, inference or both
  • 10. Structured Sparsity Our focus – Tree Structured Sparsity! Tree Sparsity in Wavelets Grid Sparsity in Networks Graph Sparsity – background subtraction 9 (a) Wavelet Tree Sparsity (b) Background Subtracted Image: Graph Sparsity Figure 1.3: Structured sparsity. (a) The brain image has tree sparsity after wavelet transfor- mation; (b) The background subtracted image has graph sparsity. From above introductions, we know that there exists literature on structured sparsity, with empirical evidence showing that one can achieve better performance by imposing additional structures. However, none of the previous work was able to establish a general theoretical framework for structured sparsity that can quantify its effectiveness. The goal of this thesis is to develop such a general theory that addresses the following issues, where we pay special attention to the benefit of structured sparsity over the standard non-structured sparsity: • Quantifying structured sparsity; • The minimal number of measurements required in compressive sensing; • locations of nonzeros are inter-dependent • structure knowledge can be used during sensing, inference or both
  • 11. Tree Structured Sparsity 1 52 3 4 6 7 Characteristics of tree structure 1 2 3 4 5 6 7
  • 12. Tree Structured Sparsity – Why? Wavelets! •  Tree sparsity naturally arises in the wavelet coefficients of many signals •  for e.g. natural images •  Several prior efforts that examined wavelet tree structure specialized sensing techniques •  for e.g. in dynamic MRI [*] and compressive imaging [**] •  Previous work was either experimental or analyzed only in noise-free settings [*]  L.  P.  Panych  and  F.  A.  Jolesz,  “A  dynamically  adap9ve  imaging  algorithm  for  wavelet-­‐encoded  MRI,”  Magne9c  Resonance  in  Medicine,  vol.  32,   no.  6,  pp.  738–748,  1994.     [**]  M.  W.  Seeger  and  H.  Nickisch,  “Compressed  sensing  and  Bayesian  experimental  design,”  in  Proc.  ICML,  2008,  pp.  912–919.   [**]  S.  Deutsch,  A.  Averbuch,  and  S.  Dekel,  “Adap9ve  compressed  image  sensing  based  on  wavelet  modeling  and  direct  sampling,”  in  Proc.  Intl.  Conf   on  Sampling  Theory  and  Applica9ons,  2009.  
  • 13. -- Sensing Sparse Signals -- Noisy Linear Measurement Model
  • 15. Sensing Strategies Non-Adaptive Sensing Adaptive Sensing • j-th measurement vector aj is a function of {al, yl}j 1 l=1 for each j = 2, 3, . . . , m. Measurement vectors y y1 y2 yj ym
  • 16. Exact Support Recovery (ESR) 1 2 3 4 5 6 7 so that |xi| µ > 0, i 2 S, Task of Interest:
  • 17. Primary questions: Exact Support Recovery (ESR) 1 2 3 4 5 6 7 so that |xi| µ > 0, i 2 S, Task of Interest:
  • 18. -- Adaptive Sensing of Tree-Sparse Signals -- A Simple Algorithm with Guarantees
  • 19. Few Tree Specifics •  Signal components are coefficients in an orthonormal representation (canonical basis without loss of generality) •  We consider binary trees (all results may be extended to trees with any degree) 1 52 3 4 6 7
  • 20. Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7  
  • 21. Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7  
  • 22. Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7  
  • 23. Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5}
  • 24. Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5} (can also measure each location r 1 times and average to reduce e↵ective noise)
  • 25. Theorem (2011 & 2013): AS & J. Haupt Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5} Choose any 2 (0, 1) and set ⌧ = p 2 2 log(4k/ ). If the signal x being acquired by our procedure is k-tree sparse, and the nonzero components of x satisfy |xi| s 24  1 + log ✓ 4 ◆ s 2 ✓ k m ◆ log k, for every i 2 S(x), then with probability at least 1 , a “repeated measure- ment” variant of algorithm to the left that acquires r measurements at each observed location terminates after collecting m  r(2k + 1) measurements, and produces support estimate ˆS satisfying ˆS = S(x)
  • 26. Question: Can any other “smart” scheme recover support of a tree- sparse signal having “significantly” smaller magnitude? i.e., is this the best one can hope for? Theorem (2011 & 2013): AS & J. Haupt Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5} Choose any 2 (0, 1) and set ⌧ = p 2 2 log(4k/ ). If the signal x being acquired by our procedure is k-tree sparse, and the nonzero components of x satisfy |xi| s 24  1 + log ✓ 4 ◆ s 2 ✓ k m ◆ log k, for every i 2 S(x), then with probability at least 1 , a “repeated measure- ment” variant of algorithm to the left that acquires r measurements at each observed location terminates after collecting m  r(2k + 1) measurements, and produces support estimate ˆS satisfying ˆS = S(x)
  • 27. -- Our Investigation in Context -- Fundamental Limits for ESR
  • 28. The Big Picture: Minimum Signal Amplitudes for ESR Let’s identify necessary conditions for ESR in each case… Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse  
  • 29. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   [*]  S.  Aeron,  V.  Saligrama,  and  M.  Zhao,  "Informa9on  Theore9c  Bounds  for  Compressed  Sensing,"  IEEE  Transac9ons  on  Informa9on  Theory,   vol.56,  no.10,  pp.5111-­‐5130,  2010   [*]  M.  J.  Wainwright,  ”Sharp  thresholds  for  high-­‐dimensional  and  noisy  sparsity  recovery  using  l1-­‐constrained  quadra9c  programming  (lasso),  "   IEEE  Transac9ons  on  Informa9on  Theory,  vol.55,  no.5,  pp.2183-­‐2202,  2009     [*]  M.  J.  Wainwright,  ”Informa9on-­‐theore9c  limita9ons  on  sparsity  recovery  in  the  high-­‐dimensional  and  noisy  sehng,  "  IEEE  Transac9ons  on   Informa9on  Theory,  vol.55,  no.12,    2009     [*]   W.   Wang,   M.   J.   Wainwright   and   K.   Ramchandran,   ”Informa9on-­‐theore9c   limits   on   sparse   signal   recovery:   Dense   versus   sparse   measurement  matrices,  "  IEEE  Transac9ons  on  Informa9on  Theory,  vol.56,  no.6,  pp.2967-­‐2979,  2010    
  • 30. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   [*]  S.  Aeron,  V.  Saligrama,  and  M.  Zhao,  "Informa9on  Theore9c  Bounds  for  Compressed  Sensing,"  IEEE  Transac9ons  on  Informa9on  Theory,   vol.56,  no.10,  pp.5111-­‐5130,  2010   [*]  M.  J.  Wainwright,  ”Sharp  thresholds  for  high-­‐dimensional  and  noisy  sparsity  recovery  using  l1-­‐constrained  quadra9c  programming  (lasso),  "   IEEE  Transac9ons  on  Informa9on  Theory,  vol.55,  no.5,  pp.2183-­‐2202,  2009     [*]  M.  J.  Wainwright,  ”Informa9on-­‐theore9c  limita9ons  on  sparsity  recovery  in  the  high-­‐dimensional  and  noisy  sehng,  "  IEEE  Transac9ons  on   Informa9on  Theory,  vol.55,  no.12,    2009     [*]   W.   Wang,   M.   J.   Wainwright   and   K.   Ramchandran,   ”Informa9on-­‐theore9c   limits   on   sparse   signal   recovery:   Dense   versus   sparse   measurement  matrices,  "  IEEE  Transac9ons  on  Informa9on  Theory,  vol.56,  no.6,  pp.2967-­‐2979,  2010     uncompressed or compressed
  • 31. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   [*]   M.   Malloy   and   R.   Nowak,   “Sequen9al   analysis   in   high-­‐dimensional   mul9ple   tes9ng   and   sparse   recovery,”   in   Proc.   IEEE   Intl.   Symp.   on   Informa9on  Theory,  2011,  pp.  2661-­‐2665.   Adaptivity may at best improve log(n) to log(k)!
  • 32. -- Problem Formulation -- Tree-Sparse Model
  • 33. Signal Model: Sensing Strategies: Observations: 1 52 3 4 6 7 {Am, ym} : short hand for {aj, yj}m j=1 Notations: Adaptive : aj depends on {al, yl}j 1 l=1 , subject to constraint kajk2 2 = 1 8 j Support estimate: amplitude parameter (>=0) Set of all k-node rooted sub-trees (in underlying tree) Non Adaptive : here Gaussian; row aj of A is independent and aj ⇠ N(0, I/n) Mm : class of all adaptive (or non-adaptive) sensing strategies based on m measurements a mapping from observations ! subset of {1, 2, . . . , n}
  • 34. (Maximum) Risk of a support estimator: Element whose support is most difficult to estimate Minimax Risk: Our aim – quantify errors corresponding to these hard cases! Preliminaries: for estimators and sensing strategies M 2 M In words, error of the best estimator when estimating the support of the “most di cult” If R⇤ Xµ,k,M > 0 =) regardless of and M 2 M, we have at least one signal x 2 Xµ,k for Note In words, worst-case performance of when estimating the “most di cult”
  • 35. -- Non-Adaptive Tree-Structured Sensing -- Fundamental Limits
  • 36. Theorem (2013): AS & J. Haupt Non-Adaptive Tree-Structured Sensing – fundamental limits Implications: no uniform guarantees can be made for any estimation procedure for recovering the support of tree-sparse signals when signal amplitude is “too small”. For ESR with non-adaptive sensing a necessary condition is:
  • 37. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   [*]  AS  and  J.  Haupt,  “On  the  Fundamental  Limits  of  Recovering  Tree  Sparse  Vectors  from  Noisy  Linear  Measurement,”  IEEE  Transac9ons  on   Informa9on  Theory,      2013  (accepted  for  publica9on).  
  • 38. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Same necessary conditions as for adaptive + unstructured! Structure or Adaptivity in isolation may at best improve log(n) to log(k) [*]  AS  and  J.  Haupt,  “On  the  Fundamental  Limits  of  Recovering  Tree  Sparse  Vectors  from  Noisy  Linear  Measurement,”  IEEE  Transac9ons  on   Informa9on  Theory,      2013  (accepted  for  publica9on).  
  • 39. Proof Idea – Non-Adaptive + Tree-Sparse Restrict to a “Smaller Set”: Convert to a Multiple-Hypothesis testing problem: We can get a lower bound on minimax risk over a smaller subset of signals! minimax prob. of error for multiple hypothesis testing problem Introduc9on  to  Nonparametric  Es9ma9on  –  A.B.  Tsybokov   sup x2Xµ,k Prx ( (Am, ym; M) 6= S(x)) sup x2X0 µ,k Prx ( (Am, ym; M) 6= S(x)) For any X0 µ,k ✓ Xµ,k, =) • get lower bound on pe,L using Fano’s inequality (or similar ideas)
  • 40. -- Adaptive Tree-Structured Sensing -- Fundamental Limits
  • 41. Theorem (2013): AS & J. Haupt Adaptive Tree-Structured Sensing – fundamental limits For ESR with non-adaptive sensing a necessary condition is: Proof Idea: this problem is as hard as recovering the location of one nonzero given all other k-1 nonzero locations.
  • 42. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   [*]  AS  and  J.  Haupt,  “On  the  Fundamental  Limits  of  Recovering  Tree  Sparse  Vectors  from  Noisy  Linear  Measurement,”  IEEE  Transac9ons  on   Informa9on  Theory,    2013  (accepted  for  publica9on).  
  • 43. Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Recall, for our simple tree-structured adaptive algorithm the sufficient condition for ESR was which is only log(k) factor away from the lower bound. We cannot do much better than the simple proposed algorithm! µ q 2 k m log k, The Big Picture:
  • 44. Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   (when m > n) Note: for adaptive + unstructured, our proof ideas can show in case of m < n, a necessary condition for ESR is µ q 2 n k+1 m The Big Picture:
  • 45. The Big Picture: Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Related Works: [*]  A.  Krishnamurthy,  J.  Sharpnack,  and  A.  Singh,  “Recovering  block-­‐structured  ac9va9ons  using  compressive  measurements,”  Submi0ed  2012.    
  • 46. Question: Can any other “smart” scheme recover support of a tree- sparse signal having “significantly” smaller magnitude? Theorem (2011 & 2013): AS & J. Haupt Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5} Choose any 2 (0, 1) and set ⌧ = p 2 2 log(4k/ ). If the signal x being acquired by our procedure is k-tree sparse, and the nonzero components of x satisfy |xi| s 24  1 + log ✓ 4 ◆ s 2 ✓ k m ◆ log k, for every i 2 S(x), then with probability at least 1 , a “repeated measure- ment” variant of algorithm to the left that acquires r measurements at each observed location terminates after collecting m  r(2k + 1) measurements, and produces support estimate ˆS satisfying ˆS = S(x)
  • 47. Answer: No! We’re within log(k) of minimax optimal Question: Can any other “smart” scheme recover support of a tree- sparse signal having “significantly” smaller magnitude? Theorem (2011 & 2013): AS & J. Haupt Tree Structured Adaptive Support Recovery 1   5  2   3   4   6   7   Q[1] = {5} y5 = eT 5 x + w ˆS ˆS [ {5} Q {6, 7} [ Q{5} Q[1] = {6} ˆS = {1, 5} y6 = eT 6 x + w suppose |y5| > ⌧ suppose |y6| < ⌧ Q Q{6} Q[1] = {7} ˆS = {1, 5} y7 = eT 7 x + w suppose |y7| < ⌧ Q Q{7} Q[1] = {;} ˆS = {1, 5} Choose any 2 (0, 1) and set ⌧ = p 2 2 log(4k/ ). If the signal x being acquired by our procedure is k-tree sparse, and the nonzero components of x satisfy |xi| s 24  1 + log ✓ 4 ◆ s 2 ✓ k m ◆ log k, for every i 2 S(x), then with probability at least 1 , a “repeated measure- ment” variant of algorithm to the left that acquires r measurements at each observed location terminates after collecting m  r(2k + 1) measurements, and produces support estimate ˆS satisfying ˆS = S(x)
  • 49. Simulation Setup Non-adaptive + unstructured: Non-adaptive + tree sparsity: Adaptive + unstructured: Adaptive + tree sparsity: 10 −1 10 0 10 1 10 2 0 0.2 0.4 0.6 0.8 1 Amplitude parameter µ Prob.Error n=28 −1 10 −1 10 0 10 1 10 2 0 0.2 0.4 0.6 0.8 1 Amplitude parameter µ Prob.Error n=2 10 −1 10 −1 10 0 10 1 10 2 0 0.2 0.4 0.6 0.8 1 Amplitude parameter µProb.Error n=212 −1 4 orders of magnitude [*]   M.   Malloy   and   R.   Nowak,   “Near-­‐op9mal   adap9ve   compressive   sensing,”   in   Proc.   Asilomar   Conf.   on   Signals,   Systems,  and  Computers,  2012.  
  • 50. -- Next Step -- 1) MSE estimation implications?
  • 51. MSE estimation implications Unstructured + Non-Adaptive: If the measurement matrix Am satisfies the norm constraint kAmk2 F  m, then we have minimax MSE bound Unstructured + Adaptive: infbx,M2Mna supx:|S(x)|=k E ⇥ kbx(Am, ym; M) xk2 2 ⇤ c 2 n m k log n, [   *   ]   E.   J.   Cand`es   and   M.   A.   Davenport,   “How   well   can   we   es9mate   a   sparse   vector?,”   Applied   and   Computa9onal  Harmonic  Analysis,  vol.  34,  no.  2,  pp.  317–323,  2013     [  **  ]  E.  Arias-­‐Castro,  E.  J.  Candes,  and  M.  Davenport,  “On  the  fundamental  limits  of  adap9ve  sensing,”   Submi0ed,  2011,  online  at  arxiv.org/abs/1111.4646.   c > 0 is a constant. [ * ] c0 > 0 is another constant. [ ** ]
  • 52. MSE estimation implications Unstructured + Non-Adaptive: If the measurement matrix Am satisfies the norm constraint kAmk2 F  m, then we have minimax MSE bound Unstructured + Adaptive: c > 0 is a constant. infbx,M2Mna supx:|S(x)|=k E ⇥ kbx(Am, ym; M) xk2 2 ⇤ c 2 n m k log n, c0 > 0 is another constant. Tree Structured + Non-Adaptive: Tree Structured + Adaptive:
  • 53. MSE estimation implications Unstructured + Non-Adaptive: If the measurement matrix Am satisfies the norm constraint kAmk2 F  m, then we have minimax MSE bound Unstructured + Adaptive: c > 0 is a constant. infbx,M2Mna supx:|S(x)|=k E ⇥ kbx(Am, ym; M) xk2 2 ⇤ c 2 n m k log n, c0 > 0 is another constant. Tree-sparse + our adaptive procedure: There exists a two-stage (support recovery followed by direct measurements) adaptive compressive sensing procedure for k-tree sparse signals that produces, from O(k) measurements, an estimate ˆx satisfying kˆx xk2 2 = O ✓ 2 ✓ k m ◆ k ◆ , with high probability, provided the nonzero signal component amplitudes exceed a constant times q 2 k m log k.
  • 54. -- Next Step -- 2) Learning Adaptive Sensing Representations (LASeR)
  • 55. LASeR Use Dictionary Learning and training data to learn tree-sparse representationsLearning Adaptive Sensing !"#$%$%&' (#)#' *)"+,)+"-.' */#"0$)1' 2.#/34-' *-%0$%&' 52*-6' 52*-67'5-#"%$%&'2.#/34-'*-%0$%&'6-/"-0-%)#38%0' Example images (128 ⇥ 128) 52*-6' 52*-67'5-#"%$%&'2.#/34-'*-%0$%&'6-/"-0-%)#38%0' (PICS) http://pics.psych.stir.ac.uk/ Example images (128 ⇥ 128) Learn representation for 163 images from Psychological Image Collection at Stirling Wavelet Tree Sensing PCA CS LASSO CS Tree LASSO LASeR m = 50 m = 80m = 20 R = 128⇥128 32 “Sensing Energy” Qualitative Results Details   &   examples   of   LASeR   in   ac9on:   AS   and   J.   Haupt,   “Efficient   adap9ve   compressive   sensing   using   sparse   hierarchical   learned   dic9onaries,”  in  Proc.  Asilomar  Conf.  on  Signals,  Systems  and  Computers,  2011,  pp.  1250-­‐1254.   Tree Elements Present in Sparse Representation Original Image
  • 56. Overall Taxonomy Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Sufficient condition for ESR for our algorithm: =)nearly optimal!!   µ q 2 k m log k
  • 57. Overall Taxonomy Non-Adaptive   Adaptive   Non-Adaptive   Adaptive   Unstructured   Unstructured   TreeSparse   TreeSparse   Thank You! Akshay Soni University of Minnesota sonix022@umn.edu Sufficient condition for ESR for our algorithm: =)nearly optimal!!   µ q 2 k m log k