Probabilistic Control of Switched Linear Systems with Chance ConstraintsLeo Asselborn
An approach to algorithmically synthesize control
strategies for set-to-set transitions of uncertain discrete-time
switched linear systems based on a combination of tree search
and reachable set computations in a stochastic setting is
proposed in this presentation. The initial state and disturbances
are assumed to be Gaussian distributed, and a time-variant
hybrid control law stabilizes the system towards a goal set.
The algorithmic solution computes sequences of discrete states
via tree search and the continuous controls are obtained
from solving embedded semi-definite programs (SDP). These
program taking polytopic input constraints as well as timevarying
probabilistic state constraints into account. An example
for demonstrating the principles of the solution procedure with
focus on handling the chance constraints is included.
Wave-packet Treatment of Neutrinos and Its Quantum-mechanical ImplicationsCheng-Hsien Li
This is a presentation about one of my PhD dissertation projects. By modeling neutrinos as Gaussian wave-packets, whose propagation satisfy the wave equation for quantum particle, wave-particle duality can be demonstrated. The issue of fermionic exchange symmetry of neutrinos from intense source such as core-collapse supernovae is further discussed.
Slides of a talk at CMU Theory lunch (http://www.cs.cmu.edu/~theorylunch/20111116.html) and Capital Area Theory seminar (http://www.cs.umd.edu/areas/Theory/CATS/#Grigory).
Probabilistic Control of Switched Linear Systems with Chance ConstraintsLeo Asselborn
An approach to algorithmically synthesize control
strategies for set-to-set transitions of uncertain discrete-time
switched linear systems based on a combination of tree search
and reachable set computations in a stochastic setting is
proposed in this presentation. The initial state and disturbances
are assumed to be Gaussian distributed, and a time-variant
hybrid control law stabilizes the system towards a goal set.
The algorithmic solution computes sequences of discrete states
via tree search and the continuous controls are obtained
from solving embedded semi-definite programs (SDP). These
program taking polytopic input constraints as well as timevarying
probabilistic state constraints into account. An example
for demonstrating the principles of the solution procedure with
focus on handling the chance constraints is included.
Wave-packet Treatment of Neutrinos and Its Quantum-mechanical ImplicationsCheng-Hsien Li
This is a presentation about one of my PhD dissertation projects. By modeling neutrinos as Gaussian wave-packets, whose propagation satisfy the wave equation for quantum particle, wave-particle duality can be demonstrated. The issue of fermionic exchange symmetry of neutrinos from intense source such as core-collapse supernovae is further discussed.
Slides of a talk at CMU Theory lunch (http://www.cs.cmu.edu/~theorylunch/20111116.html) and Capital Area Theory seminar (http://www.cs.umd.edu/areas/Theory/CATS/#Grigory).
Control of Discrete-Time Piecewise Affine Probabilistic Systems using Reachab...Leo Asselborn
This presentation proposes an algorithmic approach to
synthesize stabilizing control laws for discrete-time piecewise
affine probabilistic (PWAP) systems based on computations of
probabilistic reachable sets. The considered class of systems
contains probabilistic components (with Gaussian distribution)
modeling additive disturbances and state initialization. The
probabilistic reachable state sets contain all states that are
reachable with a given confidence level under the effect of
time-variant control laws. The control synthesis uses principles
of the ellipsoidal calculus, and it considers that the system
parametrization depends on the partition of the state space. The
proposed algorithm uses LMI-constrained semi-definite programming
(SDP) problems to compute stabilizing controllers,
while polytopic input constraints and transitions between regions
of the state space are considered. The formulation of
the SDP is adopted from a previous work in [1] for switched
systems, in which the switching of the continuous dynamics
is triggered by a discrete input variable. Here, as opposed
to [1], the switching occurs autonomously and an algorithmic
procedure is suggested to synthesis a stabilizing controller. An
example for illustration is included.
Time-Frequency Representation of Microseismic Signals using the SSTUT Technology
Resonance frequencies could provide useful information on the deformation occurring during fracturing experiments or CO2 management, complementary to the microseismic events distribution. An accurate time-frequency representation is of crucial importance to interpret the cause of resonance frequencies during microseismic experiments. The popular methods of Short-Time Fourier Transform (STFT) and wavelet analysis have limitations in representing close frequencies and dealing with fast varying instantaneous frequencies and this is often the nature of microseismic signals. The synchrosqueezing transform (SST) is a promising tool to track these resonant frequencies and provide a detailed time-frequency representation. Here we apply the synchrosqueezing transform to microseismic signals and also show its potential to general seismic signal processing applications.
SOCG: Linear-Size Approximations to the Vietoris-Rips FiltrationDon Sheehy
The Vietoris-Rips filtration is a versatile tool in topological data analysis.
Unfortunately, it is often too large to construct in full.
We show how to construct an $O(n)$-size filtered simplicial complex on an $n$-point metric space such that the persistence diagram is a good approximation to that of the Vietoris-Rips filtration.
The filtration can be constructed in $O(n\log n)$ time.
The constants depend only on the doubling dimension of the metric space and the desired tightness of the approximation.
For the first time, this makes it computationally tractable to approximate the persistence diagram of the Vietoris-Rips filtration across all scales for large data sets.
Our approach uses a hierarchical net-tree to sparsify the filtration.
We can either sparsify the data by throwing out points at larger scales to give a zigzag filtration,
or sparsify the underlying graph by throwing out edges at larger scales to give a standard filtration.
Both methods yield the same guarantees.
Transformations of continuous-variable entangled states of lightOndrej Cernotik
Gaussian states and Gaussian transformations represent an interesting counterpart to two-level photonic systems in the field of quantum information processing. On the theoretical side, Gaussian states are easily described using first and second moments of the quadrature operators; from the experimental point of view, Gaussian operations can be implemented using linear optics and optical parametric amplifiers. The biggest advantage compared to two-level photonic systems, is deterministic generation of entangled states in parametric amplifiers and highly efficient homodyne detection. In this presentation, we propose new protocols for manipulation of entanglement of Gaussian states.
Firstly, we study entanglement concentration of split single-mode squeezed vacuum states by photon subtraction enhanced by local coherent displacements. These states can be obtained by mixing a single-mode squeezed vacuum state with vacuum on a beam splitter and are, therefore, generated more easily than two-mode squeezed vacuum states. We show that performing local coherent displacements prior to photon subtraction can lead to an enhancement of the output entanglement. This is seen in weak-squeezing approximation where destructive quantum interference of dominant Fock states occurs, while for arbitrarily squeezed input states, we analyze a realistic scenario, including limited transmittance of tap-off beam splitters and limited efficiency of heralding detectors.
Next, motivated by results obtained for bipartite Gaussian states, we study symmetrization of multipartite Gaussian states by local Gaussian operations. Namely, we analyze strategies based on addition of correlated noise and on quantum non-demolition interaction. We use fidelity of assisted quantum teleportation as a figure of merit to characterize entanglement of the state before and after the symmetrization procedure. Analyzing the teleportation protocol and considering more general transformations of multipartite Gaussian states, we show that the fidelity can be improved significantly.
MVPA with SpaceNet: sparse structured priorsElvis DOHMATOB
The GraphNet (aka S-Lasso), as well as other “sparsity + structure” priors like TV (Total-Variation), TV-L1, etc., are not easily applicable to brain data because of technical problems
relating to the selection of the regularization parameters. Also, in
their own right, such models lead to challenging high-dimensional optimization problems. In this manuscript, we present some heuristics for speeding up the overall optimization process: (a) Early-stopping, whereby one halts the optimization process when the test score (performance on leftout data) for the internal cross-validation for model-selection stops improving, and (b) univariate feature-screening, whereby irrelevant (non-predictive) voxels are detected and eliminated before the optimization problem is entered, thus reducing the size of the problem. Empirical results with GraphNet on real MRI (Magnetic Resonance Imaging) datasets indicate that these heuristics are a win-win strategy, as they add speed without sacrificing the quality of the predictions. We expect the proposed heuristics to work on other models like TV-L1, etc.
Control of Discrete-Time Piecewise Affine Probabilistic Systems using Reachab...Leo Asselborn
This presentation proposes an algorithmic approach to
synthesize stabilizing control laws for discrete-time piecewise
affine probabilistic (PWAP) systems based on computations of
probabilistic reachable sets. The considered class of systems
contains probabilistic components (with Gaussian distribution)
modeling additive disturbances and state initialization. The
probabilistic reachable state sets contain all states that are
reachable with a given confidence level under the effect of
time-variant control laws. The control synthesis uses principles
of the ellipsoidal calculus, and it considers that the system
parametrization depends on the partition of the state space. The
proposed algorithm uses LMI-constrained semi-definite programming
(SDP) problems to compute stabilizing controllers,
while polytopic input constraints and transitions between regions
of the state space are considered. The formulation of
the SDP is adopted from a previous work in [1] for switched
systems, in which the switching of the continuous dynamics
is triggered by a discrete input variable. Here, as opposed
to [1], the switching occurs autonomously and an algorithmic
procedure is suggested to synthesis a stabilizing controller. An
example for illustration is included.
Time-Frequency Representation of Microseismic Signals using the SSTUT Technology
Resonance frequencies could provide useful information on the deformation occurring during fracturing experiments or CO2 management, complementary to the microseismic events distribution. An accurate time-frequency representation is of crucial importance to interpret the cause of resonance frequencies during microseismic experiments. The popular methods of Short-Time Fourier Transform (STFT) and wavelet analysis have limitations in representing close frequencies and dealing with fast varying instantaneous frequencies and this is often the nature of microseismic signals. The synchrosqueezing transform (SST) is a promising tool to track these resonant frequencies and provide a detailed time-frequency representation. Here we apply the synchrosqueezing transform to microseismic signals and also show its potential to general seismic signal processing applications.
SOCG: Linear-Size Approximations to the Vietoris-Rips FiltrationDon Sheehy
The Vietoris-Rips filtration is a versatile tool in topological data analysis.
Unfortunately, it is often too large to construct in full.
We show how to construct an $O(n)$-size filtered simplicial complex on an $n$-point metric space such that the persistence diagram is a good approximation to that of the Vietoris-Rips filtration.
The filtration can be constructed in $O(n\log n)$ time.
The constants depend only on the doubling dimension of the metric space and the desired tightness of the approximation.
For the first time, this makes it computationally tractable to approximate the persistence diagram of the Vietoris-Rips filtration across all scales for large data sets.
Our approach uses a hierarchical net-tree to sparsify the filtration.
We can either sparsify the data by throwing out points at larger scales to give a zigzag filtration,
or sparsify the underlying graph by throwing out edges at larger scales to give a standard filtration.
Both methods yield the same guarantees.
Transformations of continuous-variable entangled states of lightOndrej Cernotik
Gaussian states and Gaussian transformations represent an interesting counterpart to two-level photonic systems in the field of quantum information processing. On the theoretical side, Gaussian states are easily described using first and second moments of the quadrature operators; from the experimental point of view, Gaussian operations can be implemented using linear optics and optical parametric amplifiers. The biggest advantage compared to two-level photonic systems, is deterministic generation of entangled states in parametric amplifiers and highly efficient homodyne detection. In this presentation, we propose new protocols for manipulation of entanglement of Gaussian states.
Firstly, we study entanglement concentration of split single-mode squeezed vacuum states by photon subtraction enhanced by local coherent displacements. These states can be obtained by mixing a single-mode squeezed vacuum state with vacuum on a beam splitter and are, therefore, generated more easily than two-mode squeezed vacuum states. We show that performing local coherent displacements prior to photon subtraction can lead to an enhancement of the output entanglement. This is seen in weak-squeezing approximation where destructive quantum interference of dominant Fock states occurs, while for arbitrarily squeezed input states, we analyze a realistic scenario, including limited transmittance of tap-off beam splitters and limited efficiency of heralding detectors.
Next, motivated by results obtained for bipartite Gaussian states, we study symmetrization of multipartite Gaussian states by local Gaussian operations. Namely, we analyze strategies based on addition of correlated noise and on quantum non-demolition interaction. We use fidelity of assisted quantum teleportation as a figure of merit to characterize entanglement of the state before and after the symmetrization procedure. Analyzing the teleportation protocol and considering more general transformations of multipartite Gaussian states, we show that the fidelity can be improved significantly.
MVPA with SpaceNet: sparse structured priorsElvis DOHMATOB
The GraphNet (aka S-Lasso), as well as other “sparsity + structure” priors like TV (Total-Variation), TV-L1, etc., are not easily applicable to brain data because of technical problems
relating to the selection of the regularization parameters. Also, in
their own right, such models lead to challenging high-dimensional optimization problems. In this manuscript, we present some heuristics for speeding up the overall optimization process: (a) Early-stopping, whereby one halts the optimization process when the test score (performance on leftout data) for the internal cross-validation for model-selection stops improving, and (b) univariate feature-screening, whereby irrelevant (non-predictive) voxels are detected and eliminated before the optimization problem is entered, thus reducing the size of the problem. Empirical results with GraphNet on real MRI (Magnetic Resonance Imaging) datasets indicate that these heuristics are a win-win strategy, as they add speed without sacrificing the quality of the predictions. We expect the proposed heuristics to work on other models like TV-L1, etc.
PROGRAMMA ATTIVITA’ DIDATTICA A.A. 2016/17
DOTTORATO DI RICERCA IN INGEGNERIA STRUTTURALE E GEOTECNICA
____________________________________________________________
STOCHASTIC DYNAMICS AND MONTE CARLO SIMULATION IN EARTHQUAKE ENGINEERING APPLICATIONS
Lecture Series by
Agathoklis Giaralis, Ph.D., M.ASCE., P.E. City, University of London
Visiting Professor Sapienza University of Rome
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...Pioneer Natural Resources
This paper studies the application of bit allocation using JPEG2000 for compressing multi-dimensional remote sensing data. Past experiments have shown that the Karhunen- Lo
`
e
ve transform (KLT) along with rate distortion optimal(RDO) bit allocation produces good compression perfor-mance. However, this model has the unavoidable disadvan-tage of paying a price in terms of implementation complex-ity. In this research we address this complexity problem byusing the discrete wavelet transform (DWT) instead of theKLT as the decorrelator. Further, we have incorporated amixed model (MM) to find the rate distortion curves instead of the prior method of using experimental rate distortioncurves for RDO bit allocation. We compared our results tothe traditional high bit rate quantizer bit allocation modelbased on the logarithm of variances among the bands. Our comparisons show that by using the MM-RDO bit rate al-location method result in lower mean squared error (MSE)compared to the traditional bit allocation scheme. Our ap- proach also has an additional advantage of using DWT asa computationally efficient decorrelator when compared tothe KLT
International Conference on Monte Carlo techniques
Closing conference of thematic cycle
Paris July 5-8th 2016
Campus les cordeliers
Jere Koskela's slides
Stochastic reaction networks (SRNs) are a particular class of continuous-time Markov chains used to model a wide range of phenomena, including biological/chemical reactions, epidemics, risk theory, queuing, and supply chain/social/multi-agents networks. In this context, we explore the efficient estimation of statistical quantities, particularly rare event probabilities, and propose two alternative importance sampling (IS) approaches [1,2] to improve the Monte Carlo (MC) estimator efficiency. The key challenge in the IS framework is to choose an appropriate change of probability measure to achieve substantial variance reduction, which often requires insights into the underlying problem. Therefore, we propose an automated approach to obtain a highly efficient path-dependent measure change based on an original connection between finding optimal IS parameters and solving a variance minimization problem via a stochastic optimal control formulation. We pursue two alternative approaches to mitigate the curse of dimensionality when solving the resulting dynamic programming problem. In the first approach [1], we propose a learning-based method to approximate the value function using a neural network, where the parameters are determined via a stochastic optimization algorithm. As an alternative, we present in [2] a dimension reduction method, based on mapping the problem to a significantly lower dimensional space via the Markovian projection (MP) idea. The output of this model reduction technique is a low dimensional SRN (potentially one dimension) that preserves the marginal distribution of the original high-dimensional SRN system. The dynamics of the projected process are obtained via a discrete $L^2$ regression. By solving a resulting projected Hamilton-Jacobi-Bellman (HJB) equation for the reduced-dimensional SRN, we get projected IS parameters, which are then mapped back to the original full-dimensional SRN system, and result in an efficient IS-MC estimator of the full-dimensional SRN. Our analysis and numerical experiments verify that both proposed IS (learning based and MP-HJB-IS) approaches substantially reduce the MC estimator’s variance, resulting in a lower computational complexity in the rare event regime than standard MC estimators. [1] Ben Hammouda, C., Ben Rached, N., and Tempone, R., and Wiechert, S. Learning-based importance sampling via stochastic optimal control for stochastic reaction net-works. Statistics and Computing 33, no. 3 (2023): 58. [2] Ben Hammouda, C., Ben Rached, N., and Tempone, R., and Wiechert, S. (2023). Automated Importance Sampling via Optimal Control for Stochastic Reaction Networks: A Markovian Projection-based Approach. To appear soon.
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
ICML 2021 tutorial on random matrix theory and machine learning.
Part 3 covers: 1. Motivation: Average-case versus worst-case in high dimensions 2. Algorithm halting times (runtimes) 3. Outlook
Subgradient Methods for Huge-Scale Optimization Problems - Юрий Нестеров, Cat...Yandex
We consider a new class of huge-scale problems, the problems with sparse subgradients. The most important functions of this type are piecewise linear. For optimization problems with uniform sparsity of corresponding linear operators, we suggest a very efficient implementation of subgradient iterations, the total cost of which depends logarithmically in the dimension. This technique is based on a recursive update of the results of matrix/vector products and the values of symmetric functions. It works well, for example, for matrices with few nonzero diagonals and for max-type functions.
We show that the updating technique can be efficiently coupled with the simplest subgradient methods. Similar results can be obtained for a new non-smooth random variant of a coordinate descent scheme. We also present promising results of preliminary computational experiments.
Bayesian modelling and computation for Raman spectroscopyMatt Moores
Raman spectroscopy can be used to identify molecules by the characteristic scattering of light from a laser. Each Raman-active dye label has a unique spectral signature, comprised by the locations and amplitudes of the peaks. The Raman spectrum is discretised into a multivariate observation that is highly collinear, hence it lends itself to a reduced-rank representation. We introduce a sequential Monte Carlo (SMC) algorithm to separate this signal into a series of peaks plus a smoothly-varying baseline, corrupted by additive white noise. By incorporating this representation into a Bayesian functional regression, we can quantify the relationship between dye concentration and peak intensity. We also estimate the model evidence using SMC to investigate long-range dependence between peaks. These methods have been implemented as an R package, using RcppEigen and OpenMP.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
Salas, V. (2024) "John of St. Thomas (Poinsot) on the Science of Sacred Theol...Studia Poinsotiana
I Introduction
II Subalternation and Theology
III Theology and Dogmatic Declarations
IV The Mixed Principles of Theology
V Virtual Revelation: The Unity of Theology
VI Theology as a Natural Science
VII Theology’s Certitude
VIII Conclusion
Notes
Bibliography
All the contents are fully attributable to the author, Doctor Victor Salas. Should you wish to get this text republished, get in touch with the author or the editorial committee of the Studia Poinsotiana. Insofar as possible, we will be happy to broker your contact.
3. Sparse Vector Estimation
Gene Expressions
{
{X
n
p
k sparse
?
Samples
Genes
Few Active Genes
Disease Susceptibility =
Response
4. Low Rank Matrices
The Netflix Problem
Movie Ratings
{
{
thousands of
movies
millions of
subscribers
movie profile
user profile
Action
+ a handful of other features...
Decomposition:
5. Low Rank Matrices
Topic Models
Term Document
Matrix
{
{
thousands of
documents
millions of
words
degree of topicality
topic profile
Politics
+ a handful of other features...
Decomposition:
6. Simple objects
Objects Notion of Simplicity Atoms
Vectors Sparsity Canonical unit vectors
Matrices Rank Rank-1 matrices
Bandlimited signals No. of frequencies Complex sinusoids
Linear systems Number of poles Single pole systems
x =
X
a2A
caax =
X
a2A
caa
atomic set
atoms
“features”
nonnegative weights
7. Universal Regularizer
Chandrasekaran, Recht, Parillo,Wilsky (2010)
Objects Notion of Simplicity Atomic Norm
Vectors Sparsity
Matrices Rank nuclear norm
Bandlimited signals No. of frequencies ?
Linear systems McMillan degree ?
1
x A = {t 0 : x t (A)}
=
a
ca : x =
a A
caa, ca > 0
t
xx A = {t 0 : x t (A)}
8. Moment Problems
A
atoms
+ +
⇤ ⇥ ⌅
x
measure space
moments of a discrete measure
atomic moments
System Identification
−1
0
1
−1
0
1
0
2
4
6
−0.5 0 0.5
2
3
4
5
6
7
Line Spectral Estimation
9. Decomposition
How many atoms do you
need to represent x?
composing atoms are on an
exposed face: decomposition is
easy to find.
find a supporting hyperplane
to certify
concentrate on recovering
“good” objects with atoms on
exposed faces
t
x
10. Dual Certificate
1 2 3 4 5 6 7
0
1
2
3
4
5
6
X Axis
YAxis
Gradient
Subgradients
optimum value is atomic norm.
solutions are subgradients at x
semi-infinite problem
maximize
q
hq, x?
i
subject to kqk⇤
A 1
Dual Characterization of atomic norm
kqk⇤
A = sup
a2A
hq, ai
(under mild conditions, Bonsall)
hq, ai = 1, a 2 support
hq, ai < 1, otherwise.
x
t
11. Denoising
minimize
x
1
2
ky xk2
2 + ⌧kxkA
Tradeoff parameter
AST (Atomic Soft Thresholding)
Dual AST
maximize
q
hq, yi ⌧kqk2
2
subject to kqk⇤
A 1
y = x + w
Cn
N(0, 2
In)
12. AST Results
⌧ ˆq + ˆx = y
kˆqk⇤
A 1
hˆq, ˆxi = kˆxkA
Optimality Conditions
suggests
⌧ ˆq ⇡ w ) ⌧ ⇡ Ekwk⇤
A
means
ˆq 2 @kˆxkA
Can find composing atoms!
Tradeoff = Gaussian Width
No cross validation needed - estimate
Gaussian width
Gaussian width also important to
characterize convergence rates
13. AST Convergence Rates
Fast Rate
s sparse
1
n
X ˆ X 2
2 C 2 s log(p)
n
Slow RateLasso n
1
n
X ˆ X 2
2
log(p)
n
1
Choose appropriate tradeoff
= E ( w A) , 1
Universal, but slow rate
Minimal assumptions on atoms
1
n
ˆx x 2
2
n
x A
With a “cone condition”, faster rates are possible
Weaker than RIP,
Coherence, RSC, RE 1
n
ˆx x 2
2 C
2
1/2n
14. Summary
Notion of simple models
Atomic norm unifies treatment of many models
Recovers tight published bounds
Bounds depend on Gaussian width
Faster rates with a local cone condition
16. Super resolution
Time frequency dual
Extrapolate high frequency information from
low frequency samples
Just exchange time and frequency: same as line
spectral estimation
17. Classical Line Spectrum Estimation
(a crash course)
Tn(x) =
2
6
6
6
4
x1 x2 . . . xn
x⇤
2 x1 . . . xn 1
...
...
...
...
x⇤
n x⇤
n 1 . . . x1
3
7
7
7
5
Define for any vector x,
Nonlinear Parameter Estimation xk =
s
j=1
cje2 ifj k
Low rank and Toeplitz
FACT:
Tn(x ) =
s
j=1
cj
1
ei2 fj
...
ei2 (n 1)fj
1 e i2 fj
. . . e i2 (n 1)fj
18. Classical techniques
PRONY estimate s, root finding
CADZOW alternating projections
MUSIC low rank, plot pseudospectrum
Sensitive to model order
No rigorous theory
Only optimal asymptotically
Low rank and Toeplitz
FACT:
Tn(x ) =
s
j=1
cj
1
ei2 fj
...
ei2 (n 1)fj
1 e i2 fj
. . . e i2 (n 1)fj
19. Convex Perspective
0 2 4 6 8 10 12 14
−10
−5
0
5
10
15
x =
s
j=1
cja(fj)
amplitudes
frequencies
x0
...
xk
...
xn 1
=
s
j=1
cjei2 fj k
1
...
ei2 fj k
...
ei2 fj (n 1)
AtomsLine Spectrum
A+ = {a(f) | f [0, 1]}Positive Amplitudes:
A = a(f)ei
f [0, 1], [0, 2 ]Complex:
20. Convergence rates
With high probability, for n>4,
⌧ ⇡
p
n log(n)Choose tradeoff as
Dudley’s inequality
Bernstein’s polynomial theorem
For a signal with n samples
of the form,
Using the global AST guarantee, we get,
E
1
n
ˆx x 2
2 C
log(n)
n
s
j=1
|cj|
yk =
s
j=1
cje2 ifj k
+ wk
n log(n) n
2 log(4 log(n)) w A 1 +
1
log(n)
n log(n) + n log(16 3/2 log(n))
21. Fast Rates
E
✓
1
n
kˆx x?
k2
2
◆
C 2
s
log(n)
n
min
p6=q
d(up, uq)
4
n
If every two frequencies are far enough apart (minimum separation condition)
For s frequencies and n samples, MSE is
Not a global condition on the dictionary, local to the signal
Proof uses many properties of trigonometric polynomials and Fejer
kernels derived by Candes and Fernandez Granda.
⌧ = ⌘
p
n log(n), ⌘ 2 (1, 1) large enough.
and the tradeoff parameter is chosen correctly,
22. Can’t do much better
Also nearly matches minimax bound on best prediction rate for
signals with 4/n minimum separation.
E
✓
1
n
kˆx x?
k2
2
◆
C 2 s log(n/4s)
n
Only logarithmic factor from “oracle rate”
E
✓
1
n
kˆx x?
k2
2
◆
C 2
s
log(n)
n
Our rate:
E
1
n
ˆx x 2
2 C 2 s
n
23. Minimum Separation and Exact Recovery
Result of Candes and Fernandez Granda
4/n separation = can construct explicit dual
certificate of exact recovery.
Convex hull of far enough vertices are exposed
faces.
Prony does not have this limitation. (Then why
bother? Robustness guarantee)
Minimum separation is necessary
24. except for the positive case...
Convex hull of every collection of n/2
vertices is an exposed face
Infinite dimensional version of cyclic
polytope, which is neighbourly
No separation condition needed.
25. Localizing Frequencies
0 0.2 0.4 0.6 0.8 1
0
0.2
0.4
0.6
0.8
1
Frequency
DualPolynomialValue
0.34 0.36 0.38 0.4 0.42 0.44 0.46
0.85
0.9
0.95
1
Frequency
DualPolynomialValue
Dual Polynomial
y = ˆx + ⌧ ˆq
ˆQ(f) =
n 1
j=0
ˆqje i2 jf
hq, ai = 1, a 2 support
hq, ai < 1, otherwise.
x
t
26. Localization Guarantees
Spurious Amplitudes
Weighted Frequency Deviation
Near region approximation
10 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
1
0
Frequency
DualPolynomialMagnitude
True FrequencyEstimated Frequency
0.16/n
Spurious Frequency
l: ˆfl F
|ˆcl| C1
k2 (n)
n
cj
l: ˆfl Nj
ˆcl C3
k2 (n)
n
l: ˆfl Nj
|ˆcl|
fj T
d(fj, ˆfl)
2
C2
k2 (n)
n
27. How to actually solve AST
Semidefinite Characterizations
minimize
x
1
2
ky xk2
2 + ⌧kxkA
conv(A+) conv(A)
kxkA+
=
(
x0, Tn(x) ⌫ 0,
+1, otherwise.
Bounded Real Lemma
Caratheodory Toeplitz
A+ = {a(f) | f [0, 1]}Positive Amplitudes: A = a(f)ei
f [0, 1], [0, 2 ]Complex:
x A = inf
1
2n
tr(Tn(u)) +
1
2
t
Tn(u) x
x t
0
28. Alternating Direction Method of Multipliers
minimizet,u,x,Z
1
2 kx yk2
2 + ⌧
2 (t + u1)
subject to Z =
T(u) x
x⇤
t
Z ⌫ 0.
AST SDP Formulation
(tl+1
, ul+1
, xl+1
) arg min
t,u,x
L⇢(t, u, x, Zl
, ⇤l
)
Zl+1
arg min
Z⌫0
L⇢(tl+1
, ul+1
, xl+1
, Z, ⇤l
)
⇤l+1
⇤l
+ ⇢
✓
Zl+1
T(ul+1
) xl+1
xl+1⇤
tl+1
◆
.
Alternating Direction Iterations
quadratic objective, linear
projection on psd cone
linear dual update
L⇢(t, u, x, Z, ⇤) =
1
2
kx yk2
2 +
⌧
2
(t + u1) +
⌧
⇤, Z
T(u) x
x⇤
t
+
⇢
2
Z
T(u) x
x⇤
t
2
F
Form Augmented Lagrangian
Lagrangian Term Augmented Lagrangian
Momentum Parameter
| {z } | {z }
29. Alternative Discretization Method
Can use a grid AN = {a(f, ) : f = j/N, j = 1, . . . , N, 2 [0, 2⇡)}
We show (1
2⇡n
N
)kxkAN
kxkA kxkAN
AST is approximated by Lasso
{
{
n
frequencies on a N grid
amplitudes
c?
+ w
atomic moments
Via Bernstein’s Polynomial
Theorem or Fritz John
Works for general epsilon nets
Fast shrinkage algorithms
Fourier projections are cheap
Coherence is irrelevant
30. Experimental setup
n=64 or 128 or 256 samples
no. of frequencies: n/16, n/8, n/4
SNR: -5 dB, 0 dB, ... , 20 dB
(Root) MUSIC, Matrix pencil, Cadzow: All fed true model
order.
Empirical estimated (not fed) noise variance for AST
Compared mean-squared-error and localization error
metrics, averaged over 20 trials.
31. Mean Squared Error
MSE vs SNR
10 5 0 5 10 15 20
30
20
10
0
10
SNR(dB)
MSE(dB)
AST
Cadzow
MUSIC
Lasso
128 samples, 8 frequencies
32. Mean Squared Error
Performance Profiles
2 4 6 8 10 12 14
0
0.2
0.4
0.6
0.8
1
β
Ps(β)
AST
Lasso 215
MUSIC
Cadzow
Matrix Pencil
P( ) = Fraction of experiments with MSE less than ⇥ minimum MSE.
Average MSE in 20 trials
n=64,128,256
s=n/16,n/8,n/4
SNR=-5, 0, ..., 20 dB
34. Frequency Localization
20 40 60 80 100
0
0.2
0.4
0.6
0.8
1
β
Ps(β)
AST
MUSIC
Cadzow
100 200 300 400 500
0
0.2
0.4
0.6
0.8
1
β
Ps(β)
AST
MUSIC
Cadzow
5 10 15 20
0
0.2
0.4
0.6
0.8
1
β
Ps(β)
AST
MUSIC
Cadzow
−10 −5 0 5 10 15 20
0
20
40
60
80
100
k = 32
SNR (dB)
m1
AST
MUSIC
Cadzow
−10 −5 0 5 10 15 20
0
0.01
0.02
0.03
0.04
k = 32
SNR (dB)
m2
AST
MUSIC
Cadzow
−10 −5 0 5 10 15 20
0
1
2
3
4
5
k = 32
SNR (dB)
m3
AST
MUSIC
Cadzow
Spurious Amplitudes Weighted Frequency Deviation Near region approximation
35. Simple LTI Systems
Vibration u(t)
u(t) y(t)
?
x[t + 1] = Ax[t] + Bu[t]
y[t] = Cx[t] + Du[t]
State Space
Find a simple linear system from time series data
Nonlinear Kalman filter based estimation
Parametric methods based on Hankel Low rank
formulations
36. Subspace Methods
2
6
6
6
6
6
4
y1 y2 · · · yk
y2 y3 · · · yk+1
y3 y4 · · · yk+2
...
...
...
y` y`+1 · · · y`+k 1
3
7
7
7
7
7
5
=
2
6
6
6
6
6
4
C
CA
CA2
...
CA` 1
3
7
7
7
7
7
5
⇥
x1 x2 . . . xk
⇤
+
2
6
6
6
6
6
4
D 0 . . . 0
CB D 0 . . . 0
CAB CB D . . . 0
...
...
...
CA` 2
B . . . D
3
7
7
7
7
7
5
2
6
6
6
6
6
4
u1 u2 · · · uk
u2 u3 · · · uk+1
u3 u4 · · · uk+2
...
...
...
u` u`+1 · · · u`+k 1
3
7
7
7
7
7
5
With a little computation,
| {z }
| {z }
| {z }
Y
X
U
Y = OX + TU Y U?
= OXU?
Output
State
Input
System Matrix T
Observability
SVD Subspace ID (n4sid)
Nuclear Norm (Vandenberghe et al)
Low Rank
37. SISO
⇥ 2 1. . .
u[t]
G
0 1 2 . . .
Past inputs Hankel Operator Future Outputs
G : 2( ⇤, 0) ⇥ 2[0, ⇤)
Fact: Rank of Hankel Operator = McMillan Degree
Relax and use Hankel nuclear norm?
Not clear how to compute...
38. Convex Perspective
SISO Systems
G(z) =
2(1 0.72
)
z 0.7
+
5(1 0.52
)
z (0.3 + 0.4i)
+
5(1 0.52
)
z (0.3 0.4i)
−1
0
1
−1
0
1
0
2
4
6
A =
⇢
1 |w|2
z w
w 2 D⇢
D
y = L (G(z)) + w
−0.5 0 0.5
2
3
4
5
6
7
minimize
G
ky L(G)k2
2 + µkGkA
Regularize with atomic norms:
⇡
8
kGkA k Gk1 kGkA
Theorem:
39. Atomic Norm Regularization
How to solve this?
Equivalent to:
Discretize and Set
k` := Lk
✓
1 |w|2
z w`
◆
minimize
c
1
2 ky ck2
2 + µkck1 Lasso
minimize
G
ky L(G)k2
2 + µkGkA
minimizec,x ky xk2
2 + µ
P
w2D⇢
|cw|
1 |w|2
subject to x =
P
w2D⇢
cwL
⇣
1
z w
⌘
40. DAST Analysis
constant Hankel nuclear
norm
stability
radius
number of
measurements
Measure uniform spaced samples of
frequency response:
kG ˆGk2
H2
0k Gk1
(1 ⇢)
p
n
ˆG(z) =
X
`
ˆc`
1 |w`|2
z w`
Set:
Then we have,
k` =
1 |w`|2
e2⇡ik/n w`
Solve: ˆx = arg min
x
1
2 ky xk2
2 + µkxkA✏
over a net on
the disk
where the coefficients
correspond to the support of ˆx
42. Summary
Optimal bounds for denoising using convex
methods.
Better empirical performance than classical
approaches
Local conditions on signals instead of global
dictionary conditions
Discretization works!
Acknowledgements: Results derived in
collaboration with Tang, Shah and Prof. Recht.
43.
44. Future Work
• Better convergence results for discretization.
• Why does weighted mean heuristic work so well for
discretization?
• Better algorithms for System Identification
• Greedy techniques
• More atomic sets. General localization guarantees.
45. Referenced Publications
B, Recht. “Atomic Norm Denoising for Line Spectral Estimation” in 49th
Allerton Conference on Controls and Systems, 2011
B, Tang, Recht. “Atomic Norm Denoising for Line Spectral Estimation”,
submitted to IEEE Transactions on Signal Processing
Tang, B, Recht. “Near Minimax Line Spectral Estimation”, To be submitted
Shah, B, Tang, Recht. “Linear System Identification using Atomic Norm
Minimization”, Control and Decision Conference, 2012
46. Other Publications
Tang, B, Shah, Recht. “Compressed Sensing off the grid”, submitted to
IEEE Transactions on Information Theory
Dasarathy, Shah, B, Nowak. “Covariance Sketching”, 50thAllerton
Conference on Controls and Systems, 2012
Gubner, B, Hao “Multipath-Cluster Channel Models”, IEEE Conference on
Ultrawideband Systems 2012
Wittenberg, Alimadhi, B, Lau “ei. RxC: Hierarchical Multinomial-
Dirichlet Ecological Inference Model” in Zelig: Everyone’s Statistical Software.
Tang, B, Recht. “Sparse recovery over continuous dictionaries: Just
discretize” Submitted toAsilomar Conference 2013
47. Cone Condition
=
z=0
z 2
z A
z C (x )
C (x ) = {z | x + z A x A + z A}
Inflated Descent Cone
No direction in Cᵧ with zero atomic norm
C0
(0 1)
48. Minimum separation is necessary
Using two sinusoids:
ka(f1) a(f2)kA n1/2
ka(f1) a(f2)k2
2⇡n2
|f1 f2|
So,
When separation is lower than 1/4n2 atomic norm can’t recover it.
ka(f1) a(f2)k2 = 2⇡n3/2
|f1 f2|
Can empirically show failure with less than 1/n separation and adversarial signal