Compressive Sensing with Random Matrices

Size: px
Start display at page:

Download "Compressive Sensing with Random Matrices"

Transcription

1 Compressive Sensing with Random Matrices Lucas Connell University of Georgia 9 November 017 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

2 Overview We will discuss compressive sensing, and, in particular, the viability of using random matrices. An outline: 1 Introduce some vector notation and terminology Discuss Compressive Sensing 3 Review probability 4 Define random matrices and concentration of measure 5 Define an ɛ-covering number, and prove a short lemma 6 Use these tools to prove our main result Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November 017 / 18

3 Vector Notation The l N p norm of a vector x R N is defined as ( N ) 1 i=1 x l N p := x i p p if 0 < p < max x i if p = i=1,...,n Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

4 Vector Notation The l N p norm of a vector x R N is defined as ( N ) 1 i=1 x l N p := x i p p if 0 < p < max x i if p = i=1,...,n Definition (Sparsity) We say a vector x R N is k-sparse if #{i : x i 0} k Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

5 Vector Notation The l N p norm of a vector x R N is defined as ( N ) 1 i=1 x l N p := x i p p if 0 < p < max x i if p = i=1,...,n Definition (Sparsity) We say a vector x R N is k-sparse if #{i : x i 0} k In R N, with a fixed k N and given a set T = {i 1, i,..., i k } {1,,..., N} of indices, we define X T := {[x 1, x,..., x N ] T : x i = 0 for i T } which is a k-dimensional subspace of R N We similarly define x T Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

6 Examples Example (Sparse Vectors) Let 3 4 x = and y = x is 3-sparse. y is 1-sparse. Example (X T Subspace) Consider R 5, and set T = {1,, 4}. X T := {[x 1, x, 0, x 4, 0] T : x 1, x, x 4 R} Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

7 Motivation In many applications, one wants to solve Φx = b where x R N is k - sparse, Φ R n N, and n << N Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

8 Motivation In many applications, one wants to solve Φx = b where x R N is k - sparse, Φ R n N, and n << N This can be quite hard, and we instead solve: argmin x l N, such that Φx = b (1) 1 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

9 Motivation In many applications, one wants to solve Φx = b where x R N is k - sparse, Φ R n N, and n << N This can be quite hard, and we instead solve: argmin x l N, such that Φx = b (1) 1 Theorem (Candes and Tao, Decoding by linear programming, 005) Suppose b = Φx, with x k-sparse. If Φ has the Restricted Isometry Property of order k, then x is the unique solution to (1) Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

10 Motivation In many applications, one wants to solve Φx = b where x R N is k - sparse, Φ R n N, and n << N This can be quite hard, and we instead solve: argmin x l N, such that Φx = b (1) 1 Theorem (Candes and Tao, Decoding by linear programming, 005) Suppose b = Φx, with x k-sparse. If Φ has the Restricted Isometry Property of order k, then x is the unique solution to (1) Theorem (Candes and Tao, Near-optimal signal recovery from random projections: Universal encoding strategies?, 006) If Φ R n N is a subgaussian random matrix with n cklog ( ) N k, then Φ satisfies the RIP of order k with overwhelming probability. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

11 Probability Review Definition (Random Variable) A (real) random variable returns a real number with some probability determined by a probability distribution function. Example: Let X N (0, 1) (standard normal distribution) Figure: Pr [X 1] = Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

12 Probability Review Subgaussian Random Variable: A subgaussian random variable X is a random variable whose distribution function s tails decay exponentially fast. Or, more precisely, for any t R: Pr [ X µ t] e t σ Figure: Blue is standard normal; orange is subgaussian. Notice the red area (subgaussian) is less than the green area (standard normal) Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

13 Probability Review Union Bounds: Given random variables X 1, X, we know Pr [X 1 t or X t] Pr [X 1 t] + Pr [X t]. The same inequality holds for any finite number of random variables. Example Let X 1, X,..., X m N (0, 1) be independent. Using a finite union bound: Pr [X i t for at least one i = 1,,..., m] m Pr [X i t] = mpr [X 1 t] i=1 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

14 Concentration of Measure The sum of (independent) random variables concentrates around the mean. Figure: Blue is standard normal; orange is where X 1,..., X 5 N (0, 1) 5 X i i=1 5 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

15 Random Matrix Definition (Random Matrix) A random matrix, Φ R n N, is a matrix whose entries, φ i,j, are each (independent) random variables. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

16 Random Matrix Definition (Random Matrix) A random matrix, Φ R n N, is a matrix whose entries, φ i,j, are each (independent) random variables. Theorem (Subgaussian Random Matrix) Letting q R N, ɛ (0, 1), and Φ be an n N matrix with entries φ i,j N (0, 1 n ), this concentration results in c 0(ɛ) > 0 with: [ Pr (1 ɛ) q l N Φq l n (1 + ɛ) q l N ] fails e nc 0(ɛ) Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

17 Covering Numbers Definition ( -Covering Number -Packing Number) 1 -covering number: N(U, ) = min n such that {q1, q,..., qn } with the property n [ U B (qi ) i=1 -packing number: P(U, ) = max m such that {q1, q,..., qm } with the property d(qi, qj ) >, i 6= j Figure: A space, U Lucas Connell (University of Georgia) Figure: A space, V Compressive Sensing with Random Matrices 9 November / 18

18 Covering Numbers Definition ( -Covering Number -Packing Number) 1 -covering number: N(U, ) = min n such that {q1, q,..., qn } with the property n [ U B (qi ) i=1 -packing number: P(U, ) = max m such that {q1, q,..., qm } with the property d(qi, qj ) >, i 6= j Figure: An -cover of our space U Lucas Connell (University of Georgia) Figure: An -packing in our space V Compressive Sensing with Random Matrices 9 November / 18

19 An Upper Bound for the ɛ-covering Number of a Unit Ball Lemma (Upper Bound for Covering of a Unit Ball in R k ) If U is the unit ball in R k and ɛ (0, 1), N(U, ɛ) ( 3 ɛ )k Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

20 An Upper Bound for the ɛ-covering Number of a Unit Ball Lemma (Upper Bound for Covering of a Unit Ball in R k ) If U is the unit ball in R k and ɛ (0, 1), N(U, ɛ) ( 3 ɛ )k It is known that N(U, ɛ) P(U, ɛ), so we bound the ɛ-packing number. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

21 An Upper Bound for the ɛ-covering Number of a Unit Ball Lemma (Upper Bound for Covering of a Unit Ball in R k ) If U is the unit ball in R k and ɛ (0, 1), N(U, ɛ) ( 3 ɛ )k It is known that N(U, ɛ) P(U, ɛ), so we bound the ɛ-packing number. Let {q 1, q,..., q m } be an optimal ɛ-packing. Hence m B ɛ (q i) B 1+ ɛ (0) i=1 Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

22 An Upper Bound for the ɛ-covering Number of a Unit Ball Lemma (Upper Bound for Covering of a Unit Ball in R k ) If U is the unit ball in R k and ɛ (0, 1), N(U, ɛ) ( 3 ɛ )k It is known that N(U, ɛ) P(U, ɛ), so we bound the ɛ-packing number. Let {q 1, q,..., q m } be an optimal ɛ-packing. Hence m B ɛ (q i) B 1+ ɛ (0) i=1 Since d(q i, q j ) > ɛ, i j, these balls are actually disjoint. Taking volume [ m ] m [ ] ( ɛ ) k Vol B ɛ (q i) = Vol B ɛ (q i) = mc k i=1 i=1 [ ] ( Vol B 1+ ɛ (0) = C k 1 + ɛ ) k Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

23 An Upper Bound for the ɛ-covering Number of a Unit Ball Lemma (Upper Bound for Covering of a Unit Ball in R k ) If U is the unit ball in R k and ɛ (0, 1), N(U, ɛ) ( 3 ɛ )k It is known that N(U, ɛ) P(U, ɛ), so we bound the ɛ-packing number. Let {q 1, q,..., q m } be an optimal ɛ-packing. Hence m B ɛ (q i) B 1+ ɛ (0) i=1 Since d(q i, q j ) > ɛ, i j, these balls are actually disjoint. Taking volume [ m ] m [ ] ( ɛ ) k Vol B ɛ (q i) = Vol B ɛ (q i) = mc k i=1 i=1 [ ] ( Vol B 1+ ɛ (0) = C k 1 + ɛ ) k Rearranging these results ultimately yields: m ( ) +ɛ k ( ɛ 3 ) k ɛ Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

24 The Restricted Isometry Property (RIP) Definition We say a matrix Φ satisfies the RIP of order k if δ k (0, 1) such that, for any T with #(T ) k (1 δ k ) x T l N Φx T l n (1 + δ k) x T l N () Our goal is now to show that a random matrix, Φ with entries from a subgaussian distribution, satisfies the RIP of order k with overwhelming probability. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

25 Major Theorem (See also 1 ) Theorem (Candes and Tao, Near-optimal signal recovery from random projections: Universal encoding strategies?, 006) Given δ (0, 1), c 1, c R + such that Φ satisfies the RIP of order k with probability greater than 1 e cn, provided n c 1 k log ( ) N k 1 Baraniuk et al., A Simple Proof of the Restricted Isometry Property for Random Matrices. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

26 Major Theorem (See also 1 ) Theorem (Candes and Tao, Near-optimal signal recovery from random projections: Universal encoding strategies?, 006) Given δ (0, 1), c 1, c R + such that Φ satisfies the RIP of order k with probability greater than 1 e cn, provided n c 1 k log ( ) N k Fix a k satisfying the hypothesis, let T be a set of indices with #(T ) = k. Note here that there are ( ) N k choices for T. Now consider B T := X T B 1 (0) 1 Baraniuk et al., A Simple Proof of the Restricted Isometry Property for Random Matrices. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

27 Major Theorem (See also 1 ) Theorem (Candes and Tao, Near-optimal signal recovery from random projections: Universal encoding strategies?, 006) Given δ (0, 1), c 1, c R + such that Φ satisfies the RIP of order k with probability greater than 1 e cn, provided n c 1 k log ( ) N k Fix a k satisfying the hypothesis, let T be a set of indices with #(T ) = k. Note here that there are ( ) N k choices for T. Now consider B T := X T B 1 (0) By our Lemma on covering numbers, can pick δ 4 -covering Q T = {q 1,..., q α } with α ( ) 1 k δ 1 Baraniuk et al., A Simple Proof of the Restricted Isometry Property for Random Matrices. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

28 Via concentration of measure, for a given q i Q T : [( Pr 1 δ ) ( q i l N Φq i l n 1 + δ ) ] q i l N fails e nc 0( δ ) Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

29 Via concentration of measure, for a given q i Q T : [( Pr 1 δ ) ( q i l N Φq i l n 1 + δ ) ] q i l N fails e nc 0( δ ) Using a union bound on our finite number of points: [( Pr 1 δ ) ( q l N Φq l n 1 + δ ) ] q l fails for a q Q T N ( ) 1 k e nc 0( δ ) δ Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

30 Via concentration of measure, for a given q i Q T : [( Pr 1 δ ) ( q i l N Φq i l n 1 + δ ) ] q i l N fails e nc 0( δ ) Using a union bound on our finite number of points: [( Pr 1 δ ) ( q l N Φq l n 1 + δ ) ] q l fails for a q Q T N ( ) 1 k e nc 0( δ ) δ From this, one can show: [ Pr (1 δ) x l N Φx l n (1 + δ) x l N fails for an x X T ] ( ) 1 k e nc 0( δ ) δ Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

31 Since there are ( ) ( N k en ) k k such k-dimensional subspaces, use a union bound to see [ Pr (1 δ) x l N Φx l n (1 + δ) x l N fails for an x ] X T T ( 1 δ ) k ( ) en k e nc 0( δ ) = e nc 0( δ )+k[log( en k )+log( 1 δ )] k Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

32 Since there are ( ) ( N k en ) k k such k-dimensional subspaces, use a union bound to see [ Pr (1 δ) x l N Φx l n (1 + δ) x l N fails for an x ] X T T ( 1 δ ) k ( ) en k e nc 0( δ ) = e nc 0( δ )+k[log( en k )+log( 1 δ )] k By picking some constant c and taking the complement, [ Pr x ] X T, (1 δ) x l N Φx l n (1 + δ) x l N T 1 e nc Which is to say: Φ satisfies the RIP of order k specified in the theorem with at least the above probability. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

33 Summary Want to compress a high-sparse vector, in a way where we can recover it with little-to-no error. How do we compress it? Compressing with a subgaussian matrix Φ allows for unique l 1 -minimization; an accurate recovery How small can we compress? We can pick n logarithmic to N. Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

34 References Baraniuk, Richard et al. A Simple Proof of the Restricted Isometry Property for Random Matrices. In: Constructive Approximation 8.3 (Dec. 008), pp issn: doi: /s x. url: Candes, Emmanuel J and Terence Tao. Decoding by linear programming. In: IEEE transactions on information theory 51.1 (005), pp Near-optimal signal recovery from random projections: Universal encoding strategies?. In: IEEE transactions on information theory 5.1 (006), pp Lucas Connell (University of Georgia) Compressive Sensing with Random Matrices 9 November / 18

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 13 October 6, 2016 Scribe: Kiyeon Jeon and Loc Hoang 1 Overview In the last lecture we covered the lower bound for p th moment (p > 2) and

More information

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5 CS 229r: Algorithms for Big Data Fall 215 Prof. Jelani Nelson Lecture 19 Nov 5 Scribe: Abdul Wasay 1 Overview In the last lecture, we started discussing the problem of compressed sensing where we are given

More information

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad

More information

A New Estimate of Restricted Isometry Constants for Sparse Solutions

A New Estimate of Restricted Isometry Constants for Sparse Solutions A New Estimate of Restricted Isometry Constants for Sparse Solutions Ming-Jun Lai and Louis Y. Liu January 12, 211 Abstract We show that as long as the restricted isometry constant δ 2k < 1/2, there exist

More information

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Constructing Explicit RIP Matrices and the Square-Root Bottleneck Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry

More information

Compressed Sensing: Lecture I. Ronald DeVore

Compressed Sensing: Lecture I. Ronald DeVore Compressed Sensing: Lecture I Ronald DeVore Motivation Compressed Sensing is a new paradigm for signal/image/function acquisition Motivation Compressed Sensing is a new paradigm for signal/image/function

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit

Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit Uniform Uncertainty Principle and signal recovery via Regularized Orthogonal Matching Pursuit arxiv:0707.4203v2 [math.na] 14 Aug 2007 Deanna Needell Department of Mathematics University of California,

More information

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing Afonso S. Bandeira April 9, 2015 1 The Johnson-Lindenstrauss Lemma Suppose one has n points, X = {x 1,..., x n }, in R d with d very

More information

AN INTRODUCTION TO COMPRESSIVE SENSING

AN INTRODUCTION TO COMPRESSIVE SENSING AN INTRODUCTION TO COMPRESSIVE SENSING Rodrigo B. Platte School of Mathematical and Statistical Sciences APM/EEE598 Reverse Engineering of Complex Dynamical Networks OUTLINE 1 INTRODUCTION 2 INCOHERENCE

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html Acknowledgement: this slides is based on Prof. Emmanuel Candes and Prof. Wotao Yin

More information

Recent Developments in Compressed Sensing

Recent Developments in Compressed Sensing Recent Developments in Compressed Sensing M. Vidyasagar Distinguished Professor, IIT Hyderabad m.vidyasagar@iith.ac.in, www.iith.ac.in/ m vidyasagar/ ISL Seminar, Stanford University, 19 April 2018 Outline

More information

Z Algorithmic Superpower Randomization October 15th, Lecture 12

Z Algorithmic Superpower Randomization October 15th, Lecture 12 15.859-Z Algorithmic Superpower Randomization October 15th, 014 Lecture 1 Lecturer: Bernhard Haeupler Scribe: Goran Žužić Today s lecture is about finding sparse solutions to linear systems. The problem

More information

Solution Recovery via L1 minimization: What are possible and Why?

Solution Recovery via L1 minimization: What are possible and Why? Solution Recovery via L1 minimization: What are possible and Why? Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Eighth US-Mexico Workshop on Optimization

More information

Limitations in Approximating RIP

Limitations in Approximating RIP Alok Puranik Mentor: Adrian Vladu Fifth Annual PRIMES Conference, 2015 Outline 1 Background The Problem Motivation Construction Certification 2 Planted model Planting eigenvalues Analysis Distinguishing

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Noisy Signal Recovery via Iterative Reweighted L1-Minimization Noisy Signal Recovery via Iterative Reweighted L1-Minimization Deanna Needell UC Davis / Stanford University Asilomar SSC, November 2009 Problem Background Setup 1 Suppose x is an unknown signal in R d.

More information

Near Optimal Signal Recovery from Random Projections

Near Optimal Signal Recovery from Random Projections 1 Near Optimal Signal Recovery from Random Projections Emmanuel Candès, California Institute of Technology Multiscale Geometric Analysis in High Dimensions: Workshop # 2 IPAM, UCLA, October 2004 Collaborators:

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER On the Performance of Sparse Recovery IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 11, NOVEMBER 2011 7255 On the Performance of Sparse Recovery Via `p-minimization (0 p 1) Meng Wang, Student Member, IEEE, Weiyu Xu, and Ao Tang, Senior

More information

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011

Introduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011 Compressed Sensing Huichao Xue CS3750 Fall 2011 Table of Contents Introduction From News Reports Abstract Definition How it works A review of L 1 norm The Algorithm Backgrounds for underdetermined linear

More information

Conditions for a Unique Non-negative Solution to an Underdetermined System

Conditions for a Unique Non-negative Solution to an Underdetermined System Conditions for a Unique Non-negative Solution to an Underdetermined System Meng Wang and Ao Tang School of Electrical and Computer Engineering Cornell University Ithaca, NY 14853 Abstract This paper investigates

More information

Combining geometry and combinatorics

Combining geometry and combinatorics Combining geometry and combinatorics A unified approach to sparse signal recovery Anna C. Gilbert University of Michigan joint work with R. Berinde (MIT), P. Indyk (MIT), H. Karloff (AT&T), M. Strauss

More information

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit

Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit Claremont Colleges Scholarship @ Claremont CMC Faculty Publications and Research CMC Faculty Scholarship 6-5-2008 Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit

More information

Sparse Optimization Lecture: Sparse Recovery Guarantees

Sparse Optimization Lecture: Sparse Recovery Guarantees Those who complete this lecture will know Sparse Optimization Lecture: Sparse Recovery Guarantees Sparse Optimization Lecture: Sparse Recovery Guarantees Instructor: Wotao Yin Department of Mathematics,

More information

Geometry of log-concave Ensembles of random matrices

Geometry of log-concave Ensembles of random matrices Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann

More information

A Simple Proof of the Restricted Isometry Property for Random Matrices

A Simple Proof of the Restricted Isometry Property for Random Matrices DOI 10.1007/s00365-007-9003-x A Simple Proof of the Restricted Isometry Property for Random Matrices Richard Baraniuk Mark Davenport Ronald DeVore Michael Wakin Received: 17 May 006 / Revised: 18 January

More information

Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery

Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery Sparse analysis Lecture VII: Combining geometry and combinatorics, sparse matrices for sparse signal recovery Anna C. Gilbert Department of Mathematics University of Michigan Sparse signal recovery measurements:

More information

sample lectures: Compressed Sensing, Random Sampling I & II

sample lectures: Compressed Sensing, Random Sampling I & II P. Jung, CommIT, TU-Berlin]: sample lectures Compressed Sensing / Random Sampling I & II 1/68 sample lectures: Compressed Sensing, Random Sampling I & II Peter Jung Communications and Information Theory

More information

Sparse Interactions: Identifying High-Dimensional Multilinear Systems via Compressed Sensing

Sparse Interactions: Identifying High-Dimensional Multilinear Systems via Compressed Sensing Sparse Interactions: Identifying High-Dimensional Multilinear Systems via Compressed Sensing Bobak Nazer and Robert D. Nowak University of Wisconsin, Madison Allerton 10/01/10 Motivation: Virus-Host Interaction

More information

Multipath Matching Pursuit

Multipath Matching Pursuit Multipath Matching Pursuit Submitted to IEEE trans. on Information theory Authors: S. Kwon, J. Wang, and B. Shim Presenter: Hwanchol Jang Multipath is investigated rather than a single path for a greedy

More information

Supremum of simple stochastic processes

Supremum of simple stochastic processes Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there

More information

Sparse Recovery with Pre-Gaussian Random Matrices

Sparse Recovery with Pre-Gaussian Random Matrices Sparse Recovery with Pre-Gaussian Random Matrices Simon Foucart Laboratoire Jacques-Louis Lions Université Pierre et Marie Curie Paris, 75013, France Ming-Jun Lai Department of Mathematics University of

More information

Compressive Sensing of Sparse Tensor

Compressive Sensing of Sparse Tensor Shmuel Friedland Univ. Illinois at Chicago Matheon Workshop CSA2013 December 12, 2013 joint work with Q. Li and D. Schonfeld, UIC Abstract Conventional Compressive sensing (CS) theory relies on data representation

More information

On the l 1 -Norm Invariant Convex k-sparse Decomposition of Signals

On the l 1 -Norm Invariant Convex k-sparse Decomposition of Signals On the l 1 -Norm Invariant Convex -Sparse Decomposition of Signals arxiv:1305.6021v2 [cs.it] 11 Nov 2013 Guangwu Xu and Zhiqiang Xu Abstract Inspired by an interesting idea of Cai and Zhang, we formulate

More information

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar

More information

An Overview of Compressed Sensing

An Overview of Compressed Sensing An Overview of Compressed Sensing Nathan Schneider November 18, 2009 Abstract In a large number of applications, the system will be designed to sample at a rate equal to at least the frequency bandwidth

More information

Randomness-in-Structured Ensembles for Compressed Sensing of Images

Randomness-in-Structured Ensembles for Compressed Sensing of Images Randomness-in-Structured Ensembles for Compressed Sensing of Images Abdolreza Abdolhosseini Moghadam Dep. of Electrical and Computer Engineering Michigan State University Email: abdolhos@msu.edu Hayder

More information

Compressed Sensing and Sparse Recovery

Compressed Sensing and Sparse Recovery ELE 538B: Sparsity, Structure and Inference Compressed Sensing and Sparse Recovery Yuxin Chen Princeton University, Spring 217 Outline Restricted isometry property (RIP) A RIPless theory Compressed sensing

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

Generalized Orthogonal Matching Pursuit- A Review and Some

Generalized Orthogonal Matching Pursuit- A Review and Some Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents

More information

Dimensionality Reduction Notes 3

Dimensionality Reduction Notes 3 Dimensionality Reduction Notes 3 Jelani Nelson minilek@seas.harvard.edu August 13, 2015 1 Gordon s theorem Let T be a finite subset of some normed vector space with norm X. We say that a sequence T 0 T

More information

Lecture 22: More On Compressed Sensing

Lecture 22: More On Compressed Sensing Lecture 22: More On Compressed Sensing Scribed by Eric Lee, Chengrun Yang, and Sebastian Ament Nov. 2, 207 Recap and Introduction Basis pursuit was the method of recovering the sparsest solution to an

More information

Sparse Solutions of an Undetermined Linear System

Sparse Solutions of an Undetermined Linear System 1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research

More information

Signal Recovery from Permuted Observations

Signal Recovery from Permuted Observations EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,

More information

Does Compressed Sensing have applications in Robust Statistics?

Does Compressed Sensing have applications in Robust Statistics? Does Compressed Sensing have applications in Robust Statistics? Salvador Flores December 1, 2014 Abstract The connections between robust linear regression and sparse reconstruction are brought to light.

More information

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina

INDUSTRIAL MATHEMATICS INSTITUTE. B.S. Kashin and V.N. Temlyakov. IMI Preprint Series. Department of Mathematics University of South Carolina INDUSTRIAL MATHEMATICS INSTITUTE 2007:08 A remark on compressed sensing B.S. Kashin and V.N. Temlyakov IMI Preprint Series Department of Mathematics University of South Carolina A remark on compressed

More information

Recovery of Compressible Signals in Unions of Subspaces

Recovery of Compressible Signals in Unions of Subspaces 1 Recovery of Compressible Signals in Unions of Subspaces Marco F. Duarte, Chinmay Hegde, Volkan Cevher, and Richard G. Baraniuk Department of Electrical and Computer Engineering Rice University Abstract

More information

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1

The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1 The Sparsest Solution of Underdetermined Linear System by l q minimization for 0 < q 1 Simon Foucart Department of Mathematics Vanderbilt University Nashville, TN 3784. Ming-Jun Lai Department of Mathematics,

More information

Compressed Sensing and Related Learning Problems

Compressed Sensing and Related Learning Problems Compressed Sensing and Related Learning Problems Yingzhen Li Dept. of Mathematics, Sun Yat-sen University Advisor: Prof. Haizhang Zhang Advisor: Prof. Haizhang Zhang 1 / Overview Overview Background Compressed

More information

Optimisation Combinatoire et Convexe.

Optimisation Combinatoire et Convexe. Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix

More information

5406 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER 2006

5406 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER 2006 5406 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER 2006 Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? Emmanuel J. Candes and Terence Tao Abstract

More information

Lecture 16: Compressed Sensing

Lecture 16: Compressed Sensing Lecture 16: Compressed Sensing Introduction to Learning and Analysis of Big Data Kontorovich and Sabato (BGU) Lecture 16 1 / 12 Review of Johnson-Lindenstrauss Unsupervised learning technique key insight:

More information

Sensing systems limited by constraints: physical size, time, cost, energy

Sensing systems limited by constraints: physical size, time, cost, energy Rebecca Willett Sensing systems limited by constraints: physical size, time, cost, energy Reduce the number of measurements needed for reconstruction Higher accuracy data subject to constraints Original

More information

The Pros and Cons of Compressive Sensing

The Pros and Cons of Compressive Sensing The Pros and Cons of Compressive Sensing Mark A. Davenport Stanford University Department of Statistics Compressive Sensing Replace samples with general linear measurements measurements sampled signal

More information

Analog-to-Information Conversion

Analog-to-Information Conversion Analog-to-Information Conversion Sergiy A. Vorobyov Dept. Signal Processing and Acoustics, Aalto University February 2013 Winter School on Compressed Sensing, Ruka 1/55 Outline 1 Compressed Sampling (CS)

More information

MAKING DO WITH LESS: AN INTRODUCTION TO COMPRESSED SENSING

MAKING DO WITH LESS: AN INTRODUCTION TO COMPRESSED SENSING MAKING DO WITH LESS: AN INTRODUCTION TO COMPRESSED SENSING KURT BRYAN AND TANYA LEISE Abstract. This article offers an accessible but rigorous and essentially self-contained account of some of the central

More information

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016 Lecture notes 5 February 9, 016 1 Introduction Random projections Random projections are a useful tool in the analysis and processing of high-dimensional data. We will analyze two applications that use

More information

Sparse solutions of underdetermined systems

Sparse solutions of underdetermined systems Sparse solutions of underdetermined systems I-Liang Chern September 22, 2016 1 / 16 Outline Sparsity and Compressibility: the concept for measuring sparsity and compressibility of data Minimum measurements

More information

Least singular value of random matrices. Lewis Memorial Lecture / DIMACS minicourse March 18, Terence Tao (UCLA)

Least singular value of random matrices. Lewis Memorial Lecture / DIMACS minicourse March 18, Terence Tao (UCLA) Least singular value of random matrices Lewis Memorial Lecture / DIMACS minicourse March 18, 2008 Terence Tao (UCLA) 1 Extreme singular values Let M = (a ij ) 1 i n;1 j m be a square or rectangular matrix

More information

Phase Transition Phenomenon in Sparse Approximation

Phase Transition Phenomenon in Sparse Approximation Phase Transition Phenomenon in Sparse Approximation University of Utah/Edinburgh L1 Approximation: May 17 st 2008 Convex polytopes Counting faces Sparse Representations via l 1 Regularization Underdetermined

More information

COMPRESSIVE SAMPLING USING EM ALGORITHM. Technical Report No: ASU/2014/4

COMPRESSIVE SAMPLING USING EM ALGORITHM. Technical Report No: ASU/2014/4 COMPRESSIVE SAMPLING USING EM ALGORITHM ATANU KUMAR GHOSH, ARNAB CHAKRABORTY Technical Report No: ASU/2014/4 Date: 29 th April, 2014 Applied Statistics Unit Indian Statistical Institute Kolkata- 700018

More information

Shannon-Theoretic Limits on Noisy Compressive Sampling Mehmet Akçakaya, Student Member, IEEE, and Vahid Tarokh, Fellow, IEEE

Shannon-Theoretic Limits on Noisy Compressive Sampling Mehmet Akçakaya, Student Member, IEEE, and Vahid Tarokh, Fellow, IEEE 492 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 1, JANUARY 2010 Shannon-Theoretic Limits on Noisy Compressive Sampling Mehmet Akçakaya, Student Member, IEEE, Vahid Tarokh, Fellow, IEEE Abstract

More information

On State Estimation with Bad Data Detection

On State Estimation with Bad Data Detection On State Estimation with Bad Data Detection Weiyu Xu, Meng Wang, and Ao Tang School of ECE, Cornell University, Ithaca, NY 4853 Abstract We consider the problem of state estimation through observations

More information

The convex algebraic geometry of rank minimization

The convex algebraic geometry of rank minimization The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming

More information

Optimization for Compressed Sensing

Optimization for Compressed Sensing Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve

More information

Sparse analysis Lecture V: From Sparse Approximation to Sparse Signal Recovery

Sparse analysis Lecture V: From Sparse Approximation to Sparse Signal Recovery Sparse analysis Lecture V: From Sparse Approximation to Sparse Signal Recovery Anna C. Gilbert Department of Mathematics University of Michigan Connection between... Sparse Approximation and Compressed

More information

Deterministic constructions of compressed sensing matrices

Deterministic constructions of compressed sensing matrices Journal of Complexity 23 (2007) 918 925 www.elsevier.com/locate/jco Deterministic constructions of compressed sensing matrices Ronald A. DeVore Department of Mathematics, University of South Carolina,

More information

A Generalized Restricted Isometry Property

A Generalized Restricted Isometry Property 1 A Generalized Restricted Isometry Property Jarvis Haupt and Robert Nowak Department of Electrical and Computer Engineering, University of Wisconsin Madison University of Wisconsin Technical Report ECE-07-1

More information

Rui ZHANG Song LI. Department of Mathematics, Zhejiang University, Hangzhou , P. R. China

Rui ZHANG Song LI. Department of Mathematics, Zhejiang University, Hangzhou , P. R. China Acta Mathematica Sinica, English Series May, 015, Vol. 31, No. 5, pp. 755 766 Published online: April 15, 015 DOI: 10.1007/s10114-015-434-4 Http://www.ActaMath.com Acta Mathematica Sinica, English Series

More information

Stochastic geometry and random matrix theory in CS

Stochastic geometry and random matrix theory in CS Stochastic geometry and random matrix theory in CS IPAM: numerical methods for continuous optimization University of Edinburgh Joint with Bah, Blanchard, Cartis, and Donoho Encoder Decoder pair - Encoder/Decoder

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Error Correction via Linear Programming

Error Correction via Linear Programming Error Correction via Linear Programming Emmanuel Candes and Terence Tao Applied and Computational Mathematics, Caltech, Pasadena, CA 91125 Department of Mathematics, University of California, Los Angeles,

More information

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Forty-Fifth Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 26-28, 27 WeA3.2 Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Benjamin

More information

On the Observability of Linear Systems from Random, Compressive Measurements

On the Observability of Linear Systems from Random, Compressive Measurements On the Observability of Linear Systems from Random, Compressive Measurements Michael B Wakin, Borhan M Sanandaji, and Tyrone L Vincent Abstract Recovering or estimating the initial state of a highdimensional

More information

Compressed Sensing and Linear Codes over Real Numbers

Compressed Sensing and Linear Codes over Real Numbers Compressed Sensing and Linear Codes over Real Numbers Henry D. Pfister (joint with Fan Zhang) Texas A&M University College Station Information Theory and Applications Workshop UC San Diego January 31st,

More information

Compressed Sensing Using Bernoulli Measurement Matrices

Compressed Sensing Using Bernoulli Measurement Matrices ITSchool 11, Austin Compressed Sensing Using Bernoulli Measurement Matrices Yuhan Zhou Advisor: Wei Yu Department of Electrical and Computer Engineering University of Toronto, Canada Motivation Motivation

More information

IN order to digitize a signal, two discretization steps are

IN order to digitize a signal, two discretization steps are IEEE TRANSACTIONS ON INFORMATION THEORY to appear 1 Universal Rate-Efficient Scalar Quantization Petros T. Boufounos Member IEEE Abstract Scalar quantization is the most practical and straightforward approach

More information

Guaranteed Rank Minimization via Singular Value Projection: Supplementary Material

Guaranteed Rank Minimization via Singular Value Projection: Supplementary Material Guaranteed Rank Minimization via Singular Value Projection: Supplementary Material Prateek Jain Microsoft Research Bangalore Bangalore, India prajain@microsoft.com Raghu Meka UT Austin Dept. of Computer

More information

Analysis of Greedy Algorithms

Analysis of Greedy Algorithms Analysis of Greedy Algorithms Jiahui Shen Florida State University Oct.26th Outline Introduction Regularity condition Analysis on orthogonal matching pursuit Analysis on forward-backward greedy algorithm

More information

Gradient Descent with Sparsification: An iterative algorithm for sparse recovery with restricted isometry property

Gradient Descent with Sparsification: An iterative algorithm for sparse recovery with restricted isometry property : An iterative algorithm for sparse recovery with restricted isometry property Rahul Garg grahul@us.ibm.com Rohit Khandekar rohitk@us.ibm.com IBM T. J. Watson Research Center, 0 Kitchawan Road, Route 34,

More information

Lecture 16 Oct. 26, 2017

Lecture 16 Oct. 26, 2017 Sketching Algorithms for Big Data Fall 2017 Prof. Piotr Indyk Lecture 16 Oct. 26, 2017 Scribe: Chi-Ning Chou 1 Overview In the last lecture we constructed sparse RIP 1 matrix via expander and showed that

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 3: Sparse signal recovery: A RIPless analysis of l 1 minimization Yuejie Chi The Ohio State University Page 1 Outline

More information

Fast Hard Thresholding with Nesterov s Gradient Method

Fast Hard Thresholding with Nesterov s Gradient Method Fast Hard Thresholding with Nesterov s Gradient Method Volkan Cevher Idiap Research Institute Ecole Polytechnique Federale de ausanne volkan.cevher@epfl.ch Sina Jafarpour Department of Computer Science

More information

GREEDY SIGNAL RECOVERY REVIEW

GREEDY SIGNAL RECOVERY REVIEW GREEDY SIGNAL RECOVERY REVIEW DEANNA NEEDELL, JOEL A. TROPP, ROMAN VERSHYNIN Abstract. The two major approaches to sparse recovery are L 1-minimization and greedy methods. Recently, Needell and Vershynin

More information

Thresholds for the Recovery of Sparse Solutions via L1 Minimization

Thresholds for the Recovery of Sparse Solutions via L1 Minimization Thresholds for the Recovery of Sparse Solutions via L Minimization David L. Donoho Department of Statistics Stanford University 39 Serra Mall, Sequoia Hall Stanford, CA 9435-465 Email: donoho@stanford.edu

More information

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm

Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm J. K. Pant, W.-S. Lu, and A. Antoniou University of Victoria August 25, 2011 Compressive Sensing 1 University

More information

Stability and robustness of l 1 -minimizations with Weibull matrices and redundant dictionaries

Stability and robustness of l 1 -minimizations with Weibull matrices and redundant dictionaries Stability and robustness of l 1 -minimizations with Weibull matrices and redundant dictionaries Simon Foucart, Drexel University Abstract We investigate the recovery of almost s-sparse vectors x C N from

More information

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora Scribe: Today we continue the

More information

An algebraic perspective on integer sparse recovery

An algebraic perspective on integer sparse recovery An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:

More information

Compressed Sensing Using Reed- Solomon and Q-Ary LDPC Codes

Compressed Sensing Using Reed- Solomon and Q-Ary LDPC Codes Compressed Sensing Using Reed- Solomon and Q-Ary LDPC Codes Item Type text; Proceedings Authors Jagiello, Kristin M. Publisher International Foundation for Telemetering Journal International Telemetering

More information

Performance Analysis for Sparse Support Recovery

Performance Analysis for Sparse Support Recovery Performance Analysis for Sparse Support Recovery Gongguo Tang and Arye Nehorai ESE, Washington University April 21st 2009 Gongguo Tang and Arye Nehorai (Institute) Performance Analysis for Sparse Support

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

Compressed sensing techniques for hyperspectral image recovery

Compressed sensing techniques for hyperspectral image recovery Compressed sensing techniques for hyperspectral image recovery A. Abrardo, M. Barni, C. M. Carretti, E. Magli, S. Kuiteing Kamdem, R. Vitulli ABSTRACT Compressed Sensing (CS) theory is progressively gaining

More information

Tutorial: Sparse Signal Recovery

Tutorial: Sparse Signal Recovery Tutorial: Sparse Signal Recovery Anna C. Gilbert Department of Mathematics University of Michigan (Sparse) Signal recovery problem signal or population length N k important Φ x = y measurements or tests:

More information