Compressive Distilled Sensing: Sparse Recovery Using Adaptivity in Compressive Measurements

Size: px
Start display at page:

Download "Compressive Distilled Sensing: Sparse Recovery Using Adaptivity in Compressive Measurements"

Transcription

1 1 Copressive Distilled Sensing: Sparse Recovery Using Adaptivity in Copressive Measureents Jarvis D. Haupt 1 Richard G. Baraniuk 1 Rui M. Castro 2 and Robert D. Nowak 3 1 Dept. of Electrical and Coputer Engineering Rice University Houston TX Dept. of Electrical Engineering Colubia University New York NY Dept. of Electrical and Coputer Engineering University of Wisconsin Madison WI Abstract The recently-proposed theory of distilled sensing establishes that adaptivity in sapling can draatically iprove the perforance of sparse recovery in noisy settings. In particular it is now known that adaptive point sapling enables the detection and/or support recovery of sparse signals that are otherwise too weak to be recovered using any ethod based on non-adaptive point sapling. In this paper the theory of distilled sensing is extended to highly-undersapled regies as in copressive sensing. A siple adaptive sapling-and-refineent procedure called copressive distilled sensing is proposed where each step of the procedure utilizes inforation fro previous observations to focus subsequent easureents into the proper signal subspace resulting in a significant iproveent in effective easureent SNR on the signal subspace. As a result for the sae budget of sensing resources copressive distilled sensing can result in significantly iproved error bounds copared to those for traditional copressive sensing. I. INTRODUCTION Let x R n be a sparse vector supported on the set S = {i : x i 0} where S = s n and consider observing x according to the linear observation odel y = Ax + w 1 where A is an n real-valued atrix possibly rando that satisfies E [ ] A 2 iid F n and where wi N 0 σ 2 for soe σ 0. This odel is central to the eerging field of copressive sensing CS which deals priarily with recovery of x in highly-underdeterined settings that is where the nuber of easureents n. Initial results in CS establish a rather surprising result using certain observation atrices A for which the nuber of rows is a constant ultiple of s log n it is possible to recover x exactly fro {y A} and in addition the recovery can be accoplished by solving a tractable convex optiization [1] [3]. Matrices A for which this exact recovery is possible are easy to construct in practice. For exaple atrices whose entries are i.i.d. realizations of certain zero-ean distributions Gaussian syetric Bernoulli etc. are sufficient to allow this recovery with high probability [2] [4]. In practice however it is rarely the case that observations are perfectly noise-free. In these settings rather than attept This work was partially supported by the ARO grant no. W911NF the NSF grant no. CCF and the AFOSR grant no. FA to recover x exactly the goal becoes to estiate x to high accuracy in soe etric such as l 2 nor [5] [6]. One such estiation procedure is the Dantzig selector proposed in [6] which establishes that CS recovery reains stable in the presence of noise. We state the result here as a lea. Lea 1 Dantzig selector. For = Ωs log n generate a rando n atrix A whose entries are i.i.d. N 0 1/ and collect observations y according to 1. The estiate x = arg in z R n z l 1 subject to A T y Az l < λ where λ = Θσ log n satisfies x x 2 l 2 = Osσ 2 log n with probability 1 On C 0 for soe constant C 0 > 0. Reark 1. The constants in the above can be specified explicitly or bounded appropriately but we choose to present the results here and where appropriate in the sequel in ters of scaling relationships 1 in the interest of siplicity. On the other hand suppose that an oracle were to identify the locations of the nonzero signal coponents or equivalently the support S prior to recovery. Then one could construct the least-squares estiate x LS = A T S A S 1 A T S y where A S denotes the subatrix of A fored fro the coluns indexed by the eleents of S. The error of this estiate is x LS x 2 l 2 = Osσ 2 with probability 1 On C 1 for soe C 1 > 0 as shown in [6]. Coparing this oracleassisted bound with the result of Lea 1 we see that the priary difference is the presence of the logarithic ter in the error bound of the latter which can be interpreted as the searching penalty associated with having to learn the correct signal subspace. Of course the signal subspace will rarely if ever be known a priori. But suppose that it were possible to learn the signal subspace fro the data in a sequential adaptive fashion as the data are collected. In this case sensing energy could be focused only into the true signal subspace gradually iproving the effective easureent SNR. Intuitively one ight expect that this type of procedure could ultiately yield an estiate whose accuracy would be closer to that of 1 Recall that for functions f = fn and g = gn f = Og eans f cg for soe constant c for all n sufficiently large f = Ωg eans f c g for a constant c for all n sufficiently large and f = Θg eans that f = Og and f = Ωg. In addition we will use the notation f = og to indicate that li n f/g = 0.

2 2 the oracle-assisted estiator since the effective observation atrix would begin to assue the structure of A S. Such adaptive copressive sapling ethods have been proposed and exained epirically [7] [9] but to date the perforance benefits of adaptivity in copressive sapling have not been established theoretically. In this paper we take a step in that direction by analyzing the perforance of a ulti-step adaptive saplingand-refineent procedure called copressive distilled sensing CDS extending our own prior work in distilled sensing where the theoretical advantages of adaptive sapling in uncopressed settings were quantified [10] [11]. Our ain results here guarantee that for signals having not too any nonzero entries and for which the dynaic range is not too large a total of Os log n adaptively-collected easureents yield an estiator that with high probability achieves the Osσ 2 error bound of the oracle-assisted estiator. The reainder of the paper is organized as follows. The CDS procedure is described in Sec. II and its perforance is quantified as a theore in Sec. III. Extensions and conclusions are briefly described in Sec. IV and a sketch of the proof of the ain result and associated leata appear in the Appendix. Algorith 1: Copressive distilled sensing CDS. Input: Nuber of observation steps k; R j j = 1... k such that k j=1 Rj n; j j = 1... k such that k j=1 j ; Initialize: Initial index set I 1 = { n}; Distillation: for j = 1 to k do Copute τ j = R j / I j ; Construct { A j where A j N 0 τ j j uv iid u {1... j } v I j ; 0 u {1... j } v / I j Collect y j = A j x + w j ; Copute x j = A j T y j ; Refine I j+1 = {i I j : x j i > 0}; end Output: Distilled observations { y j A j} k j=1 ; II. COMPRESSIVE DISTILLED SENSING In this section we describe the copressive distilled sensing CDS procedure which is a natural generalization of the distilled sensing DS procedure [10] [11]. The CDS procedure given in Algorith 1 is an adaptive procedure coprised of an alternating sequence of sapling or observation steps and refineent or distillation steps and for which the observations are subject to a global budget of sensing resources or sensing energy that effectively quantifies the average easureent precision. The key point is that the adaptive nature of the procedure allows for sensing resources to be allocated nonuniforly; in particular proportionally ore of the resources can be devoted to subspaces of interest as they are identified. In the jth sapling step for j = 1... k we collect easureents only at locations of x corresponding to indices in a set I j where I 1 = {1... n} initially. The jth refineent step for j = 1... k 1 identifies the set of locations I j+1 I j for which the corresponding signal coponents are to be easured in step j + 1. It is clear that in order to leverage the benefit of adaptivity the distillation step should have the property that I j+1 contains ost or ideally all of the indices in I j that correspond to true signal coponents. In addition and perhaps ore iportantly we also want the set I j+1 to be significantly saller than I j since in that case we can realize an SNR iproveent fro focusing our sensing resources into the appropriate subspace. In the DS procedure exained in [10] [11] observations were in the for of noisy saples of x at any location i {1... n} at each step j. In that case it was shown a siple refineent operation identifying all locations for which the corresponding observation exceeded a threshold was sufficient to ensure that with high probability I j+1 would contain ost of the indices in I j corresponding to true signal coponents but only about half of the reaining indices even when the signal is very weak. On the other hand here we utilize a copressive sensing observation odel where at each step the observations are in the for of a lowdiensional vector y R with n. In an attept to iic the uncopressed case here we propose a siilar refineent step applied to the back-projection estiate A j T y j = x j R n which can essentially be thought of as one of any possible estiates or reconstructions of x that can be obtained fro y j and A j. The results in the next section quantify the iproveents that can be achieved using this approach. III. MAIN RESULTS To state our ain results we set the input paraeters of Algorith 1 as follows. Choose α 0 1/3 let b = 1 α/1 2α and let k = 1 + log b log n. Allocate sensing resources according to R j = { αn 1 2α 1 α αn j 1 j = 1... k 1 j = k and note that this allocation guarantees that R j+1 /R j > 1/2 and k j=1 Rj n. The latter inequality ensures that the total sensing energy does not exceed the total sensing energy used in conventional CS. The nuber of easureents acquired in each step is { } j ρ0 s log n/k 1 j = 1... k 1 = ρ 1 s log n j = k for soe constants ρ 0 which depends on the dynaic range and ρ 1 sufficiently large so that the results of Lea 1 hold. Note that = Os log n the sae order as the iniu nuber of easureents required by conventional CS. }

3 3 Our ain result of the paper stated below and proved in the Appendix quantifies the error perforance of one particular estiate obtained fro adaptive observations collected using the CDS procedure. Theore 1. Assue that x R n is sparse with s = n β/ log log n for soe constant 0 < β < 1. Furtheore assue that each non-zero coponent of x satisfies σµ x i Dσµ for soe µ > 0. Here σ is the noise standard deviation D > 1 is the dynaic range of the signal and µ 2 is the SNR. Adaptively easure x according to Algorith 1 with the input paraeters as specified above and construct the estiator x CDS by applying the Dantzig selector with λ = Θσ to the output of the algorith i.e. with A = A k and y = y k. 1 There exists µ 0 = Ω log n/ log log n such that if µ µ 0 then x CDS x 2 l 2 = Osσ 2 with probability 1 On C 0 / log log n for soe C 0 > 0. 2 There exists µ 1 = Ω log log log n such that if µ 1 µ < µ 0 then x CDS x 2 l 2 = Osσ 2 with probability 1 Oe C 1 µ2 for soe C 1 > 0. 3 If µ < µ 1 then x CDS x 2 l 2 = Osσ 2 log log log n with probability 1 On C 2 for soe C 2 > 0. In words when the SNR is sufficiently large the estiate achieves the error perforance of the oracle-assisted estiator albeit with a lower slightly sub-polynoial convergence rate. For a class of slightly weaker signals the oracle-assisted error perforance is still achieved but with a rate of convergence that is inversely proportional to the SNR. Note that we ay suarize the results of the theore with the general clai x CDS x 2 l 2 = Osσ 2 log log log n with probability 1 o1. It is worth pointing out that for any probles of practical interest the log log log n ter can be negligible whereas log n is not; for exaple log log log10 6 < 1 but log IV. EXTENSIONS AND CONCLUSIONS Although the CDS procedure was specified under the assuption that the nonzero signal coponents were positive it can be easily extended to signals having negative entries as well. In that case one could split the budget of sensing resources in half executing the procedure once as written and again replacing the refineent step by I j+1 = {i I j : x j i < 0}. In addition the results presented here also apply if the signal is sparse another basis. To ipleent the procedure in that case one would generate the A j as above but observations of x would be obtained using A j T where T R n n is an appropriate orthonoral transforation atrix discrete wavelet or cosine transfor for exaple. In either case the qualitative behavior is the sae observations are collected by projecting x onto a superposition of basis eleents fro the appropriate basis. We have shown here that the copressive distilled sensing procedure can significantly iprove the theoretical perforance of copressive sensing. In experients not shown here due to space liitations we have found that CDS can perfor significantly better than CS in practice like siilar previously proposed adaptive ethods [7] [9]. We reark that our theoretical analysis shows that CDS is sensitive to the dynaic range of the signal. This is an artifact of the ethod for obtaining the signal estiate x j at each step. As alluded at the end of Section II x j could be obtained using any of a nuber of ethods including for exaple Dantzig selector estiation with a saller value of λ or other ixed-nor reconstruction techniques such as LASSO with sufficiently sall regularization paraeters. Such extensions will be explored in future work. A. Leata V. APPENDIX We first establish several key leata that will be used in the sketch of the proof of the ain result. In particular the first two results presented below quantify the effects of each refineent step. Lea 2. Let x R n be supported on S with S = s and let x S denote the subvector of x coposed of entries of x whose indices are in S. Let A be an n atrix whose entries are i.i.d. N 0 τ/ for soe 0 < τ in τ and let A S and A S c be subatrices of A coposed of the coluns of A corresponding to the indices in the sets S and S c respectively. Let w R be independent of A and have i.i.d. N 0 σ 2 entries. For the z 1 vector U = A T S ca Sx S + A T Scw where z = S c = n s we have 1/2 ɛ z z j=1 1 {U i>0} 1/2 + ɛ z for any ɛ 0 1/2 with probability at least 1 2 exp 2ɛ 2 z. Proof: Define Y = Ax + w = A S x S + w and note that given Y the entries of U = A T S cy are i.i.d. N 0 Y 2 2τ/. Thus when Y 0 we have PrU i > 0 = 1/2 for all i = 1... z. Let T i = 1 {Ui>0} and apply Hoeffding s inequality to obtain that for any ɛ 0 1/2 z i z 2 > ɛz Y : Y 0 2 exp 2ɛ 2 z. Now we integrate to obtain z i z 2 > ɛz 2 exp 2ɛ 2 z dp Y + Y :Y 0 2 exp 2ɛ 2 z. Y :Y =0 1 dp Y The last result follows fro the fact that the event Y = 0 has probability zero since Y is Gaussian-distributed. Lea 3. Let x S x S A A S and w be as defined in the previous lea. Assue further that the entries of x satisfy σµ x i Dσµ for i S for soe µ > 0 and fixed D > 1. Define = exp 32 sd 2 + µ 2 < 1 /τ in then for the s 1 vector V = A T S A Sx S + A T S w either of the following bounds are valid: s Pr 1 {Vi>0} s 2s 2

4 4 or s Pr 1 {Vi>0} < s Proof: Given A i the ith colun of A we have V i N A i 2 l 2 x i A i 2 l 2 τ s x 2 j + σ 2 j=1 j i and so by a standard Gaussian tail bound A i l2 x i PrV i 0 A i = Pr N 0 1 > τ s j=1 x 2 j + σ2 j i A i 2 l exp 2 x 2 i 2τ x 2 / + σ 2 Now we can leverage a result on the tails of a chi-squared rando variable fro [12] to obtain that for any γ 0 1 Pr A i 2 1 γτ exp γ 2 /4. Again we eploy conditioning to obtain PrV i 0 1 dp Ai A i : A i 2 1 γτ + PrV i 0 A i dp Ai A i : A i 2 >1 γτ exp γ2 + exp τ1 γx2 i 4 exp γ2 4 + exp 2τ x 2 / + σ 2 τ1 γµ 2 2τsD 2 µ 2 / + 1 where the last bound follows fro the conditions on the x i. Now to siplify we choose γ = γ 0 1 to balance the two ters obtaining γ = sd τµ sd 2 + τµ 2 1. Using the fact that 1 + 2t 1 t > 1 2 t for t > 1 we can conclude γ > 1 sd 2 + 1/2 2 τµ 2 since s > 1 by assuption. Now using the fact that τ τ in we have that PrV i where = exp 32 sd 2 + µ 2. /τ in The first result follows fro s s Pr 1 {Vi >0} s = Pr {V i 0} s ax i {1...s} Pr V i 0 2s 2. For the second result let us siplify notation by introducing the variables T i = 1 {Vi>0} and t i = E [T i ]. By Markov s Inequality we have [ ] s s s s i t i > p p 1 E T i t i Now note that T i t i = s p 1 E [ T i t i ] p 1 s ax i {1...s} E [ T i t i ]. { 1 P Vi > 0 V i > 0 P V i > 0 V i 0 and so E [ T i t i ] 2P V i 0. Thus we have that ax i {1...s} E [ T i t i ] = 2 2 and so s s i t i > p 4p 1 s 2. Now let p = s to obtain s i < s t i s 4. Since t i = 1 Pr V i 0 we have s t i s1 2 2 and thus s i < s The result follows fro the fact that < 3. Lea 4. For 0 < p < 1 and q > 0 we have 1 p q 1 qp/1 p. Proof: We have log 1 p q = q log 1 p = q log 1 + p/1 p qp/1 p where the last bound follows fro the fact that log 1 + t t for t 0. Thus 1 p q exp qp/1 p 1 qp/1 p the last bound following fro the fact e t 1 + t for all t R. B. Sketch of Proof of Theore 1 A j A j S j T w j and To establish the ain results of the paper we will first show that the final set of observations of the CDS procedure is with high probability equivalent in distribution to a set of observations of the for 1 but with different paraeters saller effective diension n eff and effective noise power σeff 2 and for which soe fraction of the original signal coponents ay be absent. To that end let S j = S I j and Z j = S c I j for j = 1... k denote the subsets of indices of S and its copleent respectively that reain to be easured in step j. Note that at each step of the procedure the back-projection estiate x j = A j T A j x + A j T w j can be deco- T j posed into x S j = S A x j S j S j + T x Z j = A j j Z A x j S j S j + subvectors are precisely of the for specified in the conditions of Leas 2 and 3. A j Z j T w j and that these

5 5 Let z j = Z j and s j = S j and in particular note that s 1 = s and z 1 = z = n s. Choose the paraeters of the CDS algorith as specified in Section III. Iteratively applying Lea 2 we have that for any fixed ɛ 0 1/2 the bounds 1/2 ɛ j 1 z z j 1/2 + ɛ j 1 z hold siultaneously for all j = k with probability at least 1 2k 1 exp 2zɛ 2 1/2 ɛ k 2 which is no less than 1 O exp c 0 n/ log c 1 n for soe constants c 0 > 0 and c 1 > 0 for n sufficiently large 2. As a result with the sae probability the total nuber of locations in the set I j satisfies I j s 1 +z ɛ j 1 for all j = k. Thus we can lower bound τ j = R j / I j at each step by τ j αn1 2α/1 α j 1 s+z1+2ɛ/2 j 1 j = 1... k 1 αn s+z1+2ɛ/2 j 1 j = k. Now note that when n is sufficiently large 3 we have s z 1/2 + ɛ j 1 holding for all j = 1... k. Letting ɛ = 1 3α/2 2α we can siplify the bounds on τ j to obtain that τ j α/2 for j = 1... k 1 and τ k α log n/2. The salient point to note here is the value of τ k and in particular its dependence on the signal diension n. This essentially follows fro the fact that the set of indices to easure decreases by a fixed factor with each distillation step and so after Olog log n steps the nuber of indices to easure is saller than in the initial step by a factor of about log n. Thus for the sae allocation of resources R 1 = R k the SNR of the final set of observations is larger than that of the first set by a factor of log n. Now the final set of observations is y k = A k x k +w k where x k R n eff for soe n eff < n is supported on the set S k = S I k A k is an k n eff atrix and the w i are i.i.d. N 0 σ 2. We can divide throughout by τ k to obtain the equivalent stateent ỹ = à x + w where now the entries of à are i.i.d. N 0 1/ and the w i are i.i.d. N 0 σ 2 where σ 2 2σ 2 /α log n. To bound the overall squared error we consider the variance associated with estiating the coponents of x using the Dantzig selector cf. Lea 1 as well as the squared bias arising fro the fact that soe signal coponents ay not be present in the final support set S k. In particular a bound for the overall error is given by x x 2 l 2 = x x + x x 2 l 2 2 x x 2 l x x 2 l 2. We can bound the first ter by applying the result of Lea 1 to obtain that for ρ 1 sufficiently large x x 2 l 2 = Osσ 2 holds with probability 1 On C0 for soe C 0 > 0. Now let δ = S S k /s denote the fraction of true signal coponents that are rejected by the CDS procedure. Then we have x x 2 l 2 = Osσ 2 δµ 2 and so overall we have x x 2 l 2 = Osσ 2 + sσ 2 δµ 2 with probability 1 On C0. The ethod for bounding the second ter in the error bound varies 2 In particular we require n c 0 log log log nlog nc 1/1 n c 2 / log log n 1 where c 0 c 1 and c 2 are positive functions of ɛ and β. 3 In particular we require n 1 + log n log log n/log log n β. depending on the signal aplitude µ; we consider three cases below. 1 µ 8D 3/α log n/ log log n: Conditioned on the event that the stated lower-bounds for τ j are valid we can iteratively apply Lea 3 taking τ in = α/2. For ρ 0 = 96D 2 / log b where b is the paraeter fro the expression for k let j = ρ 0 s log n/ log b log n. Then we obtain that for all n sufficiently large δ = 0 with probability at least 1 On C 0 / log log n for soe constant C 0 > 0. Since this ter governs the rate we have overall that x x 2 l 2 = Osσ 2 holds with probability 1 On C 0 / log log n as claied /α log b log log log n µ < 8D 3/α log n/ log log n: For this range of signal aplitude we will need to control δ explicitly. Conditioned on the event that the lower-bounds for τ j hold we iteratively apply Lea 3 where for ρ 0 = 96D 2 / log b we let j = ρ 0 s log n/ log b log n. Now we invoke Lea 4 to obtain that for n sufficiently large δ = k 1 = Oe C 1 µ2 with probability at least 1 Oe C 1 µ2 for soe C 1 > 0. It follows that δµ 2 is O1 and so overall x x 2 l 2 = Osσ 2 with probability 1 Oe C 1 µ2. 3 µ < 16 2/α log b log log log n: Invoking the trivial bound δ 1 it follows fro above that for n sufficiently large the error satisfies x x 2 l 2 = Osσ 2 log log log n with probability 1 On C 2 for soe constant C 2 > 0 as claied. REFERENCES [1] E. J. Candès J. Roberg and T. Tao Robust uncertainty principles: Exact signal reconstruction fro highly incoplete frequency inforation IEEE Trans. Infor. Theory vol. 52 no. 2 pp Feb [2] D. L. Donoho Copressed sensing IEEE Trans. Infor. Theory vol. 52 no. 4 pp Apr [3] E. J. Candès and T. Tao Near-optial signal recovery fro rando projections: Universal encoding strategies? IEEE Trans. Infor. Theory vol. 52 no. 12 pp Dec [4] R. Baraniuk M. Davenport R. A. DeVore and M. Wakin A siple proof of the restricted isoetry property for rando atrices Constructive Approxiation [5] J. Haupt and R. Nowak Signal reconstruction fro noisy rando projections IEEE Trans. Infor. Theory vol. 52 no. 9 pp Sept [6] E. J. Candès and T. Tao The Dantzig selector: Statistical estiation when p is uch larger than n Ann. Statist. vol. 35 no. 6 pp Dec [7] S. Ji Y. Xue and L. Carin Bayesian copressive sensing IEEE Trans. Signal Processing vol. 56 no. 6 pp June [8] R. Castro J. Haupt R. Nowak and G. Raz Finding needles in noisy haystacks in Proc. IEEE Conf. Acoustics Speech and Signal Proc. Honolulu HI Apr pp [9] J. Haupt R. Castro and R. Nowak Adaptive sensing for sparse signal recovery in Proc. IEEE 13th Digital Sig. Proc./5th Sig. Proc. Education Workshop Marco Island FL Jan pp [10] J. Haupt R. Castro and R. Nowak Adaptive discovery of sparse signals in noise in Proc. 42nd Asiloar Conf. on Signals Systes and Coputers Pacific Grove CA Oct pp [11] J. Haupt R. Castro and R. Nowak Distilled sensing: Selective sapling for sparse signal recovery in Proc. 12th International Conference on Artificial Intelligence and Statistics AISTATS Clearwater Beach FL Apr pp [12] B. Laurent and P. Massart Adaptive estiation of a quadratic functional by odel selection Ann. Statist. vol. 28 no. 5 pp Oct

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

Weighted- 1 minimization with multiple weighting sets

Weighted- 1 minimization with multiple weighting sets Weighted- 1 iniization with ultiple weighting sets Hassan Mansour a,b and Özgür Yılaza a Matheatics Departent, University of British Colubia, Vancouver - BC, Canada; b Coputer Science Departent, University

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Support recovery in compressed sensing: An estimation theoretic approach

Support recovery in compressed sensing: An estimation theoretic approach Support recovery in copressed sensing: An estiation theoretic approach Ain Karbasi, Ali Horati, Soheil Mohajer, Martin Vetterli School of Coputer and Counication Sciences École Polytechnique Fédérale de

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Hamming Compressed Sensing

Hamming Compressed Sensing Haing Copressed Sensing Tianyi Zhou, and Dacheng Tao, Meber, IEEE Abstract arxiv:.73v2 [cs.it] Oct 2 Copressed sensing CS and -bit CS cannot directly recover quantized signals and require tie consuing

More information

On the Use of A Priori Information for Sparse Signal Approximations

On the Use of A Priori Information for Sparse Signal Approximations ITS TECHNICAL REPORT NO. 3/4 On the Use of A Priori Inforation for Sparse Signal Approxiations Oscar Divorra Escoda, Lorenzo Granai and Pierre Vandergheynst Signal Processing Institute ITS) Ecole Polytechnique

More information

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion Suppleentary Material for Fast and Provable Algoriths for Spectrally Sparse Signal Reconstruction via Low-Ran Hanel Matrix Copletion Jian-Feng Cai Tianing Wang Ke Wei March 1, 017 Abstract We establish

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

Highly Robust Error Correction by Convex Programming

Highly Robust Error Correction by Convex Programming Highly Robust Error Correction by Convex Prograing Eanuel J. Candès and Paige A. Randall Applied and Coputational Matheatics, Caltech, Pasadena, CA 9115 Noveber 6; Revised Noveber 7 Abstract This paper

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

SPECTRUM sensing is a core concept of cognitive radio

SPECTRUM sensing is a core concept of cognitive radio World Acadey of Science, Engineering and Technology International Journal of Electronics and Counication Engineering Vol:6, o:2, 202 Efficient Detection Using Sequential Probability Ratio Test in Mobile

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

A Simple Homotopy Algorithm for Compressive Sensing

A Simple Homotopy Algorithm for Compressive Sensing A Siple Hootopy Algorith for Copressive Sensing Lijun Zhang Tianbao Yang Rong Jin Zhi-Hua Zhou National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, China Departent of Coputer

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

Recovering Block-structured Activations Using Compressive Measurements

Recovering Block-structured Activations Using Compressive Measurements Recovering Block-structured Activations Using Copressive Measureents Sivaraan Balakrishnan, Mladen Kolar, Alessandro Rinaldo, and Aarti Singh Abstract We consider the probles of detection and support recovery

More information

A PROBABILISTIC AND RIPLESS THEORY OF COMPRESSED SENSING. Emmanuel J. Candès Yaniv Plan. Technical Report No November 2010

A PROBABILISTIC AND RIPLESS THEORY OF COMPRESSED SENSING. Emmanuel J. Candès Yaniv Plan. Technical Report No November 2010 A PROBABILISTIC AND RIPLESS THEORY OF COMPRESSED SENSING By Eanuel J Candès Yaniv Plan Technical Report No 200-0 Noveber 200 Departent of Statistics STANFORD UNIVERSITY Stanford, California 94305-4065

More information

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate The Siplex Method is Strongly Polynoial for the Markov Decision Proble with a Fixed Discount Rate Yinyu Ye April 20, 2010 Abstract In this note we prove that the classic siplex ethod with the ost-negativereduced-cost

More information

Interactive Markov Models of Evolutionary Algorithms

Interactive Markov Models of Evolutionary Algorithms Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary

More information

Lecture 20 November 7, 2013

Lecture 20 November 7, 2013 CS 229r: Algoriths for Big Data Fall 2013 Prof. Jelani Nelson Lecture 20 Noveber 7, 2013 Scribe: Yun Willia Yu 1 Introduction Today we re going to go through the analysis of atrix copletion. First though,

More information

An l 1 Regularized Method for Numerical Differentiation Using Empirical Eigenfunctions

An l 1 Regularized Method for Numerical Differentiation Using Empirical Eigenfunctions Journal of Matheatical Research with Applications Jul., 207, Vol. 37, No. 4, pp. 496 504 DOI:0.3770/j.issn:2095-265.207.04.0 Http://jre.dlut.edu.cn An l Regularized Method for Nuerical Differentiation

More information

Optimal Jamming Over Additive Noise: Vector Source-Channel Case

Optimal Jamming Over Additive Noise: Vector Source-Channel Case Fifty-first Annual Allerton Conference Allerton House, UIUC, Illinois, USA October 2-3, 2013 Optial Jaing Over Additive Noise: Vector Source-Channel Case Erah Akyol and Kenneth Rose Abstract This paper

More information

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Tie-Varying Jaing Links Jun Kurihara KDDI R&D Laboratories, Inc 2 5 Ohara, Fujiino, Saitaa, 356 8502 Japan Eail: kurihara@kddilabsjp

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup) Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

Compressive Sensing Over Networks

Compressive Sensing Over Networks Forty-Eighth Annual Allerton Conference Allerton House, UIUC, Illinois, USA Septeber 29 - October, 200 Copressive Sensing Over Networks Soheil Feizi MIT Eail: sfeizi@it.edu Muriel Médard MIT Eail: edard@it.edu

More information

On the theoretical analysis of cross validation in compressive sensing

On the theoretical analysis of cross validation in compressive sensing MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.erl.co On the theoretical analysis of cross validation in copressive sensing Zhang, J.; Chen, L.; Boufounos, P.T.; Gu, Y. TR2014-025 May 2014 Abstract

More information

Introduction to Machine Learning. Recitation 11

Introduction to Machine Learning. Recitation 11 Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,

More information

A Probabilistic and RIPless Theory of Compressed Sensing

A Probabilistic and RIPless Theory of Compressed Sensing A Probabilistic and RIPless Theory of Copressed Sensing Eanuel J Candès and Yaniv Plan 2 Departents of Matheatics and of Statistics, Stanford University, Stanford, CA 94305 2 Applied and Coputational Matheatics,

More information

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. IX Uncertainty Models For Robustness Analysis - A. Garulli, A. Tesi and A. Vicino

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. IX Uncertainty Models For Robustness Analysis - A. Garulli, A. Tesi and A. Vicino UNCERTAINTY MODELS FOR ROBUSTNESS ANALYSIS A. Garulli Dipartiento di Ingegneria dell Inforazione, Università di Siena, Italy A. Tesi Dipartiento di Sistei e Inforatica, Università di Firenze, Italy A.

More information

An RIP-based approach to Σ quantization for compressed sensing

An RIP-based approach to Σ quantization for compressed sensing An RIP-based approach to Σ quantization for copressed sensing Joe-Mei Feng and Felix Kraher October, 203 Abstract In this paper, we provide new approach to estiating the error of reconstruction fro Σ quantized

More information

Distributed Subgradient Methods for Multi-agent Optimization

Distributed Subgradient Methods for Multi-agent Optimization 1 Distributed Subgradient Methods for Multi-agent Optiization Angelia Nedić and Asuan Ozdaglar October 29, 2007 Abstract We study a distributed coputation odel for optiizing a su of convex objective functions

More information

Lecture 21. Interior Point Methods Setup and Algorithm

Lecture 21. Interior Point Methods Setup and Algorithm Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and

More information

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths

More information

Exact tensor completion with sum-of-squares

Exact tensor completion with sum-of-squares Proceedings of Machine Learning Research vol 65:1 54, 2017 30th Annual Conference on Learning Theory Exact tensor copletion with su-of-squares Aaron Potechin Institute for Advanced Study, Princeton David

More information

3.3 Variational Characterization of Singular Values

3.3 Variational Characterization of Singular Values 3.3. Variational Characterization of Singular Values 61 3.3 Variational Characterization of Singular Values Since the singular values are square roots of the eigenvalues of the Heritian atrices A A and

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory Proble sets 5 and 6 Due: Noveber th Please send your solutions to learning-subissions@ttic.edu Notations/Definitions Recall the definition of saple based Radeacher

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

A Generalized Restricted Isometry Property

A Generalized Restricted Isometry Property 1 A Generalized Restricted Isometry Property Jarvis Haupt and Robert Nowak Department of Electrical and Computer Engineering, University of Wisconsin Madison University of Wisconsin Technical Report ECE-07-1

More information

Detection and Estimation Theory

Detection and Estimation Theory ESE 54 Detection and Estiation Theory Joseph A. O Sullivan Sauel C. Sachs Professor Electronic Systes and Signals Research Laboratory Electrical and Systes Engineering Washington University 11 Urbauer

More information

Principal Components Analysis

Principal Components Analysis Principal Coponents Analysis Cheng Li, Bingyu Wang Noveber 3, 204 What s PCA Principal coponent analysis (PCA) is a statistical procedure that uses an orthogonal transforation to convert a set of observations

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

arxiv: v5 [cs.it] 16 Mar 2012

arxiv: v5 [cs.it] 16 Mar 2012 ONE-BIT COMPRESSED SENSING BY LINEAR PROGRAMMING YANIV PLAN AND ROMAN VERSHYNIN arxiv:09.499v5 [cs.it] 6 Mar 0 Abstract. We give the first coputationally tractable and alost optial solution to the proble

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

arxiv: v1 [cs.ds] 17 Mar 2016

arxiv: v1 [cs.ds] 17 Mar 2016 Tight Bounds for Single-Pass Streaing Coplexity of the Set Cover Proble Sepehr Assadi Sanjeev Khanna Yang Li Abstract arxiv:1603.05715v1 [cs.ds] 17 Mar 2016 We resolve the space coplexity of single-pass

More information

On Conditions for Linearity of Optimal Estimation

On Conditions for Linearity of Optimal Estimation On Conditions for Linearity of Optial Estiation Erah Akyol, Kuar Viswanatha and Kenneth Rose {eakyol, kuar, rose}@ece.ucsb.edu Departent of Electrical and Coputer Engineering University of California at

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

Efficient Filter Banks And Interpolators

Efficient Filter Banks And Interpolators Efficient Filter Banks And Interpolators A. G. DEMPSTER AND N. P. MURPHY Departent of Electronic Systes University of Westinster 115 New Cavendish St, London W1M 8JS United Kingdo Abstract: - Graphical

More information

Testing equality of variances for multiple univariate normal populations

Testing equality of variances for multiple univariate normal populations University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Inforation Sciences 0 esting equality of variances for ultiple univariate

More information

Using a De-Convolution Window for Operating Modal Analysis

Using a De-Convolution Window for Operating Modal Analysis Using a De-Convolution Window for Operating Modal Analysis Brian Schwarz Vibrant Technology, Inc. Scotts Valley, CA Mark Richardson Vibrant Technology, Inc. Scotts Valley, CA Abstract Operating Modal Analysis

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

Weighted Superimposed Codes and Constrained Integer Compressed Sensing

Weighted Superimposed Codes and Constrained Integer Compressed Sensing Weighted Superiposed Codes and Constrained Integer Copressed Sensing Wei Dai and Olgica Milenovic Dept. of Electrical and Coputer Engineering University of Illinois, Urbana-Chapaign Abstract We introduce

More information

Hybrid System Identification: An SDP Approach

Hybrid System Identification: An SDP Approach 49th IEEE Conference on Decision and Control Deceber 15-17, 2010 Hilton Atlanta Hotel, Atlanta, GA, USA Hybrid Syste Identification: An SDP Approach C Feng, C M Lagoa, N Ozay and M Sznaier Abstract The

More information

Physics 215 Winter The Density Matrix

Physics 215 Winter The Density Matrix Physics 215 Winter 2018 The Density Matrix The quantu space of states is a Hilbert space H. Any state vector ψ H is a pure state. Since any linear cobination of eleents of H are also an eleent of H, it

More information

A Nonlinear Sparsity Promoting Formulation and Algorithm for Full Waveform Inversion

A Nonlinear Sparsity Promoting Formulation and Algorithm for Full Waveform Inversion A Nonlinear Sparsity Prooting Forulation and Algorith for Full Wavefor Inversion Aleksandr Aravkin, Tristan van Leeuwen, Jaes V. Burke 2 and Felix Herrann Dept. of Earth and Ocean sciences University of

More information

Computational and Statistical Learning Theory

Computational and Statistical Learning Theory Coputational and Statistical Learning Theory TTIC 31120 Prof. Nati Srebro Lecture 2: PAC Learning and VC Theory I Fro Adversarial Online to Statistical Three reasons to ove fro worst-case deterinistic

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Randomized Accuracy-Aware Program Transformations For Efficient Approximate Computations

Randomized Accuracy-Aware Program Transformations For Efficient Approximate Computations Randoized Accuracy-Aware Progra Transforations For Efficient Approxiate Coputations Zeyuan Allen Zhu Sasa Misailovic Jonathan A. Kelner Martin Rinard MIT CSAIL zeyuan@csail.it.edu isailo@it.edu kelner@it.edu

More information

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters journal of ultivariate analysis 58, 96106 (1996) article no. 0041 The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Paraeters H. S. Steyn

More information

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay A Low-Coplexity Congestion Control and Scheduling Algorith for Multihop Wireless Networks with Order-Optial Per-Flow Delay Po-Kai Huang, Xiaojun Lin, and Chih-Chun Wang School of Electrical and Coputer

More information

Lecture 9 November 23, 2015

Lecture 9 November 23, 2015 CSC244: Discrepancy Theory in Coputer Science Fall 25 Aleksandar Nikolov Lecture 9 Noveber 23, 25 Scribe: Nick Spooner Properties of γ 2 Recall that γ 2 (A) is defined for A R n as follows: γ 2 (A) = in{r(u)

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE Proceedings of ICIPE rd International Conference on Inverse Probles in Engineering: Theory and Practice June -8, 999, Port Ludlow, Washington, USA : RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS

More information

Supplementary to Learning Discriminative Bayesian Networks from High-dimensional Continuous Neuroimaging Data

Supplementary to Learning Discriminative Bayesian Networks from High-dimensional Continuous Neuroimaging Data Suppleentary to Learning Discriinative Bayesian Networks fro High-diensional Continuous Neuroiaging Data Luping Zhou, Lei Wang, Lingqiao Liu, Philip Ogunbona, and Dinggang Shen Proposition. Given a sparse

More information

The linear sampling method and the MUSIC algorithm

The linear sampling method and the MUSIC algorithm INSTITUTE OF PHYSICS PUBLISHING INVERSE PROBLEMS Inverse Probles 17 (2001) 591 595 www.iop.org/journals/ip PII: S0266-5611(01)16989-3 The linear sapling ethod and the MUSIC algorith Margaret Cheney Departent

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Estimating Parameters for a Gaussian pdf

Estimating Parameters for a Gaussian pdf Pattern Recognition and achine Learning Jaes L. Crowley ENSIAG 3 IS First Seester 00/0 Lesson 5 7 Noveber 00 Contents Estiating Paraeters for a Gaussian pdf Notation... The Pattern Recognition Proble...3

More information

Complexity reduction in low-delay Farrowstructure-based. filters utilizing linear-phase subfilters

Complexity reduction in low-delay Farrowstructure-based. filters utilizing linear-phase subfilters Coplexity reduction in low-delay Farrowstructure-based variable fractional delay FIR filters utilizing linear-phase subfilters Air Eghbali and Håkan Johansson Linköping University Post Print N.B.: When

More information

Fixed-to-Variable Length Distribution Matching

Fixed-to-Variable Length Distribution Matching Fixed-to-Variable Length Distribution Matching Rana Ali Ajad and Georg Böcherer Institute for Counications Engineering Technische Universität München, Gerany Eail: raa2463@gail.co,georg.boecherer@tu.de

More information

CHAPTER 19: Single-Loop IMC Control

CHAPTER 19: Single-Loop IMC Control When I coplete this chapter, I want to be able to do the following. Recognize that other feedback algoriths are possible Understand the IMC structure and how it provides the essential control features

More information

PAC-Bayes Analysis Of Maximum Entropy Learning

PAC-Bayes Analysis Of Maximum Entropy Learning PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E

More information

Identical Maximum Likelihood State Estimation Based on Incremental Finite Mixture Model in PHD Filter

Identical Maximum Likelihood State Estimation Based on Incremental Finite Mixture Model in PHD Filter Identical Maxiu Lielihood State Estiation Based on Increental Finite Mixture Model in PHD Filter Gang Wu Eail: xjtuwugang@gail.co Jing Liu Eail: elelj20080730@ail.xjtu.edu.cn Chongzhao Han Eail: czhan@ail.xjtu.edu.cn

More information

Multi-Scale/Multi-Resolution: Wavelet Transform

Multi-Scale/Multi-Resolution: Wavelet Transform Multi-Scale/Multi-Resolution: Wavelet Transfor Proble with Fourier Fourier analysis -- breaks down a signal into constituent sinusoids of different frequencies. A serious drawback in transforing to the

More information

In this chapter, we consider several graph-theoretic and probabilistic models

In this chapter, we consider several graph-theoretic and probabilistic models THREE ONE GRAPH-THEORETIC AND STATISTICAL MODELS 3.1 INTRODUCTION In this chapter, we consider several graph-theoretic and probabilistic odels for a social network, which we do under different assuptions

More information

A Smoothed Boosting Algorithm Using Probabilistic Output Codes

A Smoothed Boosting Algorithm Using Probabilistic Output Codes A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu

More information

An Approximate Model for the Theoretical Prediction of the Velocity Increase in the Intermediate Ballistics Period

An Approximate Model for the Theoretical Prediction of the Velocity Increase in the Intermediate Ballistics Period An Approxiate Model for the Theoretical Prediction of the Velocity... 77 Central European Journal of Energetic Materials, 205, 2(), 77-88 ISSN 2353-843 An Approxiate Model for the Theoretical Prediction

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS An Iproved Method for Bandwidth Selection When Estiating ROC Curves Peter G Hall and Rob J Hyndan Working Paper 11/00 An iproved

More information

MEASUREMENT MATRIX DESIGN FOR COMPRESSIVE SENSING WITH SIDE INFORMATION AT THE ENCODER

MEASUREMENT MATRIX DESIGN FOR COMPRESSIVE SENSING WITH SIDE INFORMATION AT THE ENCODER MEASUREMENT MATRIX DESIGN FOR COMPRESSIVE SENSING WITH SIDE INFORMATION AT THE ENCODER Pingfan Song João F. C. Mota Nikos Deligiannis Miguel Raul Dias Rodrigues Electronic and Electrical Engineering Departent,

More information

Recovery of Sparsely Corrupted Signals

Recovery of Sparsely Corrupted Signals TO APPEAR IN IEEE TRANSACTIONS ON INFORMATION TEORY 1 Recovery of Sparsely Corrupted Signals Christoph Studer, Meber, IEEE, Patrick Kuppinger, Student Meber, IEEE, Graee Pope, Student Meber, IEEE, and

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

arxiv: v3 [quant-ph] 18 Oct 2017

arxiv: v3 [quant-ph] 18 Oct 2017 Self-guaranteed easureent-based quantu coputation Masahito Hayashi 1,, and Michal Hajdušek, 1 Graduate School of Matheatics, Nagoya University, Furocho, Chikusa-ku, Nagoya 464-860, Japan Centre for Quantu

More information

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information