Solving Most Systems of Random Quadratic Equations

Size: px
Start display at page:

Download "Solving Most Systems of Random Quadratic Equations"

Transcription

1 Solving Most Systes of Rando Quadratic Equations Gang Wang, Georgios B. Giannakis Yousef Saad Jie Chen Digital Tech. Center & Dept. of Electrical and Coputer Eng., Univ. of Minnesota Key Lab of Intell. Contr. and Decision of Coplex Syst., Beijing Inst. of Technology Departent of Coputer Science and Engineering, Univ. of Minnesota {gangwang, georgios, Abstract This paper deals with finding an n-diensional solution x to a syste of quadratic equations y i = a i, x 2, 1 i, which in general is known to be NP-hard. We put forth a novel procedure, that starts with a weighted axial correlation initialization obtainable with a few power iterations, followed by successive refineents based on iteratively reweighted gradient-type iterations. The novel techniques distinguish theselves fro prior works by the inclusion of a fresh (re)weighting regularization. For certain rando easureent odels, the proposed procedure returns the true solution x with high probability in tie proportional to reading the data {(a i ; y i )} 1 i, provided that the nuber of equations is soe constant c > 0 ties the nuber n of unknowns, that is, cn. Epirically, the upshots of this contribution are: i) perfect signal recovery in the high-diensional regie given only an inforation-theoretic liit nuber of equations; and, ii) (near-)optial statistical accuracy in the presence of additive noise. Extensive nuerical tests using both synthetic data and real iages corroborate its iproved signal recovery perforance and coputational efficiency relative to state-of-the-art approaches. 1 Introduction One is often faced with solving quadratic equations of the for y i = a i, x 2, or equivalently, ψ i = a i, x, 1 i (1) where x R n /C n (hereafter, sybol A/B denotes either A or B) is the wanted unknown n 1 vector, while given observations ψ i and feature vectors a i R n /C n that are collectively stacked in the data vector ψ := [ψ i ] 1 i and the n sensing atrix A := [a i ] 1 i, respectively. Put differently, given inforation about the (squared) odulus of the inner products of the signal vector x and several known design vectors a i, can one reconstruct exactly (up to a global phase factor) x, or alternatively, the issing phase of a i, x? In fact, uch effort has been devoted to deterining the nuber of such equations necessary and/or sufficient for the uniqueness of the solution x; see e.g., [1, 8]. It has been proved that 2n 1 ( 4n 4) generic 1 (which includes the case of rando vectors) real (coplex) vectors a i are sufficient for uniquely deterining an n-diensional real (coplex) vector x [1, Theore 2.8], [8], while in the real case = 2n 1 is shown also necessary [1]. In this sense, the nuber = 2n 1 of equations as in (1) can be regarded as the inforation-theoretic liit for such a quadratic syste to be uniquely solvable. 1 It is out of the scope of the present paper to explain the eaning of generic vectors, whereas interested readers are referred to [1]. 31st Conference on Neural Inforation Processing Systes (NIPS 2017), Long Beach, CA, USA.

2 In diverse physical sciences and engineering fields, it is ipossible or very difficult to record phase easureents. The proble of recovering the signal or phase fro agnitude easureents only, also coonly known as phase retrieval, eerges naturally [11, 12]. Relevant application doains include e.g., X-ray crystallography, astronoy, icroscopy, ptychography, and coherent diffraction iaging [22]. In such setups, optical easureent and detection systes record solely the photon flux, which is proportional to the (squared) agnitude of the field, but not the phase. Proble (1) in its squared for, on the other hand, can be readily recast as an instance of nonconvex quadratically constrained quadratic prograing, that subsues as special cases several well-known cobinatorial optiization probles involving Boolean variables, e.g., the NP-coplete stone proble [2, Sec ]. A related task of this kind is that of estiating the ixture of linear regressions, where the latent ebership indicators can be converted into the issing phases [30]. Although of siple for and practical relevance across different fields, solving systes of nonlinear equations is arguably the ost difficult proble in all of the nuerical coputations [20, Page 355]. Notation: Lower- (upper-) case boldface letters denote vectors (atrices), e.g., a R n (A R n ). Calligraphic letters are reserved for sets. The floor operation c gives the largest integer no greater than the given real quantity c > 0, the cardinality S counts the nuber of eleents in set S, and x denotes the Euclidean nor of x. Since for any phase φ R, vectors x C n and e jφ x are indistinguishable given {ψ i } in (1), let dist(z, x) := in φ [0,2π) z xe jφ be the Euclidean distance of any estiate z C n to the solution set {e jφ x} 0 φ<2π of (1); in particular, φ = 0/π in the real case. 1.1 Prior contributions Following the least-squares (LS) criterion (which coincides with the axiu likelihood (ML) one assuing additive white Gaussian noise), the proble of solving quadratic equations can be naturally recast as an epirical loss iniization iniize z R n /C n L(z) := 1 l(z; ψ i /y i ) (2) where one can choose to work with the aplitude-based loss l(z; ψ i ) := (ψi ai,z )2 /2 [29, 31], or the intensity-based one l(z; y i ) := (yi ai,z 2 ) 2 /2 [3], and its related Poisson likelihood l(z; y i ) := y i log( a i, z 2 ) a i, z 2 [7]. Either way, the objective functional L(z) is nonconvex; hence, it is generally NP-hard and coputationally intractable to copute the ML or LS estiate. Miniizing the squared odulus-based LS loss in (2), several nuerical polynoial-tie algoriths have been devised via convex prograing for certain choices of design vectors a i [4, 26]. Such convex paradigs first rely on the atrix-lifting technique to express all squared odulus ters into linear ones in a new rank-1 atrix variable, followed by solving a convex sei-definite progra (SDP) after dropping the rank constraint. It has been established that perfect recovery and (near-)optial statistical accuracy are achieved in noiseless and noisy settings respectively with an optial-order nuber of easureents [4]. In ters of coputational efficiency however, such lifting-based convex approaches entail storing and solving for an n n sei-definite atrix fro general SDP constraints, whose coputational coplexity in the worst case scales as n 4.5 log 1 /ɛ for n [26], which is not scalable. Another recent line of convex relaxation [13], [14] reforulated the proble of phase retrieval as that of sparse signal recovery, and solved a linear progra in the natural paraeter vector doain. Although exact signal recovery can be established assuing an accurate enough anchor vector, its epirical perforance is in general not copetitive with state-of-the-art phase retrieval approaches. Recent proposals advocate suitably initialized iterative procedures for coping with certain nonconvex forulations directly; see e.g., algoriths abbreviated as AltMinPhase, (R/P)WF, (M)TWF, (S)TAF [17, 3, 7, 27, 29, 28, 31, 23, 6, 25], as well as (stochastic) prox-linear algoriths [9, 10]. These nonconvex approaches operate directly upon vector optiization variables, thus leading to significant coputational advantages over their convex counterparts. With rando features, they can be interpreted as perforing stochastic optiization over acquired exaples {(a i ; ψ i /y i )} 1 i to approxiately iniize the population risk functional L(z) := E (ai,ψ i/y i)[l(z; ψ i /y i )]. It is well docuented that iniizing nonconvex functionals is generally intractable due to existence of ultiple critical points [18]. Assuing Gaussian sensing vectors however, such nonconvex paradigs can provably locate the global optiu, several of which also achieve optial (statistical) guarantees. 2

3 Specifically, starting with a judiciously designed initial guess, successive iproveent is effected by eans of a sequence of (truncated) (generalized) gradient-type iterations given by z t+1 := z t µt i T t+1 l i (z t ; ψ i /y i ), t = 0, 1,... (3) where z t denotes the estiate returned by the algorith at the t-th iteration, µ t > 0 is learning rate that can be pre-selected or found via e.g., the backtracking line search strategy, and l(z t, ψ i /y i ) represents the (generalized) gradient of the odulus- or squared odulus-based LS loss evaluated at z t. Here, T t+1 denotes soe tie-varying index set signifying the per-iteration gradient truncation. Although they achieve optial statistical guarantees in both noiseless and noisy settings, state-of-theart (convex and nonconvex) approaches studied under Gaussian designs, epirically require stable recovery of a nuber of equations (several) ties larger than the inforation-theoretic liit [7, 3, 31]. As a atter of fact, when there are nuerously enough easureents (on the order of n up to soe polylog factors), the squared odulus-based LS functional adits benign geoetric structure in the sense that [24]: i) all local iniizers are also global; and, ii) there always exists a negative directional curvature at every saddle point. In a nutshell, the grand challenge of tackling systes of rando quadratic equations reains to develop algoriths capable of achieving perfect recovery and statistical accuracy when the nuber of easureents approaches the inforation liit. 1.2 This work Building upon but going beyond the scope of the aforeentioned nonconvex paradigs, the present paper puts forward a novel iterative linear-tie schee, naely, tie proportional to that required by the processor to scan all the data {(a i ; ψ i )} 1 i, that we ter reweighted aplitude flow, and henceforth, abbreviate as RAF. Our ethodology is capable of solving noiseless rando quadratic equations exactly, yielding an estiate of (near)-optial statistical accuracy fro noisy odulus observations. Exactness and accuracy hold with high probability and without extra assuption on the unknown signal vector x, provided that the ratio /n of the nuber of equations to that of the unknowns is larger than a certain constant. Epirically, our approach is shown able to ensure exact recovery of high-diensional unstructured signals given a inial nuber of equations, where /n in the real case can be as sall as 2. The new twist here is to leverage judiciously designed yet conceptually siple (re)weighting regularization techniques to enhance existing initializations and also gradient refineents. An inforal depiction of our RAF ethodology is given in two stages as follows, with rigorous details deferred to Section 3: S1) Weighted axial correlation initialization: Obtain an initializer z 0 axially correlated with a carefully selected subset S M := {1, 2,..., } of feature vectors a i, whose contributions toward constructing z 0 are judiciously weighted by suitable paraeters {wi 0 > 0} i S. S2) Iteratively reweighted gradient-like iterations: Loop over 0 t T : z t+1 = z t µt wi t l(z t ; ψ i ) (4) for soe tie-varying weighting paraeters {wi t 0}, each possibly relying on the current iterate z t and the datu (a i ; ψ i ). Two attributes of the novel approach are worth highlighting next. First, albeit being a variant of the spectral initialization devised in [29], the initialization here [cf. S1)] is distinct in the sense that different iportance is attached to each selected datu (a i ; ψ i ). Likewise, the gradient flow [cf. S2)] weighs judiciously the search direction suggested by each datu (a i ; ψ i ). In this anner, ore robust initializations and ore stable overall search directions can be constructed even based solely on a rather liited nuber of data saples. Moreover, with particular choices of the weights w t i s (e.g., taking 0/1 values), the developed ethodology subsues as special cases the recently proposed algoriths RWF [31] and TAF [29]. 2 Algorith: Reweighted Aplitude Flow This section explains the intuition and basic principles behind each stage of the advocated RAF algorith in detail. For analytical concreteness, we focus on the real Gaussian odel with x R n, 3

4 and independent sensing vectors a i R n N (0, I) for all 1 i. Nonetheless, the presented approach can be directly applied when the coplex Gaussian and the coded diffraction pattern (CDP) odels are considered. 2.1 Weighted axial correlation initialization A key enabler of general nonconvex iterative heuristics success in finding the global optiu is to seed the with an excellent starting point [15]. Indeed, several sart initialization strategies have been advocated for iterative phase retrieval algoriths; see e.g., the spectral initialization [17], [3] as well as its truncated variants [7], [29], [9], [31], [16]. One proising approach is the one pursued in [29], which is also shown robust to outliers in [9]. To hopefully approach the inforationtheoretic liit however, its perforance ay need further enhanceent. Intuitively, it is increasingly challenging to iprove the initialization (over state-of-the-art) as the nuber of acquired data saples approaches the inforation-theoretic liit. In this context, we develop a ore flexible initialization schee based on the correlation property (as opposed to the orthogonality in [29]), in which the added benefit is the inclusion of a flexible weighting regularization technique to better balance the useful inforation exploited in the selected data. Siilar to related approaches of the sae kind, our strategy entails estiating both the nor x and the direction x / x of x. Leveraging the strong law of large nubers and the rotational invariance of Gaussian a i vectors (the latter suffices to assue x = x e 1, with e 1 being the first canonical vector in R n ), it is clear that 1 ψi 2 = 1 ai, x e 1 ( 2 1 = a 2 i,1 ) x 2 x 2 (5) whereby x can be estiated to be ψ2 i/. This estiate proves very accurate even with a liited nuber of data saples because a2 i,1/ is unbiased and tightly concentrated. The challenge thus lies in accurately estiating the direction of x, or seeking a unit vector axially aligned with x. Toward this end, let us first present a variant of the initialization in [29]. Note that the larger the odulus ψ i of the inner-product between a i and x is, the known design vector a i is deeed ore correlated to the unknown solution x, hence bearing useful directional inforation of x. Inspired by this fact and having available data {(a i ; ψ i )} 1 i, one can sort all (absolute) correlation coefficients {ψ i } 1 i in an ascending order, yielding ordered coefficients 0 < ψ [] ψ [2] ψ [1]. Sorting records takes tie proportional to O( log ). 2 Let S M denote the set of selected feature vectors a i to be used for coputing the initialization, which is to be designed next. Fix a priori the cardinality S to soe integer on the order of, say, S := 3 /13. It is then natural to define S to collect the a i vectors that correspond to one of the largest S correlation coefficients {ψ [i] } 1 i S, each of which can be thought of as pointing to (roughly) the direction of x. Approxiating the direction of x therefore boils down to finding a vector to axiize its correlation with the subset S of selected directional vectors a i. Succinctly, the wanted approxiation vector can be efficiently found as the solution of 1 axiize ai, z 2 = z ( 1 ) a i a i z (6) z =1 S S i S where the superscript represents the transpose or the conjugate transpose that will be clear fro the context. Upon scaling the unity-nor solution of (6) by the nor estiate obtained ψ2 i/ in (5), to atch the agnitude of x, we will develop what we will henceforth refer to as axial correlation initialization. As long as S is chosen on the order of, the axial correlation ethod outperfors the spectral ones in [3, 17, 7], and has coparable perforance to the orthogonality-prooting ethod [29]. Its perforance around the inforation-liit however, is still not the best that we can hope for. Recall fro (6) that all selected directional vectors {a i } i S are treated the sae in ters of their contributions to constructing the initialization. Nevertheless, according to our starting principle, this ordering inforation carried by the selected a i vectors is not exploited by the initialization schee in (6) and [29]. In other words, if for i, j S, the correlation coefficient of ψ i with a i is larger 2 f() = O(g()) eans that there exists a constant C > 0 such that f() C g(). i S 4

5 Relative error than that of ψ j with a j, then a i is deeed ore correlated (with x) than a j is, hence bearing ore useful inforation about the direction of x. It is thus prudent to weigh ore the selected a i vectors associated with larger ψ i values. Given the ordering inforation ψ [ S ] ψ [2] ψ [1] available fro the sorting procedure, a natural way to achieve this goal is weighting each a i vector with siple onotonically increasing functions of ψ i, say e.g., taking the weights wi 0 := ψγ i, i S with the exponent paraeter γ 0 chosen to aintain the wanted ordering w S 0 w0 [2] w0 [1]. In a nutshell, a ore flexible initialization strategy, that we refer to as weighted axial correlation, can be suarized as follows z 0 := arg ax z =1 z ( 1 S ψ γ i a ia i i S ) z. (7) For any given ɛ > 0, the power ethod or the Lanczos algorith can be called for to find an ɛ-accurate solution to (7) in tie proportional to O(n S ) [21], assuing a positive eigengap between the largest and the second largest eigenvalues of the atrix (1/ S ) i S ψγ i a ia i, which is often true when {a i } are sapled fro continuous distribution. The proposed initialization can be obtained upon scaling z 0 fro (7) by the nor estiate in (5), to yield z 0 := ( ψ2 i/) z 0. By default, we take γ := 1 /2 in all reported nuerical ipleentations, yielding wi 0 := a i, x for all i S. Regarding the initialization procedure in (7), we next highlight two features, whereas technical details and theoretical perforance guarantees are provided in Section 3: F1) The weights {wi 0 } in the axial correlation schee enable leveraging useful inforation that each feature vector a i ay bear regarding the direction of x. F2) Taking wi 0 := ψγ i for all i S and 0 otherwise, proble (7) can be equivalently rewritten as z 0 := arg ax z =1 z ( 1 wi 0 a i a i ) z (8) which subsues previous initialization schees with particular selections of weights {wi 0 }. For instance, the spectral initialization in [17, 3] is recovered by choosing S := M, and wi 0 := ψ2 i for all 1 i Reweight. ax. correlation Spectral initialization Trunc. spectral in TWF Orthogonality prooting Trunc. spectral in RWF 0.9 1,000 2,000 3,000 4,000 5,000 n: signal diension (=2n-1) Figure 1: Relative initialization error for i.i.d. a i N (0, I 1,000 ), 1 i 1, 999. For coparison, define dist(z, x) Relative error :=. x Throughout the paper, all siulated results were averaged over 100 Monte Carlo (MC) realizations, and each siulated schee was ipleented with their pertinent default paraeters. Figure 1 evaluates the perforance of the developed initialization relative to several state-of-the-art strategies, and also with the inforation liit nuber of data bencharking the inial nuber of saples required. It is clear that our initialization is: i) consistently better than the state-of-the-art; and, ii) stable as n grows, which is in contrast to the instability encountered by the spectral ones [17, 3, 7, 31]. It is worth stressing that the ore than 5% epirical advantage (relative to the best) at the challenging inforation-theoretic benchark is nontrivial, and is one of the ain RAF upshots. This advantage becoes increasingly pronounced as the ratio /n grows. 2.2 Iteratively reweighted gradient flow For independent data obeying the real Gaussian odel, the direction that TAF oves along in stage S2) presented earlier is given by the following (generalized) gradient [29]: 1 l(z; ψ i ) = 1 ( a a i i z ψ z ) i a i z a i (9) i T i T 5

6 where the dependence on the iterate count t is neglected for notational brevity, and the convention a i z / a i z := 0 is adopted when a i z = 0. Unfortunately, the (negative) gradient of the average in (9) generally ay not point towards the true solution x unless the current iterate z is already very close to x. Therefore, oving along such a descent direction ay not drag z closer to x. To see this, consider an initial guess z 0 that has already been in a basin of attraction (i.e., a region within which there is only a unique stationary point) of x. Certainly, there are suands (a i z ψ i a i z a z )a i in (9), that could give rise to bad/isleading i gradient directions due to the erroneously estiated signs a i z a i z a i x a i x [29], or (a i z)(a i x) < 0 [31]. Those gradients as a whole ay drag z away fro x, and hence out of the basin of attraction. Such an effect becoes increasingly severe as approaches the inforation-theoretic liit of 2n 1, thus rendering past approaches less effective in this case. Although this issue is soewhat reedied by TAF with a truncation procedure, its efficacy is liited due to isses of bad gradients and is-rejections of eaningful ones around the inforation liit. To address this challenge, reweighted aplitude flow effecting suitable gradient directions fro all data saples {(a i ; ψ i )} 1 i will be adopted in a (tiely) adaptive fashion, naely introducing appropriate weights for all gradients to yield the update z t+1 = z t µ t l rw (z t ; ψ i ), t = 0, 1,... (10) The reweighted gradient l rw (z t ) evaluated at the current point z t is given as l rw (z):= 1 w i l(z; ψ i ) (11) for suitable weights {w i } 1 i to be designed next. To that end, we observe that the truncation criterion [29] { T := 1 i : a i z } a i x α with soe given paraeter α > 0 suggests to include only gradient coponents associated with a i z of relatively large sizes. This is because gradients of sizable a i z / a i x offer reliable and eaningful directions pointing to the truth x with large probability [29]. As such, the ratio a i z / a i x can be soewhat viewed as a confidence score about the reliability or eaningfulness of the corresponding gradient l(z; ψ i ). Recognizing that confidence can vary, it is natural to distinguish the contributions that different gradients ake to the overall search direction. An easy way is to attach large weights to the reliable gradients, and sall weights to the spurious ones. Assue without loss of generality that 0 w i 1 for all 1 i ; otherwise, lup the noralization factor achieving this into the learning rate µ t. Building upon this observation and leveraging the gradient reliability confidence score a i z / a x, the weight per gradient l(z; ψ i i) in RAF is designed to be 1 w i := 1 i (13) 1 + βi /( a i z / a i x ), in which {β i > 0} 1 i are soe pre-selected paraeters. Regarding the proposed weighting criterion in (13), three rearks are in order, followed by the RAF algorith suarized in Algorith 1. R1) The weights {wi t} 1 i are tie adapted to z t. One can also interpret the reweighted gradient flow z t+1 in (10) as perforing a single gradient step to iniize the sooth reweighted loss 1 wt i l(z; ψ i) with starting point z t ; see also [4] for related ideas successfully exploited in the iteratively reweighted least-squares approach to copressive sapling. R2) Note that the larger a i z / a i x is, the larger w i will be. More iportance will be attached to reliable gradients than to spurious ones. Gradients fro alost all data points are are judiciously accounted for, which is in sharp contrast to [29], where withdrawn gradients do not contribute the inforation they carry. R3) At the points {z} where a i z = 0 for certain i M, the corresponding weight will be w i = 0. That is, the losses l(z; ψ i ) in (2) that are nonsooth at points z will be eliinated, to prevent their contribution to the reweighted gradient update in (10). Hence, the convergence analysis of RAF can be considerably siplified because it does not have to cope with the nonsoothness of the objective function in (2). (12) 6

7 2.3 Algorithic paraeters To optiize the epirical perforance and facilitate nuerical ipleentations, choice of pertinent algorithic paraeters of RAF is independently discussed here. It is obvious that the RAF algorith entails four paraeters. Our theory and all experients are based on: i) S / 0.25; ii) 0 β i 10 for all 1 i ; and, iii) 0 γ 1. For convenience, a constant step size µ t µ > 0 is suggested, but other step size rules such as backtracking line search with the reweighted objective work as well. As will be foralized in Section 3, RAF converges if the constant µ is not too large, with the upper bound depending in part on the selection of {β i } 1 i. In the nuerical tests presented in Sections 2 and 4, we take S := 3 /13, β i β := 10, γ := 0.5, and µ := 2 (14) while larger step sizes µ > 0 can be afforded for larger /n values. Algorith 1 Reweighted Aplitude Flow 1: Input: Data {(a i ; ψ i } 1 i ; axiu nuber of iterations T ; step size µ t = 2/6 and weighting paraeter β i = 10/5 for real/coplex Gaussian odel; S = 3 /13, and γ = : Construct S to include indices associated with the S largest entries aong {ψ i } 1 i. 3: Initialize z 0 := ψ2 i/ z 0 with z 0 being the unit principal eigenvector of { ψ where wi 0 γ := i, i S M 0, otherwise 4: Loop: for t = 0 to T 1 where w t i := 5: Output: z T. z t+1 = z t µt Y := 1 wi 0 a i a i (15) for all 1 i. a i zt /ψ i a i zt /ψ i +β i for all 1 i. w t i (a i z t a ) i ψ zt i a a i (16) i zt 3 Main results Our ain results suarized in Theore 1 next establish exact recovery under the real Gaussian odel, whose proof is provided in the suppleentary aterial. Our RAF approach however can be generalized readily to the coplex Gaussian and CDP odels. Theore 1 (Exact recovery) Consider noiseless easureents ψ = Ax for an arbitrary x R n. If the data size c 0 S c 1 n and the step size µ µ 0, then with probability at least 1 c 3 e c2, the reweighted aplitude flow s estiates z t in Algorith 1 obey dist(z t, x) 1 10 (1 ν)t x, t = 0, 1,... (17) where c 0, c 1, c 2, c 3 > 0, 0 < ν < 1, and µ 0 > 0 are certain nuerical constants depending on the choice of algorithic paraeters S, β, γ, and µ. According to Theore 1, a few interesting properties of our RAF algorith are worth highlighting. To start, RAF recovers the true solution exactly with high probability whenever the ratio /n of the nuber of equations to the unknowns exceeds soe nuerical constant. Expressed differently, RAF achieves the inforation-theoretic optial order of saple coplexity, which is consistent with the state-of-the-art including TWF [7], TAF [29], and RWF [31]. Notice that (17) also holds at t = 0, naely, dist(z 0, x) x /10, therefore providing perforance guarantees for the proposed initialization schee (cf. Step 3 in Algorith 1). Moreover, starting fro this initial estiate, RAF converges linearly to the true solution x. That is, to reach any ɛ-relative solution accuracy (i.e., dist(z T, x) ɛ x ), it suffices to run at ost T = O(log 1 /ɛ) RAF iterations (cf. Step 4). This in 7

8 ! log 10 L(z T ) conjunction with the per-iteration coplexity O(n) confirs that RAF solves exactly a quadratic syste in tie O(n log 1 /ɛ), which is linear in O(n), the tie required to read the entire data {(a i ; ψ i )} 1 i. Given the fact that the initialization stage can be perfored in tie O(n S ) and S <, the overall linear-tie coplexity of RAF is order-optial. Proof of Theore 1 is provided in the suppleentary aterial. 4 Siulated tests Our theoretical findings about RAF have been corroborated with coprehensive nuerical tests, a saple of which are discussed next. Perforance of RAF is evaluated relative to the state-of-the-art (T)WF, RWF, and TAF in ters of the epirical success rate aong 100 MC trials, where a success will be declared for a trial if the returned estiate incurs error ψ Az T 10 5 x where the odulus operator is understood eleent-wise. The real Gaussian odel and the physically realizable CDPs were siulated in this section. For fairness, all schees were ipleented with their suggested paraeter values. The true signal vector x was randoly generated using x N (0, I), and the i.i.d. sensing vectors a i a i N (0, I). Each schee obtained the initial guess based on 200 power iterations, followed by Realization nuber Figure 2: Function value L(z T ) by RAF for 100 MC realizations when = 2n 1. a series of T = 2, 000 (truncated/reweighted) gradient iterations. All experients were perfored using MATLAB on an Intel 3.4 GHz (32 GB RAM) coputer. For reproducibility, the Matlab code of the RAF algorith is publicly available at To deonstrate the power of RAF in the high-diensional regie, the function value L(z) in (2) evaluated at the returned estiate z T for 100 independent trials is plotted (in negative logarithic scale) in Fig. 2, where = 2n 1 = 9, 999. It is self-evident that RAF succeeded in all trials even at this challenging inforation liit. To the best of our knowledge, RAF is the first algorith that epirically recovers any solution exactly fro a inial nuber of rando quadratic equations. Left panel in Fig. 3 further copares the epirical success rate of five schees under the real Gaussian odel with n = 1, 000 and /n varying by 0.1 fro 1 to 5. Evidently, the developed RAF achieves perfect recovery as soon as is about 2n, where its copeting alternatives do not work well. To deonstrate the stability and robustness of RAF in the presence of additive noise, the right panel in Fig. 3 depicts the noralized ean-square error NMSE := dist2 (z T, x) x 2 as a function of the signal-to-noise ratio (SNR) for /n taking values {3, 4, 5}. The noise odel ψ i = a i, x + η i, 1 i with η := [η i ] 1 i N (0, σ 2 I ) was eployed, where σ 2 was set such that certain SNR := 10 log 10 ( Ax 2 /σ 2 ) values on the x-axis were achieved. To exaine the efficacy and scalability of RAF in real-world conditions, the last experient entails the Galaxy iage 3 depicted by a three-way array X R 1,080 1,920 3, whose first two coordinates encode the pixel locations, and the third the RGB color bands. Consider the physically realizable CDP odel with rando asks [3]. Letting x R n (n ) be a vectorization of a certain band of X, the CDP odel with K asks is ψ (k) = F D (k) x, 1 k K, 3 Downloaded fro 8

9 10 0 =3n =4n =5n NMSE Epirical success rate RAF TAF TWF RWF WF , SNR (db) for x2 R1,000 /n for x2 R Figure 3: Real Gaussian odel: Epirical success rate (Left); and, Relative MSE vs. SNR (Right). where F Cn n is a discrete Fourier transfor atrix, and diagonal atrices D (k) have their diago nal entries sapled uniforly at rando fro {1, 1, j, j} with j := 1. Each D (k) represents a rando ask placed after the object to odulate the illuination patterns [5]. Ipleenting K = 4 asks, each algorith perfors independently over each band 100 power iterations for an initial guess, which was refined by 100 gradient iterations. Recovered iages of TAF (left) and RAF (right) are displayed in Fig. 4, whose relative errors were and , respectively. WF and TWF returned iages of corresponding relative error and , which are far away fro the ground truth. Figure 4: Recovered Galaxy iages after 100 gradient iterations of TAF (Left); and of RAF (Right). 5 Conclusion This paper developed a linear-tie algorith called RAF for solving systes of rando quadratic equations. Our procedure consists of two stages: a weighted axial correlation initializer attainable with a few power or Lanczos iterations, and a sequence of scalable reweighted gradient refineents for a nonconvex nonsooth least-squares loss function. It was deonstrated that RAF achieves the optial saple and coputational coplexity. Judicious nuerical tests showcase its superior perforance over state-of-the-art alternatives. Epirically, RAF solves a set of rando quadratic equations with high probability so long as a unique solution exists. Proising extensions include studying robust and/or sparse phase retrieval and (sei-definite) atrix recovery via (stochastic) reweighted aplitude flow counterparts, and in particular exploiting the power of (re)weighting regularization techniques to enable ore general nonconvex optiization such as training deep neural networks [19]. Acknowledgents Work in this paper was partially supported by NSF grants , , and

10 References [1] R. Balan, P. Casazza, and D. Edidin, On signal reconstruction without phase, Appl. Coput. Haron. Anal., vol. 20, no. 3, pp , May [2] A. Ben-Tal and A. Neirovski, Lectures on Modern Convex Optiization: Analysis, Algoriths, and Engineering Applications. SIAM, 2001, vol. 2. [3] E. J. Candès, X. Li, and M. Soltanolkotabi, Phase retrieval via Wirtinger flow: Theory and algoriths, IEEE Trans. Inf. Theory, vol. 61, no. 4, pp , Apr [4] E. J. Candès, T. Stroher, and V. Voroninski, PhaseLift: Exact and stable signal recovery fro agnitude easureents via convex prograing, Appl. Coput. Haron. Anal., vol. 66, no. 8, pp , Nov [5] E. J. Candès, X. Li, and M. Soltanolkotabi, Phase retrieval fro coded diffraction patterns, Appl. Coput. Haron. Anal., vol. 39, no. 2, pp , Sept [6] J. Chen, L. Wang, X. Zhang, and Q. Gu, Robust Wirtinger flow for phase retrieval with arbitrary corruption, arxiv: , [7] Y. Chen and E. J. Candès, Solving rando quadratic systes of equations is nearly as easy as solving linear systes, in Adv. on Neural Inf. Process. Syst., Montréal, Canada, 2015, pp [8] A. Conca, D. Edidin, M. Hering, and C. Vinzant, An algebraic characterization of injectivity in phase retrieval, Appl. Coput. Haron. Anal., vol. 38, no. 2, pp , Mar [9] J. C. Duchi and F. Ruan, Solving (ost) of a set of quadratic equalities: Coposite optiization for robust phase retrieval, arxiv: , [10] J. Duchi and F. Ruan, Stochastic ethods for coposite optiization probles, arxiv: , [11] J. R. Fienup, Phase retrieval algoriths: A coparison, Appl. Opt., vol. 21, no. 15, pp , Aug [12] R. W. Gerchberg and W. O. Saxton, A practical algorith for the deterination of phase fro iage and diffraction, Optik, vol. 35, pp , Nov [13] T. Goldstein and S. Studer, PhaseMax: Convex phase retrieval via basis pursuit, arxiv: v1, [14] P. Hand and V. Voroninski, An eleentary proof of convex phase retrieval in the natural paraeter space via the linear progra phaseax, arxiv: , [15] R. H. Keshavan, A. Montanari, and S. Oh, Matrix copletion fro a few entries, IEEE Trans. Inf. Theory, vol. 56, no. 6, pp , Jun [16] Y. M. Lu and G. Li, Phase transitions of spectral initialization for high-diensional nonconvex estiation, arxiv: , [17] P. Netrapalli, P. Jain, and S. Sanghavi, Phase retrieval using alternating iniization, in Adv. on Neural Inf. Process. Syst., Stateline, NV, 2013, pp [18] P. M. Pardalos and S. A. Vavasis, Quadratic prograing with one negative eigenvalue is NP-hard, J. Global Opti., vol. 1, no. 1, pp , [19] G. Pereyra, G. Tucker, J. Chorowski, Ł. Kaiser, and G. Hinton, Regularizing neural networks by penalizing confident output distributions, arxiv: , [20] J. R. Rice, Nuerical Methods in Software and Analysis. Acadeic Press, [21] Y. Saad, Nuerical Methods for Large Eigenvalue Probles: Revised Edition. SIAM,

11 [22] Y. Shechtan, Y. C. Eldar, O. Cohen, H. N. Chapan, J. Miao, and M. Segev, Phase retrieval with application to optical iaging: A conteporary overview, IEEE Signal Process. Mag., vol. 32, no. 3, pp , May [23] M. Soltanolkotabi, Structured signal recovery fro quadratic easureents: Breaking saple coplexity barriers via nonconvex optiization, arxiv: , [24] J. Sun, Q. Qu, and J. Wright, A geoetric analysis of phase retrieval, Found. Coput. Math., 2017 (to appear); see also arxiv: , [25] I. Waldspurger, Phase retrieval with rando Gaussian sensing vectors by alternating projections, axiv: , [26] I. Waldspurger, A. d Aspreont, and S. Mallat, Phase recovery, axcut and coplex seidefinite prograing, Math. Progra., vol. 149, no. 1, pp , [27] G. Wang and G. B. Giannakis, Solving rando systes of quadratic equations via truncated generalized gradient flow, in Adv. on Neural Inf. Process. Syst., Barcelona, Spain, 2016, pp [28] G. Wang, G. B. Giannakis, and J. Chen, Scalable solvers of rando quadratic equations via stochastic truncated aplitude flow, IEEE Trans. Signal Process., vol. 65, no. 8, pp , Apr [29] G. Wang, G. B. Giannakis, and Y. C. Eldar, Solving systes of rando quadratic equations via truncated aplitude flow, IEEE Trans. Inf. Theory, 2017 (to appear); see also arxiv: , [30] X. Yi, C. Caraanis, and S. Sanghavi, Alternating iniization for ixed linear regression, in Proc. Intl. Conf. on Mach. Learn., Beijing, China, 2014, pp [31] H. Zhang, Y. Zhou, Y. Liang, and Y. Chi, Reshaped Wirtinger flow and increental algorith for solving quadratic syste of equations, J. Mach. Learn. Res., 2017 (to appear); see also arxiv: ,

Solving Large-scale Systems of Random Quadratic Equations via Stochastic Truncated Amplitude Flow

Solving Large-scale Systems of Random Quadratic Equations via Stochastic Truncated Amplitude Flow Solving Large-scale Systems of Random Quadratic Equations via Stochastic Truncated Amplitude Flow Gang Wang,, Georgios B. Giannakis, and Jie Chen Dept. of ECE and Digital Tech. Center, Univ. of Minnesota,

More information

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup) Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Hybrid System Identification: An SDP Approach

Hybrid System Identification: An SDP Approach 49th IEEE Conference on Decision and Control Deceber 15-17, 2010 Hilton Atlanta Hotel, Atlanta, GA, USA Hybrid Syste Identification: An SDP Approach C Feng, C M Lagoa, N Ozay and M Sznaier Abstract The

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Provable Non-convex Phase Retrieval with Outliers: Median Truncated Wirtinger Flow

Provable Non-convex Phase Retrieval with Outliers: Median Truncated Wirtinger Flow Provable Non-convex Phase Retrieval with Outliers: Median runcated Wirtinger Flow Huishuai Zhang Departent of EECS, Syracuse University, Syracuse, NY 3244 USA Yuejie Chi Departent of ECE, he Ohio State

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

Structured signal recovery from quadratic measurements: Breaking sample complexity barriers via nonconvex optimization

Structured signal recovery from quadratic measurements: Breaking sample complexity barriers via nonconvex optimization Structured signal recovery fro quadratic easureents: Breaking saple coplexity barriers via nonconvex optiization Mahdi Soltanolkotabi Ming Hsieh Departent of Electrical Engineering University of Southern

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical IEEE TRANSACTIONS ON INFORMATION THEORY Large Alphabet Source Coding using Independent Coponent Analysis Aichai Painsky, Meber, IEEE, Saharon Rosset and Meir Feder, Fellow, IEEE arxiv:67.7v [cs.it] Jul

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

GLOBALLY CONVERGENT LEVENBERG-MARQUARDT METHOD FOR PHASE RETRIEVAL

GLOBALLY CONVERGENT LEVENBERG-MARQUARDT METHOD FOR PHASE RETRIEVAL GLOBALLY CONVERGENT LEVENBERG-MARQUARDT METHOD FOR PHASE RETRIEVAL CHAO MA, XIN LIU, AND ZAIWEN WEN Abstract. In this paper, we consider a nonlinear least squares odel for the phase retrieval proble. Since

More information

Weighted- 1 minimization with multiple weighting sets

Weighted- 1 minimization with multiple weighting sets Weighted- 1 iniization with ultiple weighting sets Hassan Mansour a,b and Özgür Yılaza a Matheatics Departent, University of British Colubia, Vancouver - BC, Canada; b Coputer Science Departent, University

More information

HIGH RESOLUTION NEAR-FIELD MULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR MACHINES

HIGH RESOLUTION NEAR-FIELD MULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR MACHINES ICONIC 2007 St. Louis, O, USA June 27-29, 2007 HIGH RESOLUTION NEAR-FIELD ULTIPLE TARGET DETECTION AND LOCALIZATION USING SUPPORT VECTOR ACHINES A. Randazzo,. A. Abou-Khousa 2,.Pastorino, and R. Zoughi

More information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information Cite as: Straub D. (2014). Value of inforation analysis with structural reliability ethods. Structural Safety, 49: 75-86. Value of Inforation Analysis with Structural Reliability Methods Daniel Straub

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 65, NO. 8, APRIL 15,

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 65, NO. 8, APRIL 15, IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 65, NO. 8, APRIL 15, 2017 1961 Scalable Solvers of Rando Quadratic Equations via Stochastic Truncated Aplude Flow Gang Wang, Student Meber, IEEE, Georgios B.

More information

Nonconvex Methods for Phase Retrieval

Nonconvex Methods for Phase Retrieval Nonconvex Methods for Phase Retrieval Vince Monardo Carnegie Mellon University March 28, 2018 Vince Monardo (CMU) 18.898 Midterm March 28, 2018 1 / 43 Paper Choice Main reference paper: Solving Random

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

Data-Driven Imaging in Anisotropic Media

Data-Driven Imaging in Anisotropic Media 18 th World Conference on Non destructive Testing, 16- April 1, Durban, South Africa Data-Driven Iaging in Anisotropic Media Arno VOLKER 1 and Alan HUNTER 1 TNO Stieltjesweg 1, 6 AD, Delft, The Netherlands

More information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department

More information

SPECTRUM sensing is a core concept of cognitive radio

SPECTRUM sensing is a core concept of cognitive radio World Acadey of Science, Engineering and Technology International Journal of Electronics and Counication Engineering Vol:6, o:2, 202 Efficient Detection Using Sequential Probability Ratio Test in Mobile

More information

On the Communication Complexity of Lipschitzian Optimization for the Coordinated Model of Computation

On the Communication Complexity of Lipschitzian Optimization for the Coordinated Model of Computation journal of coplexity 6, 459473 (2000) doi:0.006jco.2000.0544, available online at http:www.idealibrary.co on On the Counication Coplexity of Lipschitzian Optiization for the Coordinated Model of Coputation

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths

More information

Introduction to Machine Learning. Recitation 11

Introduction to Machine Learning. Recitation 11 Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,

More information

Hamming Compressed Sensing

Hamming Compressed Sensing Haing Copressed Sensing Tianyi Zhou, and Dacheng Tao, Meber, IEEE Abstract arxiv:.73v2 [cs.it] Oct 2 Copressed sensing CS and -bit CS cannot directly recover quantized signals and require tie consuing

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

Compressive Distilled Sensing: Sparse Recovery Using Adaptivity in Compressive Measurements

Compressive Distilled Sensing: Sparse Recovery Using Adaptivity in Compressive Measurements 1 Copressive Distilled Sensing: Sparse Recovery Using Adaptivity in Copressive Measureents Jarvis D. Haupt 1 Richard G. Baraniuk 1 Rui M. Castro 2 and Robert D. Nowak 3 1 Dept. of Electrical and Coputer

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Convex Programming for Scheduling Unrelated Parallel Machines

Convex Programming for Scheduling Unrelated Parallel Machines Convex Prograing for Scheduling Unrelated Parallel Machines Yossi Azar Air Epstein Abstract We consider the classical proble of scheduling parallel unrelated achines. Each job is to be processed by exactly

More information

Incremental Reshaped Wirtinger Flow and Its Connection to Kaczmarz Method

Incremental Reshaped Wirtinger Flow and Its Connection to Kaczmarz Method Incremental Reshaped Wirtinger Flow and Its Connection to Kaczmarz Method Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 3244 hzhan23@syr.edu Yingbin Liang Department of EECS Syracuse

More information

On the Use of A Priori Information for Sparse Signal Approximations

On the Use of A Priori Information for Sparse Signal Approximations ITS TECHNICAL REPORT NO. 3/4 On the Use of A Priori Inforation for Sparse Signal Approxiations Oscar Divorra Escoda, Lorenzo Granai and Pierre Vandergheynst Signal Processing Institute ITS) Ecole Polytechnique

More information

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo

More information

Sparse beamforming in peer-to-peer relay networks Yunshan Hou a, Zhijuan Qi b, Jianhua Chenc

Sparse beamforming in peer-to-peer relay networks Yunshan Hou a, Zhijuan Qi b, Jianhua Chenc 3rd International Conference on Machinery, Materials and Inforation echnology Applications (ICMMIA 015) Sparse beaforing in peer-to-peer relay networs Yunshan ou a, Zhijuan Qi b, Jianhua Chenc College

More information

Fundamental Limits of Weak Recovery with Applications to Phase Retrieval

Fundamental Limits of Weak Recovery with Applications to Phase Retrieval Proceedings of Machine Learning Research vol 75:1 6, 2018 31st Annual Conference on Learning Theory Fundamental Limits of Weak Recovery with Applications to Phase Retrieval Marco Mondelli Department of

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Research Article Robust ε-support Vector Regression

Research Article Robust ε-support Vector Regression Matheatical Probles in Engineering, Article ID 373571, 5 pages http://dx.doi.org/10.1155/2014/373571 Research Article Robust ε-support Vector Regression Yuan Lv and Zhong Gan School of Mechanical Engineering,

More information

Figure 1: Equivalent electric (RC) circuit of a neurons membrane

Figure 1: Equivalent electric (RC) circuit of a neurons membrane Exercise: Leaky integrate and fire odel of neural spike generation This exercise investigates a siplified odel of how neurons spike in response to current inputs, one of the ost fundaental properties of

More information

Extension of CSRSM for the Parametric Study of the Face Stability of Pressurized Tunnels

Extension of CSRSM for the Parametric Study of the Face Stability of Pressurized Tunnels Extension of CSRSM for the Paraetric Study of the Face Stability of Pressurized Tunnels Guilhe Mollon 1, Daniel Dias 2, and Abdul-Haid Soubra 3, M.ASCE 1 LGCIE, INSA Lyon, Université de Lyon, Doaine scientifique

More information

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS BIT Nuerical Matheatics 43: 459 466, 2003. 2003 Kluwer Acadeic Publishers. Printed in The Netherlands 459 RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS V. SIMONCINI Dipartiento di

More information

arxiv: v1 [cs.ds] 3 Feb 2014

arxiv: v1 [cs.ds] 3 Feb 2014 arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/

More information

Analyzing Simulation Results

Analyzing Simulation Results Analyzing Siulation Results Dr. John Mellor-Cruey Departent of Coputer Science Rice University johnc@cs.rice.edu COMP 528 Lecture 20 31 March 2005 Topics for Today Model verification Model validation Transient

More information

Detection and Estimation Theory

Detection and Estimation Theory ESE 54 Detection and Estiation Theory Joseph A. O Sullivan Sauel C. Sachs Professor Electronic Systes and Signals Research Laboratory Electrical and Systes Engineering Washington University 11 Urbauer

More information

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS

EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS EMPIRICAL COMPLEXITY ANALYSIS OF A MILP-APPROACH FOR OPTIMIZATION OF HYBRID SYSTEMS Jochen Till, Sebastian Engell, Sebastian Panek, and Olaf Stursberg Process Control Lab (CT-AST), University of Dortund,

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

Asynchronous Gossip Algorithms for Stochastic Optimization

Asynchronous Gossip Algorithms for Stochastic Optimization Asynchronous Gossip Algoriths for Stochastic Optiization S. Sundhar Ra ECE Dept. University of Illinois Urbana, IL 680 ssrini@illinois.edu A. Nedić IESE Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu

More information

Exact tensor completion with sum-of-squares

Exact tensor completion with sum-of-squares Proceedings of Machine Learning Research vol 65:1 54, 2017 30th Annual Conference on Learning Theory Exact tensor copletion with su-of-squares Aaron Potechin Institute for Advanced Study, Princeton David

More information

Optimal Jamming Over Additive Noise: Vector Source-Channel Case

Optimal Jamming Over Additive Noise: Vector Source-Channel Case Fifty-first Annual Allerton Conference Allerton House, UIUC, Illinois, USA October 2-3, 2013 Optial Jaing Over Additive Noise: Vector Source-Channel Case Erah Akyol and Kenneth Rose Abstract This paper

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

A NEW ROBUST AND EFFICIENT ESTIMATOR FOR ILL-CONDITIONED LINEAR INVERSE PROBLEMS WITH OUTLIERS

A NEW ROBUST AND EFFICIENT ESTIMATOR FOR ILL-CONDITIONED LINEAR INVERSE PROBLEMS WITH OUTLIERS A NEW ROBUST AND EFFICIENT ESTIMATOR FOR ILL-CONDITIONED LINEAR INVERSE PROBLEMS WITH OUTLIERS Marta Martinez-Caara 1, Michael Mua 2, Abdelhak M. Zoubir 2, Martin Vetterli 1 1 School of Coputer and Counication

More information

Machine Learning Basics: Estimators, Bias and Variance

Machine Learning Basics: Estimators, Bias and Variance Machine Learning Basics: Estiators, Bias and Variance Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Basics

More information

Bootstrapping Dependent Data

Bootstrapping Dependent Data Bootstrapping Dependent Data One of the key issues confronting bootstrap resapling approxiations is how to deal with dependent data. Consider a sequence fx t g n t= of dependent rando variables. Clearly

More information

RANDOM GRADIENT EXTRAPOLATION FOR DISTRIBUTED AND STOCHASTIC OPTIMIZATION

RANDOM GRADIENT EXTRAPOLATION FOR DISTRIBUTED AND STOCHASTIC OPTIMIZATION RANDOM GRADIENT EXTRAPOLATION FOR DISTRIBUTED AND STOCHASTIC OPTIMIZATION GUANGHUI LAN AND YI ZHOU Abstract. In this paper, we consider a class of finite-su convex optiization probles defined over a distributed

More information

Multi-view Discriminative Manifold Embedding for Pattern Classification

Multi-view Discriminative Manifold Embedding for Pattern Classification Multi-view Discriinative Manifold Ebedding for Pattern Classification X. Wang Departen of Inforation Zhenghzou 450053, China Y. Guo Departent of Digestive Zhengzhou 450053, China Z. Wang Henan University

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab

Support Vector Machines. Machine Learning Series Jerry Jeychandra Blohm Lab Support Vector Machines Machine Learning Series Jerry Jeychandra Bloh Lab Outline Main goal: To understand how support vector achines (SVMs) perfor optial classification for labelled data sets, also a

More information

arxiv: v1 [cs.lg] 8 Jan 2019

arxiv: v1 [cs.lg] 8 Jan 2019 Data Masking with Privacy Guarantees Anh T. Pha Oregon State University phatheanhbka@gail.co Shalini Ghosh Sasung Research shalini.ghosh@gail.co Vinod Yegneswaran SRI international vinod@csl.sri.co arxiv:90.085v

More information

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x)

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x) 7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

Identical Maximum Likelihood State Estimation Based on Incremental Finite Mixture Model in PHD Filter

Identical Maximum Likelihood State Estimation Based on Incremental Finite Mixture Model in PHD Filter Identical Maxiu Lielihood State Estiation Based on Increental Finite Mixture Model in PHD Filter Gang Wu Eail: xjtuwugang@gail.co Jing Liu Eail: elelj20080730@ail.xjtu.edu.cn Chongzhao Han Eail: czhan@ail.xjtu.edu.cn

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

A remark on a success rate model for DPA and CPA

A remark on a success rate model for DPA and CPA A reark on a success rate odel for DPA and CPA A. Wieers, BSI Version 0.5 andreas.wieers@bsi.bund.de Septeber 5, 2018 Abstract The success rate is the ost coon evaluation etric for easuring the perforance

More information

Support Vector Machines. Goals for the lecture

Support Vector Machines. Goals for the lecture Support Vector Machines Mark Craven and David Page Coputer Sciences 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Soe of the slides in these lectures have been adapted/borrowed fro aterials developed

More information

A Nonlinear Sparsity Promoting Formulation and Algorithm for Full Waveform Inversion

A Nonlinear Sparsity Promoting Formulation and Algorithm for Full Waveform Inversion A Nonlinear Sparsity Prooting Forulation and Algorith for Full Wavefor Inversion Aleksandr Aravkin, Tristan van Leeuwen, Jaes V. Burke 2 and Felix Herrann Dept. of Earth and Ocean sciences University of

More information

Reshaped Wirtinger Flow for Solving Quadratic System of Equations

Reshaped Wirtinger Flow for Solving Quadratic System of Equations Reshaped Wirtinger Flow for Solving Quadratic System of Equations Huishuai Zhang Department of EECS Syracuse University Syracuse, NY 344 hzhan3@syr.edu Yingbin Liang Department of EECS Syracuse University

More information

Tracking using CONDENSATION: Conditional Density Propagation

Tracking using CONDENSATION: Conditional Density Propagation Tracking using CONDENSATION: Conditional Density Propagation Goal Model-based visual tracking in dense clutter at near video frae rates M. Isard and A. Blake, CONDENSATION Conditional density propagation

More information

Statistical clustering and Mineral Spectral Unmixing in Aviris Hyperspectral Image of Cuprite, NV

Statistical clustering and Mineral Spectral Unmixing in Aviris Hyperspectral Image of Cuprite, NV CS229 REPORT, DECEMBER 05 1 Statistical clustering and Mineral Spectral Unixing in Aviris Hyperspectral Iage of Cuprite, NV Mario Parente, Argyris Zynis I. INTRODUCTION Hyperspectral Iaging is a technique

More information

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points

More information

Support recovery in compressed sensing: An estimation theoretic approach

Support recovery in compressed sensing: An estimation theoretic approach Support recovery in copressed sensing: An estiation theoretic approach Ain Karbasi, Ali Horati, Soheil Mohajer, Martin Vetterli School of Coputer and Counication Sciences École Polytechnique Fédérale de

More information

When Short Runs Beat Long Runs

When Short Runs Beat Long Runs When Short Runs Beat Long Runs Sean Luke George Mason University http://www.cs.gu.edu/ sean/ Abstract What will yield the best results: doing one run n generations long or doing runs n/ generations long

More information

Solving Corrupted Quadratic Equations, Provably

Solving Corrupted Quadratic Equations, Provably Solving Corrupted Quadratic Equations, Provably Yuejie Chi London Workshop on Sparse Signal Processing September 206 Acknowledgement Joint work with Yuanxin Li (OSU), Huishuai Zhuang (Syracuse) and Yingbin

More information

Supplementary Information for Design of Bending Multi-Layer Electroactive Polymer Actuators

Supplementary Information for Design of Bending Multi-Layer Electroactive Polymer Actuators Suppleentary Inforation for Design of Bending Multi-Layer Electroactive Polyer Actuators Bavani Balakrisnan, Alek Nacev, and Elisabeth Sela University of Maryland, College Park, Maryland 074 1 Analytical

More information

arxiv: v1 [cs.ds] 17 Mar 2016

arxiv: v1 [cs.ds] 17 Mar 2016 Tight Bounds for Single-Pass Streaing Coplexity of the Set Cover Proble Sepehr Assadi Sanjeev Khanna Yang Li Abstract arxiv:1603.05715v1 [cs.ds] 17 Mar 2016 We resolve the space coplexity of single-pass

More information

OPTIMIZATION in multi-agent networks has attracted

OPTIMIZATION in multi-agent networks has attracted Distributed constrained optiization and consensus in uncertain networks via proxial iniization Kostas Margellos, Alessandro Falsone, Sione Garatti and Maria Prandini arxiv:603.039v3 [ath.oc] 3 May 07 Abstract

More information

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES

TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES TEST OF HOMOGENEITY OF PARALLEL SAMPLES FROM LOGNORMAL POPULATIONS WITH UNEQUAL VARIANCES S. E. Ahed, R. J. Tokins and A. I. Volodin Departent of Matheatics and Statistics University of Regina Regina,

More information

Topic 5a Introduction to Curve Fitting & Linear Regression

Topic 5a Introduction to Curve Fitting & Linear Regression /7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline

More information

Fixed-to-Variable Length Distribution Matching

Fixed-to-Variable Length Distribution Matching Fixed-to-Variable Length Distribution Matching Rana Ali Ajad and Georg Böcherer Institute for Counications Engineering Technische Universität München, Gerany Eail: raa2463@gail.co,georg.boecherer@tu.de

More information

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE

MSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE Proceeding of the ASME 9 International Manufacturing Science and Engineering Conference MSEC9 October 4-7, 9, West Lafayette, Indiana, USA MSEC9-8466 MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space

Grafting: Fast, Incremental Feature Selection by Gradient Descent in Function Space Journal of Machine Learning Research 3 (2003) 1333-1356 Subitted 5/02; Published 3/03 Grafting: Fast, Increental Feature Selection by Gradient Descent in Function Space Sion Perkins Space and Reote Sensing

More information