ON THE ADVERSARIAL ROBUSTNESS OF ROBUST ESTIMATORS

Size: px
Start display at page:

Download "ON THE ADVERSARIAL ROBUSTNESS OF ROBUST ESTIMATORS"

Transcription

1 O THE ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS BY ERHA BAYAKTAR,, LIFEG LAI, University of Michigan and University of California, Davis Motivated by recent data analytics alications, we study the adversarial robustness of robust estimators. Instead of assuming that only a fraction of the data oints are outliers as considered in the classic robust estimation setu, in this aer, we consider an adversarial setu in which an attacker can observe the whole dataset and can modify all data samles in an adversarial manner so as to maximize the estimation error caused by his attack. We characterize the attacker s otimal attack strategy, and further introduce adversarial influence function (AIF) to quantify an estimator s sensitivity to such adversarial attacks. We rovide an aroach to characterize AIF for any given robust estimator, and then design otimal estimator that minimizes AIF, which imlies it is least sensitive to adversarial attacks and hence is most robust against adversarial attacks. From this characterization, we identify a tradeoff between AIF (i.e., robustness against adversarial attack) and influence function, a quantity used in classic robust estimators to measure robustness against outliers, and design estimators that strike a desirable tradeoff between these two quantities.. Introduction. Robust estimation is a classic toic that addresses the outlier or model uncertainty issues. In the existing setu, a certain ercentage of the data oints are assumed to be outliers. Various concets such as influence function (IF), breakdown oint, and change of variance etc were develoed to quantify the robustness of estimators against the resence of outliers, lease see [, 2, 3] and references therein for details. These concets are very useful for the classic setu where a fraction (u to 50%) of data oints are outliers while the remaining data come from the true distribution. In this aer, motivated by recent interest in data analytics, we address the issue of adversarial robustness. In a tyical data analytics setu, a dataset is stored in a database. If an attacker has access to the database, he can modify all data oints (i.e., u to 00%) in an adversarial manner, and hence the existing results on robust statistics are not directly alicable anymore. This scenario also arises in the adversarial examle henomena in dee neural networks that have attracted signif- The work of E. Bayraktar was suorted in art by the ational Science Foundation under grant DMS-6370 and by the Susan M. Smith Professorshi. The work of L. Lai was suorted by the ational Science Foundation under grants CCF and ECCS MSC 200 subject classifications: Primary 62F35; secondary 62F0,62F2 Keywords and hrases: Robust estimators, adversarial robustness, M-estimator, non-convex otimization.

2 2 E. BAYAKTAR AD L. LAI icant recent research interests [4, 5, 6]. In the adversarial examle in dee neural networks, by making small but carefully chosen changes on the image, the attacker can mislead neural network to make wrong decisions, even though a human will hardly notice changes on the modified image. Certainly, if the attacker can modify all data and no further restrictions on attacker s caability are imosed, then no meaningful estimator can be constructed (this can be viewed as 00% of the data are modified in the classic setu). In this aer, we investigate the scenario that the total amount of change measured by l norm is limited, and we will study how these quantities will affect the estimation erformance. This tye of constraints are reasonable and are motivated by real life examles. For examle, in generating adversarial examles in images [4], the total distortion should be limited, otherwise human eyes will be able to detect such changes. Towards this goal, we introduce the concet of adversarial influence function (AIF) to quantify how sensitive an estimator is to adversarial attacks. We first focus on the scenario with a given data set. For this scenario, we characterize the otimal attack vector that the attacker, who observes the whole data set, can emloy to maximize the estimation error. Using this characterization, we can then analyze AIF of any given estimator. This analysis enables us to design estimators that are robust to adversarial attacks. In articular, from the estimator s ersective, one would like to design an estimator that minimizes AIF, which imlies that such an estimator is least sensitive to adversarial attacks and hence is most robust against adversarial attacks. We derive universal lower bounds on AIF and characterize the conditions under which an estimator can achieve this lower bound (and hence is most robustness against adversarial attacks). We then illustrate these results for two secific models: location estimators and scale estimators. With the results in the given samle scenario, we then extend our study to the oulation scenario, in which we investigate the behavior of AIF as the number of samles increases. For this case, we identify a tradeoff between robustness against adversarial attacks vs robustness against outliers. In articular, we first characterize the otimal estimator that minimizes AIF. However, the estimator that minimizes AIF has a oor erformance in term of IF [3, 7], a quantity that measures robustness against outliers. Realizing this fact, we then formulate otimization roblems to design estimators that strike a desirable tradeoff between AIF (i.e., robustness against adversarial attack) and IF (i.e., robustness against outliers). Using tools from calculus of variations [8, 9], we are able to exloit the unique structure of our roblems and obtain analytical form of the otimal solution. The obtained solution share similar interretation as classic robust estimators that it will carefully trim data oints that are from the distribution tails. However, the detailed form and thresholds are determined by different criteria. In the above discussion, we mainly focus on a class of widely used robust esti-

3 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 3 mators: M-estimator. However, the develoed tools and analysis can be extended to analyze other tyes of robust estimators. In this aer, we will use L-estimator as an examle to discuss how to extend the analysis to other tyes of estimators. Our aer is related to a growing list of recent work on adversarial machine learning. Here we give several examles on data oisoning attack that is related to our work. For examle, [0] considers an adversarial rincial comonent analysis (PCA) roblem in which an attacker adds an extra data oint in an adversarial manner so as to maximize the error of subsace estimated by PCA. [] investigates data oisoning attack in regression roblems, in which the attacker adds data oints to the training dataset with the goal of introducing errors into or guiding the results of regression models. [2] studies an attack that inserts carefully chosen data oints to the training set for suort vector machine. [3] considers learning roblems from untrusted data. In articular, under the assumtion that at least α ercent of data oints are drawn from a distribution of interest, [3] considers two frameworks: ) list-decodable learning, whose goal is to return a list of answers, with the guarantee that at least one of them is accurate; and 2) semi-verified learning, in which one has a small dataset of trusted data that can be leveraged to enable the accurate extraction of information from a much larger but untrusted dataset. A major difference between our work and these interesting work is that the existing work assume that a certain ercentage of data oints are not comromised, while in our work all data oints could be comromised. The remainder of the aer is organized as follows. In Section 2, we introduce our roblem formulation. In Section 3, we introduce the necessary background. In Section 4, we investigate AIF for the given samle scenario. In Section 5, we consider the oulation scenario. We extend the study to L-estimator in Section 6. umerical examles are given in Section 7. Finally, we offer concluding remarks in Section Model. We consider an adversarially robust arameter estimation roblem in which the adversary has access to the whole dataset. In articular, we have a given data set x = {x,, x }, in which x n are i.i.d realizations of random variable X that has cumulative density function (cdf) F θ (x) with unknown arameter θ. We will use f θ (x) to denote the corresonding robability density function (df). From this given data set, we would like to estimate the unknown arameter θ. However, as the adversary has access to the whole dataset, it will modify the data to x = x + x := {x + x,, x + x }, in which x = { x,, x } is the attack vector chosen by the adversary after observing x. We will discuss the attacker s otimal attack strategy in choosing x in the sequel. In the classic robust estimation setu, it is tyically assume that some ercentage (u to 50%)

4 4 E. BAYAKTAR AD L. LAI of the data oints are outliers, that is some entries in x are nonzero while the remainders are zero. In this work, we consider the case where the attacker can modify all data oints, which is a more suitable setu for recent data analytical alications. However, certain restrictions need to be ut on x, otherwise the estimation roblem will not be meaningful. In this aer, we assume that x η, in which is the l norm. The normalization factor imlies that the erdimension change (on average) is uer-bound by η. As mentioned in the introduction, this tye of constraints are reasonable and are motivated by real life examles. The classic setu can be viewed as a secial case of our formulation by letting 0, i.e., the classic setu has constraint on the total number of data oints that the attacker can modify. Following notation used in robust statistics [2, 3], we will use T (x) to denote an estimator. For a given estimator T, we would like to characterize how sensitive the estimator is with resect to the adversarial attack. In this aer, we consider a scenario where the goal of the attacker is to maximize the estimation error caused by the attack. In articular, the attacker aims to choose x by solving the following otimization roblem (2.) max x s.t. T (x + x) T (x), x η. We use T (x) to denote the otimal value obtained from the otimization roblem (2.), and define the adversarial influence function (AIF) of estimator T at x under l norm constraint as T (x) AIF(T, x, ) = lim. η 0 η This quantity, a generalization of the concet of IF used in classic robust estimation (we will briefly review IF in Section 3), quantifies the asymtotic rate at which the attacker can introduce estimation error through its attack. From the defender s ersective, the smaller AIF is, the more robust the estimator is. In this aer, building on the characterization of AIF(T, x, ), we will characterize the otimal estimator T, among a certain class of estimators T, that minimizes AIF(T, x, ). In articular, we will investigate min AIF(T, x, ). T T

5 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 5 We will show that, for certain class of T, the otimal T is indeendent of x and, which is a very desirable roerty. ote that AIF(T, x, ) deends on the data realization x. Based on the characterization of AIF for a given data realization x of length, we will then study the oulation version of AIF where each entry of X = {X,, X } is i.i.d generated by F θ. We will examine the behavior of AIF(T, X, ) as increases. Following the convention in robust statistics, we will assume that there exists a functional T such that (2.2) T (X) T (F θ ) in robability as. We will see that for a large class of estimators AIF(T, X, ) has a well-defined limit as. We will use AIF(T, F θ, ) to denote this limit when it exists. Similarly, from the defense s ersective, we would like to design an estimator that is least sensitive to the adversarial attack. Again, we will characterize the otimal estimator T, among a certain class of estimators T, that minimizes AIF(T, F θ, ). That is, for a certain class of estimators T, we will solve (2.3) min AIF(T, F θ, ). T T It will be clear in the sequel that the solution to the otimization roblem (2.3), even though is robust against adversarial attacks, has oor erformance in guarding against outliers. This motivates us to design estimators that strike a desirable tradeoff between these two robustness measures. In articular, we will solve (2.3) with an additional constraint on IF. We will need to use tools from calculus of variations for this urose. 3. Background. In this section, we briefly review results from classic robust estimator literature that are closely related to our study. 3.. Influence Function (IF). As mentioned above, in the classic robust estimation setu, it is assumed that a fraction η of data oints are outliers, while the remainder of data oints are generated from the true distribution F θ. For a given estimator T, the concet of IF introduced by Hamer [7] is defined T (( η)f θ + ηδ x ) T (F θ ) IF(x, T, F θ ) = lim. η 0 η In this definition, δ x is a distribution that uts mass at oint x, T (F θ ), introduced in (2.2), is the obtained estimate when all data oints are generated i.i.d from F θ, and T (( η)f θ + ηδ x ) is the obtained estimate when η fraction of data oints

6 6 E. BAYAKTAR AD L. LAI are generated i.i.d from F θ while η fraction of the data oints are at x. Hence, IF(x, T, F θ ) measures the asymtotic influence of having outliers at oint x as η 0. To measure the influence of the worst outliers, [7] then further introduced the concet of gross-error sensitivity of T by taking su over the absolute value of IF(x, T, F θ ): γ (T, F θ ) = su IF(x, T, F θ ). x Intuitively seaking, γ (T, F θ ) can be viewed as the solution of our roblem setu for the secial case of = 0. The values of IF(x, T, F θ ) and γ (T, F θ ) have been characterized for various class of estimators. Furthermore, under certain conditions, otimal estimator T that minimizes these quantities have been established. Some of these results will be introduced in later sections. More details can be found in [2, 3] M-Estimator. In this aer, we will mainly focus on a class of commonly used estimator in robust statistic: M-estimator [], in which one obtains an estimate T (x) of θ by solving (3.) ψ(x n, T ) = 0. n= Here ψ(x n, θ) is a function of data x n and arameter θ to be estimated. Different choices of ψ lead to different robust estimators. For examle, the most likely estimator (MLE) can be obtained by setting ψ = f θ /f θ. As the form of ψ determines T, in the remainder of the aer, we will use ψ and T interchangeably. For examle, we will denote AIF(T, x, ) as AIF(ψ, x, ). Similarly, we will denote IF(x, T, F θ ) as IF(x, ψ, F θ ). It is tyically assumed that ψ(x, θ) is continuous and almost everywhere differentiable. This assumtion is valid for all ψ s that are commonly used. It is also tyically assume the estimator is Fisher consistent [3]: (3.2) E Fθ [ψ(x, θ)] = 0, in which E Fθ means exectation under F θ. Intuitively seaking, this imlies that the true arameter θ is the solution of the M-estimator if there are increasingly more i.i.d. data oints generated from F θ. For M-estimator, IF(x, ψ, F θ ) was shown to be see (2.3.5) of [3]. ψ(x, T (F θ )) IF(x, ψ, F θ ) = θ [ψ(y, θ)] θ=t (F θ )df θ (y),

7 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 7 4. The Fixed Samle Case. In this section, we focus on analyzing AIF(ψ, x, ) for a given dataset x. We will extend the study to the oulation case and analyze AIF(ψ, F θ, ) in Section General ψ. We will first characterize AIF(ψ, x, ) for general ψ, and will then secialize the results to secific roblems in later sections. For any given ψ that is continuous and almost everywhere differentiable, we have the following theorem that characterizes AIF(ψ, x, ). THEOREM 4.. When =, AIF(ψ, x, ) = x [ψ] x=x n,θ=t, θ [ψ] x=x n,θ=t n= where (4.) n = arg max n x [ψ] x=x n,θ=t. For >, we have AIF(ψ, x, ) = ( n= x [ψ] x=x n,θ=t n= ). θ [ψ] x=x n,θ=t PROOF. Please see Aendix A for detailed roof. From Theorem 4., we can characterize the form of ψ that leads to the smallest AIF, i.e., the most robust M-estimator against adversarial attacks. COROLLARY 4.. AIF(ψ, x, ) n= n= x [ψ] x=x n,θ=t, θ [ψ] x=x n,θ=t and the equality holds when x [ψ] x=x,θ=t = = x [ψ] x=x,θ=t.

8 8 E. BAYAKTAR AD L. LAI PROOF. For >, it is easy to check that x ( )/ is a concave function when x 0. Hence, using Jensen s inequality, we have ( x [ψ] x=x n,θ=t n= ) x [ψ] x=x n,θ=t, and the equality holds when x [ψ] x=x n,θ=t is a constant with resect to n. This corollary imlies that, from defender s ersective, we should design ψ(x, θ) such that x [ψ] is constant in x. It is also interesting that, this result holds for any value of. And hence we can design an estimator without knowledge about which constraint the attacker is using Secific Estimators. To illustrate the results obtained above, we secialize results to location estimators and scale estimators Location Estimator. For location estimator models, F θ (x) = F 0 (x θ), and hence it is natural to use ψ(x, θ) = ψ(x θ), see [2, 3]. For this model, it is easy to check that x [ψ] x=x n,θ=t = ψ (x n T ), θ [ψ] x=x n,θ=t = ψ (x n T ). Plugging these two equations in the AIF exressions in Theorem 4., for the case with =, we have ψ (x n T ) AIF(ψ, x, ) =, ψ (x n T ) n= for which the equality holds when ψ (x n T ) is a constant with resect to n. For the case with >, we have (4.2) AIF(ψ, x, ) = (a) ( n= ψ (x n T ) n= ψ (x n T ) n= ψ (x n T ), ψ (x n T ) n= n= )

9 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 9 in which (a) is due to Jensen s inequality. Both inequalities will hold if ψ (x n T ) is a constant in n. EXAMPLE 4.. Consider an estimator with ψ(x n T ) = x n T. This estimator is simly the emirical samle mean. It is easy to see that ψ (x) is a constant in n, which imlies that this choice of ψ has AIF(ψ, x, ) =. It achieves the lower bound established above, regardless of the value of x and. Hence, it is the most robust estimator against adversarial attacks. However, as we will discuss in Section 5, this choice of ψ is not robust against outliers. In Section 5, we will design estimators that strike a desirable balance between robustness against outliers and robustness against adversarial attacks. EXAMPLE 4.2. Consider the Huber estimator [] with ψ(x n T ) = min{b, max{x n T, b}}, arameterized by a arameter 0 < b <. Using (4.2), it is easy to check that AIF(ψ, x, ) = /β, in which β is the ercentage of oints in x such that x n T < b. It is clear that Huber estimator, while being more robust against outliers [2], is less robust against than adversarial attacks than the emirical mean estimator Scale Estimator. The scale model [2, 3] is given by F θ (x) = F (x/θ), and it is tyical to consider ψ(x, θ) = ψ(x/θ). It is easy to check that for ψ with this form, we have x [ψ] x=x n,θ=t = ψ (x n /T ), T θ [ψ] x=x n,θ=t = x nψ (x n /T ). Using Theorem 4., for the case with =, we obtain ψ (x n /T ) T ψ (x n /T ) (4.3) AIF(ψ, x, ) = =, x n /T ψ (x n /T ) n= in which n is defined in (4.). x nψ (x n/t ) T 2 T 2 n=

10 0 E. BAYAKTAR AD L. LAI When >, we have AIF(ψ, x, ) = (4.4) = ( ( ψ (x n/t ) T n= n= n= n= ) x nψ (x n/t ) T 2 ) ψ (x n /T ) x n /T ψ (x n /T ). EXAMPLE 4.3. Consider MLE for variance of zero-mean Gaussian random variables, which corresonds to ψ(x) = x(φ (x)/φ(x)) = x 2. Here, φ(x) is the df of zero mean variance one Gaussian random variable. For this choice of ψ, we have T = n= x2 n and ψ (x n /T ) = 2x n /T. Plugging these values into (4.3), we obtain Using (4.4), we have AIF(ψ, x, ) = x n. ( ) AIF(ψ, x, ) = x n, from which we know that when = 2, AIF(ψ, x, 2) = T = n= x2 n. n= 5. Poulation Case. With the results on the fixed dataset case, we now consider the oulation version where X n are i.i.d from F θ, and analyze the behavior of AIF as. Following the convention in classic robust statistic literature, we will focus on the case in which the estimator is Fisher consistent as defined a.s. in (3.2). Under a broad range of regularity conditions, [2] shows that T θ. As a result, following similar arguments in Proosition 3 of [4], as in Theorem 4., we have, (5.) AIF(ψ, x, ) = a.s. x [ψ] x=x n,θ=t n= θ [ψ] x=x n,θ=t max x [ E Fθ θ [ψ](x, θ)] := AIF(ψ, F θ, ). x [ψ](x, θ)

11 For >, we have ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS (5.2) AIF(ψ, x, ) = ( a.s. x n [ψ] x=xn,θ=t θ [ψ] x=x n,θ=t n= n= ( [ E Fθ x [ψ](x, θ) [ EFθ θ [ψ](x, θ)] ) ]) := AIF(ψ, F θ, ). 5.. Location Estimator. We now secialize the results to the location model mentioned above. We will first characterize ψ that minimizes AIF(ψ, F θ, ). We will then discuss the tradeoff between the robustness to outliers and robustness to adversarial attacks, and will characterize the otimal ψ that achieves this tradeoff. In the location estimator, we will assume ψ(x, θ) is monotonic in θ, which will satisfy the regularity conditions established in [2] Minimizing AIF(ψ, F θ, ). For =, using (5.), we have max ψ (x θ) x AIF(ψ, F θ, ) = E Fθ [ψ (X θ)]. (5.3) For >, using (5.2), we obtain ( [ ψ E Fθ (X θ) AIF(ψ, F θ, ) = E Fθ [ψ (X θ)] ]) In articular, for = 2, we have E Fθ [ψ (X θ) AIF(ψ, F θ, 2) = 2 ] (E Fθ [ψ (X θ)]) 2. From (5.3) and using Jensen s equality, we have AIF(ψ, F θ, ), for which the equality holds when ψ (x θ) is constant in x..

12 2 E. BAYAKTAR AD L. LAI Tradeoff between AIF (ψ, F θ, ) and γ (ψ, F θ ). From (2.3.2) of [3], we know that the influence function of the location estimator secified by ψ is IF(x, ψ, F θ ) = ψ(x θ) E Fθ [ψ (X θ)], and hence γ (ψ, F θ ) = su x ψ(x θ) E Fθ [ψ (X θ)]. As a result, if ψ (X θ) is a constant that minimizes AIF(ψ, F θ, ) as discussed in Section 5.., then γ (ψ, F θ ) might go to, esecially for those distributions with unbounded suort. To achieve a desirable tradeoff between robustness to outliers (i.e., γ (ψ, F θ ) is small) and robustness to adversarial attacks (i.e., AIF(ψ, F θ, ) is small), in the following, we characterize the otimal estimator that minimizes AIF(ψ, F θ, ) subject to a constraint on γ (ψ, F θ ). (5.4) (5.5) (5.6) min AIF(ψ, F, 2) s.t. γ (ψ, F θ ) ξ, E Fθ [ψ(x θ)] = 0, ψ (x) 0, in which constraint (5.5) imlies that γ (ψ, F θ ) is uerbounded by a ositive constant ξ, constraint (5.6) imlies that ψ is Fisher consistent, and the last constraint comes from the condition that ψ is monotonic in θ. For location estimator, f θ (x) = f 0 (x θ), so all quantities in (5.4) remain the same by assuming θ = 0 [3]. Hence, in the following, we will solve this otimization roblem assuming θ = 0. THEOREM 5.. The solution to the otimization roblem (5.4) has the following structure: ψ (x) satisfies (5.7) ψ (x) = { ν ϑ 2 +(ϑ ϑ 2 )F 0(x) f 0 (x), ν f 0 (x) > ϑ F 0(x) + ϑ 2 ( F 0(x)); 0, otherwise, in which the arameters ν, ϑ 0 and ϑ 2 0 are arameters chosen to

13 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 3 satisfy the following conditions (5.8) (5.9) ϑ ( E F0 [ψ (X)] =, ) ψ (x)f 0 (x)dx ξ = 0, [ X ] ) (5.0) ϑ 2 (E F0 ψ (t)dt ξ = 0, along with [ ] ψ X (x)f 0 (x)dx ξ and E F0 ψ (t)dt ξ. [ ] X ψ() is set as E F0 ψ (t)dt. PROOF. Please see Aendix B for details. The condition ν f 0 (x) > ϑ F 0(x) + ϑ 2 ( F 0(x)) has a natural interretation. It will trim data oints from the tails. In articular, when x is left tail (i.e. F 0 is small), F 0 will be close to. On the other hand, when x is in the right tail (i.e. F 0 is small), F 0 will be close to. In these regions, ψ = 0 if the corresonding f 0 (x) is small. Figure illustrates the scenario for estimating the mean of Gaussian variables for the case assuming ϑ > ϑ 2. It is easy to check that, in this examle, if ν > 2π(ϑ + ϑ 2 ), there exist a and b such that ψ (x) = 0 when x < a or x > b. Corresondingly, ψ(x) is given as ξ ψ(x) = ξ + x a ψ (t)dt ξ x b a < x < b x < a. a b FIG. Gaussian mean examle

14 4 E. BAYAKTAR AD L. LAI 5.2. Scale Estimator. We now secialize the results to the scale model where F θ (x) = F (x/θ). For this model, it is natural to consider ψ(x, θ) = ψ(x/θ) [2, 3]. Similar to the location model, we will first characterize ψ that minimizes AIF(ψ, F θ, ). We will then discuss the tradeoff between the robustness to outliers and robustness to adversarial attacks, and will characterize the otimal ψ that achieves this tradeoff. For the case with =, using (5.), we obtain AIF(ψ, x, ) = T ψ (x n /T ) x n ψ (x n /T ) n= For >, using (5.), we have a.s. max θψ (x/θ) x E θ [Xψ (X/θ)] := AIF(ψ, F θ, ). AIF(ψ, x, ) = ( a.s. n= ) ψ (x n /T ) n= x n /T ψ (x n /T ) ( [ ψ E θ (X/θ) E θ [X/θψ (X/θ)] ]) := AIF(ψ, F θ, ). ( Since in scale model F θ (x) = F (x/θ), we have f θ (x) = f x ) θ θ, and hence AIF(ψ, F θ, ) = For = 2, we have ( [ ψ E F (X) E F [Xψ (X)] ]) := AIF(ψ, F, ). (5.) AIF(ψ, F, 2) = (E F [ψ (X) 2 ]) 2 E F [Xψ (X)] Minimizing AIF (ψ, F θ, ). In the following, among Fisher consistent estimators, we aim to design ψ that minimizes AIF(ψ, F, 2). THEOREM 5.2. structure: The otimal ψ that minimizes AIF(ψ, F, 2) has the following

15 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 5 For x in the range of f (x), ψ satisfies ψ() is chosen as ψ (x) = x E F [X 2 ]. ψ() = E F [ X ] ψ (t)dt. With this choice of ψ(x), the minimal value of AIF(ψ, F, 2) is / E F [X 2 ]. PROOF. Please see Aendix C for details. We note that for scale estimator ψ θ = ψ (x/θ)x/θ 2, hence for this articular choice of ψ in Theorem 5.2, ψ θ = x2 /(θ 3 E F [X 2 ]), which means ψ(x, θ) is monotone in θ. This ensures that the obtained ψ(x) satisfies the regularity conditions [2] mentioned at the beginning of this section Tradeoff between AIF (ψ, F θ, ) and γ (ψ, F θ ). Similar to the location estimation case, we can also design ψ to minimize AIF(ψ, F θ, ) with a constraint on γ (ψ, F θ ). From (2.3.7) of [3], we know that for scale estimators IF(x, ψ, F θ ) = ψ(x/θ)θ E Fθ [X/θψ (X/θ)]. To facilitate the analysis, we will focus on ψ that is monotonic. Since in scale model, ψ(x, θ) = ψ(x/θ), we can simly focus on the case of θ =. Hence, we will solve the following otimization roblem to strike a desirable tradeoff between robustness against outliers and robustness against adversarial attacks. [ ] E F ψ (X) 2 (5.2) min (E F [Xψ (X)]) 2, (5.3) (5.4) (5.5) s.t. γ (ψ, F ) = su x E F [ψ] = 0, ψ (x) 0. ψ(x) E F [Xψ (X)] ξ, Here, constraint (5.3) is a constraint on the outliers influence, (5.4) imlies that ψ is Fisher consistent. THEOREM 5.3. The solution to (5.2) has the following structure:

16 6 E. BAYAKTAR AD L. LAI ψ has the following form (5.6) ψ (x) = { ν x ϑ 2 +(ϑ ϑ 2 )F (x) f (x), ν xf (x) > ϑ F (x) + ϑ 2 ( F (x)); 0, otherwise, in which ν, ϑ 0 and ϑ 2 0 are chosen to satisfy ϑ ϑ 2 ( E F [Xψ (X)] ) =, = 0, ψ (x)f (x)dx ξ ] (E F [ X ψ (t)dt ) ξ = 0, along with [ ] X ψ (x)f (x)dx ξ and E F ψ (t)dt ξ. [ ] X ψ() is set to be E F ψ (t)dt. PROOF. The roof follows similar strategy as that of the roof of Theorem 5. and 5.2. Details can be found in Aendix D. Similar to the location estimator case, the condition ν xf (x) > ϑ F (x) + ϑ 2 ( F (x)) will limit the influences of data oints at the tails. 6. Extension: L-estimator. In this section, we briefly discuss how to extend the analysis above to other class of estimators. We will L-estimator as an examle. L-estimator has the following form [2, 3]: T (x) = a n x (n), n= where x () x () are the ordered sequence of x, and a n s are coefficients. For examle, for location estimator, a natural choice of a n is (6.) a n = n/ (n )/ h(t)dt 0 h(t)dt, for a given function h(t) such that 0 h(t)dt 0. For examle, setting h(t) = δ(t /2) leads to the median estimator.

17 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 7 We first look at the given samle scenario. Let x = x + x, and let x () x () be the ordered sequence of x. Hence, (6.2) T (x + x) = a n x (n). n= For general x, the ordering of x + x may not necessarily be the same as the ordering of x. For examle, x () might come from x (2), i.e., x () = x (2) + (x (2) ). This ossibility could make the following analysis messy. However, it is easy to show that for any ˆx that makes the ordering of x+ ˆx different from the ordering of x, we can find another x such that the ordering of x + x is the same as the ordering of x and x ˆx. With this, we can limit (6.2) to the following form Hence and (2.) becomes T (x + x) = a n (x (n) + (x (n) )). n= T (x + x) T (x) = min s.t. a n x (n), n= a n x (n), n= x η. Using the exactly same aroach as those in the roof of Theorem 4., we have the following characterization. For =, let n = arg max a n, n and x (n) = 0, n n. Hence, For >, we have x (n ) = sign {a n } η, AIF(T, x, ) = a n. x (n) = a n /( ) () / ( a n /( ) ) / sign(a n)η.

18 8 E. BAYAKTAR AD L. LAI Hence, (6.3) AIF(ψ, x, ) = a n a n /( ) () / ( a n /( ) ) / sign(a n) = ( a n /( ) n= ) /. a n /( ) n= When = 2, this can be simlified to (6.4) AIF(T, x, 2) = a 2 n n= a 2 n n= = a 2 n. n= For examle, for α-trimmed estimator [3] defined by T α (x) = 2 α α n= α + x (n), for a given arameter 0 < α < /2. For this α-trimmed estimator, using (6.3), we obtain AIF(T, α / x, ) = ( 2 α ) /. If a n s are chosen as (6.), then (6.4) simlifies to AIF(T, x, 2) = n=, ( ( n/ (n )/ h(t)dt ) 2 ( n= ( 0 h(t)dt ) 2 n/ (n )/ h(t)dt ) 2 0 h(t)dt ) 2 in which the first inequality is due to Jensen s inequality, and both inequalities become equality when a n = n/ (n )/ h(t)dt is a constant in n.

19 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 9 7. umerical Examles. In this section, we rovide numerical examles to illustrate results obtained. We consider location estimation and illustrate the otimal estimator obtained in Theorem 5. for the case when f 0 is exonential random variable f 0 (x) = e x, x 0, hence f θ is shifted exonential random variable f θ = e (x θ), x θ and the goal is to estimate θ. As the exonential random variable has a unbounded suort, choosing ψ to be a constant, which minimizes AIF, will lead to an infinite IF. Hence, we use Theorem 5. to characterize the otimal ψ that minimizes AIF while satisfying the condition that IF ξ. For this articular class of distribution, the condition ν f 0 (x) > ϑ F 0(x) + ϑ 2 ( F 0(x)) becomes 0 x < a with the arameter a chosen as (7.) Hence we have e a = ϑ ν + ϑ. ϑ 2 ψ (x) = { ν + ϑ ϑ 2 ϑ ex, 0 x < a; 0, otherwise, for which the arameters ν, ϑ, ϑ 2 are chosen to satisfy the conditions secified in Theorem 5.. After tedious calculation, conditions (5.8) - (5.0) can be simlified to (ν + ϑ ϑ 2)( e a ) ϑ a =, ϑ ((ν + ϑ ϑ 2)(a + e a ) ϑ (e a ) + aϑ ξ) = 0, ϑ 2((ν + ϑ ϑ 2)( e a ) ϑ a ξ) = 0. From here, we know that if ξ >, ϑ 2 = 0, using this fact along with (7.), we have that the conditions are simlified to ν aϑ =, 2aϑ + (a 2)ν = ξ, ϑ 2 = 0. Using these, we can exress ν and ϑ in terms of a: ν = ξ + 2 a, ϑ = ξ + 2 a a 2.

20 20 E. BAYAKTAR AD L. LAI a FIG 2. The solution of a Finally, for any given ξ >, the value of a can be determined by (7.), which is simlified to (7.2) e a = ϑ ν + ϑ ϑ 2 = ξ + 2 a (ξ + )a + ξ + 2. It is easy to check that, for any given ξ >, there is always a unique ositive solution to (7.2). For examle, Figure 2 illustrates the solution for a when ξ = 3. In this figure, the dotted curve is the right side of (7.2) and the solid curve is the left side of (7.2). From the figure, we know that these two curves have two intersections a = 0 and a = 4.8. With these arameters, we know that (7.3) ψ (x) = { e x, 0 x 4.8; 0, otherwise, hence the otimal ψ is { ψ ξ, x 4.8; (7.4) (x) =.047x (e x ), 0 x 4.8. Figure 3 illustrates the obtained ψ(x) for the case with ξ = 3. Figure 4 illustrates the tradeoff curve between AIF and IF. We obtain this curve by solving (7.2) and other arameters using different values of ξ. As we can see from the curve, as ξ increases, AIF decreases. Furthermore, the value of AIF converges to, the lower bound established in Section Conclusion. Motivated by recent data analytics alications, we have studied adversarial robustness of robust estimators. We have introduced the concet

21 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS (x) x FIG 3. ψ that minimizes AIF when IF AIF IF FIG 4. Tradeoff between AIF and IF of location estimator for exonential random variables. of AIF to quantify an estimator s sensitivity to such adversarial attacks and have rovided an aroach to characterize AIF for given robust estimator. We have further designed otimal estimators that minimize AIF. From this characterization, we have identified a tradeoff between AIF and IF, and have designed estimators that strike a desirable tradeoff between these two quantities. APPEDIX A: PROOF OF THEOREM 4. From (3.), we know that T and x satisfy ψ(x n, T ) = 0. n=

22 22 E. BAYAKTAR AD L. LAI Hence, we have (A.) x n T = x [ψ] x=x n,θ=t. θ [ψ] x=x n,θ=t n= Based on Taylor exansion, we have T (x + x) T (x) = x n T + higher order terms. x n n= When η is small, the adversary can solve the following roblem and obtain an o(η) otimal solution (A.2) in which min x x n c n, n= s.t. x η, c n := x n T. For =, this is a linear rograming roblem, whose solution is simle. In articular, let n = arg max n x n T, which is the same as arg max n x [ψ] x=x n,θ=t due to (A.), it is easy to check that we have { } x n = sign T η, x n and x n = 0, n n. Hence, AIF(ψ, x, ) = T = x n x [ψ] x=x n,θ=t. θ [ψ] x=x n,θ=t n= For >, (A.2) is a convex otimization roblem. To solve this, we form Lagrange L( x, λ) = x n c n + λ ( x η ). n=

23 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 23 The corresonding otimality conditions are: (A.3) c n + λ sign( x n) x n = 0, n λ 0, λ ( x η ) = 0. From (A.3), we know that λ 0, hence (A.4) x = η, and (A.5) sign( x n) x n = c n λ. From (A.5) and the fact that λ is ositive, we know sign( x n) = sign(c n ), and hence we have which can be simlified further to x n = x n = c n λ, ( ) cn /( ) λ sign(c n ). Combining these with (A.4), we obtain the value of λ : λ = c n /( ) n= η ( )/. As the result, we have Hence, x n = c n /( ) () / ( c n /( ) ) / sign(c n)η. AIF(ψ, x, ) = c n c n /( ) () / ( c n /( ) ) / sign(c n) = c n /( ) n= ( cn /( )) /.

24 24 E. BAYAKTAR AD L. LAI Using (A.), we can further simlify the exression to AIF(ψ, x, ) = ( x n [ψ] x=xn,θ=t. θ [ψ] x=x n,θ=t n= n= ) APPEDIX B: PROOF OF THEOREM 5. As ψ (x) 0, we have E F0 [ψ (X)] > 0, and su ψ(x) is either ψ( ) or ψ(). Hence for = 2, the otimization roblem (5.4) is equivalent to x min s.t. E F0 [ψ (X) 2 ] (E F0 [ψ (X)]) 2 ψ() + ψ (x)dx ξ, E F0 [ψ (X)] ψ() E F0 [ψ (X)] ξ, [ X ] ψ() + E F0 ψ (t)dt = 0, ψ 0. As the objective function does not involve ψ(), we can first solve min s.t. E F0 [ψ (X) 2 ] (E F0 [ψ (X)]) 2, [ ] X E F0 ψ (t)dt E F0 [ψ (X)] [ ] X E F0 ψ (t)dt ξ, E F0 [ψ (X)] + ψ (x)dx ξ, ψ (x) 0. [ ] X After obtaining the solution, we can simly set ψ() = E F0 ψ (t)dt to make ψ Fisher consistent. To simlify the notation, in the remainder of the roof, we will use g(x) to

25 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 25 denote ψ (x). We now further simlify the otimization roblem. First, we have (B.) E F0 [ X ] g(t)dt = = = [ x f 0 (x) [ g(t) t ] g(t)dt dx ] f 0 (x)dx dt g(t) [ F 0 (t)] dt. Couled with the fact that g(x) 0 and f 0 (x) 0, the otimization above is equivalent to min s.t. g2 (x)f 0 (x)dx ( ) 2, g(x)f 0(x)dx g(x) 0. g(x)f 0 (x)dx ξ g(x) [ F 0 (x)] dx ξ g(x)f 0 (x)dx, g(x)f 0 (x)dx, It is clear that the otimization roblem is scale invariant in the sense that if g (x) is a solution to this roblem, then for any ositive constant c, cg (x) is also a solution to this roblem. As a result, without loss of generality, we can assume g(x)f 0(x)dx =. Using this, we can further simlify the otimization roblem to min s.t. 2 g(x) 0. g 2 (x)f 0 (x)dx, g(x)f 0 (x)dx =, g(x)f 0 (x)dx ξ, g(x) [ F 0 (x)] dx ξ, To solve this functional minimization roblem, we first form the Lagrangian

26 26 E. BAYAKTAR AD L. LAI function L = ( g 2 (x)f 0 (x)dx + ν 2 ( ) ( +ϑ g(x)f 0 (x)dx ξ + ϑ 2 ) g(x)f 0 (x)dx + λ(x)g(x) ). g(x) [ F 0 (x)] dx ξ Let H = 2 g2 (x)f 0 (x) νg(x)f 0 (x) + ϑ g(x)f 0 (x) + ϑ 2 g(x)( F 0 (x)) λ(x)g(x). As no derivative g (x) is involved in H, the otimality condition Euler- Lagrange equation [9] simlifies to H g d dx ( ) H = 0 g (B.2) g (x)f 0 (x) ν f 0 (x) + ϑ 2 + (ϑ ϑ 2)F 0 (x) λ (x) = 0, in which the arameters ϑ 0, ϑ 2 0, λ (x) 0 satisfy [8] (B.3) g (x)f 0 (x)dx =, ( ) ϑ g (x)f 0 (x)dx ξ = 0, ( ) ϑ 2 g (x) [ F 0 (x)] dx ξ = 0, λ (x)g(x) 0. From (B.2), for x in the range of f 0 (x), we have g (x) = λ (x) + ν f 0 (x) ϑ 2 (ϑ ϑ 2 )F 0(x). f 0 (x) Combining this with the condition (B.3), we know that if ν f 0 (x) ϑ 2 (ϑ ϑ 2 )F 0(x) > 0, then λ (x) = 0. On the other hand, if ν f 0 (x) ϑ 2 (ϑ ϑ 2 )F 0(x) < 0, then g (x) = 0. As a result, we have g (x) = { ν ϑ 2 +(ϑ ϑ 2 )F 0(x) f 0 (x), ν f 0 (x) > ϑ F 0(x) + ϑ 2 ( F 0(x)); 0, otherwise, which comletes the roof.

27 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 27 APPEDIX C: PROOF OF THEOREM 5.2 First of all, minimizing (5.) is same as solving [ ] E F ψ (X) 2 (C.) min (E F [Xψ (X)]) 2, (C.2) s.t. E F [ψ(x)] = ψ() + E F [ X ] ψ (t)dt = 0, in which the condition E F [ψ(x)] = 0 ensures that the estimator is Fisher consistent. As ψ() does not aear in the objective function, we can solve (C.) without the constraint (C.2) first. After that, we can simly set ψ() = E F [ X ] ψ (t)dt so that the constraint (C.2) will be satisfied. Furthermore, similar to the roof of Theorem 5., to simlify the notation, we will use g(x) to denote ψ (x). It is clear from (C.) that the cost function is scale-invariant. Hence, without loss of generality, we can assume (E F [Xg(X)]) 2 =, for which we can further focus on E F [Xg(X)] =. Combining all these together, the otimization roblem can be converted to min s.t. 2 g 2 (x)f (x)dx, xg(x)f (x)dx =. For this calculus of variations roblem, we form Lagrange function ( ) L = 2 g2 (x)f (x)dx + ν xg(x)f (x)dx. The corresonding Euler-Lagrange equation can be simlified to (C.3) g (x)f (x) ν xf (x) = 0, and the otimal value of ν is selected to satisfy the condition (C.4) xg (x)f (x)dx =.

28 28 E. BAYAKTAR AD L. LAI From (C.3), we know that in the range of X where f (x) > 0, g (x) = ν x. Plugging this into (C.4), we obtain ν = x2 f (x)dx. As the result, for x in the range of f (x), the otimal g (x) is g x (x) = E F [X 2 ], and ψ() = E F [ X g(t)dt ]. APPEDIX D: PROOF OF THEOREM 5.3 Following the same strategy as those in the roof of Theorem 5., we can first solve the following roblem [ ] E F ψ (X) 2 min (E F [Xψ (X)]) 2, [ X ] s.t. E F ψ (t)dt + ψ (t)dt ξ E F [Xψ (X)], E F [ X ] ψ (t)dt ξ E F [Xψ (X)], ψ (x) 0, [ ] X and then set ψ() = E F ψ (t)dt to satisfy the Fisher consistent constraint (5.4). ow, we consider two different cases deending on whether E F [Xψ (X)] is ositive or negative. In the following, to simlify notation, we will use g(x) to denote ψ (x). We will solve the case with E F [Xg(X)] > 0 in detail. The case E F [Xg(X)] < 0 can be solved in the similar manner. With E F [Xg(X)] > 0, the otimization roblem is same as min s.t. E F [ g 2 (X) ] (E F [Xg(X)]) 2, [ X ] E F g(t)dt E F [ X g(x) 0. + g(t)dt ξe F [Xg(X)], ] g(t)dt ξe F [Xg(X)],

29 ADVERSARIAL ROBUSTESS OF ROBUST ESTIMATORS 29 Similar to the otimization roblems in Theorem 5. and 5.2, the otimization roblem is scale-invariant, and hence without loss of generality, [ we can focus on ] X E F [Xg(X)] =. Furthermore, similar to (B.), we have E F g(t)dt = g(t) [ F (t)] dt. The roblem is then converted to min s.t. g(x) 0. g 2 (x)f (x)dx, xg(x)f (x)dx =, g(x)f (x)dx ξ, g(x)[ F (x)]dx ξ, To solve this functional minimization roblem, we first form the Lagrangian function L = ( ) g 2 (x)f (x)dx + ν xg(x)f (x)dx + λ(x)g(x) 2 ( ) ( ) +ϑ g(x)f (x)dx ξ + ϑ 2 g(x) [ F (x)] dx ξ. Let F = 2 g2 (x)f (x) νxg(x)f (x) + ϑ g(x)f (x) + ϑ 2 g(x)( F (x)) λ(x)g(x). As no derivative g (x) is involved in F, the Euler-Lagrange equation F g d ( ) F = 0 dx g simlifies to (D.) g (x)f (x) ν xf (x) + ϑ 2 + (ϑ ϑ 2)F (x) λ (x) = 0, in which the arameters ϑ 0, ϑ 2 0, λ (x) 0 satisfy (D.2) xg (x)f (x)dx =, ( ) ϑ g (x)f (x)dx ξ = 0, ( ) ϑ 2 g (x) [ F (x)] dx ξ = 0, λ (x)g (x) 0.

30 30 E. BAYAKTAR AD L. LAI From (D.), for those x with f (x) > 0, we have g (x) = λ (x) + ν xf (x) ϑ 2 (ϑ ϑ 2 )F (x). f (x) Combining this with the condition (D.2), we know that if ν xf (x) ϑ 2 (ϑ ϑ 2 )F (x) > 0, then λ (x) = 0. On the other hand, if ν f (x) ϑ 2 (ϑ ϑ 2 )F (x) < 0, then g (x) = 0. As the result, we have g (x) = { ν x ϑ 2 +(ϑ ϑ 2 )F (x) f (x), ν xf (x) > ϑ F (x) + ϑ 2 ( F (x)); 0, otherwise. REFERECES [] Huber, P. J. (964). Robust estimation of a location arameter. Ann. Math. Statist., [2] Huber, P. J. and Ronchetti, E. (2009). Robust statistics. Wiley, ew York. [3] Hamel, F., Ronchetti, E., Rousseeuw, P., and Stahel, W. (2009). Robust statistics: The Aroach Based on Influence Functions. Wiley, Hoboken, J. [4] Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I. J. and Fergus, R. (203). Intriguing roerties of neural networks. CoRR, vol. abs/ [5] Goodfellow, I. J., Shlens, J. and Szegedy, C. (205). Exlaining and harnessing adversarial examles. in Proc. International Conference on Learning Reresentations, San Diego, CA. [6] Carlini,. and Wagner, D. (207). Towards evaluating the robustness of neural networks. In Proc. IEEE Intl. Symosium on Security and Privacy, San Jose, CA. [7] Hamel, F. (968). Contributions to the theory of robust estimation. Ph.D. thesis, University of California, Berkeley. [8] Burger, M. (2003). Infinite-dimensional Otimization and Otimal Design. Lecture notes, available at ft://ft.math.ucla.edu/ub/camreort/cam04-.df. [9] Kot, M. (204). A first course in the calculus of variations. Providence, Rhode Island: American Mathematical Society. [0] Pimentel-Alarcon, D. L., Biswas, A. and Solis-Lemus, C. R. (207). Adversarial rincial comonent analysis. In Proc. IEEE Intl. Symosium on Inform. Theory, [] Jagielski, M., Orea, A., Biggio, B., Liu, C., ita-rotaru, C. and Li, B. (208). Maniulating machine learning: Poisoning attacks and countermeasures for regression learning. In Proc. IEEE Intl. Symosium on Security and Privacy, San Francisco, CA. [2] Biggio, B., elson, B. and Laskov, P. (202). Poisoning attacks against suort vector machines. In Proc. International Conference on Machine Learning, Edinburgh, Scotland. [3] Charikar, M., Steinhardt, J., and Valiant, G. (207). Learning from untrusted data. In Proc. Annual ACM SIGACT Symosium on Theory of Comuting. [4] Croux, C. (998). Limit behavior of the emirical influence function of the median. Statistics & Probability Letter, DEPARTMET OF MATHEMATICS UIVERSITY OF MICHIGA A ARBOR, MI erhan@umich.edu DEPARTMET OF ELECTRICAL AD COMPUTER EGIEERIG UIVERSITY OF CALIFORIA DAVIS, CA, lflai@ucdavis.edu

On the Adversarial Robustness of Robust Estimators

On the Adversarial Robustness of Robust Estimators On the Adversarial Robustness of Robust Estimators Lifeng Lai and Erhan Bayraktar Abstract Motivated by recent data analytics applications, we study the adversarial robustness of robust estimators. Instead

More information

General Linear Model Introduction, Classes of Linear models and Estimation

General Linear Model Introduction, Classes of Linear models and Estimation Stat 740 General Linear Model Introduction, Classes of Linear models and Estimation An aim of scientific enquiry: To describe or to discover relationshis among events (variables) in the controlled (laboratory)

More information

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)]

LECTURE 7 NOTES. x n. d x if. E [g(x n )] E [g(x)] LECTURE 7 NOTES 1. Convergence of random variables. Before delving into the large samle roerties of the MLE, we review some concets from large samle theory. 1. Convergence in robability: x n x if, for

More information

Convex Optimization methods for Computing Channel Capacity

Convex Optimization methods for Computing Channel Capacity Convex Otimization methods for Comuting Channel Caacity Abhishek Sinha Laboratory for Information and Decision Systems (LIDS), MIT sinhaa@mit.edu May 15, 2014 We consider a classical comutational roblem

More information

Some Results on the Generalized Gaussian Distribution

Some Results on the Generalized Gaussian Distribution Some Results on the Generalized Gaussian Distribution Alex Dytso, Ronit Bustin, H. Vincent Poor, Daniela Tuninetti 3, Natasha Devroye 3, and Shlomo Shamai (Shitz) Abstract The aer considers information-theoretic

More information

ECON 4130 Supplementary Exercises 1-4

ECON 4130 Supplementary Exercises 1-4 HG Set. 0 ECON 430 Sulementary Exercises - 4 Exercise Quantiles (ercentiles). Let X be a continuous random variable (rv.) with df f( x ) and cdf F( x ). For 0< < we define -th quantile (or 00-th ercentile),

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

Introduction to Probability and Statistics

Introduction to Probability and Statistics Introduction to Probability and Statistics Chater 8 Ammar M. Sarhan, asarhan@mathstat.dal.ca Deartment of Mathematics and Statistics, Dalhousie University Fall Semester 28 Chater 8 Tests of Hyotheses Based

More information

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley Elements of Asymtotic Theory James L. Powell Deartment of Economics University of California, Berkeley Objectives of Asymtotic Theory While exact results are available for, say, the distribution of the

More information

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material Robustness of classifiers to uniform l and Gaussian noise Sulementary material Jean-Yves Franceschi Ecole Normale Suérieure de Lyon LIP UMR 5668 Omar Fawzi Ecole Normale Suérieure de Lyon LIP UMR 5668

More information

Estimation of the large covariance matrix with two-step monotone missing data

Estimation of the large covariance matrix with two-step monotone missing data Estimation of the large covariance matrix with two-ste monotone missing data Masashi Hyodo, Nobumichi Shutoh 2, Takashi Seo, and Tatjana Pavlenko 3 Deartment of Mathematical Information Science, Tokyo

More information

Statics and dynamics: some elementary concepts

Statics and dynamics: some elementary concepts 1 Statics and dynamics: some elementary concets Dynamics is the study of the movement through time of variables such as heartbeat, temerature, secies oulation, voltage, roduction, emloyment, rices and

More information

On a class of Rellich inequalities

On a class of Rellich inequalities On a class of Rellich inequalities G. Barbatis A. Tertikas Dedicated to Professor E.B. Davies on the occasion of his 60th birthday Abstract We rove Rellich and imroved Rellich inequalities that involve

More information

VARIANTS OF ENTROPY POWER INEQUALITY

VARIANTS OF ENTROPY POWER INEQUALITY VARIANTS OF ENTROPY POWER INEQUALITY Sergey G. Bobkov and Arnaud Marsiglietti Abstract An extension of the entroy ower inequality to the form N α r (X + Y ) N α r (X) + N α r (Y ) with arbitrary indeendent

More information

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek

Use of Transformations and the Repeated Statement in PROC GLM in SAS Ed Stanek Use of Transformations and the Reeated Statement in PROC GLM in SAS Ed Stanek Introduction We describe how the Reeated Statement in PROC GLM in SAS transforms the data to rovide tests of hyotheses of interest.

More information

Shadow Computing: An Energy-Aware Fault Tolerant Computing Model

Shadow Computing: An Energy-Aware Fault Tolerant Computing Model Shadow Comuting: An Energy-Aware Fault Tolerant Comuting Model Bryan Mills, Taieb Znati, Rami Melhem Deartment of Comuter Science University of Pittsburgh (bmills, znati, melhem)@cs.itt.edu Index Terms

More information

Asymptotically Optimal Simulation Allocation under Dependent Sampling

Asymptotically Optimal Simulation Allocation under Dependent Sampling Asymtotically Otimal Simulation Allocation under Deendent Samling Xiaoing Xiong The Robert H. Smith School of Business, University of Maryland, College Park, MD 20742-1815, USA, xiaoingx@yahoo.com Sandee

More information

Improved Bounds on Bell Numbers and on Moments of Sums of Random Variables

Improved Bounds on Bell Numbers and on Moments of Sums of Random Variables Imroved Bounds on Bell Numbers and on Moments of Sums of Random Variables Daniel Berend Tamir Tassa Abstract We rovide bounds for moments of sums of sequences of indeendent random variables. Concentrating

More information

On Z p -norms of random vectors

On Z p -norms of random vectors On Z -norms of random vectors Rafa l Lata la Abstract To any n-dimensional random vector X we may associate its L -centroid body Z X and the corresonding norm. We formulate a conjecture concerning the

More information

On Wald-Type Optimal Stopping for Brownian Motion

On Wald-Type Optimal Stopping for Brownian Motion J Al Probab Vol 34, No 1, 1997, (66-73) Prerint Ser No 1, 1994, Math Inst Aarhus On Wald-Tye Otimal Stoing for Brownian Motion S RAVRSN and PSKIR The solution is resented to all otimal stoing roblems of

More information

Analysis of some entrance probabilities for killed birth-death processes

Analysis of some entrance probabilities for killed birth-death processes Analysis of some entrance robabilities for killed birth-death rocesses Master s Thesis O.J.G. van der Velde Suervisor: Dr. F.M. Sieksma July 5, 207 Mathematical Institute, Leiden University Contents Introduction

More information

1 Extremum Estimators

1 Extremum Estimators FINC 9311-21 Financial Econometrics Handout Jialin Yu 1 Extremum Estimators Let θ 0 be a vector of k 1 unknown arameters. Extremum estimators: estimators obtained by maximizing or minimizing some objective

More information

On the minimax inequality and its application to existence of three solutions for elliptic equations with Dirichlet boundary condition

On the minimax inequality and its application to existence of three solutions for elliptic equations with Dirichlet boundary condition ISSN 1 746-7233 England UK World Journal of Modelling and Simulation Vol. 3 (2007) No. 2. 83-89 On the minimax inequality and its alication to existence of three solutions for ellitic equations with Dirichlet

More information

Advanced Calculus I. Part A, for both Section 200 and Section 501

Advanced Calculus I. Part A, for both Section 200 and Section 501 Sring 2 Instructions Please write your solutions on your own aer. These roblems should be treated as essay questions. A roblem that says give an examle requires a suorting exlanation. In all roblems, you

More information

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning TNN-2009-P-1186.R2 1 Uncorrelated Multilinear Princial Comonent Analysis for Unsuervised Multilinear Subsace Learning Haiing Lu, K. N. Plataniotis and A. N. Venetsanooulos The Edward S. Rogers Sr. Deartment

More information

STA 250: Statistics. Notes 7. Bayesian Approach to Statistics. Book chapters: 7.2

STA 250: Statistics. Notes 7. Bayesian Approach to Statistics. Book chapters: 7.2 STA 25: Statistics Notes 7. Bayesian Aroach to Statistics Book chaters: 7.2 1 From calibrating a rocedure to quantifying uncertainty We saw that the central idea of classical testing is to rovide a rigorous

More information

4. Score normalization technical details We now discuss the technical details of the score normalization method.

4. Score normalization technical details We now discuss the technical details of the score normalization method. SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules

More information

Khinchine inequality for slightly dependent random variables

Khinchine inequality for slightly dependent random variables arxiv:170808095v1 [mathpr] 7 Aug 017 Khinchine inequality for slightly deendent random variables Susanna Sektor Abstract We rove a Khintchine tye inequality under the assumtion that the sum of Rademacher

More information

A Note on Guaranteed Sparse Recovery via l 1 -Minimization

A Note on Guaranteed Sparse Recovery via l 1 -Minimization A Note on Guaranteed Sarse Recovery via l -Minimization Simon Foucart, Université Pierre et Marie Curie Abstract It is roved that every s-sarse vector x C N can be recovered from the measurement vector

More information

Journal of Mathematical Analysis and Applications

Journal of Mathematical Analysis and Applications J. Math. Anal. Al. 44 (3) 3 38 Contents lists available at SciVerse ScienceDirect Journal of Mathematical Analysis and Alications journal homeage: www.elsevier.com/locate/jmaa Maximal surface area of a

More information

Chapter 7: Special Distributions

Chapter 7: Special Distributions This chater first resents some imortant distributions, and then develos the largesamle distribution theory which is crucial in estimation and statistical inference Discrete distributions The Bernoulli

More information

Improved Capacity Bounds for the Binary Energy Harvesting Channel

Improved Capacity Bounds for the Binary Energy Harvesting Channel Imroved Caacity Bounds for the Binary Energy Harvesting Channel Kaya Tutuncuoglu 1, Omur Ozel 2, Aylin Yener 1, and Sennur Ulukus 2 1 Deartment of Electrical Engineering, The Pennsylvania State University,

More information

Finite Mixture EFA in Mplus

Finite Mixture EFA in Mplus Finite Mixture EFA in Mlus November 16, 2007 In this document we describe the Mixture EFA model estimated in Mlus. Four tyes of deendent variables are ossible in this model: normally distributed, ordered

More information

Sums of independent random variables

Sums of independent random variables 3 Sums of indeendent random variables This lecture collects a number of estimates for sums of indeendent random variables with values in a Banach sace E. We concentrate on sums of the form N γ nx n, where

More information

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK

Towards understanding the Lorenz curve using the Uniform distribution. Chris J. Stephens. Newcastle City Council, Newcastle upon Tyne, UK Towards understanding the Lorenz curve using the Uniform distribution Chris J. Stehens Newcastle City Council, Newcastle uon Tyne, UK (For the Gini-Lorenz Conference, University of Siena, Italy, May 2005)

More information

On Isoperimetric Functions of Probability Measures Having Log-Concave Densities with Respect to the Standard Normal Law

On Isoperimetric Functions of Probability Measures Having Log-Concave Densities with Respect to the Standard Normal Law On Isoerimetric Functions of Probability Measures Having Log-Concave Densities with Resect to the Standard Normal Law Sergey G. Bobkov Abstract Isoerimetric inequalities are discussed for one-dimensional

More information

Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various Families of R n Norms and Some Open Problems

Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various Families of R n Norms and Some Open Problems Int. J. Oen Problems Comt. Math., Vol. 3, No. 2, June 2010 ISSN 1998-6262; Coyright c ICSRS Publication, 2010 www.i-csrs.org Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various

More information

Stochastic integration II: the Itô integral

Stochastic integration II: the Itô integral 13 Stochastic integration II: the Itô integral We have seen in Lecture 6 how to integrate functions Φ : (, ) L (H, E) with resect to an H-cylindrical Brownian motion W H. In this lecture we address the

More information

The non-stochastic multi-armed bandit problem

The non-stochastic multi-armed bandit problem Submitted for journal ublication. The non-stochastic multi-armed bandit roblem Peter Auer Institute for Theoretical Comuter Science Graz University of Technology A-8010 Graz (Austria) auer@igi.tu-graz.ac.at

More information

arxiv:cond-mat/ v2 25 Sep 2002

arxiv:cond-mat/ v2 25 Sep 2002 Energy fluctuations at the multicritical oint in two-dimensional sin glasses arxiv:cond-mat/0207694 v2 25 Se 2002 1. Introduction Hidetoshi Nishimori, Cyril Falvo and Yukiyasu Ozeki Deartment of Physics,

More information

Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process

Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process P. Mantalos a1, K. Mattheou b, A. Karagrigoriou b a.deartment of Statistics University of Lund

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell October 25, 2009 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

Elementary Analysis in Q p

Elementary Analysis in Q p Elementary Analysis in Q Hannah Hutter, May Szedlák, Phili Wirth November 17, 2011 This reort follows very closely the book of Svetlana Katok 1. 1 Sequences and Series In this section we will see some

More information

MATH 2710: NOTES FOR ANALYSIS

MATH 2710: NOTES FOR ANALYSIS MATH 270: NOTES FOR ANALYSIS The main ideas we will learn from analysis center around the idea of a limit. Limits occurs in several settings. We will start with finite limits of sequences, then cover infinite

More information

arxiv: v2 [stat.me] 3 Nov 2014

arxiv: v2 [stat.me] 3 Nov 2014 onarametric Stein-tye Shrinkage Covariance Matrix Estimators in High-Dimensional Settings Anestis Touloumis Cancer Research UK Cambridge Institute University of Cambridge Cambridge CB2 0RE, U.K. Anestis.Touloumis@cruk.cam.ac.uk

More information

Lecture 3 January 16

Lecture 3 January 16 Stats 3b: Theory of Statistics Winter 28 Lecture 3 January 6 Lecturer: Yu Bai/John Duchi Scribe: Shuangning Li, Theodor Misiakiewicz Warning: these notes may contain factual errors Reading: VDV Chater

More information

Approximating min-max k-clustering

Approximating min-max k-clustering Aroximating min-max k-clustering Asaf Levin July 24, 2007 Abstract We consider the roblems of set artitioning into k clusters with minimum total cost and minimum of the maximum cost of a cluster. The cost

More information

An Estimate For Heilbronn s Exponential Sum

An Estimate For Heilbronn s Exponential Sum An Estimate For Heilbronn s Exonential Sum D.R. Heath-Brown Magdalen College, Oxford For Heini Halberstam, on his retirement Let be a rime, and set e(x) = ex(2πix). Heilbronn s exonential sum is defined

More information

Prospect Theory Explains Newsvendor Behavior: The Role of Reference Points

Prospect Theory Explains Newsvendor Behavior: The Role of Reference Points Submitted to Management Science manuscrit (Please, rovide the mansucrit number! Authors are encouraged to submit new aers to INFORMS journals by means of a style file temlate, which includes the journal

More information

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests 009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 0-, 009 FrB4. System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests James C. Sall Abstract

More information

Principal Components Analysis and Unsupervised Hebbian Learning

Principal Components Analysis and Unsupervised Hebbian Learning Princial Comonents Analysis and Unsuervised Hebbian Learning Robert Jacobs Deartment of Brain & Cognitive Sciences University of Rochester Rochester, NY 1467, USA August 8, 008 Reference: Much of the material

More information

arxiv: v1 [physics.data-an] 26 Oct 2012

arxiv: v1 [physics.data-an] 26 Oct 2012 Constraints on Yield Parameters in Extended Maximum Likelihood Fits Till Moritz Karbach a, Maximilian Schlu b a TU Dortmund, Germany, moritz.karbach@cern.ch b TU Dortmund, Germany, maximilian.schlu@cern.ch

More information

COMMUNICATION BETWEEN SHAREHOLDERS 1

COMMUNICATION BETWEEN SHAREHOLDERS 1 COMMUNICATION BTWN SHARHOLDRS 1 A B. O A : A D Lemma B.1. U to µ Z r 2 σ2 Z + σ2 X 2r ω 2 an additive constant that does not deend on a or θ, the agents ayoffs can be written as: 2r rθa ω2 + θ µ Y rcov

More information

IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES

IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES OHAD GILADI AND ASSAF NAOR Abstract. It is shown that if (, ) is a Banach sace with Rademacher tye 1 then for every n N there exists

More information

Positive decomposition of transfer functions with multiple poles

Positive decomposition of transfer functions with multiple poles Positive decomosition of transfer functions with multile oles Béla Nagy 1, Máté Matolcsi 2, and Márta Szilvási 1 Deartment of Analysis, Technical University of Budaest (BME), H-1111, Budaest, Egry J. u.

More information

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO)

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO) Combining Logistic Regression with Kriging for Maing the Risk of Occurrence of Unexloded Ordnance (UXO) H. Saito (), P. Goovaerts (), S. A. McKenna (2) Environmental and Water Resources Engineering, Deartment

More information

Distributed Rule-Based Inference in the Presence of Redundant Information

Distributed Rule-Based Inference in the Presence of Redundant Information istribution Statement : roved for ublic release; distribution is unlimited. istributed Rule-ased Inference in the Presence of Redundant Information June 8, 004 William J. Farrell III Lockheed Martin dvanced

More information

Positivity, local smoothing and Harnack inequalities for very fast diffusion equations

Positivity, local smoothing and Harnack inequalities for very fast diffusion equations Positivity, local smoothing and Harnack inequalities for very fast diffusion equations Dedicated to Luis Caffarelli for his ucoming 60 th birthday Matteo Bonforte a, b and Juan Luis Vázquez a, c Abstract

More information

Quantitative estimates of propagation of chaos for stochastic systems with W 1, kernels

Quantitative estimates of propagation of chaos for stochastic systems with W 1, kernels oname manuscrit o. will be inserted by the editor) Quantitative estimates of roagation of chaos for stochastic systems with W, kernels Pierre-Emmanuel Jabin Zhenfu Wang Received: date / Acceted: date Abstract

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell February 10, 2010 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

CMSC 425: Lecture 4 Geometry and Geometric Programming

CMSC 425: Lecture 4 Geometry and Geometric Programming CMSC 425: Lecture 4 Geometry and Geometric Programming Geometry for Game Programming and Grahics: For the next few lectures, we will discuss some of the basic elements of geometry. There are many areas

More information

DEPARTMENT OF ECONOMICS ISSN DISCUSSION PAPER 20/07 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES

DEPARTMENT OF ECONOMICS ISSN DISCUSSION PAPER 20/07 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES DEPARTMENT OF ECONOMICS ISSN 1441-549 DISCUSSION PAPER /7 TWO NEW EXPONENTIAL FAMILIES OF LORENZ CURVES ZuXiang Wang * & Russell Smyth ABSTRACT We resent two new Lorenz curve families by using the basic

More information

1 Probability Spaces and Random Variables

1 Probability Spaces and Random Variables 1 Probability Saces and Random Variables 1.1 Probability saces Ω: samle sace consisting of elementary events (or samle oints). F : the set of events P: robability 1.2 Kolmogorov s axioms Definition 1.2.1

More information

Lecture 6. 2 Recurrence/transience, harmonic functions and martingales

Lecture 6. 2 Recurrence/transience, harmonic functions and martingales Lecture 6 Classification of states We have shown that all states of an irreducible countable state Markov chain must of the same tye. This gives rise to the following classification. Definition. [Classification

More information

Journal of Inequalities in Pure and Applied Mathematics

Journal of Inequalities in Pure and Applied Mathematics Journal of Inequalities in Pure and Alied Mathematics htt://jiam.vu.edu.au/ Volume 3, Issue 5, Article 8, 22 REVERSE CONVOLUTION INEQUALITIES AND APPLICATIONS TO INVERSE HEAT SOURCE PROBLEMS SABUROU SAITOH,

More information

Slash Distributions and Applications

Slash Distributions and Applications CHAPTER 2 Slash Distributions and Alications 2.1 Introduction The concet of slash distributions was introduced by Kafadar (1988) as a heavy tailed alternative to the normal distribution. Further literature

More information

Unobservable Selection and Coefficient Stability: Theory and Evidence

Unobservable Selection and Coefficient Stability: Theory and Evidence Unobservable Selection and Coefficient Stability: Theory and Evidence Emily Oster Brown University and NBER August 9, 016 Abstract A common aroach to evaluating robustness to omitted variable bias is to

More information

WALD S EQUATION AND ASYMPTOTIC BIAS OF RANDOMLY STOPPED U-STATISTICS

WALD S EQUATION AND ASYMPTOTIC BIAS OF RANDOMLY STOPPED U-STATISTICS PROCEEDINGS OF HE AMERICAN MAHEMAICAL SOCIEY Volume 125, Number 3, March 1997, Pages 917 925 S 0002-9939(97)03574-0 WALD S EQUAION AND ASYMPOIC BIAS OF RANDOMLY SOPPED U-SAISICS VICOR H. DE LA PEÑA AND

More information

Department of Mathematics

Department of Mathematics Deartment of Mathematics Ma 3/03 KC Border Introduction to Probability and Statistics Winter 209 Sulement : Series fun, or some sums Comuting the mean and variance of discrete distributions often involves

More information

CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules

CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules. Introduction: The is widely used in industry to monitor the number of fraction nonconforming units. A nonconforming unit is

More information

substantial literature on emirical likelihood indicating that it is widely viewed as a desirable and natural aroach to statistical inference in a vari

substantial literature on emirical likelihood indicating that it is widely viewed as a desirable and natural aroach to statistical inference in a vari Condence tubes for multile quantile lots via emirical likelihood John H.J. Einmahl Eindhoven University of Technology Ian W. McKeague Florida State University May 7, 998 Abstract The nonarametric emirical

More information

Real Analysis 1 Fall Homework 3. a n.

Real Analysis 1 Fall Homework 3. a n. eal Analysis Fall 06 Homework 3. Let and consider the measure sace N, P, µ, where µ is counting measure. That is, if N, then µ equals the number of elements in if is finite; µ = otherwise. One usually

More information

On Doob s Maximal Inequality for Brownian Motion

On Doob s Maximal Inequality for Brownian Motion Stochastic Process. Al. Vol. 69, No., 997, (-5) Research Reort No. 337, 995, Det. Theoret. Statist. Aarhus On Doob s Maximal Inequality for Brownian Motion S. E. GRAVERSEN and G. PESKIR If B = (B t ) t

More information

Elementary theory of L p spaces

Elementary theory of L p spaces CHAPTER 3 Elementary theory of L saces 3.1 Convexity. Jensen, Hölder, Minkowski inequality. We begin with two definitions. A set A R d is said to be convex if, for any x 0, x 1 2 A x = x 0 + (x 1 x 0 )

More information

Semiparametric Efficiency in GMM Models with Nonclassical Measurement Error

Semiparametric Efficiency in GMM Models with Nonclassical Measurement Error Semiarametric Efficiency in GMM Models with Nonclassical Measurement Error Xiaohong Chen New York University Han Hong Duke University Alessandro Tarozzi Duke University August 2005 Abstract We study semiarametric

More information

Feedback-error control

Feedback-error control Chater 4 Feedback-error control 4.1 Introduction This chater exlains the feedback-error (FBE) control scheme originally described by Kawato [, 87, 8]. FBE is a widely used neural network based controller

More information

The spectrum of kernel random matrices

The spectrum of kernel random matrices The sectrum of kernel random matrices Noureddine El Karoui Deartment of Statistics, University of California, Berkeley Abstract We lace ourselves in the setting of high-dimensional statistical inference,

More information

On the capacity of the general trapdoor channel with feedback

On the capacity of the general trapdoor channel with feedback On the caacity of the general tradoor channel with feedback Jui Wu and Achilleas Anastasooulos Electrical Engineering and Comuter Science Deartment University of Michigan Ann Arbor, MI, 48109-1 email:

More information

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM JOHN BINDER Abstract. In this aer, we rove Dirichlet s theorem that, given any air h, k with h, k) =, there are infinitely many rime numbers congruent to

More information

Conversions among Several Classes of Predicate Encryption and Applications to ABE with Various Compactness Tradeoffs

Conversions among Several Classes of Predicate Encryption and Applications to ABE with Various Compactness Tradeoffs Conversions among Several Classes of Predicate Encrytion and Alications to ABE with Various Comactness Tradeoffs Nuttaong Attraadung, Goichiro Hanaoka, and Shota Yamada National Institute of Advanced Industrial

More information

Interactive Hypothesis Testing Against Independence

Interactive Hypothesis Testing Against Independence 013 IEEE International Symosium on Information Theory Interactive Hyothesis Testing Against Indeendence Yu Xiang and Young-Han Kim Deartment of Electrical and Comuter Engineering University of California,

More information

Almost All Palindromes Are Composite

Almost All Palindromes Are Composite Almost All Palindromes Are Comosite William D Banks Det of Mathematics, University of Missouri Columbia, MO 65211, USA bbanks@mathmissouriedu Derrick N Hart Det of Mathematics, University of Missouri Columbia,

More information

A New Perspective on Learning Linear Separators with Large L q L p Margins

A New Perspective on Learning Linear Separators with Large L q L p Margins A New Persective on Learning Linear Searators with Large L q L Margins Maria-Florina Balcan Georgia Institute of Technology Christoher Berlind Georgia Institute of Technology Abstract We give theoretical

More information

Bayesian Spatially Varying Coefficient Models in the Presence of Collinearity

Bayesian Spatially Varying Coefficient Models in the Presence of Collinearity Bayesian Satially Varying Coefficient Models in the Presence of Collinearity David C. Wheeler 1, Catherine A. Calder 1 he Ohio State University 1 Abstract he belief that relationshis between exlanatory

More information

How to Estimate Expected Shortfall When Probabilities Are Known with Interval or Fuzzy Uncertainty

How to Estimate Expected Shortfall When Probabilities Are Known with Interval or Fuzzy Uncertainty How to Estimate Exected Shortfall When Probabilities Are Known with Interval or Fuzzy Uncertainty Christian Servin Information Technology Deartment El Paso Community College El Paso, TX 7995, USA cservin@gmail.com

More information

Universal Finite Memory Coding of Binary Sequences

Universal Finite Memory Coding of Binary Sequences Deartment of Electrical Engineering Systems Universal Finite Memory Coding of Binary Sequences Thesis submitted towards the degree of Master of Science in Electrical and Electronic Engineering in Tel-Aviv

More information

Sharp gradient estimate and spectral rigidity for p-laplacian

Sharp gradient estimate and spectral rigidity for p-laplacian Shar gradient estimate and sectral rigidity for -Lalacian Chiung-Jue Anna Sung and Jiaing Wang To aear in ath. Research Letters. Abstract We derive a shar gradient estimate for ositive eigenfunctions of

More information

MATH 250: THE DISTRIBUTION OF PRIMES. ζ(s) = n s,

MATH 250: THE DISTRIBUTION OF PRIMES. ζ(s) = n s, MATH 50: THE DISTRIBUTION OF PRIMES ROBERT J. LEMKE OLIVER For s R, define the function ζs) by. Euler s work on rimes ζs) = which converges if s > and diverges if s. In fact, though we will not exloit

More information

#A37 INTEGERS 15 (2015) NOTE ON A RESULT OF CHUNG ON WEIL TYPE SUMS

#A37 INTEGERS 15 (2015) NOTE ON A RESULT OF CHUNG ON WEIL TYPE SUMS #A37 INTEGERS 15 (2015) NOTE ON A RESULT OF CHUNG ON WEIL TYPE SUMS Norbert Hegyvári ELTE TTK, Eötvös University, Institute of Mathematics, Budaest, Hungary hegyvari@elte.hu François Hennecart Université

More information

State Estimation with ARMarkov Models

State Estimation with ARMarkov Models Deartment of Mechanical and Aerosace Engineering Technical Reort No. 3046, October 1998. Princeton University, Princeton, NJ. State Estimation with ARMarkov Models Ryoung K. Lim 1 Columbia University,

More information

Lower Confidence Bound for Process-Yield Index S pk with Autocorrelated Process Data

Lower Confidence Bound for Process-Yield Index S pk with Autocorrelated Process Data Quality Technology & Quantitative Management Vol. 1, No.,. 51-65, 15 QTQM IAQM 15 Lower onfidence Bound for Process-Yield Index with Autocorrelated Process Data Fu-Kwun Wang * and Yeneneh Tamirat Deartment

More information

An Analysis of Reliable Classifiers through ROC Isometrics

An Analysis of Reliable Classifiers through ROC Isometrics An Analysis of Reliable Classifiers through ROC Isometrics Stijn Vanderlooy s.vanderlooy@cs.unimaas.nl Ida G. Srinkhuizen-Kuyer kuyer@cs.unimaas.nl Evgueni N. Smirnov smirnov@cs.unimaas.nl MICC-IKAT, Universiteit

More information

Lecture 4: Law of Large Number and Central Limit Theorem

Lecture 4: Law of Large Number and Central Limit Theorem ECE 645: Estimation Theory Sring 2015 Instructor: Prof. Stanley H. Chan Lecture 4: Law of Large Number and Central Limit Theorem (LaTeX reared by Jing Li) March 31, 2015 This lecture note is based on ECE

More information

On Code Design for Simultaneous Energy and Information Transfer

On Code Design for Simultaneous Energy and Information Transfer On Code Design for Simultaneous Energy and Information Transfer Anshoo Tandon Electrical and Comuter Engineering National University of Singaore Email: anshoo@nus.edu.sg Mehul Motani Electrical and Comuter

More information

Probability Estimates for Multi-class Classification by Pairwise Coupling

Probability Estimates for Multi-class Classification by Pairwise Coupling Probability Estimates for Multi-class Classification by Pairwise Couling Ting-Fan Wu Chih-Jen Lin Deartment of Comuter Science National Taiwan University Taiei 06, Taiwan Ruby C. Weng Deartment of Statistics

More information

Optimal Design of Truss Structures Using a Neutrosophic Number Optimization Model under an Indeterminate Environment

Optimal Design of Truss Structures Using a Neutrosophic Number Optimization Model under an Indeterminate Environment Neutrosohic Sets and Systems Vol 14 016 93 University of New Mexico Otimal Design of Truss Structures Using a Neutrosohic Number Otimization Model under an Indeterminate Environment Wenzhong Jiang & Jun

More information

Session 5: Review of Classical Astrodynamics

Session 5: Review of Classical Astrodynamics Session 5: Review of Classical Astrodynamics In revious lectures we described in detail the rocess to find the otimal secific imulse for a articular situation. Among the mission requirements that serve

More information

Inference for Empirical Wasserstein Distances on Finite Spaces: Supplementary Material

Inference for Empirical Wasserstein Distances on Finite Spaces: Supplementary Material Inference for Emirical Wasserstein Distances on Finite Saces: Sulementary Material Max Sommerfeld Axel Munk Keywords: otimal transort, Wasserstein distance, central limit theorem, directional Hadamard

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Introduction to Optimization (Spring 2004) Midterm Solutions

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Introduction to Optimization (Spring 2004) Midterm Solutions MASSAHUSTTS INSTITUT OF THNOLOGY 15.053 Introduction to Otimization (Sring 2004) Midterm Solutions Please note that these solutions are much more detailed that what was required on the midterm. Aggregate

More information

A Special Case Solution to the Perspective 3-Point Problem William J. Wolfe California State University Channel Islands

A Special Case Solution to the Perspective 3-Point Problem William J. Wolfe California State University Channel Islands A Secial Case Solution to the Persective -Point Problem William J. Wolfe California State University Channel Islands william.wolfe@csuci.edu Abstract In this aer we address a secial case of the ersective

More information