A Characterization of Guesswork on Swiftly Tilting Curves

Size: px
Start display at page:

Download "A Characterization of Guesswork on Swiftly Tilting Curves"

Transcription

1 A Characterization of Guesswork on Swiftly Tilting Curves Ahma Beirami, Robert Calerbank, Mark Christiansen, Ken Duffy, an Muriel Méar Abstract arxiv: v [cs.it] 27 Jan 208 Given a collection of strings, each with an associate probability of occurrence, the guesswork of each of them is their position in an orere list from most likely to least likely, breaking ties arbitrarily. Guesswork is central to several applications in information theory: Average guesswork provies a lower boun on the expecte computational cost of a sequential ecoer to ecoe successfully the intene message; the complementary cumulative istribution function of guesswork gives the error probability in list ecoing; the logarithm of guesswork is the number of bits neee in optimal lossless oneto-one source coing; an guesswork is the number of trials require of an aversary to breach a passwor protecte system in a brute-force attack. In this paper, we consier memoryless string-sources that generate strings consisting of i.i.. characters rawn from a finite alphabet, an characterize their corresponing guesswork. Our main tool is the tilt operation on a memoryless string-source. We show that the tilt operation on a memoryless string-source parametrizes an exponential family of memoryless stringsources, which we refer to as the tilte family of the string-source. We provie an operational meaning to the tilte families by proving that two memoryless string-sources result in the same guesswork on all strings of all lengths if an only if their respective categorical istributions belong to the same tilte family. Establishing some general properties of the tilt operation, we generalize the notions of weakly typical set an asymptotic equipartition property to tilte weakly typical sets of ifferent orers. We use This work was presente in part in the 53r Annual Allerton Conference on Communication, Control, an Computing Allerton 205 []. This work was supporte in part by the Air Force Office of Scientific Reseah AFOSR uner awars FA an FA ; the work of K. Duffy an M. Méar was also supporte in part by Netapp through a Faculty Fellowship. A. Beirami is with the Research Laboratory of Electronics, Massachusetts Institute of Technology, USA beirami@mit.eu. R. Calerbank is with the Department of Electrical an Computer Engineering, Duke University, USA robert.calerbank@uke.eu. M. Christiansen an K. Duffy are with the Hamilton Institute, National University of Irelan Maynooth, Irelan mark.christiansen, ken.uffy@nuim.ie. M. Méar is with the Research Laboratory of Electronics, Massachusetts Institute of Technology, USA mear@mit.eu.

2 this new efinition to characterize the large eviations for all atypical strings an characterize the volume of weakly typical sets of ifferent orers. We subsequently buil on this characterization to prove large eviation bouns on guesswork an provie an accurate approximation of its probability mass function. Inex Terms Tilting; Orering; Guesswork; One-to-One Coes; Rényi Entropy; Information Geometry; Weakly Typical Set. I. INTRODUCTION A. Motivation An early motivation for the stuy of guesswork is to provie lower bouns on the computational complexity of sequential ecoing [2] where the ecoer sequentially examines several paths until it fins the correct coe sequence. Guesswork was first stuie by Massey [3], who showe that average guesswork is lower boune in terms of the Shannon entropy of the source. Arıkan [4] consiere guesswork on memoryless sources an prove that the moments of guesswork are relate to the Rényi entropy rates of the source in the it as the strings become long. This has been generalize to ergoic Markov chains [5] an a wie range of stationary sources [6]. It has also been stuie subject to an allowable istortion [7], an subject to constraine Shannon entropy [8]. Hanawal an Sunaresan [9] reerive the moments of guesswork assuming that the logarithm of guesswork satisfies a large eviations principle LDP. Christiansen an Duffy [0] establishe that guesswork satisfies a LDP an characterize the rate function. They also provie an approximation to the istribution of guesswork using the rate function associate with the LDP. In list ecoing, rather than eclaring one coewor message, the ecoer returns a list of size L of possible coewors [] [3]. The error event in list ecoing is the event where the message coewor is not returne in the list of the L returne coewors. For strings of fixe length, as the size of the list grows larger, the probability of error ecreases. The list ecoer can be use as an inner coe for an outer coe that eventually ecies the intene message from the returne list of messages by the inner list ecoer. The optimal list ecoer returns the L most likely messages given the receive coewor. Therefore, the probability of the error event in the optimal list ecoer is characterize by the complementary cumulative istribution function of guesswork conitione on the receive coewor see [4], [5]. Guesswork can also be use to quantify computational security against brute-force attack [6]. Suppose that a secret string is rawn from a given process on a finite alphabet, an is use to secure a passwor-

3 2 protecte system. Further suppose that the system only grants access if the correct secret string is provie exactly, an oes not reveal any information otherwise. If a brute-force attacker aversary ecies to guess the secret string by query, the best strategy at an aversary s isposal is to query the possible secret strings from the most likely to the least likely in terms of minimizing the average number of queries or maximizing the probability of success using a finite number of queries. In this case, guesswork gives the number of queries that it woul take an attacker to fin the secret string using the best strategy. Christiansen et al. [7] stuie guesswork over the weakly typical set an prove that the exponent is strictly smaller than that of a uniform set with the same support size; they showe that the average guesswork of a passwor over an erasure channel oes not relate to the average noise in general [8]; Christiansen et al. also consiere the setting where an attacker wants to guess one or more out of many secret strings rawn inepenently from not necessarily ientical string-sources [6]. Finally, the iea of guesswork has been extene to the setup where the probability istribution is unknown [8], [9], [20]. In the context of source coing, it is known that the length of the optimal one-to-one source coe that nee not satisfy Kraft s inequality is within one bit of the logarithm of guesswork. The average length of a one-to-one coe is relate to the normalize zeroth moment of guesswork or alternatively the expecte logarithm of guesswork see [20]. The source coing problem without prefix constraint ates back to Wyner [2], who showe that the average coewor length of one-to-one coes is upper boune by the entropy. Alon an Orlistky erive a lower boun on the average coewor length in terms of the Shannon entropy [22], which was recently revisite for other moments of guesswork [23]. Szpankowski [24] erive the asymptotic average coewor length of one-to-one coes on binary memoryless sources, which was subsequently generalize to finite-alphabet inepenent an ientically istribute i.i.. processes [25], [26], an later stuie uner a universal setup [20], [27]. B. Problem setup, efinitions, an notation Let X := a,..., a X be an orere set that enotes a finite alphabet of size X. We will subsequently use the orering on X to break ties in guesswork an reverse guesswork. Denote an n-string over X by x n+k k = x k x k+... x n+k X n. Further, let x n = x n an for i > n, xn i =, where enotes the null string. Let µ n enote a multinomial probability istribution on the sample space X n, where µ is the corresponing categorical istribution on X, i.e., µ n x n = n i= µ x i. We refer to µ := µ n : n N as a memoryless string-source, also calle source. Let θ = θ,..., θ X be the parameter vector corresponing to the categorical istribution µ. That is θ i := µ a i = P [X = a i ] for all i [ X ]. By efinition, θ is an element of the X imensional simplex of all stochastic vectors of size X. Denote Θ X as the X -ary probability simplex, i.e., Θ X := θ R X + : i [ X ] θ i =. We reserve

4 3 upper case letters for ranom variables while lower case letters correspon to their realizations. For example, X n enotes an n-string ranomly rawn from µ while x n X n is a realization of the ranom n-string. Note that as string lengths become large, the logarithm of the guesswork normalize by string length has been shown to satisfy a LDP for a more general set of sources [0]. While existing results on guesswork consier the orering of strings from most likely to least likely, either unconitionally or conitionally starting at [4], or universal guessing orers [9], the more restrictive assumptions use here enable us to treat, for the first time, the reverse guesswork process. In reverse guessing, strings are orere from least likely to most likely, breaking ties arbitrarily. At first glance, consieration of the worst possible strategy seems to be of ite value, but we will show that is not the case. Forwar guesswork provies goo estimates of the likelihoo of quickly iscovere strings, but we shall emonstrate reverse guesswork focuses on better estimates on the likelihoo that large numbers of guesses occur. We also establish an inherent geometrical corresponence between string-sources whose orers are reverse. Amongst other results, we prove that logarithm of reverse guesswork scale by string length satisfies a LDP with a concave rate function. The concavity captures the fact that as one moves along the reverse orer the rate of acquisition of probability increases initially. Due to it, the approach employe in [0], which leverage earlier results on the scaling behavior of moments of guesswork, has no hope of success an we instea use istinct means. Definition information: Let information ı n µ : X n R + be efine as: ı n µx n := log µ n x n. Note that throughout this paper we use log to enote the natural logarithm an hence information is measure in nats as efine above. Definition 2 Shannon entropy: Let H n µ enote the entropy, which is efine as the expecte information of a ranom n-string, an quantifies the average uncertainty that is reveale when a ranom n-string rawn from µ is observe: H n µ := E µ ı n µx n = x n X n µ n x n ı n µx n, 2 where E µ enotes the expectation operator with respect to the measure µ. The entropy rate, is enote by Hµ an is given by Hµ := n n Hn µ = H µ. Definition 3: For any n N, we efine [n] :=,..., n an [n] 0 := 0,..., n. R + := x R x > 0.

5 4 In what follows we put forth two assumptions on the string-source for us to prove that the logarithm of reverse guesswork satisfies a LDP. We will also later characterize the set of string-sources that satisfy the assumptions. Assumption : We assume that µ x > 0 for all x X, i.e., θ i > 0 for all i [ X ]. Define ε := min x X µ x, 3 ε := max x X µ x. 4 Assumption exclues the bounaries of the probability simplex an ensures that the likelihoo of the least likely n-string scales exponentially with n. In other wors, for all n N, an all x n X n n log µ n x n log <. 5 ε Assumption 2: We assume that arg min x X µ x an arg max x X µ x are unique. Assumption 2 ensures that the most likely an the least likely wors are unambiguously etermine so that tilting the string-source with large positive an negative orers woul lea to a eterministic stringsource that only outputs a single string with high probability. As such, for each istinct i, j [ X ], the exclue set of parameter vectors will be a subset of the set of all X 2-imensional hyperplanes that satisfy θ i = θ j, an the number of such hyperplanes is X 2. As such, the set of parameter vectors that o not satisfy Assumptions an 2 can be characterize by X 2 -imensional hyperplanes in the X -imensional probability simplex Θ X. 2 The probability simplex for a ternary alphabet, Θ 3, is epicte in Fig., where the green ashe an green soli lines represent the set of string-sources in Θ 3 that o not satisfy Assumption 2. Note that in this case, X 2 = an X 2 = 3 resulting in three -imensional hyperplanes. Definition 4 set M X of string-sources: We enote by M X, the set of all memoryless string-sources on alphabet X that satisfy Assumptions an 2. In this paper, we focus our attention to string-sources in M X. Definition 5 guesswork/optimal orering for µ: A one-to-one function G n µ : X n [ X n ] is sai to be the optimal orering for µ on n-strings if µ n x n > µ n y n G n µx n < G n µy n, 6 an further if µ n x n = µ n y n, then G n µx n < G n µy n if an only if x n appears prior to y n in lexicographic orering. 2 In other wors, the exclue set is of zero Lebesgue volume in the probability simplex.

6 5 Definition 6 reverse guesswork/reverse orering for µ: We efine R n µx n as the reverse guesswork, given by R n µx n := X n G n µx n 7 While guesswork correspons to the best guessing strategy for recovering the string x n, reverse guesswork correspons to the worst guessing strategy for the string-source µ as it procees from the least likely string all the way to the most likely. Hence, the statistical performance of any other guessing strategy orering on n-strings lies somewhere between that of guesswork an reverse guesswork. As will become clear later in the paper, our interest in reverse guesswork is mainly ue to its usefulness in characterizing the PMF of guesswork. Definition 7 logarithm of forwar/reverse guesswork: We efine the logarithm of forwar guesswork g n µ : X n [ X n ] an the logarithm of reverse guesswork r n µ : X n [ X n ] as g n µx n := log G n µx n, 8 r n µx n := log R n µx n, 9 where G n µ an R n µ are guesswork an the reverse guesswork efine in 6 an 7, respectively. Definition 8 varentropy [28]: Let the varentropy, enote by V n µ, be efine as V n µ = var µ ı n µ X n 0 where var µ enotes the variance operator with respect to the measure µ. The varentropy rate is enote by V µ an is given by V µ := n n V n µ = V µ. Definition 9 relative entropy: Denote D n µ ρ as the relative entropy, also referre to as the Kullback-Leibler KL ivergence, between the measures µ n an ρ n, given by D n µ ρ := µ µ n x n n x n log ρ n x n = nd µ ρ. 2 x n X n Definition 0 cross entropy: Let the cross entropy, enote by H n µ ρ, be efine as H n µ ρ :=E µ log ρ n X n = µ n x n log ρ n x n. 3 x n X n =H n µ + D n µ ρ 4 =nh µ + nd µ ρ. 5

7 6 Observe that H n µ µ = H n µ recovers the Shannon entropy. As a natural generalization of varentropy an cross entropy, we efine cross varentropy in the following efinition. Definition cross varentropy: Let the cross varentropy, enote by V n µ ρ, be efine as V n µ ρ :=var µ log ρ n X n = [ 2 µ n x n log ρ n x n H µ ρ] n. 6 x n X n Note that V n µ µ = V n µ simplifies to the varentropy. Further note that V n µ ρ = nv µ ρ because of our restriction to i.i.. string-sources. An operational interpretation of the cross entropy H n µ ρ arises in prefix-free source coing, where it is known that, ignoring the integer constraints on the coewor length, the length of the coewor for the string x n that minimizes the average coewor length equals ı n µx n = log µ n x n. In this case, the resulting average coewor length is given by E µ ı n µx n = E µ log µ n X n = H n µ. 7 Therefore, if a source coe is esigne for the mismatche string-source ρ, the length of the coewor for x n woul be chosen to be ı n ρx n. Hence, the average coewor length of the optimal prefix-free source coe that is esigne to minimize the average coewor length uner the mismatche istribution ρ is equal to the cross entropy: E µ ı n ρx n = E µ log ρ n X n = H n µ ρ. 8 Definition 2 Rényi entropy: For any istribution µ n, let the Rényi entropy of orer α, enote by H n α, be efine as [29] Hαµ n := α log [µ n x n ] α. 9 x n X n Further, efine the Rényi entropy rate of orer α if it exists as H α µ := n n Hn αµ = Hαµ. 20 Note that α H α µ = Hµ, which recovers the Shannon entropy. C. The tilt The primary objective of this paper is to evelop a tool, calle the tilt, for the characterization of guesswork, its moments, an its large eviations behavior. The tilt operation T on any string-source µ is formally efine below.

8 7 Definition 3 tilte string-source of orer α: Let the map T operate on a string-source µ an a real-value parameter α an output a string-source calle tilte µ of orer α, which is enote by T µ, α an is efine as T µ, α := T µ n, α : n N, 2 where T µ n, α with overloaing the notation for all n N an all x n X n is efine as T µ n, αx n := [µ n x n ] α y n X n[µn y n ] α. 22 Note that the tilt is a real analytic function of α in R for all µ M X an all n N, formally state in Lemma 7. Tilte µ of orer is the string-source µ itself, i.e., T µ, = µ. Further, for any µ M X, T µ, 0 = u X, where u X is efine as the uniform string-source on alphabet X such that for all x n X n Notice that by efinition u X M X. u n X x n := X n. 23 Observe that for all n N, T µ n, α is a multinomial istribution with the corresponing categorical istribution given by τθ, α = τ θ, α,..., τ X θ, α, where τ i : Θ X R Θ X for all i [ X ] is given by Let Γ + θ τ i θ, α := θ α i X ı= θα i Θ X enote the tilte family of θ an be given by. 24 Γ + θ := τθ, α : α R Observe that Γ + θ Θ X is a set of stochastic vectors in the probability simplex. Thus, the tilte family of a memoryless string-source with parameter vector θ is comprise of a set of memoryless string-sources whose parameter vectors belong to the tilte family of the vector θ, i.e., Γ + θ. Definition 4 reverse string-source: We efine the reverse of the string-source µ also calle µ- reverse, enote by T µ,. Note that the guesswork corresponing to µ-reverse is exactly the reverse guesswork. As an example, the reverse of a binary memoryless string-source with parameter vector θ = p, p is a binary memoryless string-source with τθ, = p, p. Observe that Rµ n is the optimal orering guesswork for the reverse string source T µ,. Further, G n µ is the reverse orering for T µ,. Definition 5 tilte/reverse family of a string-source: Let the set T µ + enote the tilte family of the string-source µ, an be given by T µ + := T µ, α : α R +. 26

9 8 0,0, θ - θ θ +,0,0 0,,0 θ 0 Fig. : The soli black lines epict the bounaries of the probability simplex Θ 3 of all ternary stochastic vectors, which are exclue from our analysis as they o not satisfy Assumption. The ashe green lines represent the string-sources that o not satisfy 29 in Assumption 2; an the soli green lines represent the string-sources that o not satisfy 30 in Assumption 2. The re pentagon marker is the ternary source parameter vector θ = 0.2, 0.3, 0.5; the blue soli curve epicts Γ + θ, i.e., the tilte family of θ; the re otte curve epicts Γ θ, i.e., the reverse family of θ; the yellow iamon marker correspons to the uniform parameter vector θ 0 = 3, 3, 3 ; an θ =, 0, 0 an θ = 0, 0,. Note that θ an θ are only achieve in the it an o not belong to either Γ + θ or Γ θ. Further, efine the reverse family of the string-source as T µ := T µ, α : α R. 27 The first notable property of the tilte family of µ is that it parametrizes a stanar exponential family of string-sources with α. In other wors, for any n N, the family of istributions T µ n, α : α R parametrize by α characterizes an exponential family of istributions on X n. As a working example throughout the paper, we consier a memoryless string-source on a ternary alphabet X = 3 with the parameter vector θ = 0.2, 0.3, 0.5. The ternary alphabet is general enough to encompass the ifficulties of the problem while simple enough for us to emonstrate the geometric techniques evelope in this paper. The ternary probability simplex is epicte in Fig., where the tilte family Γ + θ is plotte for θ = 0.2, 0.3, 0.5. As can be seen, the tilte family of θ is a parametric curve that lives in the 2-imensional simplex of stochastic vectors. The curve starts arbitrarily close to the maximum entropy uniform parameter vector θ 0 for α = 0 + an then moves towars one of the

10 9 zero-entropy corner points of the simplex corresponing to the most likely symbol as α. We shall show that these properties hol generally for all string-sources in M X. As a final note, the tilt is intimately relate to the concept of information geometry see [30], where several geometric properties of the Kullback-Leibler ivergence relative entropy have been stuie an their relevance to problems in information theory have been investigate. In particular, Borae an Zheng applie concepts from information geometry to the analysis of error exponents [3]; an Huang et al. performe a geometric stuy of hypothesis testing [32]. This work also buils upon the insights on the geometry of the probability simplex in the context of information theory an statistics pioneere by Barron, Clarke, an Cover [33], [34]. The preinary version of this work relie on the metho of types see [] while in this paper we buil on a generalize notion of the weakly typical set, calle the tilte weakly typical sets. We believe that the current treatment may be generalize to a class of stationary ergoic string-sources as iscusse in the concluing remarks. D. Organization an contribution In Section II, we provie the main properties of the tilt operation. In particular, we characterize the entropy H n T µ, α, the cross entropy H n T µ, α µ, the relative entropy D n T µ, α µ of the tilite string-source as well as the Rényi entropy Hαµ. n These quantities are use in Section V to implicitly characterize the large eviations rate function of the logarithm of reverse guesswork, the logarithm of forwar guesswork, an information. We further show that the erivatives with respect to the parameter α of these quantities can be expresse in terms of the varentropy V n T µ, α an the cross varentropy V n T µ, α µ of the tilte string-source. In Section III, we efine the equivalent orer class of a string-source as the set of all string-sources that result in the same guesswork for all n, an enote this set of string-sources by C µ. In Theorem, we show that the equivalent orer class of µ is equal to the tilte family, i.e., C µ = T µ +, proviing an operational meaning to the tilte family. Further, the set of all string-sources whose optimal orerings are inverse for µ is equal to Tµ, i.e., C T µ, = Tµ. This result raws an important istinction between information an guesswork. The string-source µ is uniquely etermine if information ı n µx n is provie for all x n X n. On the other han, it is not possible to recover uniquely the string-source µ if one is provie with the guesswork G n µx n for all strings x n X n. Our result shows that one can only etermine the string-source up to equivalence within the tilte family that the source belongs to given the guesswork for all x n X n leaing to one less egree of freeom. This result has implications in terms of universal source coing [20], [27] an universal guesswork [8], [9], where the egree of freeom in the moel class etermines the penalty of universality, which is calle reunancy in source coing.

11 0 In Section IV, we generalize the notion of the weakly typical set. The weakly typical set comprises of all strings x n X n such that e Hn µ nɛ < µ n x n < e Hn µ+nɛ. An important implication of the weakly typical set known as the asymptotic equipartition property is that the typical set roughly contains e Hn µ strings each of probability e Hn µ. For sufficiently large n, the probability of the typical set is close to one, i.e., if a string is ranomly rawn from the source it is in the weakly typical set with high probability. The weakly typical set is use to prove key results in information theory that are ominate by the typical events, such as the source coing theorem, an the channel coing theorem. Despite the great intuition that typicality provies on such quantities, the arguments base on typicality fall short in proviing intuition on other key quantities that are ominate by atypical events, such as the error exponents in ecoing, an the cut-off rate in sequential ecoing. In Section IV, we consier all strings x n X n such that e Hn T µ,α µ nɛ < µ n x n < e Hn T µ,α µ+nɛ for α R, which we efine as the tilte weakly typical set of orer α an enote by A n µ,α,ɛ see Definition 8. We show that for a fixe ɛ, for any string x n X n, there exists an α such that x n is containe in the tilte weakly typical set of orer α. Hence, this notion coul be use to cluster/partition all of the strings x n X n base on their likelihoos. In Theorem 2, we erive tight bouns on the probability, size, guesswork, an reverse guesswork for all elements of the tilte weakly typical set of any orer α. Roughly speaking, for any α R +, we show that A n µ,α,ɛ contains e Hn T µ,α strings where the likelihoo of each string is e Hn T µ,α µ, an the probability of the entire set is e Dn T µ,α µ, an the guesswork for each element in the set is e Hn T µ,α. For any α R, we show that A n µ,α,ɛ contains e Hn T µ,α strings where the likelihoo of each string is e Hn T µ,α µ, an the probability of the entire set is e Dn T µ,α µ, an the reverse guesswork for each element in the set is e Hn T µ,α. Note that α = recovers the weakly typical set as H n T µ, µ = H n T µ, = H n µ, an for α, this provies a generalization of the well-known weakly typical set. This generalization coul be use to provie new insights on unerstaning an proving results on information theoretic quantities that are ominate by atypical events. In Section V, we prove that the logarithm of reverse guesswork rµx n n, the logarithm of forwar guesswork gµx n n, an information ı n µx n satisfy a LDP an provie an implicit formula for the rate function in each case base on entropy, cross entropy, an relative entropy. Although some of the results in this section have been previously prove in more generality, we believe that the machinery built upon the tilt operation in this paper as well as the information theoretic characterization of the rate functions provie useful intuition in characterizing the large eviations behavior of these quantities. In particular, the LDP for the logarithm of reverse guesswork coul not have been establishe using the previous methos by Christiansen an Duffy [0], as the resulting rate function here is concave. The most common way

12 to prove a LDP is the Gärtner-Ellis Theorem [35, Theorem 2.3.6]. There one first etermines the scale cumulant generating function of the process of interest, which ientifies how moments scale, an then leverages properties of it to establish the LDP. This approach can only work if the resulting rate function is convex. Here, that is not the case. In Section VI, we provie an approximation on the istribution of guesswork that captures the behavior even for very small string lengths. The approximation is built base on the tilt operation, guesswork, an reverse guesswork. We empirically show that the approximation can be applie to a variety of string-sources to characterize the PMF of guesswok. Finally, the conclusion is provie in Section VII, where we also present ieas on how to generalize these results beyon the setting consiere in this paper. II. PROPERTIES OF THE TILTED STRING-SOURCES In this section, we provie the main properties of the tilt operation that shall be use in proving our results on the characterization of guesswork. This section may be skippe by the reaer an only referre to whenever a particular lemma is neee. Lemma : For any α, β R, T T µ, α, β = T T µ, β, α = T µ, αβ. 28 Proof: This is obtaine from the efinition using light algebraic manipulation. Lemma 2: The optimal orering for µ is the reverse orering for T µ, an the reverse orering for µ is the optimal orering for T µ,. Proof: The statement is a consequence of the efinition. Lemma 3: For any µ M X, Proof: The Lemma is a irect consequence of Assumption 2. HT µ, α = 0, 29 α HT µ, α = α Next, we provie a lemma that relates Rényi entropy, Shannon entropy, relative entropy, an cross entropy. Lemma 4: For any µ M X, an any α R, the following relations hol: H n T µ, α µ H n αµ = α Dn T µ, α µ, 3

13 2 where α Dn T µ, α µ := 0. α= Proof: To establish 3, H n T µ, α H n αµ = H n T µ, α µ H n αµ α Dn T µ, α µ = α α α α Dn T µ, α µ, 32 x n X n T µ n, αx n log µ n x n 33 [µ n x n ] α 34 x n X n [µ n T µ n, αx n x n ] α / y log n X n[µn y n ] α µ n x n x n X n 35 = 0, 36 where 35 follows from the efinition, an 36 hols because the term y n X n[µn y n ] α in 35 is a constant, an hence 35 coul be ecompose into two terms one of which cancels the right han sie of 33, an the other cancels 34. Finally, 32 is establishe by putting together 4 an 3. Lemma 5: For any µ M X, an any α R, the following hols: log T µ n, αx n Hn T µ, α = α log µ n x n Hn T µ, α µ. 37 Proof: The lemma is prove by consiering the following chain of equations: log T µ n, αx n Hn T µ, α = α log µ n x n + log [µ n x n ] α x n X n T µ n, αx n x n X n log = α α log µ n x n Hn T µ, α µ µ n x n + log [µ n x n ] α x n X n. Lemma 6: For any µ M X, an any α R, the following hols: α 2 V n T µ, α µ = V n T µ, α. 38

14 3 Proof: We have α 2 V n T µ, α µ = x n X n T µ n, αx n = x n X n T µ n, αx n = V n T µ, α, [ α log µ n x n ] 2 H n T µ, α µ [ ] 2 log T µ n, αx n Hn T µ, α 39 where 39 follows from Lemma 5. Lemma 7: For any string-source µ M X, for any n N, T µ n, α is a real analytic function of α R. In particular, α T µn, α exists; an α 0 T µ n, α = u n X, where un X X n. is the uniform istribution on Proof: The statement follows from the facts that the exponential function is real analytic in its argument an that the enominator in 22 in the efinition of the tilt is non-zero for all α R. The it as α 0 hols since µ n x n > 0 for all x n X n by Assumption. Lemma 8: For any string-source µ M X, for any n N, H n T µ, α, D n T µ, α µ, H n T µ, α µ, V n T µ, α, an V n T µ, α µ are continuous an ifferentiable functions of α R. Proof: Note that H n ρ, D n ρ µ, H n ρ µ, V n ρ, an V n ρ µ are continuously ifferentiable with respect to ρ M X, an by Lemma 7, T µ, α is also continuous an ifferentiable with respect to α R, an hence the overall composition by consiering ρ = T µ, α is continuous an ifferentiable with respect to α R. Lemma 9: For any µ M X, an any n N, an any α R: α T µn, αx n = T µ n, αx n H n T µ, α µ log µ n x n. 40 Proof: Let T µ n, α 0 x n := α T µn, αx n. 4 α=α0 Note that by Lemma, we know that for any β > 0: T µ n, αx n = T T µ n, β, α x n. 42 β

15 4 Differentiating both sies with respect to α, for all α 0, β > 0 we have α T µn, αx n = T µ n, α 0 x n = α=α0 α T T µ n, β, α β = β T T µ n, β, α 0 β x n α=α0 43 x n 44 = α 0 T T µ n, α 0, x n, 45 where 45 follows by setting β = α 0. In other wors, 45 implies that the erivative of a tilt with respect to α can be erive by taking the erivative at α = an rescaling the erivative evaluate at the tilte istribution. Consequently, to prove the lemma it suffices to prove that α T µn, αx n = µ n x n H n µ log α= µ n x n, 46 an then it follows from 45 that for any α 0, we will arrive at the following: H n T µ, α log α T µn, αx n = α T µn, αx n T µ n, αx n. 47 On the other han, the esire result is obtaine by consiering Lemma 5, an noting the continuity an ifferentiability of T µ n, α at 0 from Lemma 7 an that of HT µ n, α at 0 from Lemma 8. To establish 46, note the efinition of the tilt in 22 an the fact that the erivative of the numerator with respect to α is [µ n x n ] α log µ n x n an the erivative of the enominator with respect to α is -H n µ. Lemma 0: For any µ M X, we have α 0 α 2 V n T µ, α = V n u X µ. 48 Proof: Observe that α 0 α 2 V n T µ, α = α 0 V n T µ, α µ 49 = V n u X µ, where 49 follows from Lemma 6, completing the proof. Lemma : For any µ M X, an any α R, α Hn T µ, α = αv n T µ, α µ. 50 Proof: First, owing to Lemma 8, the erivative of H n T µ n, α with respect to α exists.

16 5 To prove the result for all α 0, it suffices to show that α Hn T µ, α = V n µ 5 α= for any given µ M X. Then, by Lemma, T, α + α = T T, + α/α, α which for α 0 leas to see 45 in the proof of Lemma 9 for a more formal statement of this part of the claim: α Hn T µ, α = α V n T µ, α. 52 This, in conjunction with Lemma 6 an continuity an ifferentiability of H n T µ, α with respect to all α R from Lemma 8 leas to the esire result. To show that 5 hols, we have α Hn T µ, α α= = α T µn, αx n log x n X n α= µ n x n + µ n x n α log T µ n, αx n x n X n α= = µ n x n H n µ log µ n x n log µ n x n x n X n + µ n x n log µ n x n Hn µ x n X n = E µ H n µ log µ n X n log µ n X n 53 = var µ ı n µx n = V n µ, 54 where 53 follows from Lemma 9, an 54 follows from the efinition of the varentropy in 0. The it as α 0 also follow from Lemma 0. Lemma 2: For any µ M X an any α R, α Hn T µ, α µ = V n T µ, α µ. 55

17 6 Proof: We have α Hn T µ, α µ = x n X n = α T µn, αx n log x n X n T µ n, αx n µ n x n H n T µ, α µ log = E T µ,α H n T µ, α µ log µ n X n µ n x n log log µ n X n µ n x n 56 = var T µ,α ı n µx n 57 = V n T µ, α µ, where 56 follows from Lemma 9, an 57 follows from the efinition of information in an cross entropy in 3 leaing to the fact that E T µ,α log µ n X n = H n T µ, α µ. 58 Lemma 3: For any µ M X an any α R, α Dn T µ, α µ = α V n T µ, α µ. 59 Proof: Note that from 4, D n T µ, α µ = H n T µ, α µ H n T µ, α. 60 The lemma then follows by ifferentiating both sies with respect to α an applying Lemmas an 2 on the right han sie. Lemma 4: For any µ M X an any α, α Hn αµ = α 2 Dn T µ, α µ, 6 an α Hn αµ = α= 2 V n µ. 62 Proof: To establish 6, consier: α Hn αµ = α Hn T µ, α µ α α Dn T µ, α µ = α 2 V n T µ, α + α 2 V n T µ, α α 2 Dn T µ, α µ 64 = α 2 Dn T µ, α µ, 65 63

18 7 where 63 follows by invoking 3 in Lemma 4, an 64 follows from the chain rule of erivatives. To establish 62, we have α Hn αµ = α= α α Hn αµ 66 = α D n T µ, α µ α 2 67 = α α α V n T µ, α 2 2α 68 = 2 V n µ, 69 where 66 follows from the continuity of the erivative of H n αµ with respect to α, 67 follows from 6, an 68 follows from L Hôpital s rule an Lemma 3. This result recovers the fact that Rényi entropies are monotonically ecreasing with respect to the orer, an further characterizes their rate of change with respect to the parameterization. Lemma 5: For any µ M X, Further, for all α R, n n n n V n T µ, α µ = V T µ, α µ. 70 n Hn T µ, α = HT µ, α, 7 n n Hn T µ, α µ = HT µ, α µ, 72 n Dn T µ, α µ = DT µ, α µ, 73 n n Hn αµ = H α µ, 74 HT µ, α = αv T µ, α µ, α 75 HT µ, α µ = V T µ, α µ, α 76 DT µ, α µ = α V T µ, α µ. α 77 α H αµ = DT µ, α µ, 78 α 2 where DT µ, α µ is ifferentiable with respect to α, an α DT µ, α µ := 2 α= 2 V µ. Proof: The first part of the claim is straightforwar as the equalities hol for all n an hence also hol in the it as n. The claim in 75 hence follows by invoking Lemma. Equations 76 an 77 an 78 are establishe similarly consiering Lemmas 2, 3, an 4, respectively. Lemma 6: If µ M X, then, for all α 0, we have T µ, α M X. In other wors, M X is close uner all non-zero tilts, an hence T + µ, T µ M X.

19 8 Proof: We nee to show that T µ, α satisfies Assumptions an 2. To check Assumption, for α R +, we have: Further, consiering α, an for all 0 < α, n log T µ n, αx n = y n log n X n[µn y n ] α [µ n x n ] α = log [µ n y n ] α + log n [µ n x n ] α. 79 y n X n α log X α log X n log 0 n log y n X n [µ n y n ] α 0, 80 y n X n [µ n y n ] α α log X log X. 8 Thus, for all α R +, by taking the maximum of the upper bouns an the minimum of the lower bouns in 80 an 8, an applying them to 79, we have α log X + α log ε n log T µ n, αx n log X + α log ε, 82 where ε an ε are efine in 3 an 4, respectively. On the other han, for all α R, we have n log T µ n, αx n = log [µ n y n ] α + log n [µ n x n ] α y n X n Thus, for all α R, we have Next, we nee to show that for all α 0, n log y n X n ε nα 83 = log X α log ε. 84 n log T µ n, αx n log X + α log <. 85 ε HT µ, α > Due to Lemma 5, HT µ, α is ifferentiable with respect to α an its erivative is given by αv T µ, α µ. On the other han, for α > 0, the erivative is strictly positive implying that HT µ, α is strictly ecreasing with respect to α, an hence it is non-zero for all α > 0. This together with 85 emonstrate that for all α R, T µ, α satisfies Assumption.

20 9 To show that Assumption 2 hols for the tilte source, we note that for any α > 0, we have arg min T µ x α x X µ, αx = arg min x X y X µ y α 87 = arg min x X µ x α 88 = arg min x X µ x 89 which is unique by the fact that µ satisfies Assumption 2. Similarly, we can argue that arg max x X T µ, αx is also unique. On the other han, for α < 0, we have arg min T µ x α x X µ, αx = arg min x X y X µ y α 90 = arg min x X µ x α 9 = arg max x X µ x 92 which is again unique by the fact that µ satisfies Assumption 2. Finally, arg max x X T µ, αx is unique by the fact that arg min x X µ x is unique completing the proof. III. EQUIVALENT ORDER CLASS In this section, we characterize the set of all string-sources that inuce the same optimal orering of the strings from the most likely to the least likely on all strings of all lengths. Definition 6 orer equivalent string-sources: We say two string-sources µ, ρ M X are orer equivalent an enote by µ ρ, if for all n N, G n µ = G n ρ, i.e., the optimal orering for µ is also the optimal orering for ρ. Definition 7 equivalent orer class: The equivalent orer class of a string-source µ M X is enote by C µ an is given by C µ = ρ M X : ρ µ. We also nee another efinition that generalizes the weakly typical set. Definition 8 tilte weakly typical set of orer α: For any α R, enote by A n µ,α,ɛ the tilte weakly typical set of orer α with respect to µ, which is the set of all x n such that for an ɛ R + : e Hn T µ,α µ nɛ < µ n x n < e Hn T µ,α µ+nɛ, 93 where H is the cross entropy efine in 3. Note that in the case of α =, A n µ,,ɛ is the familiar weakly typical set see 3.6 in [36]; an Definition 8 generalizes the efinition of the weakly typical set for α. We shall establish some useful properties of the tilte weakly typical sets in the next section.

21 20 Our interest in fining the equivalent orer class of a string-source µ lies in the fact that such stringsources are subject to the same optimal guessing proceure, which reners them equivalent in terms of many problems, such as brute-force passwor guessing, list ecoing, sequential ecoing, an one-to-one source coing. Theorem : For any string-source µ M X, if ρ C µ, then there exists a unique α R + such that ρ = T µ, α. Consequently, the equivalent orer class of µ is equal to the tilte family, i.e., C µ = T µ +. Further, the set of all string-sources whose optimal orerings are inverse for µ is equal to Tµ, i.e., C T µ, = Tµ. Theorem characterizes the set of all string-sources that woul result in the same orering of strings for any length. As iscusse above, these string-sources are calle orer equivalent as their optimal guesswork proceure is exactly the same. Theorem establishes that any such orer equivalent stringsource is necessarily the result of a tilt proviing an operational meaning to the tilte family. To prove the theorem, we state two intermeiate lemmas an a efinition. Lemma 7: For any string-source µ, the equivalent orer class of µ contains T µ +, i.e., T µ + C µ. Proof: To prove the lemma, we nee to show that for any α R +, we have T α µ µ, i.e., T α µ C µ. To this en, we show that G n µ is also the optimal orering on n-strings for µ. We have T µ n, αx n < T µ n, αy n G n µx n > G n µy n. 94 By efinition, this establishes that G n µ is also the optimal orering for T µ n, α. It suffices to show that for any α R +, we have µ n x n > µ n y n T µ n, αx n > T µ n, αy n. 95 We show here that µ n x n > µ n y n [µ n x n ] α > [µ n y n ] α 96 [µ n x n ] α z n X n[µn z n ] α > [µnyn]α z n X n[µn z n ] α T µ n, αx n > T µ n, αy n, 97 where 96 hols because α > 0, an 97 follows from the efinition of the tilt in 22. Next, we state a simple lemma about the weakly typical set which is a special case of the tilte weakly typical sets.

22 2 Lemma 8: For any µ M X, P µ X n A n V n µ µ,,ɛ n 2 ɛ 2, 98 A n µ,,ɛ > e Hn µ nɛ V n µ n 2 ɛ 2, 99 A n µ,,ɛ < e Hn µ+nɛ. 00 Proof: P µ X n A n µ,,ɛ = Pµ ı µ X n H n µ nɛ 0 V n µ n 2 ɛ 2, 02 where 02 follows from Chebyshev s inequality. Note that Chernoff bouns coul be use to establish an exponential ecay in 02 but the weaker boun above is sufficient for our purposes. To erive the lower boun in 99 on the size of the set A µ,,ɛ, observe that The boun in 00 is establishe similarly. A µ,,ɛ P µx n A µ,,ɛ max xn A µ,,ɛ µ n x n > V n µ n 2 ɛ 2 e Hn µ+nɛ = V n µ n 2 ɛ 2 e Hn µ nɛ. Definition 9 Shannon entropy contour: Let S h M X enote the Shannon entropy contour h, which is the set of all string-sources in M X whose entropy rate equals h 0, log X, i.e., S h := µ M X : Hµ = h. 03 Next, we provie the proof of the main result of this section on the equivalent orer class of a stringsource. Proof of Theorem : To prove the theorem, we take two steps. We first show that one element of the tilte family T µ + is also containe in the intersection of the Shannon entropy contour an the equivalent orer class, i.e., T µ + C µ S h M X is not empty. We also show that for any string-source µ M X an any h 0, log X, if ρ, ρ 2 M X are in the intersection of S h see Definition 9 an C µ enote by C µ S h, then ρ = ρ 2. Putting these two steps together ensures that any other element in the intersection C µ S h is equal to a member of the tilte family, an completes the proof.

23 22 We first show that C µ S h is not empty an contains one element from T + µ. This is carrie out by noting the continuity an monotonicity of HT µ, α in the parameter α prove in Lemma 7 an noting that Hu X = log X an α H n T µ, α = 0. Next, we procee the proof of uniqueness by contraiction. Let ρ, ρ 2 C µ S h be two string-sources that are containe in the intersection of S h an C µ. Note that by Definition 7, ρ, ρ 2 M X. Therefore, by efinition as ρ, ρ 2 S h, we have Hρ = Hρ 2 = h. 04 The claim is negate to assume that ρ an ρ 2 are not equal, i.e., there exists δ R + such that for all N N there exists n > N such that E ρ2 log ρ n 2 X n ρ n > 2nδ. 05 Xn Then, we can ecompose the expectation over the typical set an the rest of the strings as follows: ρ n E ρ2 log 2 X n ρ n = ρ ρ n Xn 2 x n n log 2 x n ρ n 06 x n A n xn ρ 2,,ɛ + ρ ρ n 2 x n n log 2 x n ρ n. 07 xn x n A n ρ 2,,ɛ By Assumption for the string-source ρ M X, there exists ε such that On the other han, by 98 in Lemma 8, max log x n X n ρ n xn < n log ε. 08 x n A n ρ 2,,ɛ Putting together 08 an 09, we have ρ ρ n 2 x n n log 2 x n ρ n xn x n A n ρ 2,,ɛ x n A n ρ 2,,ɛ ρ n 2 x n V n ρ 2 n 2 ɛ 2, 09 ρ n 2 x n log ρ n V n ρ 2 xn nɛ 2 log ε = V ρ 2 ɛ 2 log ε, 0 leaing to the fact that the term in 07 is O n. Using the above in conjunction with the fact that P µ X n A n ρ 2,,ɛ V n µ n 2 ɛ, we conclue that for any N 2 0 N, there exists n > N 0 such that: ρ n E ρ2 log 2 X n ρ n X n A n Xn ρ 2,,ɛ > nδ. Consequently, there exists y n A n ρ 2,,ɛ such that log ρ n yn > log ρ n + nδ. 2 2 yn

24 23 Since y n A n ρ 2,,ɛ, by efinition, we have which in conjunction with 2 leas to log By swapping ρ an ρ 2 an noting that nh nɛ log ρ n nh + nɛ, 2 yn ρ n yn > log ρ n + nδ nh nɛ + nδ. 3 2 yn Dρ 2 ρ = 0 Dρ 2 ρ = 0, 4 an repeating similar arguments, we coul pick a ifferent sequence ŷ n X n such that an log ρ n > nh nɛ + nδ. 5 2 ŷn log ρ n nh + nɛ. ŷn Note that the above hols for any ɛ R +. By choosing ɛ to be small enough smaller than δ/2 for any N, there exists n > N such that ρ n 2 yn > ρ n 2 ŷn an hence G n ρ 2 y n < G n ρ 2 ŷ n. Similarly, we can show that for sufficiently large n, we have ρ n yn < ρ n ŷn, which leas to G n ρ y n > G n ρ ŷ n. Therefore, ρ an ρ 2 cannot be orer equivalent, which is a contraiction. Hence, it must be true that for any δ R +, there exists N 0 N such that for all n > N 0 ρ n E ρ2 log 2 X n ρ n < nδ, 6 Xn which irectly implies that Dρ ρ 2 = Dρ 2 ρ = 0, an ρ = ρ 2, i.e., any two string-sources in the intersection of C µ an S h must be equal. IV. TILTED WEAKLY TYPICAL SET The main result in this section is a generalization of the asymptotic equipartition property on the weakly typical set concerning information an guesswork. We call this generalize notion the tilte weakly typical set an efine it in Definition 8 in the previous section. Although these results resemble the asymptotic equipartition property on the weakly typical set, they are concerne with the large eviations, i.e., atyptical behavior, rather than the typical behavior. We will consequently use these results in establishing the large eviation properties of guesswork an information in the next section. We nee two more efinitions before we can provie the statement of the main result in this section. Definition 20: For any µ M X, let Bµ,α,ɛ n A n µ,α,ɛ be a subset such that B n µ,α,ɛ = 2 A n µ,α,ɛ an for any x n B n µ,α,ɛ an y n A n µ,α,ɛ B n µ,α,ɛ we have T µ n, αx n < T µ n, αy n.

25 24 Definition 2: For any µ M X, let Dµ,α,ɛ n A n µ,α,ɛ be efine as follows: Dµ,α,ɛ n := x n X n : T µ n, αx n > e Hn T µ,α n α ɛ. 7 Definition 22: For any µ M X, let Eµ,α,ɛ n A n µ,α,ɛ be efine as follows: Eµ,α,ɛ n := x n X n : T µ n, αx n < e Hn T µ,α+n α ɛ. 8 Theorem 2: For any µ M X, any ɛ R +, an any α 0, a For any x n A n µ,α,ɛ, µ n x n > e Hn T µ,α µ nɛ, 9 µ n x n < e Hn T µ,α µ+nɛ ; 20 b For any n N, A n µ,α,ɛ > V n T µ, α µ n 2 ɛ 2 e Hn T µ,α α nɛ, 2 A n µ,α,ɛ < e Hn T µ,α+ α nɛ ; 22 c For any n N, an α, 0 [,, P µ X n Dµ,α,ɛ n Pµ X n A n µ,α,ɛ > V n T µ, α µ n 2 ɛ 2 e Dn T µ,α µ α nɛ, 23 P µ X n A n µ,α,ɛ Pµ X n D n µ,α,ɛ e D n T µ,α µ+ α nɛ, 24 P µ X n E n µ,α,ɛ e D n T µ,α µ+ α nɛ ; 25 For any n N, an α 0, ], P µ X n Eµ,α,ɛ n Pµ X n A n µ,α,ɛ > V n T µ, α µ n 2 ɛ 2 e Dn T µ,α µ α nɛ, 26 P µ X n A n µ,α,ɛ Pµ X n E n µ,α,ɛ e D n T µ,α µ+ α nɛ, 27 e For any n N an any α > 0: P µ X n D n µ,α,ɛ e D n T µ,α µ+ α nɛ ; 28 x n B n µ,α,ɛ G n µx n > 2 V n T µ, α µ n 2 ɛ 2 e Hn T µ,α αnɛ, 29 x n Dµ,α,ɛ n G n µx n e Hn T µ,α+αnɛ, 30 x n Dµ,α,ɛ n G n µx n V n T µ, α µ n 2 ɛ 2 e Hn T µ,α αnɛ, 3 x n E n µ,α,ɛ G n µx n > e Hn T µ,α+αnɛ. 32

26 25 f For any n N an any α < 0: x n B n µ,α,ɛ R n µx n > 2 V n T µ, α µ n 2 ɛ 2 e Hn T µ,α α nɛ, 33 x n Dµ,α,ɛ n Rµx n n e Hn T µ,α+ α nɛ, 34 x n Dµ,α,ɛ n Rµx n n V n T µ, α µ n 2 ɛ 2 e Hn T µ,α α nɛ, 35 x n E n µ,α,ɛ R n µx n > e Hn T µ,α+ α nɛ. 36 Recall that Shannon entropy, cross entropy, an relative entropy are measure in nats throughout this paper. We first provie a few lemmas that specialize to the typical set, which we use to prove the theorem. Lemma 9: For any µ M X, let x n Bµ,,ɛ n see Definition 20. Then, G n µx n > V n µ 2 n 2 ɛ 2 e Hn µ nɛ. 37 Proof: Note that by efinition for all x n Bµ,,ɛ n, we have Gn µx n > Bµ,,ɛ n 2 An µ,,ɛ. The proof is complete noticing the boun in 99. Lemma 20: For any µ M X, let x n D n µ,,ɛ see Definition 22. Then, G n µx n e Hn µ+nɛ. 38 Proof: We have G n µx n D n µ,,ɛ 39 min xn D n µ,,ɛ µn x n 40 = e Hn µ+nɛ. 4 Lemma 2: Suppose the following hols: Then, x n D n µ,,ɛ. G n µx n Proof: We procee the proof with contraiction. Suppose that x n V n µ n 2 ɛ 2 e Hn µ nɛ. 42 D n µ,,ɛ. Hence, µn x n e nhµ nɛ. Thus, µ n x n < µ n y n for all y n A n µ,,ɛ. Consequently, G n µx n > A n µ,,ɛ > V n µ n 2 ɛ 2 e Hn µ nɛ, 43

27 26 which contraicts the supposition of the lemma in 42. Lemma 22: Suppose the following hols: G n µx n > e Hn µ+nɛ. 44 Then, x n E n µ,,ɛ. Proof: We procee the proof with contraiction. Suppose that x n e nhµ+nɛ. Thus, µ n x n > µ n y n for all y n A n µ,,ɛ. Consequently, E n µ,,ɛ. Hence, µn x n G n µx n < G n µy n e Hn µ+nɛ, 45 which contraicts the supposition of the lemma in 42. Next, we provie a lemma of equivalence between tilte weakly typical sets an weakly typical sets of a tilte source, which we will rely on in proving the main theorem of this section. Lemma 23: Let µ M X. Then, for all α 0, the following equivalences hol: A n µ,α,ɛ = A n T µ,α,, α ɛ, 46 Bµ,α,ɛ n = BT n µ,α,, α ɛ, 47 Dµ,α,ɛ n = DT n µ,α,, α ɛ, 48 Eµ,α,ɛ n = ET n µ,α,, α ɛ. 49 Proof: We have A n T µ,α,, α ɛ x = n X n : log T µ n, αx n Hn T T µ, α, T µ, α < α nɛ = x n X n : α log T µ n, αx n α Hn T µ, α < nɛ = x n X n : log µ n x n Hn T µ, α µ < nɛ = A n µ,α,ɛ, where 50 follows from Definition 8, 5 follows by noting that T µ, = µ an H n µ µ = H n µ an the fact that α 0, an 52 follows from Lemma 5. Next, note that 47 follows irectly from 46 an Definition 20.

28 27 To establish 48 note that: DT n µ,α,, α ɛ x = n X n : log T µ n, αx n Hn T T µ, α, T µ, α < α nɛ = x n X n : log T µ n, αx n Hn T µ, α < α nɛ = D n µ,α,ɛ, where 54 follows from Definition 8, 55 follows by noting that T µ, = µ an H n µ µ = H n µ an the fact that α 0. Finally, the proof for 49 is very similar to that of 48 with all the inequalities reverse, an is omitte for brevity. Next, we procee to the proof of Theorem 2. Proof of Theorem 2: Part a is simply a reiteration of Definition 8, which is presente here for the completeness of the characterization. To establish Part b, notice that by Lemma 23, for all α 0, we have A n T µ,α,, α ɛ = An µ,α,ɛ. Part b is establishe by applying the bouns in 99 an 00 in Lemma 8 to A n T µ,α,, α ɛ, an noting that α 2 V T µ, α µ = V T µ, α from Lemma 6. In Parts c an, first notice that by efinition A n µ,α,ɛ D n µ,α,ɛ an A n µ,α,ɛ E n µ,α,ɛ an hence P µ X n A n µ,α,ɛ Pµ X n D n µ,α,ɛ an Pµ X n A n µ,α,ɛ Pµ X n E n µ,α,ɛ. Hence, we only nee to prove the right han sie inequalities in the claims. To prove 23 in Part c, consier 98 in Lemma 8 applie to A n T µ,α,, α ɛ, in conjunction with 46 in Lemma 23: P T µ,α X n A n µ,α,ɛ = PT µ,α A n T µ,α,, α ɛ On the other han, we have P T µ,α X n A n µ,α,ɛ = x n A n µ,α,ɛ V n T µ, α µ n 2 ɛ T µ n, αx n 58 x = n A [µ n x n ] α n µ,α,ɛ x n X n[µn x n ] α 59 x = n A µ n x n [µ n x n ] α n µ,α,ɛ e αhnµ 60 α < e αhn T µ,α µ Hα nµ+ α nɛ µ n x n 6 x n A n µ,α,ɛ = e Dn T µ,α µ+ α nɛ P µ X n A n µ,α,ɛ, 62 where 59 follow from the efinition of the tilt in 22, 60 follows from the efinition of the Rényi entropy in 9, 6 follows from 9 or 20 applie to [µ n x n ] α epening on the sign of

Guesswork Subject to a Total Entropy Budget

Guesswork Subject to a Total Entropy Budget Guesswork Subject to a Total Entropy Budget Arman Rezaee, Ahmad Beirami, Ali Makhdoumi, Muriel Médard, and Ken Duffy arxiv:72.09082v [cs.it] 25 Dec 207 Abstract We consider an abstraction of computational

More information

Topic 7: Convergence of Random Variables

Topic 7: Convergence of Random Variables Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information

More information

Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability

Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability Ahmad Beirami, Robert Calderbank, Ken Duffy, Muriel Médard Department of Electrical and Computer Engineering,

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Math 342 Partial Differential Equations «Viktor Grigoryan

Math 342 Partial Differential Equations «Viktor Grigoryan Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

The Exact Form and General Integrating Factors

The Exact Form and General Integrating Factors 7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily

More information

Math 1271 Solutions for Fall 2005 Final Exam

Math 1271 Solutions for Fall 2005 Final Exam Math 7 Solutions for Fall 5 Final Eam ) Since the equation + y = e y cannot be rearrange algebraically in orer to write y as an eplicit function of, we must instea ifferentiate this relation implicitly

More information

A New Minimum Description Length

A New Minimum Description Length A New Minimum Description Length Soosan Beheshti, Munther A. Dahleh Laboratory for Information an Decision Systems Massachusetts Institute of Technology soosan@mit.eu,ahleh@lis.mit.eu Abstract The minimum

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Ashish Goel Michael Kapralov Sanjeev Khanna Abstract We consier the well-stuie problem of fining a perfect matching in -regular bipartite

More information

Iterated Point-Line Configurations Grow Doubly-Exponentially

Iterated Point-Line Configurations Grow Doubly-Exponentially Iterate Point-Line Configurations Grow Doubly-Exponentially Joshua Cooper an Mark Walters July 9, 008 Abstract Begin with a set of four points in the real plane in general position. A to this collection

More information

Chromatic number for a generalization of Cartesian product graphs

Chromatic number for a generalization of Cartesian product graphs Chromatic number for a generalization of Cartesian prouct graphs Daniel Král Douglas B. West Abstract Let G be a class of graphs. The -fol gri over G, enote G, is the family of graphs obtaine from -imensional

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

An Analytical Expression of the Probability of Error for Relaying with Decode-and-forward

An Analytical Expression of the Probability of Error for Relaying with Decode-and-forward An Analytical Expression of the Probability of Error for Relaying with Decoe-an-forwar Alexanre Graell i Amat an Ingmar Lan Department of Electronics, Institut TELECOM-TELECOM Bretagne, Brest, France Email:

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a

More information

A Weak First Digit Law for a Class of Sequences

A Weak First Digit Law for a Class of Sequences International Mathematical Forum, Vol. 11, 2016, no. 15, 67-702 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1288/imf.2016.6562 A Weak First Digit Law for a Class of Sequences M. A. Nyblom School of

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Error Floors in LDPC Codes: Fast Simulation, Bounds and Hardware Emulation

Error Floors in LDPC Codes: Fast Simulation, Bounds and Hardware Emulation Error Floors in LDPC Coes: Fast Simulation, Bouns an Harware Emulation Pamela Lee, Lara Dolecek, Zhengya Zhang, Venkat Anantharam, Borivoje Nikolic, an Martin J. Wainwright EECS Department University of

More information

Equilibrium in Queues Under Unknown Service Times and Service Value

Equilibrium in Queues Under Unknown Service Times and Service Value University of Pennsylvania ScholarlyCommons Finance Papers Wharton Faculty Research 1-2014 Equilibrium in Queues Uner Unknown Service Times an Service Value Laurens Debo Senthil K. Veeraraghavan University

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

On the Aloha throughput-fairness tradeoff

On the Aloha throughput-fairness tradeoff On the Aloha throughput-fairness traeoff 1 Nan Xie, Member, IEEE, an Steven Weber, Senior Member, IEEE Abstract arxiv:1605.01557v1 [cs.it] 5 May 2016 A well-known inner boun of the stability region of

More information

Logarithmic spurious regressions

Logarithmic spurious regressions Logarithmic spurious regressions Robert M. e Jong Michigan State University February 5, 22 Abstract Spurious regressions, i.e. regressions in which an integrate process is regresse on another integrate

More information

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes Leaving Ranomness to Nature: -Dimensional Prouct Coes through the lens of Generalize-LDPC coes Tavor Baharav, Kannan Ramchanran Dept. of Electrical Engineering an Computer Sciences, U.C. Berkeley {tavorb,

More information

Lecture 2 Lagrangian formulation of classical mechanics Mechanics

Lecture 2 Lagrangian formulation of classical mechanics Mechanics Lecture Lagrangian formulation of classical mechanics 70.00 Mechanics Principle of stationary action MATH-GA To specify a motion uniquely in classical mechanics, it suffices to give, at some time t 0,

More information

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1 Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of

More information

Mark J. Machina CARDINAL PROPERTIES OF "LOCAL UTILITY FUNCTIONS"

Mark J. Machina CARDINAL PROPERTIES OF LOCAL UTILITY FUNCTIONS Mark J. Machina CARDINAL PROPERTIES OF "LOCAL UTILITY FUNCTIONS" This paper outlines the carinal properties of "local utility functions" of the type use by Allen [1985], Chew [1983], Chew an MacCrimmon

More information

Permanent vs. Determinant

Permanent vs. Determinant Permanent vs. Determinant Frank Ban Introuction A major problem in theoretical computer science is the Permanent vs. Determinant problem. It asks: given an n by n matrix of ineterminates A = (a i,j ) an

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

Lyapunov Functions. V. J. Venkataramanan and Xiaojun Lin. Center for Wireless Systems and Applications. School of Electrical and Computer Engineering,

Lyapunov Functions. V. J. Venkataramanan and Xiaojun Lin. Center for Wireless Systems and Applications. School of Electrical and Computer Engineering, On the Queue-Overflow Probability of Wireless Systems : A New Approach Combining Large Deviations with Lyapunov Functions V. J. Venkataramanan an Xiaojun Lin Center for Wireless Systems an Applications

More information

Regular tree languages definable in FO and in FO mod

Regular tree languages definable in FO and in FO mod Regular tree languages efinable in FO an in FO mo Michael Beneikt Luc Segoufin Abstract We consier regular languages of labele trees. We give an effective characterization of the regular languages over

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Table of Common Derivatives By David Abraham

Table of Common Derivatives By David Abraham Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec

More information

On combinatorial approaches to compressed sensing

On combinatorial approaches to compressed sensing On combinatorial approaches to compresse sensing Abolreza Abolhosseini Moghaam an Hayer Raha Department of Electrical an Computer Engineering, Michigan State University, East Lansing, MI, U.S. Emails:{abolhos,raha}@msu.eu

More information

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS MARK SCHACHNER Abstract. When consiere as an algebraic space, the set of arithmetic functions equippe with the operations of pointwise aition an

More information

QF101: Quantitative Finance September 5, Week 3: Derivatives. Facilitator: Christopher Ting AY 2017/2018. f ( x + ) f(x) f(x) = lim

QF101: Quantitative Finance September 5, Week 3: Derivatives. Facilitator: Christopher Ting AY 2017/2018. f ( x + ) f(x) f(x) = lim QF101: Quantitative Finance September 5, 2017 Week 3: Derivatives Facilitator: Christopher Ting AY 2017/2018 I recoil with ismay an horror at this lamentable plague of functions which o not have erivatives.

More information

Implicit Differentiation

Implicit Differentiation Implicit Differentiation Thus far, the functions we have been concerne with have been efine explicitly. A function is efine explicitly if the output is given irectly in terms of the input. For instance,

More information

Ramsey numbers of some bipartite graphs versus complete graphs

Ramsey numbers of some bipartite graphs versus complete graphs Ramsey numbers of some bipartite graphs versus complete graphs Tao Jiang, Michael Salerno Miami University, Oxfor, OH 45056, USA Abstract. The Ramsey number r(h, K n ) is the smallest positive integer

More information

II. First variation of functionals

II. First variation of functionals II. First variation of functionals The erivative of a function being zero is a necessary conition for the etremum of that function in orinary calculus. Let us now tackle the question of the equivalent

More information

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent

More information

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides Reference 1: Transformations of Graphs an En Behavior of Polynomial Graphs Transformations of graphs aitive constant constant on the outsie g(x) = + c Make graph of g by aing c to the y-values on the graph

More information

Lecture 6: Calculus. In Song Kim. September 7, 2011

Lecture 6: Calculus. In Song Kim. September 7, 2011 Lecture 6: Calculus In Song Kim September 7, 20 Introuction to Differential Calculus In our previous lecture we came up with several ways to analyze functions. We saw previously that the slope of a linear

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control 19 Eigenvalues, Eigenvectors, Orinary Differential Equations, an Control This section introuces eigenvalues an eigenvectors of a matrix, an iscusses the role of the eigenvalues in etermining the behavior

More information

A. Exclusive KL View of the MLE

A. Exclusive KL View of the MLE A. Exclusive KL View of the MLE Lets assume a change-of-variable moel p Z z on the ranom variable Z R m, such as the one use in Dinh et al. 2017: z 0 p 0 z 0 an z = ψz 0, where ψ is an invertible function

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

Stopping-Set Enumerator Approximations for Finite-Length Protograph LDPC Codes

Stopping-Set Enumerator Approximations for Finite-Length Protograph LDPC Codes Stopping-Set Enumerator Approximations for Finite-Length Protograph LDPC Coes Kaiann Fu an Achilleas Anastasopoulos Electrical Engineering an Computer Science Dept. University of Michigan, Ann Arbor, MI

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

1. Aufgabenblatt zur Vorlesung Probability Theory

1. Aufgabenblatt zur Vorlesung Probability Theory 24.10.17 1. Aufgabenblatt zur Vorlesung By (Ω, A, P ) we always enote the unerlying probability space, unless state otherwise. 1. Let r > 0, an efine f(x) = 1 [0, [ (x) exp( r x), x R. a) Show that p f

More information

A Review of Multiple Try MCMC algorithms for Signal Processing

A Review of Multiple Try MCMC algorithms for Signal Processing A Review of Multiple Try MCMC algorithms for Signal Processing Luca Martino Image Processing Lab., Universitat e València (Spain) Universia Carlos III e Mari, Leganes (Spain) Abstract Many applications

More information

Differentiability, Computing Derivatives, Trig Review. Goals:

Differentiability, Computing Derivatives, Trig Review. Goals: Secants vs. Derivatives - Unit #3 : Goals: Differentiability, Computing Derivatives, Trig Review Determine when a function is ifferentiable at a point Relate the erivative graph to the the graph of an

More information

The chromatic number of graph powers

The chromatic number of graph powers Combinatorics, Probability an Computing (19XX) 00, 000 000. c 19XX Cambrige University Press Printe in the Unite Kingom The chromatic number of graph powers N O G A A L O N 1 an B O J A N M O H A R 1 Department

More information

The Principle of Least Action

The Principle of Least Action Chapter 7. The Principle of Least Action 7.1 Force Methos vs. Energy Methos We have so far stuie two istinct ways of analyzing physics problems: force methos, basically consisting of the application of

More information

Lecture 12: November 6, 2013

Lecture 12: November 6, 2013 Information an Coing Theory Autumn 204 Lecturer: Mahur Tulsiani Lecture 2: November 6, 203 Scribe: Davi Kim Recall: We were looking at coes of the form C : F k q F n q, where q is prime, k is the message

More information

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson JUST THE MATHS UNIT NUMBER 10.2 DIFFERENTIATION 2 (Rates of change) by A.J.Hobson 10.2.1 Introuction 10.2.2 Average rates of change 10.2.3 Instantaneous rates of change 10.2.4 Derivatives 10.2.5 Exercises

More information

Capacity Analysis of MIMO Systems with Unknown Channel State Information

Capacity Analysis of MIMO Systems with Unknown Channel State Information Capacity Analysis of MIMO Systems with Unknown Channel State Information Jun Zheng an Bhaskar D. Rao Dept. of Electrical an Computer Engineering University of California at San Diego e-mail: juzheng@ucs.eu,

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21 Large amping in a structural material may be either esirable or unesirable, epening on the engineering application at han. For example, amping is a esirable property to the esigner concerne with limiting

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

15.1 Upper bound via Sudakov minorization

15.1 Upper bound via Sudakov minorization ECE598: Information-theoretic methos in high-imensional statistics Spring 206 Lecture 5: Suakov, Maurey, an uality of metric entropy Lecturer: Yihong Wu Scribe: Aolin Xu, Mar 7, 206 [E. Mar 24] In this

More information

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy,

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy, NOTES ON EULER-BOOLE SUMMATION JONATHAN M BORWEIN, NEIL J CALKIN, AND DANTE MANNA Abstract We stuy a connection between Euler-MacLaurin Summation an Boole Summation suggeste in an AMM note from 196, which

More information

Math 115 Section 018 Course Note

Math 115 Section 018 Course Note Course Note 1 General Functions Definition 1.1. A function is a rule that takes certain numbers as inputs an assigns to each a efinite output number. The set of all input numbers is calle the omain of

More information

Sharp Thresholds. Zachary Hamaker. March 15, 2010

Sharp Thresholds. Zachary Hamaker. March 15, 2010 Sharp Threshols Zachary Hamaker March 15, 2010 Abstract The Kolmogorov Zero-One law states that for tail events on infinite-imensional probability spaces, the probability must be either zero or one. Behavior

More information

Computing the Longest Common Subsequence of Multiple RLE Strings

Computing the Longest Common Subsequence of Multiple RLE Strings The 29th Workshop on Combinatorial Mathematics an Computation Theory Computing the Longest Common Subsequence of Multiple RLE Strings Ling-Chih Yao an Kuan-Yu Chen Grauate Institute of Networking an Multimeia

More information

Differentiability, Computing Derivatives, Trig Review

Differentiability, Computing Derivatives, Trig Review Unit #3 : Differentiability, Computing Derivatives, Trig Review Goals: Determine when a function is ifferentiable at a point Relate the erivative graph to the the graph of an original function Compute

More information

Linear and quadratic approximation

Linear and quadratic approximation Linear an quaratic approximation November 11, 2013 Definition: Suppose f is a function that is ifferentiable on an interval I containing the point a. The linear approximation to f at a is the linear function

More information

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7.

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7. Lectures Nine an Ten The WKB Approximation The WKB metho is a powerful tool to obtain solutions for many physical problems It is generally applicable to problems of wave propagation in which the frequency

More information

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback Journal of Machine Learning Research 8 07) - Submitte /6; Publishe 5/7 An Optimal Algorithm for Banit an Zero-Orer Convex Optimization with wo-point Feeback Oha Shamir Department of Computer Science an

More information

Balancing Expected and Worst-Case Utility in Contracting Models with Asymmetric Information and Pooling

Balancing Expected and Worst-Case Utility in Contracting Models with Asymmetric Information and Pooling Balancing Expecte an Worst-Case Utility in Contracting Moels with Asymmetric Information an Pooling R.B.O. erkkamp & W. van en Heuvel & A.P.M. Wagelmans Econometric Institute Report EI2018-01 9th January

More information

Differentiation ( , 9.5)

Differentiation ( , 9.5) Chapter 2 Differentiation (8.1 8.3, 9.5) 2.1 Rate of Change (8.2.1 5) Recall that the equation of a straight line can be written as y = mx + c, where m is the slope or graient of the line, an c is the

More information

On the Surprising Behavior of Distance Metrics in High Dimensional Space

On the Surprising Behavior of Distance Metrics in High Dimensional Space On the Surprising Behavior of Distance Metrics in High Dimensional Space Charu C. Aggarwal, Alexaner Hinneburg 2, an Daniel A. Keim 2 IBM T. J. Watson Research Center Yortown Heights, NY 0598, USA. charu@watson.ibm.com

More information

A Modification of the Jarque-Bera Test. for Normality

A Modification of the Jarque-Bera Test. for Normality Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam

More information

UNDERSTANDING INTEGRATION

UNDERSTANDING INTEGRATION UNDERSTANDING INTEGRATION Dear Reaer The concept of Integration, mathematically speaking, is the "Inverse" of the concept of result, the integration of, woul give us back the function f(). This, in a way,

More information

Section 7.2. The Calculus of Complex Functions

Section 7.2. The Calculus of Complex Functions Section 7.2 The Calculus of Complex Functions In this section we will iscuss limits, continuity, ifferentiation, Taylor series in the context of functions which take on complex values. Moreover, we will

More information

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations Diophantine Approximations: Examining the Farey Process an its Metho on Proucing Best Approximations Kelly Bowen Introuction When a person hears the phrase irrational number, one oes not think of anything

More information

The Ehrenfest Theorems

The Ehrenfest Theorems The Ehrenfest Theorems Robert Gilmore Classical Preliminaries A classical system with n egrees of freeom is escribe by n secon orer orinary ifferential equations on the configuration space (n inepenent

More information

Lecture 5. Symmetric Shearer s Lemma

Lecture 5. Symmetric Shearer s Lemma Stanfor University Spring 208 Math 233: Non-constructive methos in combinatorics Instructor: Jan Vonrák Lecture ate: January 23, 208 Original scribe: Erik Bates Lecture 5 Symmetric Shearer s Lemma Here

More information

Inverse Theory Course: LTU Kiruna. Day 1

Inverse Theory Course: LTU Kiruna. Day 1 Inverse Theory Course: LTU Kiruna. Day Hugh Pumphrey March 6, 0 Preamble These are the notes for the course Inverse Theory to be taught at LuleåTekniska Universitet, Kiruna in February 00. They are not

More information

New Bounds for Distributed Storage Systems with Secure Repair

New Bounds for Distributed Storage Systems with Secure Repair New Bouns for Distribute Storage Systems with Secure Repair Ravi Tanon 1 an Soheil Mohajer 1 Discovery Analytics Center & Department of Computer Science, Virginia Tech, Blacksburg, VA Department of Electrical

More information

Second order differentiation formula on RCD(K, N) spaces

Second order differentiation formula on RCD(K, N) spaces Secon orer ifferentiation formula on RCD(K, N) spaces Nicola Gigli Luca Tamanini February 8, 018 Abstract We prove the secon orer ifferentiation formula along geoesics in finite-imensional RCD(K, N) spaces.

More information

Calculus of Variations

Calculus of Variations Calculus of Variations Lagrangian formalism is the main tool of theoretical classical mechanics. Calculus of Variations is a part of Mathematics which Lagrangian formalism is base on. In this section,

More information

On a limit theorem for non-stationary branching processes.

On a limit theorem for non-stationary branching processes. On a limit theorem for non-stationary branching processes. TETSUYA HATTORI an HIROSHI WATANABE 0. Introuction. The purpose of this paper is to give a limit theorem for a certain class of iscrete-time multi-type

More information

Range and speed of rotor walks on trees

Range and speed of rotor walks on trees Range an spee of rotor wals on trees Wilfrie Huss an Ecaterina Sava-Huss May 15, 1 Abstract We prove a law of large numbers for the range of rotor wals with ranom initial configuration on regular trees

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

Witt#5: Around the integrality criterion 9.93 [version 1.1 (21 April 2013), not completed, not proofread]

Witt#5: Around the integrality criterion 9.93 [version 1.1 (21 April 2013), not completed, not proofread] Witt vectors. Part 1 Michiel Hazewinkel Sienotes by Darij Grinberg Witt#5: Aroun the integrality criterion 9.93 [version 1.1 21 April 2013, not complete, not proofrea In [1, section 9.93, Hazewinkel states

More information

Agmon Kolmogorov Inequalities on l 2 (Z d )

Agmon Kolmogorov Inequalities on l 2 (Z d ) Journal of Mathematics Research; Vol. 6, No. ; 04 ISSN 96-9795 E-ISSN 96-9809 Publishe by Canaian Center of Science an Eucation Agmon Kolmogorov Inequalities on l (Z ) Arman Sahovic Mathematics Department,

More information

arxiv: v1 [math.mg] 10 Apr 2018

arxiv: v1 [math.mg] 10 Apr 2018 ON THE VOLUME BOUND IN THE DVORETZKY ROGERS LEMMA FERENC FODOR, MÁRTON NASZÓDI, AND TAMÁS ZARNÓCZ arxiv:1804.03444v1 [math.mg] 10 Apr 2018 Abstract. The classical Dvoretzky Rogers lemma provies a eterministic

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

arxiv: v4 [math.pr] 27 Jul 2016

arxiv: v4 [math.pr] 27 Jul 2016 The Asymptotic Distribution of the Determinant of a Ranom Correlation Matrix arxiv:309768v4 mathpr] 7 Jul 06 AM Hanea a, & GF Nane b a Centre of xcellence for Biosecurity Risk Analysis, University of Melbourne,

More information

arxiv: v2 [math.pr] 27 Nov 2018

arxiv: v2 [math.pr] 27 Nov 2018 Range an spee of rotor wals on trees arxiv:15.57v [math.pr] 7 Nov 1 Wilfrie Huss an Ecaterina Sava-Huss November, 1 Abstract We prove a law of large numbers for the range of rotor wals with ranom initial

More information

Assignment 1. g i (x 1,..., x n ) dx i = 0. i=1

Assignment 1. g i (x 1,..., x n ) dx i = 0. i=1 Assignment 1 Golstein 1.4 The equations of motion for the rolling isk are special cases of general linear ifferential equations of constraint of the form g i (x 1,..., x n x i = 0. i=1 A constraint conition

More information

PDE Notes, Lecture #11

PDE Notes, Lecture #11 PDE Notes, Lecture # from Professor Jalal Shatah s Lectures Febuary 9th, 2009 Sobolev Spaces Recall that for u L loc we can efine the weak erivative Du by Du, φ := udφ φ C0 If v L loc such that Du, φ =

More information

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d A new proof of the sharpness of the phase transition for Bernoulli percolation on Z Hugo Duminil-Copin an Vincent Tassion October 8, 205 Abstract We provie a new proof of the sharpness of the phase transition

More information