Efficient Compression of Monotone and m-modal Distributions

Size: px
Start display at page:

Download "Efficient Compression of Monotone and m-modal Distributions"

Transcription

1 Efficient Compression of Monotone and m-modal Distriutions Jayadev Acharya, Ashkan Jafarpour, Alon Orlitsky, and Ananda Theertha Suresh ECE Department, UCSD {jacharya, ashkan, alon, Astract We consider universal compression of n samples drawn independently according to a monotone or m-modal distriution over k elements. We show that for all these distriutions, the per-sample redundancy diminishes to 0 if k expon/ log n and is at least a constant if k expωn. I. INTRODUCTION Universal compression concerns encoding the output of an unknown source in a known class of distriutions. We are interested in understanding compression of i.i.d. samples from distriutions over an underlying discrete alphaet X e.g., a suset of N. Let P n def P P... P e a product distriution over X n. Shannon s source coding theorem shows that HP def X P x log P x its are necessary and sufficient to encode a known source with distriution P, where HP is the entropy of P. By Kraft s Inequality, any distriution over X n implies a code that uses log x n 1 its to encode x n 1 X n. The code implied y the underlying distriution P n uses log P n x n 1 its and the expected numer of its used is nhp. The average redundancy of with respect to P n is DP n E P n [ log P n x n 1 x n 1 the expected extra numer of its used y eyond the entropy. DP Q is called the Kullack-Lieler divergence etween P and Q. Let the underlying distriution e in a known class P, and P n def {P n : P P} ], e the class of all distriutions of the form P n. Then RP n def inf sup P n DP n is the average or min-max redundancy of P n, which corresponds to the est coding for the worst distriution. While coding over unknown distriutions, often some apriori knowledge aout the underlying distriution is known. For example, coding the last names of people from a census data is close to coding a monotone distriution, since we have a prior knowledge aout the prevalences of last names. We expect that the last name Smith is more likely than Galifianakis. In a text document we have some knowledge aout word frequencies and proailities. In such language modeling applications, Zipf distriutions are common [1]. Geometric distriutions over integers are useful in compressing residual signals in image compression []. A natural generalization is to consider distriutions with at most m modes. For example, life expectancy of a population, Poisson, and Binomial distriutions are unimodal. There has een a considerale interest over the past decade in approximating mixtures of distriutions. Many real world phenomenon can e modeled as outcomes of mixtures of simple distriutions and these have een studied since as early as the study Naples cra population y Pearson [3]. In this setup after oserving measurements say diameter over heights, the distriution that explained the data the est was a mixture of two Gaussians, predicting the presence of more than one specie. Such mixtures of m simple distriutions are typically m-modal. There has een enormous work in the past decade in learning mixtures of distriutions. Before discussing monotone distriutions, we state the results for the most well studied finite discrete distriutions, i.e., the i.i.d. distriutions. Let Ik n e the class of all i.i.d. distriutions over k elements and lock length n. Redundancy of Ik n estalished that 1 For k on has een extensively studied [4 11]. It is now well For n ok RI n k k 1 log n 1 + o1. 1 k RIk n n log k 1 + o1. n II. TERMINOLOGY We consider distriutions over the set N of positive integers. Such a distriution is non-increasing, or monotone, if P i P i + 1 for all i 1. Let M e the class of all monotone distriutions over N, and let M k e the suset of M consisting of monotone distriutions over [k] {1,..., k}. Let M n k denote the class of distriutions otained y sampling one distriution in M k independently n times, i.e., product distriutions of the form P n for P M k. Generalizing monotone distriutions, a consecutive set [l, r] {l,..., r} of integers is a mode of a distriution P if for all i, j [l, r] P i P j, and P l 1 P l P r + 1 P r > 0, representing a local minimum or maximum of P. P is m-modal if it has at most m modes. Note that a monotone distriution is 0-modal. Let M k,m e

2 the collection of all m-modal distriutions over [k], and M n k,m, the distriutions over [k] n otained y sampling the same distriution in M k,m independently n times. III. RELATED WORK Initial work on universal compression of monotone distriutions was motivated y representation of integers, hence centered on a single instance, namely n 1. [1] showed that every positive integer k can e represented using roughly log k + log log k its. [13] derived a related lower ound, and [14] constructed a method for minimizing the expected numer of its normalized y the distriution s entropy. Remarkaly, [15] found the expected redundancy exactly, k RM k log i 1. i i 1 i Since 1 1 i i e 1 and k 1 i RM k log log k + O1, log k, this shows that and in particular, RM. More recently, [16] analyzed the min-ave redundancy that replaces the redundancy of the worst distriution in the class y the average over all distriutions. They show that monotone distriutions have constant min-ave redundancy. Recent work has concerned compression of independent identical monotone distriutions, namely distriutions in M n. The results aove imply RM n, yet [17] showed that the set of all distriutions in M with finite entropy can e compressed with a diminishing per-sample relative redundancy. Specifically, there is a distriution over N n, such that for all P M with HP <, DP n nhp log lognhp. lognhp Note that the theorem holds for all distriutions with finite entropy, even those with infinite support. It therefore does not necessarily imply that the redundancy is su-linear. For example, for the uniform distriution over [exp n], the ound is n 3/, hence super-linear. [18] considered the redundancy of monotone distriutions as a function of the alphaet size k as well as the lock length n. They show tight lower and upper ounds for k On 1, RM n k Θmin{k, n 1/3 }. In particular, this implies that when n grows linearly with k, the per-sample redundancy diminishes to 0. These results can e extended using methods from [19] to prove that M n k has su-linear redundancy for k o n. Universal compression is related to the prolem of distriution estimation with Kullack-Leiler distortion. The prolem of estimating monotone and few-modes distriutions has een studied in statistics and theoretical computer science. [0] considered learning monotone and unimodal distriutions with 1 gn Θfn if fn gnpolylogn few samples. The sample complexity of learning m-modal distriutions over an alphaet of size k was considered y [1]. Testing distriutions for monotonicity has een considered in varied settings [ 6]. IV. RESULTS To egin, in Section V we consider approximating monotone distriution over k elements y step distriutions with significantly fewer steps. While such approximation results are known for l 1 distance [0], we need approximation in KL divergence, and these require more work as the KL divergence is not a metric and does not satisfy the triangle inequality. In Theorem 3 we show that for any integer 10 log k, there is a partition of [k] into fixed intervals such that for any P, the step function derived y averaging P over each of the intervals satisfies DP 10 log k. This result can e viewed as reducing the effective alphaet size from k to, and in Section VI we use it to show that RM n k 10 n log k + 1 log n. Specifically, in Theorem 4 we show that for large n, and any k, RM n k 0n log k log n. Hence in Corollary 5, we deduce that for k on/ log n, RM n k on. In Section VII we extend these results to m-modal distriutions and show in Theorem 7 that for large n and any k m k RM n k,m log + m + 1RM n m k. Consequently, in Corollary 8, we deduce that for any fixed m, and k on/ log n, RM n k,m on. Conversely, in Theorem 9, we show that for k n, RM n k Ωn. By monotonicity of redundancy, the same ound holds for all k Ωn. It follows that if k is suexponential in n/ log n then the redundancy diminishes to 0, while if it at least exponential in n, the redundancy is at least a constant. V. APPROXIMATING MONOTONE DISTRIBUTIONS We show that any monotone distriution over k elements can e approximated in KL-divergence y a step distriution with significantly fewer steps. We will use the following simple result on the average empirical variance of non-negative numers. Its proof follows from a straight-forward expansion and is omitted for revity.

3 Lemma 1: For 0 x 1 x... x n with mean x, n x i x nx n x 1 x. Let I1 I 1,..., I e a partition of [k] into consecutive intervals. P is a step distriution over this partition if for every 1 j, all i, i I j satisfy P i P i. Let I denote the numer of integers in interval I. For a distriution P, let e the step distriution over I1 whose constant value over each interval is P s average over that interval, namely for x I j x P x I j P I j. 3 I j Let P + j and P j e the highest and lowest proailities in interval I j, and let k j e the numer of non-zero proailities in interval I j. The following lemma ounds the KL Divergence etween P and. Lemma : For any P M k, DP k j P + j P j. j1 Proof: Using log x x 1 for x 0 and Lemma 1, P x log P x P x P x j j j P x j j 1j P x j k j P + j P j, and the proof follows y summing over all intervals. We now descrie a partition of [k] into intervals over which every monotone distriution P is closely approximated y the step distriution. For γ log k define { 1 for 1 j I j, 1 + γ j / for < j. Oserve that I j / + j1 j/ γ γ j γ j / 1 + γ j / 1 + γ / 1 which is larger than k for log1 + γ log k/. Hence the intervals span [k]. For j, I j 1, hence P + j P j, and for j > /, k j I j 1 + γ j /. We use natural logarithms throughout the paper. The results are a constant factor from ase., Theorem 3: Let 10 log k and the intervals I 1 as defined aove. Then for every P M, DP 10 log k. Proof: By Lemma, DP k j P + j P j j1 j/ γ j / P + j P j 1 + γp + /+1 + P + j γj+1 / P j 1 + γj / j/+1 k /+1 P + /+1 + γ j/+1 P j 1 + γj /, where the last step uses P + j+1 P j. We need to consider only non-zero k j s. If k j+1 is non-zero this implies that k j 1 + γ j /. Therefore, the summation aove can e ounded as DP k /+1 P + /+1 + γ P j k j I /+1 P + /+1 + γ. j/+1 The theorem follows y sustituting the values of γ and I /+1 and the inequality P + /+1 /. VI. UPPER BOUND ON MONOTONE REDUNDANCY Theorem 4: For large n, and any k, RM n k 0n log k log n. Proof: Recall the definition of n in 3. As efore, is the i.i.d. distriution y sampling n times. Then for any distriution over [k] n, DP n P n x n 1 log P n xn 1 x n x n 1 [k]n 1 [ P n x n 1 log P n x n 1 n x n x n 1 [k]n 1 + log n x n ] 1 x n 1 ndp + x n 1 [k]n P n x n 1 log n x n 1 x n 1, where the last step follows since the KL divergence for product distriutions is the sum of KL divergence of distriutions on each coordinate. For intervals I 1,..., I satisfying Theorem 3, y the definition of redundancy RM n k +inf sup x n 1 [k]n P n x n 1 log n x n 1 x n 1.

4 We use this inequality and Equation 1 to construct a distriution over [k] n that has a small KL divergence with respect to any distriution in M n k. We do this y showing that the second expression is upper ounded y the redundancy of n length i.i.d. sequences over an alphaet of size. Equation 1 states that for on, there is a distriution Q,n over [] n such that any distriution P over [] satisfies DP n 1 Q,n log n. 4 Using this distriution, we design a distriution over [k] n as follows. For the intervals descried earlier, and any x [k], let fx e the index of the interval that x elongs to. Therefore, f is a map from [k] to []. Then, fx n 1 def fx 1,..., x n def fx 1,..., fx n maps [k] n to [] n. Let fx n 1 j1 n def j 1,..., j n. The numer of x n 1 that map to j1 n is I j1... I jn. Define the distriution over [k] n as x n 1 def Q,nj n 1 I j i. 5 It is easy to check that it sums to 1 and the distriution is well defined. Similarly y Equation 3, n x n P I ji 1. I ji Using these two, for any P M k P n x n n x n 1 1 log Q x n n x n 1 [k]n 1 P n x n 1 log j n 1 []n Now, for j n 1 [] n, x n 1 :fxn 1 jn 1 x n 1 :fxn 1 jn 1 P n x n 1 x n 1 :xi Ij i P I j i Q,n j n 1. P x i a P I ji, where a exchanges the sum and product. Therefore, P n x n n x n 1 1 log x n x n 1 [k]n 1 P I ji log P I j i Q,n j1 n j n 1 []n A distriution P induces a distriution over I 1,..., I, an alphaet of size, and this expression is the KL divergence of the product distriution over the intervals to Q,n. By Equation 4 it is ounded y 1 log n/, hence RM n k + inf + sup Choosing sup P n x n n x n 1 1 log x n x n 1 [k]n 1 P I ji log P I j i Q,n j1 n j1 n []n 1 + log n. 0n log k log n proves the theorem. Corollary 5: For k on/ log n, RM n k on. VII. UPPER BOUND ON m-modal DISTRIBUTIONS We upper ound the redundancy of m-modal distriutions y decomposing them into m+1 monotone distriutions. First note that the redundancy of a union of distriution classes is close to the highest redundancy of a class in the union. Lemma 6 Redundancy of unions: If P 1,..., P T are T distriution classes over the same domain, then R P i max RP i + log T. 1iT 1iT Proof: Let distriution Q i achieve the redundancy of P i, and define Q T Q i T to e the average of these redundancy achieving distriutions. For any i and P i P i, DP i Q DP i Q i + log T RP i + log T. k first decompose the class of m-modal distriutions into mwe classes. Theorem 7: For large n and any k m k RM n k,m log + m + 1RM n m k. Proof: There are k m choices for the modes of the distriutions. Divide the class of all m-modal distriutions into k m classes such that the modes of all the distriutions within one class are the same. The distriutions are monotone etween the modes and there are at most m + 1 distinct regions. Each such region can e coded with redundancy at most RM n k and therefore the total extra numer of its can e ounded y m + 1RM n k. Comining with Lemma 6 with T k m proves Theorem 7. Corollary 8: For any fixed m and k on/ log n, Theorem 9: For k n, RM n k,m on. VIII. LOWER BOUND RM n k Ωn. Proof: We show the lower ound using the redundancy capacity theorem. Our ojective will e to construct a large class of distinguishale distriutions defined elow. Definition 10: A collection S P n of distriutions over X n is 1 ɛ distinguishale if there exists a function f : X n S, such that for any P n S, P rofx n P n < ɛ. In other words, a collection of distriutions is distinguishale if given a sample generated y one of the distriutions, we can identify the distriution, with error ɛ. A collection of distinguishale distriutions provides a lower ound on the redundancy y the following formulation of the redundancy-capacity theorem.

5 Lemma 11: If there is a collection S M n k ɛ distinguishale distriutions, then RM n k 1 ɛ log S 1. of 1 We now construct a class of cn distinguishale distriutions in M n k for k n, and a constant c > 0. This will give a lower ound of cn1 ɛ 1 on RM n k. Due to lack of space, we provide an overview of the construction of distriutions and a sketch the proof. We assume n is even for the ease of illustration. Divide [ n ] into n intervals as I 1 {1, }, and I j { j 1 + 1,..., j } for j,..., n. Our collection of distinguishale distriutions will e a suset of distriutions of the form P n, where P satisfies 1 For j n, and i 1, i j, j+1 ], P i 1 P i. For n/ of the intervals, P I j 3n and for the remaining n/, P I j 4 3n. Furthermore, P I 1 is always 4 3n. It is then easy to verify that any P satisfying these properties is a distriution in M k, and furthermore, the numer of such distriutions is exactly n 1 n. Consider the following ijection from the set of distriutions satisfying these conditions to the set of inary strings in {0, 1} n that have weight n/ and whose first it is 1. For j 1,..., n, if P I j 4/3n set the jth it to 1, and 0 otherwise, defining the ijection. We now show the existence of a large suset of such distriutions that map to strings with a large Hamming distance, using the Gilert-Varshamov ound [7], as follows. Lemma 1: For α < 1 def, there exists a class of M n1 hα o1 distriutions satisfying the properties and such that the Hamming distance etween the strings that they map to satisfies, SP i SP j < n1 α/. In other words for any pair of distriutions, their distriutions are different in at least a fraction 1 α/ of the intervals. Using this, we prove the following theorem. Theorem 13: The class of distriutions defined aove is 0.9 distinguishale and contains n/100 distriutions. Applying Lemma 11 to this theorem proves the lower ound. Due to space constraints, we only sketch the function f that maps [k] n to the collection of distriutions in the previous theorem and defer the complete proof to the full version of the paper. For a distriution P in the theorem, let LP [n] e the intervals where P I 4/3n. Note that LP n/, and we expect /3rd of the n symols to appear in these intervals when P is sampled n times. Given a sample in [k] n, compute the fraction of symols lying in LP for each of the n/100 such P s. Output the P that has the highest fraction of elements in its corresponding LP. The proof of distinguishaility uses Chernoff ounds and is along the lines of the proof of lower ound in [19]. REFERENCES [1] G. K. Zipf, The Psychoiology of Language. New York, NY, USA: Houghton-Mifflin, [] N. Merhav, G. Seroussi, and M. J. Weinerger, Optimal prefix codes for sources with two-sided geometric distriutions, IEEE Transactions on Information Theory, vol. 46, no. 1, pp , 000. [3] K. Pearson, Contriutions to the Mathematical Theory of Evolution, Philosophical Transactions of the Royal Society of London. A, vol. 185, pp , [4] J. Kieffer, A unified approach to weak universal source coding, IEEE Transactions on Information Theory, vol. 4, no. 6, pp , Nov [5] L. D. Davisson, Universal noiseless coding, IEEE Transactions on Information Theory, vol. 19, no. 6, pp , Nov [6] L. D. Davisson, R. J. McEliece, M. B. Pursley, and M. S. Wallace, Efficient universal noiseless source codes, IEEE Transactions on Information Theory, vol. 7, no. 3, pp , [7] F. M. J. Willems, Y. M. Shtarkov, and T. J. Tjalkens, The context-tree weighting method: asic properties. IEEE Transactions on Information Theory, vol. 41, no. 3, pp , [8] J. Rissanen, Fisher information and stochastic complexity, IEEE Transactions on Information Theory, vol. 4, no. 1, pp , January [9] W. Szpankowski, On asymptotics of certain recurrences arising in universal coding, Prolems of Information Transmission, vol. 34, no., pp , [10] Q. Xie and A. R. Barron, Asymptotic minimax regret for data compression, gamling and prediction, IEEE Transactions on Information Theory, vol. 46, no., pp , 000. [11] W. Szpankowski and M. J. Weinerger, Minimax redundancy for large alphaets, in ISIT, 010, pp [1] P. Elias, Universal codeword sets and representations of integers, IEEE Transactions on Information Theory, vol. 1, no., pp , Mar [13] B. Y. Ryako, Coding of a source with unknown ut ordered proailities, Prolemy Peredachi Informatsii, vol. 15, no., pp , [14] J. Rissanen, Minimax codes for finite alphaets, Information Theory, IEEE Transactions on, vol. 4, no. 3, pp , [15] L. D. Davisson and A. Leon-Garcia, A source matching approach to finding minimax codes. IEEE Transactions on Information Theory, vol. 6, no., pp , [16] M. Khosravifard, H. Saidi, M. Esmaeili, and T. A. Gulliver, The minimum average code for finite memoryless monotone sources. IEEE Transactions on Information Theory, vol. 53, no. 3, pp , 007. [17] D. Foster, R. Stine, and A. Wyner, Universal codes for finite sequences of integers drawn from a monotone distriution, IEEE Transactions on Information Theory, vol. 48, no. 6, pp , June 00. [18] G. I. Shamir, Universal source coding for monotonic and fast decaying monotonic distriutions, IEEE Transactions on Information Theory, vol. 59, no. 11, pp , 013. [19] J. Acharya, H. Das, and A. Orlitsky, Tight ounds on profile redundancy and distinguishaility, in Neural Information Processing Systems NIPS, 01. [0] L. Birgé, On the risk of histograms for estimating decreasing densities, Annals of Statistics, vol. 15, no. 3, pp , [1] C. Daskalakis, I. Diakonikolas, and R. A. Servedio, Learning k-modal distriutions via testing, in SODA, 01, pp [] M. Woodroofe and J. Sun, Testing uniformity versus a monotone density, The Annals of Statistics, vol. 7, no. 1, pp. pp , [3] O. Goldreich, S. Goldwasser, E. Lehman, and D. Ron, Testing monotonicity, Foundations of Computer Science, IEEE Annual Symposium on, vol. 0, p. 46, [4] T. Batu, R. Kumar, and R. Ruinfeld, in STOC. ACM, pp [5] R. Ruinfeld and R. A. Servedio, Testing monotone high-dimensional distriutions, Random Struct. Algorithms, vol. 34, no. 1, pp. 4 44, 009. [6] J. Acharya, A. Jafarpour, A. Orlitsky, and A. T. Suresh, A competitive test for uniformity of monotone distriutions, in AISTATS, 013, pp [7] R. M. Roth, Introduction to coding theory. Camridge University Press, 006.

WE start with a general discussion. Suppose we have

WE start with a general discussion. Suppose we have 646 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 43, NO. 2, MARCH 1997 Minimax Redundancy for the Class of Memoryless Sources Qun Xie and Andrew R. Barron, Member, IEEE Abstract Let X n = (X 1 ; 111;Xn)be

More information

Depth versus Breadth in Convolutional Polar Codes

Depth versus Breadth in Convolutional Polar Codes Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK

More information

Coding on Countably Infinite Alphabets

Coding on Countably Infinite Alphabets Coding on Countably Infinite Alphabets Non-parametric Information Theory Licence de droits d usage Outline Lossless Coding on infinite alphabets Source Coding Universal Coding Infinite Alphabets Enveloppe

More information

STATISTICAL CURVATURE AND STOCHASTIC COMPLEXITY

STATISTICAL CURVATURE AND STOCHASTIC COMPLEXITY 2nd International Symposium on Information Geometry and its Applications December 2-6, 2005, Tokyo Pages 000 000 STATISTICAL CURVATURE AND STOCHASTIC COMPLEXITY JUN-ICHI TAKEUCHI, ANDREW R. BARRON, AND

More information

Optimal Testing for Properties of Distributions

Optimal Testing for Properties of Distributions Optimal Testing for Properties of Distributions Jayadev Acharya, Constantinos Daskalakis, Gautam Kamath EECS, MIT {jayadev, costis, g}@mit.edu Abstract Given samples from an unknown discrete distribution

More information

STRONG NORMALITY AND GENERALIZED COPELAND ERDŐS NUMBERS

STRONG NORMALITY AND GENERALIZED COPELAND ERDŐS NUMBERS #A INTEGERS 6 (206) STRONG NORMALITY AND GENERALIZED COPELAND ERDŐS NUMBERS Elliot Catt School of Mathematical and Physical Sciences, The University of Newcastle, Callaghan, New South Wales, Australia

More information

1 Caveats of Parallel Algorithms

1 Caveats of Parallel Algorithms CME 323: Distriuted Algorithms and Optimization, Spring 2015 http://stanford.edu/ reza/dao. Instructor: Reza Zadeh, Matroid and Stanford. Lecture 1, 9/26/2015. Scried y Suhas Suresha, Pin Pin, Andreas

More information

Variable-to-Variable Codes with Small Redundancy Rates

Variable-to-Variable Codes with Small Redundancy Rates Variable-to-Variable Codes with Small Redundancy Rates M. Drmota W. Szpankowski September 25, 2004 This research is supported by NSF, NSA and NIH. Institut f. Diskrete Mathematik und Geometrie, TU Wien,

More information

Context tree models for source coding

Context tree models for source coding Context tree models for source coding Toward Non-parametric Information Theory Licence de droits d usage Outline Lossless Source Coding = density estimation with log-loss Source Coding and Universal Coding

More information

Square Hellinger Subadditivity for Bayesian Networks and its Applications to Identity Testing. Extended Abstract 1

Square Hellinger Subadditivity for Bayesian Networks and its Applications to Identity Testing. Extended Abstract 1 Proceedings of Machine Learning Research vol 65:1 7, 2017 Square Hellinger Subadditivity for Bayesian Networks and its Applications to Identity Testing Constantinos Daskslakis EECS and CSAIL, MIT Qinxuan

More information

Bringing Order to Special Cases of Klee s Measure Problem

Bringing Order to Special Cases of Klee s Measure Problem Bringing Order to Special Cases of Klee s Measure Prolem Karl Bringmann Max Planck Institute for Informatics Astract. Klee s Measure Prolem (KMP) asks for the volume of the union of n axis-aligned oxes

More information

A One-to-One Code and Its Anti-Redundancy

A One-to-One Code and Its Anti-Redundancy A One-to-One Code and Its Anti-Redundancy W. Szpankowski Department of Computer Science, Purdue University July 4, 2005 This research is supported by NSF, NSA and NIH. Outline of the Talk. Prefix Codes

More information

Testing Distributions Candidacy Talk

Testing Distributions Candidacy Talk Testing Distributions Candidacy Talk Clément Canonne Columbia University 2015 Introduction Introduction Property testing: what can we say about an object while barely looking at it? P 3 / 20 Introduction

More information

Upper Bounds for Stern s Diatomic Sequence and Related Sequences

Upper Bounds for Stern s Diatomic Sequence and Related Sequences Upper Bounds for Stern s Diatomic Sequence and Related Sequences Colin Defant Department of Mathematics University of Florida, U.S.A. cdefant@ufl.edu Sumitted: Jun 18, 01; Accepted: Oct, 016; Pulished:

More information

Asymptotic Minimax Regret for Data Compression, Gambling, and Prediction

Asymptotic Minimax Regret for Data Compression, Gambling, and Prediction IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 46, NO. 2, MARCH 2000 431 Asymptotic Minimax Regret for Data Compression, Gambling, Prediction Qun Xie Andrew R. Barron, Member, IEEE Abstract For problems

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006 MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.44 Transmission of Information Spring 2006 Homework 2 Solution name username April 4, 2006 Reading: Chapter

More information

Quantum query complexity of entropy estimation

Quantum query complexity of entropy estimation Quantum query complexity of entropy estimation Xiaodi Wu QuICS, University of Maryland MSR Redmond, July 19th, 2017 J o i n t C e n t e r f o r Quantum Information and Computer Science Outline Motivation

More information

#A50 INTEGERS 14 (2014) ON RATS SEQUENCES IN GENERAL BASES

#A50 INTEGERS 14 (2014) ON RATS SEQUENCES IN GENERAL BASES #A50 INTEGERS 14 (014) ON RATS SEQUENCES IN GENERAL BASES Johann Thiel Dept. of Mathematics, New York City College of Technology, Brooklyn, New York jthiel@citytech.cuny.edu Received: 6/11/13, Revised:

More information

Information Theory in Intelligent Decision Making

Information Theory in Intelligent Decision Making Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory

More information

Maximum Likelihood Approach for Symmetric Distribution Property Estimation

Maximum Likelihood Approach for Symmetric Distribution Property Estimation Maximum Likelihood Approach for Symmetric Distribution Property Estimation Jayadev Acharya, Hirakendu Das, Alon Orlitsky, Ananda Suresh Cornell, Yahoo, UCSD, Google Property estimation p: unknown discrete

More information

A vector version of Witsenhausen s counterexample: Towards convergence of control, communication and computation

A vector version of Witsenhausen s counterexample: Towards convergence of control, communication and computation A vector version of Witsenhausen s counterexample: Towards convergence of control, communication and computation Pulkit Grover and Anant Sahai Wireless Foundations, Department of EECS University of California

More information

Expansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2

Expansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2 Least squares: Mathematical theory Below we provide the "vector space" formulation, and solution, of the least squares prolem. While not strictly necessary until we ring in the machinery of matrix algera,

More information

Lecture 12: Grover s Algorithm

Lecture 12: Grover s Algorithm CPSC 519/619: Quantum Computation John Watrous, University of Calgary Lecture 12: Grover s Algorithm March 7, 2006 We have completed our study of Shor s factoring algorithm. The asic technique ehind Shor

More information

Upper Bounds on the Capacity of Binary Intermittent Communication

Upper Bounds on the Capacity of Binary Intermittent Communication Upper Bounds on the Capacity of Binary Intermittent Communication Mostafa Khoshnevisan and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame, Indiana 46556 Email:{mhoshne,

More information

Learning and Testing Structured Distributions

Learning and Testing Structured Distributions Learning and Testing Structured Distributions Ilias Diakonikolas University of Edinburgh Simons Institute March 2015 This Talk Algorithmic Framework for Distribution Estimation: Leads to fast & robust

More information

EECS 750. Hypothesis Testing with Communication Constraints

EECS 750. Hypothesis Testing with Communication Constraints EECS 750 Hypothesis Testing with Communication Constraints Name: Dinesh Krithivasan Abstract In this report, we study a modification of the classical statistical problem of bivariate hypothesis testing.

More information

Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression

Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression I. Kontoyiannis To appear, IEEE Transactions on Information Theory, Jan. 2000 Last revision, November 21, 1999 Abstract

More information

Determinants of generalized binary band matrices

Determinants of generalized binary band matrices Determinants of generalized inary and matrices Dmitry Efimov arxiv:17005655v1 [mathra] 18 Fe 017 Department of Mathematics, Komi Science Centre UrD RAS, Syktyvkar, Russia Astract Under inary matrices we

More information

A Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network

A Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network A Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network Chandrashekhar Thejaswi PS Douglas Cochran and Junshan Zhang Department of Electrical Engineering Arizona State

More information

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16 EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt

More information

A lower bound on compression of unknown alphabets

A lower bound on compression of unknown alphabets Theoretical Computer Science 33 005 93 3 wwwelseviercom/locate/tcs A lower bound on compression of unknown alphabets Nikola Jevtić a, Alon Orlitsky a,b,, Narayana P Santhanam a a ECE Department, University

More information

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University

More information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/40 Acknowledgement Praneeth Boda Himanshu Tyagi Shun Watanabe 3/40 Outline Two-terminal model: Mutual

More information

Minimax Redundancy for Large Alphabets by Analytic Methods

Minimax Redundancy for Large Alphabets by Analytic Methods Minimax Redundancy for Large Alphabets by Analytic Methods Wojciech Szpankowski Purdue University W. Lafayette, IN 47907 March 15, 2012 NSF STC Center for Science of Information CISS, Princeton 2012 Joint

More information

Sequential prediction with coded side information under logarithmic loss

Sequential prediction with coded side information under logarithmic loss under logarithmic loss Yanina Shkel Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Maxim Raginsky Department of Electrical and Computer Engineering Coordinated Science

More information

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION Jiangping Wang and Dapeng Wu Department of Electrical and Computer Engineering University of Florida, Gainesville, FL 3611 Correspondence author: Prof.

More information

Estimating properties of distributions. Ronitt Rubinfeld MIT and Tel Aviv University

Estimating properties of distributions. Ronitt Rubinfeld MIT and Tel Aviv University Estimating properties of distributions Ronitt Rubinfeld MIT and Tel Aviv University Distributions are everywhere What properties do your distributions have? Play the lottery? Is it independent? Is it uniform?

More information

Codes in the Damerau Distance for Deletion and Adjacent Transposition Correction

Codes in the Damerau Distance for Deletion and Adjacent Transposition Correction 1 Codes in the Damerau Distance for Deletion and Adjacent Transposition Correction Ryan Garys, Eitan Yaakoi, and Olgica Milenkovic ECE Department, University of Illinois, Urana-Champaign Technion University

More information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/41 Outline Two-terminal model: Mutual information Operational meaning in: Channel coding: channel

More information

Estimating Symmetric Distribution Properties Maximum Likelihood Strikes Back!

Estimating Symmetric Distribution Properties Maximum Likelihood Strikes Back! Estimating Symmetric Distribution Properties Maximum Likelihood Strikes Back! Jayadev Acharya Cornell Hirakendu Das Yahoo Alon Orlitsky UCSD Ananda Suresh Google Shannon Channel March 5, 2018 Based on:

More information

Asymptotic redundancy and prolixity

Asymptotic redundancy and prolixity Asymptotic redundancy and prolixity Yuval Dagan, Yuval Filmus, and Shay Moran April 6, 2017 Abstract Gallager (1978) considered the worst-case redundancy of Huffman codes as the maximum probability tends

More information

Spiking problem in monotone regression : penalized residual sum of squares

Spiking problem in monotone regression : penalized residual sum of squares Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where

More information

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria Source Coding Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Asymptotic Equipartition Property Optimal Codes (Huffman Coding) Universal

More information

The variance for partial match retrievals in k-dimensional bucket digital trees

The variance for partial match retrievals in k-dimensional bucket digital trees The variance for partial match retrievals in k-dimensional ucket digital trees Michael FUCHS Department of Applied Mathematics National Chiao Tung University January 12, 21 Astract The variance of partial

More information

Trace Reconstruction Revisited

Trace Reconstruction Revisited Trace Reconstruction Revisited Andrew McGregor 1, Eric Price 2, and Sofya Vorotnikova 1 1 University of Massachusetts Amherst {mcgregor,svorotni}@cs.umass.edu 2 IBM Almaden Research Center ecprice@mit.edu

More information

Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability

Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability Ahmad Beirami, Robert Calderbank, Ken Duffy, Muriel Médard Department of Electrical and Computer Engineering,

More information

Structuring Unreliable Radio Networks

Structuring Unreliable Radio Networks Structuring Unreliale Radio Networks Keren Censor-Hillel Seth Gilert Faian Kuhn Nancy Lynch Calvin Newport March 29, 2011 Astract In this paper we study the prolem of uilding a connected dominating set

More information

APC486/ELE486: Transmission and Compression of Information. Bounds on the Expected Length of Code Words

APC486/ELE486: Transmission and Compression of Information. Bounds on the Expected Length of Code Words APC486/ELE486: Transmission and Compression of Information Bounds on the Expected Length of Code Words Scribe: Kiran Vodrahalli September 8, 204 Notations In these notes, denotes a finite set, called the

More information

Tight Bounds on Minimum Maximum Pointwise Redundancy

Tight Bounds on Minimum Maximum Pointwise Redundancy Tight Bounds on Minimum Maximum Pointwise Redundancy Michael B. Baer vlnks Mountain View, CA 94041-2803, USA Email:.calbear@ 1eee.org Abstract This paper presents new lower and upper bounds for the optimal

More information

Bounded Queries, Approximations and the Boolean Hierarchy

Bounded Queries, Approximations and the Boolean Hierarchy Bounded Queries, Approximations and the Boolean Hierarchy Richard Chang Department of Computer Science and Electrical Engineering University of Maryland Baltimore County March 23, 1999 Astract This paper

More information

Optimal Routing in Chord

Optimal Routing in Chord Optimal Routing in Chord Prasanna Ganesan Gurmeet Singh Manku Astract We propose optimal routing algorithms for Chord [1], a popular topology for routing in peer-to-peer networks. Chord is an undirected

More information

Long non-crossing configurations in the plane

Long non-crossing configurations in the plane Long non-crossing configurations in the plane Adrian Dumitrescu Csaa D. Tóth July 4, 00 Astract We revisit some maximization prolems for geometric networks design under the non-crossing constraint, first

More information

Variable-Rate Universal Slepian-Wolf Coding with Feedback

Variable-Rate Universal Slepian-Wolf Coding with Feedback Variable-Rate Universal Slepian-Wolf Coding with Feedback Shriram Sarvotham, Dror Baron, and Richard G. Baraniuk Dept. of Electrical and Computer Engineering Rice University, Houston, TX 77005 Abstract

More information

Content Delivery in Erasure Broadcast. Channels with Cache and Feedback

Content Delivery in Erasure Broadcast. Channels with Cache and Feedback Content Delivery in Erasure Broadcast Channels with Cache and Feedack Asma Ghorel, Mari Koayashi, and Sheng Yang LSS, CentraleSupélec Gif-sur-Yvette, France arxiv:602.04630v [cs.t] 5 Fe 206 {asma.ghorel,

More information

arxiv: v1 [math.nt] 27 May 2018

arxiv: v1 [math.nt] 27 May 2018 SETTLIG SOME SUM SUPPOSITIOS TAAY WAKHARE AD CHRISTOPHE VIGAT arxiv:180510569v1 matht] 7 May 018 Astract We solve multiple conjectures y Byszewsi and Ulas aout the sum of ase digits function In order to

More information

Minimax Optimal Bayes Mixtures for Memoryless Sources over Large Alphabets

Minimax Optimal Bayes Mixtures for Memoryless Sources over Large Alphabets Proceedings of Machine Learning Research 8: 8, 208 Algorithmic Learning Theory 208 Minimax Optimal Bayes Mixtures for Memoryless Sources over Large Alphabets Elias Jääsaari Helsinki Institute for Information

More information

Worst-Case Bounds for Gaussian Process Models

Worst-Case Bounds for Gaussian Process Models Worst-Case Bounds for Gaussian Process Models Sham M. Kakade University of Pennsylvania Matthias W. Seeger UC Berkeley Abstract Dean P. Foster University of Pennsylvania We present a competitive analysis

More information

Luis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model

Luis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model Luis Manuel Santana Gallego 100 Appendix 3 Clock Skew Model Xiaohong Jiang and Susumu Horiguchi [JIA-01] 1. Introduction The evolution of VLSI chips toward larger die sizes and faster clock speeds makes

More information

Soft Covering with High Probability

Soft Covering with High Probability Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information

More information

THE multiway relay channel is a multicast network model

THE multiway relay channel is a multicast network model 1 A Multiway Relay Channel with Balanced Sources awrence Ong and Roy Timo arxiv:160509493v1 [csit] 31 May 2016 Astract We consider a joint source-channel coding prolem on a finite-field multiway relay

More information

Genetic Algorithms applied to Problems of Forbidden Configurations

Genetic Algorithms applied to Problems of Forbidden Configurations Genetic Algorithms applied to Prolems of Foridden Configurations R.P. Anstee Miguel Raggi Department of Mathematics University of British Columia Vancouver, B.C. Canada V6T Z2 anstee@math.uc.ca mraggi@gmail.com

More information

CSCI 2570 Introduction to Nanocomputing

CSCI 2570 Introduction to Nanocomputing CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication

More information

Generalized Sampling and Variance in Counterfactual Regret Minimization

Generalized Sampling and Variance in Counterfactual Regret Minimization Generalized Sampling and Variance in Counterfactual Regret Minimization Richard Gison and Marc Lanctot and Neil Burch and Duane Szafron and Michael Bowling Department of Computing Science, University of

More information

Learning Methods for Online Prediction Problems. Peter Bartlett Statistics and EECS UC Berkeley

Learning Methods for Online Prediction Problems. Peter Bartlett Statistics and EECS UC Berkeley Learning Methods for Online Prediction Problems Peter Bartlett Statistics and EECS UC Berkeley Course Synopsis A finite comparison class: A = {1,..., m}. Converting online to batch. Online convex optimization.

More information

Necessary and Sufficient Conditions for High-Dimensional Salient Feature Subset Recovery

Necessary and Sufficient Conditions for High-Dimensional Salient Feature Subset Recovery Necessary and Sufficient Conditions for High-Dimensional Salient Feature Subset Recovery Vincent Tan, Matt Johnson, Alan S. Willsky Stochastic Systems Group, Laboratory for Information and Decision Systems,

More information

SUFFIX TREE. SYNONYMS Compact suffix trie

SUFFIX TREE. SYNONYMS Compact suffix trie SUFFIX TREE Maxime Crochemore King s College London and Université Paris-Est, http://www.dcs.kcl.ac.uk/staff/mac/ Thierry Lecroq Université de Rouen, http://monge.univ-mlv.fr/~lecroq SYNONYMS Compact suffix

More information

On the Optimal Performance in Asymmetric Gaussian Wireless Sensor Networks With Fading

On the Optimal Performance in Asymmetric Gaussian Wireless Sensor Networks With Fading 436 IEEE TRANSACTIONS ON SIGNA ROCESSING, O. 58, NO. 4, ARI 00 [9] B. Chen and. K. Willett, On the optimality of the likelihood-ratio test for local sensor decision rules in the presence of nonideal channels,

More information

The Optimal Fix-Free Code for Anti-Uniform Sources

The Optimal Fix-Free Code for Anti-Uniform Sources Entropy 2015, 17, 1379-1386; doi:10.3390/e17031379 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article The Optimal Fix-Free Code for Anti-Uniform Sources Ali Zaghian 1, Adel Aghajan

More information

Taming Big Probability Distributions

Taming Big Probability Distributions Taming Big Probability Distributions Ronitt Rubinfeld June 18, 2012 CSAIL, Massachusetts Institute of Technology, Cambridge MA 02139 and the Blavatnik School of Computer Science, Tel Aviv University. Research

More information

Compactness vs Collusion Resistance in Functional Encryption

Compactness vs Collusion Resistance in Functional Encryption Compactness vs Collusion Resistance in Functional Encryption Baiyu Li Daniele Micciancio April 10, 2017 Astract We present two general constructions that can e used to comine any two functional encryption

More information

ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA)

ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) General information Availale time: 2.5 hours (150 minutes).

More information

Unified Maximum Likelihood Estimation of Symmetric Distribution Properties

Unified Maximum Likelihood Estimation of Symmetric Distribution Properties Unified Maximum Likelihood Estimation of Symmetric Distribution Properties Jayadev Acharya HirakenduDas Alon Orlitsky Ananda Suresh Cornell Yahoo UCSD Google Frontiers in Distribution Testing Workshop

More information

Prequential Plug-In Codes that Achieve Optimal Redundancy Rates even if the Model is Wrong

Prequential Plug-In Codes that Achieve Optimal Redundancy Rates even if the Model is Wrong Prequential Plug-In Codes that Achieve Optimal Redundancy Rates even if the Model is Wrong Peter Grünwald pdg@cwi.nl Wojciech Kotłowski kotlowsk@cwi.nl National Research Institute for Mathematics and Computer

More information

New Constructions of Sonar Sequences

New Constructions of Sonar Sequences INTERNATIONAL JOURNAL OF BASIC & APPLIED SCIENCES IJBAS-IJENS VOL.:14 NO.:01 12 New Constructions of Sonar Sequences Diego F. Ruiz 1, Carlos A. Trujillo 1, and Yadira Caicedo 2 1 Department of Mathematics,

More information

CHAPTER 5. Linear Operators, Span, Linear Independence, Basis Sets, and Dimension

CHAPTER 5. Linear Operators, Span, Linear Independence, Basis Sets, and Dimension A SERIES OF CLASS NOTES TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS LINEAR CLASS NOTES: A COLLECTION OF HANDOUTS FOR REVIEW AND PREVIEW OF LINEAR THEORY

More information

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables

Minimizing a convex separable exponential function subject to linear equality constraint and bounded variables Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad

More information

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function Dinesh Krithivasan and S. Sandeep Pradhan Department of Electrical Engineering and Computer Science,

More information

where L(y) = P Y X (y 0) C PPM = log 2 (M) ( For n b 0 the expression reduces to D. Polar Coding (2)

where L(y) = P Y X (y 0) C PPM = log 2 (M) ( For n b 0 the expression reduces to D. Polar Coding (2) Polar-Coded Pulse Position odulation for the Poisson Channel Delcho Donev Institute for Communications Engineering Technical University of unich Email: delchodonev@tumde Georg Böcherer athematical and

More information

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence 1 Overview We started by stating the two principal laws of large numbers: the strong

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities

Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Maxim Raginsky and Igal Sason ISIT 2013, Istanbul, Turkey Capacity-Achieving Channel Codes The set-up DMC

More information

1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE Source Coding, Large Deviations, and Approximate Pattern Matching

1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE Source Coding, Large Deviations, and Approximate Pattern Matching 1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE 2002 Source Coding, Large Deviations, and Approximate Pattern Matching Amir Dembo and Ioannis Kontoyiannis, Member, IEEE Invited Paper

More information

Generalized Geometric Series, The Ratio Comparison Test and Raabe s Test

Generalized Geometric Series, The Ratio Comparison Test and Raabe s Test Generalized Geometric Series The Ratio Comparison Test and Raae s Test William M. Faucette Decemer 2003 The goal of this paper is to examine the convergence of a type of infinite series in which the summands

More information

Quadratic-backtracking Algorithm for String Reconstruction from Substring Compositions

Quadratic-backtracking Algorithm for String Reconstruction from Substring Compositions Quadratic-backtracking Algorithm for String Reconstruction from Substring Compositions Jayadev Acharya UCSD jacharya@ucsdedu Hirakendu Das Yahoo Labs hdas@yahoo-inccom Olgica Milenkovic UIUC milenkov@uiucedu

More information

IN this paper, we consider the estimation of the frequency

IN this paper, we consider the estimation of the frequency Iterative Frequency Estimation y Interpolation on Fourier Coefficients Elias Aoutanios, MIEEE, Bernard Mulgrew, MIEEE Astract The estimation of the frequency of a complex exponential is a prolem that is

More information

Essential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March.

Essential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March. Essential Maths 1 The information in this document is the minimum assumed knowledge for students undertaking the Macquarie University Masters of Applied Finance, Graduate Diploma of Applied Finance, and

More information

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute ENEE 739C: Advanced Topics in Signal Processing: Coding Theory Instructor: Alexander Barg Lecture 6 (draft; 9/6/03. Error exponents for Discrete Memoryless Channels http://www.enee.umd.edu/ abarg/enee739c/course.html

More information

Jumping Sequences. Steve Butler 1. (joint work with Ron Graham and Nan Zang) University of California Los Angelese

Jumping Sequences. Steve Butler 1. (joint work with Ron Graham and Nan Zang) University of California Los Angelese Jumping Sequences Steve Butler 1 (joint work with Ron Graham and Nan Zang) 1 Department of Mathematics University of California Los Angelese www.math.ucla.edu/~butler UCSD Combinatorics Seminar 14 October

More information

Motivation: Can the equations of physics be derived from information-theoretic principles?

Motivation: Can the equations of physics be derived from information-theoretic principles? 10. Physics from Fisher Information. Motivation: Can the equations of physics e derived from information-theoretic principles? I. Fisher Information. Task: To otain a measure of the accuracy of estimated

More information

Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE

Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 5, MAY 2009 1945 Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE Abstract

More information

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 = Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values

More information

Faster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians. Constantinos Daskalakis, MIT Gautam Kamath, MIT

Faster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians. Constantinos Daskalakis, MIT Gautam Kamath, MIT Faster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians Constantinos Daskalakis, MIT Gautam Kamath, MIT What s a Gaussian Mixture Model (GMM)? Interpretation 1: PDF is a convex

More information

Lecture 6 January 15, 2014

Lecture 6 January 15, 2014 Advanced Graph Algorithms Jan-Apr 2014 Lecture 6 January 15, 2014 Lecturer: Saket Sourah Scrie: Prafullkumar P Tale 1 Overview In the last lecture we defined simple tree decomposition and stated that for

More information

Entropy power inequality for a family of discrete random variables

Entropy power inequality for a family of discrete random variables 20 IEEE International Symposium on Information Theory Proceedings Entropy power inequality for a family of discrete random variables Naresh Sharma, Smarajit Das and Siddharth Muthurishnan School of Technology

More information

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H. Problem sheet Ex. Verify that the function H(p,..., p n ) = k p k log p k satisfies all 8 axioms on H. Ex. (Not to be handed in). looking at the notes). List as many of the 8 axioms as you can, (without

More information

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 Please submit the solutions on Gradescope. EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 1. Optimal codeword lengths. Although the codeword lengths of an optimal variable length code

More information

16.1 Bounding Capacity with Covering Number

16.1 Bounding Capacity with Covering Number ECE598: Information-theoretic methods in high-dimensional statistics Spring 206 Lecture 6: Upper Bounds for Density Estimation Lecturer: Yihong Wu Scribe: Yang Zhang, Apr, 206 So far we have been mostly

More information

Network Coding for Computing

Network Coding for Computing Networ Coding for Computing Rathinaumar Appuswamy, Massimo Franceschetti, Nihil Karamchandani, and Kenneth Zeger Abstract The following networ computation problem is considered A set of source nodes in

More information

Gaussian Estimation under Attack Uncertainty

Gaussian Estimation under Attack Uncertainty Gaussian Estimation under Attack Uncertainty Tara Javidi Yonatan Kaspi Himanshu Tyagi Abstract We consider the estimation of a standard Gaussian random variable under an observation attack where an adversary

More information

ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria

ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA Thomas Adidaume Uge and Stephen Seastian Akpan, Department Of Mathematics/Statistics And Computer

More information