Efficient Compression of Monotone and m-modal Distributions
|
|
- Clara Malone
- 5 years ago
- Views:
Transcription
1 Efficient Compression of Monotone and m-modal Distriutions Jayadev Acharya, Ashkan Jafarpour, Alon Orlitsky, and Ananda Theertha Suresh ECE Department, UCSD {jacharya, ashkan, alon, Astract We consider universal compression of n samples drawn independently according to a monotone or m-modal distriution over k elements. We show that for all these distriutions, the per-sample redundancy diminishes to 0 if k expon/ log n and is at least a constant if k expωn. I. INTRODUCTION Universal compression concerns encoding the output of an unknown source in a known class of distriutions. We are interested in understanding compression of i.i.d. samples from distriutions over an underlying discrete alphaet X e.g., a suset of N. Let P n def P P... P e a product distriution over X n. Shannon s source coding theorem shows that HP def X P x log P x its are necessary and sufficient to encode a known source with distriution P, where HP is the entropy of P. By Kraft s Inequality, any distriution over X n implies a code that uses log x n 1 its to encode x n 1 X n. The code implied y the underlying distriution P n uses log P n x n 1 its and the expected numer of its used is nhp. The average redundancy of with respect to P n is DP n E P n [ log P n x n 1 x n 1 the expected extra numer of its used y eyond the entropy. DP Q is called the Kullack-Lieler divergence etween P and Q. Let the underlying distriution e in a known class P, and P n def {P n : P P} ], e the class of all distriutions of the form P n. Then RP n def inf sup P n DP n is the average or min-max redundancy of P n, which corresponds to the est coding for the worst distriution. While coding over unknown distriutions, often some apriori knowledge aout the underlying distriution is known. For example, coding the last names of people from a census data is close to coding a monotone distriution, since we have a prior knowledge aout the prevalences of last names. We expect that the last name Smith is more likely than Galifianakis. In a text document we have some knowledge aout word frequencies and proailities. In such language modeling applications, Zipf distriutions are common [1]. Geometric distriutions over integers are useful in compressing residual signals in image compression []. A natural generalization is to consider distriutions with at most m modes. For example, life expectancy of a population, Poisson, and Binomial distriutions are unimodal. There has een a considerale interest over the past decade in approximating mixtures of distriutions. Many real world phenomenon can e modeled as outcomes of mixtures of simple distriutions and these have een studied since as early as the study Naples cra population y Pearson [3]. In this setup after oserving measurements say diameter over heights, the distriution that explained the data the est was a mixture of two Gaussians, predicting the presence of more than one specie. Such mixtures of m simple distriutions are typically m-modal. There has een enormous work in the past decade in learning mixtures of distriutions. Before discussing monotone distriutions, we state the results for the most well studied finite discrete distriutions, i.e., the i.i.d. distriutions. Let Ik n e the class of all i.i.d. distriutions over k elements and lock length n. Redundancy of Ik n estalished that 1 For k on has een extensively studied [4 11]. It is now well For n ok RI n k k 1 log n 1 + o1. 1 k RIk n n log k 1 + o1. n II. TERMINOLOGY We consider distriutions over the set N of positive integers. Such a distriution is non-increasing, or monotone, if P i P i + 1 for all i 1. Let M e the class of all monotone distriutions over N, and let M k e the suset of M consisting of monotone distriutions over [k] {1,..., k}. Let M n k denote the class of distriutions otained y sampling one distriution in M k independently n times, i.e., product distriutions of the form P n for P M k. Generalizing monotone distriutions, a consecutive set [l, r] {l,..., r} of integers is a mode of a distriution P if for all i, j [l, r] P i P j, and P l 1 P l P r + 1 P r > 0, representing a local minimum or maximum of P. P is m-modal if it has at most m modes. Note that a monotone distriution is 0-modal. Let M k,m e
2 the collection of all m-modal distriutions over [k], and M n k,m, the distriutions over [k] n otained y sampling the same distriution in M k,m independently n times. III. RELATED WORK Initial work on universal compression of monotone distriutions was motivated y representation of integers, hence centered on a single instance, namely n 1. [1] showed that every positive integer k can e represented using roughly log k + log log k its. [13] derived a related lower ound, and [14] constructed a method for minimizing the expected numer of its normalized y the distriution s entropy. Remarkaly, [15] found the expected redundancy exactly, k RM k log i 1. i i 1 i Since 1 1 i i e 1 and k 1 i RM k log log k + O1, log k, this shows that and in particular, RM. More recently, [16] analyzed the min-ave redundancy that replaces the redundancy of the worst distriution in the class y the average over all distriutions. They show that monotone distriutions have constant min-ave redundancy. Recent work has concerned compression of independent identical monotone distriutions, namely distriutions in M n. The results aove imply RM n, yet [17] showed that the set of all distriutions in M with finite entropy can e compressed with a diminishing per-sample relative redundancy. Specifically, there is a distriution over N n, such that for all P M with HP <, DP n nhp log lognhp. lognhp Note that the theorem holds for all distriutions with finite entropy, even those with infinite support. It therefore does not necessarily imply that the redundancy is su-linear. For example, for the uniform distriution over [exp n], the ound is n 3/, hence super-linear. [18] considered the redundancy of monotone distriutions as a function of the alphaet size k as well as the lock length n. They show tight lower and upper ounds for k On 1, RM n k Θmin{k, n 1/3 }. In particular, this implies that when n grows linearly with k, the per-sample redundancy diminishes to 0. These results can e extended using methods from [19] to prove that M n k has su-linear redundancy for k o n. Universal compression is related to the prolem of distriution estimation with Kullack-Leiler distortion. The prolem of estimating monotone and few-modes distriutions has een studied in statistics and theoretical computer science. [0] considered learning monotone and unimodal distriutions with 1 gn Θfn if fn gnpolylogn few samples. The sample complexity of learning m-modal distriutions over an alphaet of size k was considered y [1]. Testing distriutions for monotonicity has een considered in varied settings [ 6]. IV. RESULTS To egin, in Section V we consider approximating monotone distriution over k elements y step distriutions with significantly fewer steps. While such approximation results are known for l 1 distance [0], we need approximation in KL divergence, and these require more work as the KL divergence is not a metric and does not satisfy the triangle inequality. In Theorem 3 we show that for any integer 10 log k, there is a partition of [k] into fixed intervals such that for any P, the step function derived y averaging P over each of the intervals satisfies DP 10 log k. This result can e viewed as reducing the effective alphaet size from k to, and in Section VI we use it to show that RM n k 10 n log k + 1 log n. Specifically, in Theorem 4 we show that for large n, and any k, RM n k 0n log k log n. Hence in Corollary 5, we deduce that for k on/ log n, RM n k on. In Section VII we extend these results to m-modal distriutions and show in Theorem 7 that for large n and any k m k RM n k,m log + m + 1RM n m k. Consequently, in Corollary 8, we deduce that for any fixed m, and k on/ log n, RM n k,m on. Conversely, in Theorem 9, we show that for k n, RM n k Ωn. By monotonicity of redundancy, the same ound holds for all k Ωn. It follows that if k is suexponential in n/ log n then the redundancy diminishes to 0, while if it at least exponential in n, the redundancy is at least a constant. V. APPROXIMATING MONOTONE DISTRIBUTIONS We show that any monotone distriution over k elements can e approximated in KL-divergence y a step distriution with significantly fewer steps. We will use the following simple result on the average empirical variance of non-negative numers. Its proof follows from a straight-forward expansion and is omitted for revity.
3 Lemma 1: For 0 x 1 x... x n with mean x, n x i x nx n x 1 x. Let I1 I 1,..., I e a partition of [k] into consecutive intervals. P is a step distriution over this partition if for every 1 j, all i, i I j satisfy P i P i. Let I denote the numer of integers in interval I. For a distriution P, let e the step distriution over I1 whose constant value over each interval is P s average over that interval, namely for x I j x P x I j P I j. 3 I j Let P + j and P j e the highest and lowest proailities in interval I j, and let k j e the numer of non-zero proailities in interval I j. The following lemma ounds the KL Divergence etween P and. Lemma : For any P M k, DP k j P + j P j. j1 Proof: Using log x x 1 for x 0 and Lemma 1, P x log P x P x P x j j j P x j j 1j P x j k j P + j P j, and the proof follows y summing over all intervals. We now descrie a partition of [k] into intervals over which every monotone distriution P is closely approximated y the step distriution. For γ log k define { 1 for 1 j I j, 1 + γ j / for < j. Oserve that I j / + j1 j/ γ γ j γ j / 1 + γ j / 1 + γ / 1 which is larger than k for log1 + γ log k/. Hence the intervals span [k]. For j, I j 1, hence P + j P j, and for j > /, k j I j 1 + γ j /. We use natural logarithms throughout the paper. The results are a constant factor from ase., Theorem 3: Let 10 log k and the intervals I 1 as defined aove. Then for every P M, DP 10 log k. Proof: By Lemma, DP k j P + j P j j1 j/ γ j / P + j P j 1 + γp + /+1 + P + j γj+1 / P j 1 + γj / j/+1 k /+1 P + /+1 + γ j/+1 P j 1 + γj /, where the last step uses P + j+1 P j. We need to consider only non-zero k j s. If k j+1 is non-zero this implies that k j 1 + γ j /. Therefore, the summation aove can e ounded as DP k /+1 P + /+1 + γ P j k j I /+1 P + /+1 + γ. j/+1 The theorem follows y sustituting the values of γ and I /+1 and the inequality P + /+1 /. VI. UPPER BOUND ON MONOTONE REDUNDANCY Theorem 4: For large n, and any k, RM n k 0n log k log n. Proof: Recall the definition of n in 3. As efore, is the i.i.d. distriution y sampling n times. Then for any distriution over [k] n, DP n P n x n 1 log P n xn 1 x n x n 1 [k]n 1 [ P n x n 1 log P n x n 1 n x n x n 1 [k]n 1 + log n x n ] 1 x n 1 ndp + x n 1 [k]n P n x n 1 log n x n 1 x n 1, where the last step follows since the KL divergence for product distriutions is the sum of KL divergence of distriutions on each coordinate. For intervals I 1,..., I satisfying Theorem 3, y the definition of redundancy RM n k +inf sup x n 1 [k]n P n x n 1 log n x n 1 x n 1.
4 We use this inequality and Equation 1 to construct a distriution over [k] n that has a small KL divergence with respect to any distriution in M n k. We do this y showing that the second expression is upper ounded y the redundancy of n length i.i.d. sequences over an alphaet of size. Equation 1 states that for on, there is a distriution Q,n over [] n such that any distriution P over [] satisfies DP n 1 Q,n log n. 4 Using this distriution, we design a distriution over [k] n as follows. For the intervals descried earlier, and any x [k], let fx e the index of the interval that x elongs to. Therefore, f is a map from [k] to []. Then, fx n 1 def fx 1,..., x n def fx 1,..., fx n maps [k] n to [] n. Let fx n 1 j1 n def j 1,..., j n. The numer of x n 1 that map to j1 n is I j1... I jn. Define the distriution over [k] n as x n 1 def Q,nj n 1 I j i. 5 It is easy to check that it sums to 1 and the distriution is well defined. Similarly y Equation 3, n x n P I ji 1. I ji Using these two, for any P M k P n x n n x n 1 1 log Q x n n x n 1 [k]n 1 P n x n 1 log j n 1 []n Now, for j n 1 [] n, x n 1 :fxn 1 jn 1 x n 1 :fxn 1 jn 1 P n x n 1 x n 1 :xi Ij i P I j i Q,n j n 1. P x i a P I ji, where a exchanges the sum and product. Therefore, P n x n n x n 1 1 log x n x n 1 [k]n 1 P I ji log P I j i Q,n j1 n j n 1 []n A distriution P induces a distriution over I 1,..., I, an alphaet of size, and this expression is the KL divergence of the product distriution over the intervals to Q,n. By Equation 4 it is ounded y 1 log n/, hence RM n k + inf + sup Choosing sup P n x n n x n 1 1 log x n x n 1 [k]n 1 P I ji log P I j i Q,n j1 n j1 n []n 1 + log n. 0n log k log n proves the theorem. Corollary 5: For k on/ log n, RM n k on. VII. UPPER BOUND ON m-modal DISTRIBUTIONS We upper ound the redundancy of m-modal distriutions y decomposing them into m+1 monotone distriutions. First note that the redundancy of a union of distriution classes is close to the highest redundancy of a class in the union. Lemma 6 Redundancy of unions: If P 1,..., P T are T distriution classes over the same domain, then R P i max RP i + log T. 1iT 1iT Proof: Let distriution Q i achieve the redundancy of P i, and define Q T Q i T to e the average of these redundancy achieving distriutions. For any i and P i P i, DP i Q DP i Q i + log T RP i + log T. k first decompose the class of m-modal distriutions into mwe classes. Theorem 7: For large n and any k m k RM n k,m log + m + 1RM n m k. Proof: There are k m choices for the modes of the distriutions. Divide the class of all m-modal distriutions into k m classes such that the modes of all the distriutions within one class are the same. The distriutions are monotone etween the modes and there are at most m + 1 distinct regions. Each such region can e coded with redundancy at most RM n k and therefore the total extra numer of its can e ounded y m + 1RM n k. Comining with Lemma 6 with T k m proves Theorem 7. Corollary 8: For any fixed m and k on/ log n, Theorem 9: For k n, RM n k,m on. VIII. LOWER BOUND RM n k Ωn. Proof: We show the lower ound using the redundancy capacity theorem. Our ojective will e to construct a large class of distinguishale distriutions defined elow. Definition 10: A collection S P n of distriutions over X n is 1 ɛ distinguishale if there exists a function f : X n S, such that for any P n S, P rofx n P n < ɛ. In other words, a collection of distriutions is distinguishale if given a sample generated y one of the distriutions, we can identify the distriution, with error ɛ. A collection of distinguishale distriutions provides a lower ound on the redundancy y the following formulation of the redundancy-capacity theorem.
5 Lemma 11: If there is a collection S M n k ɛ distinguishale distriutions, then RM n k 1 ɛ log S 1. of 1 We now construct a class of cn distinguishale distriutions in M n k for k n, and a constant c > 0. This will give a lower ound of cn1 ɛ 1 on RM n k. Due to lack of space, we provide an overview of the construction of distriutions and a sketch the proof. We assume n is even for the ease of illustration. Divide [ n ] into n intervals as I 1 {1, }, and I j { j 1 + 1,..., j } for j,..., n. Our collection of distinguishale distriutions will e a suset of distriutions of the form P n, where P satisfies 1 For j n, and i 1, i j, j+1 ], P i 1 P i. For n/ of the intervals, P I j 3n and for the remaining n/, P I j 4 3n. Furthermore, P I 1 is always 4 3n. It is then easy to verify that any P satisfying these properties is a distriution in M k, and furthermore, the numer of such distriutions is exactly n 1 n. Consider the following ijection from the set of distriutions satisfying these conditions to the set of inary strings in {0, 1} n that have weight n/ and whose first it is 1. For j 1,..., n, if P I j 4/3n set the jth it to 1, and 0 otherwise, defining the ijection. We now show the existence of a large suset of such distriutions that map to strings with a large Hamming distance, using the Gilert-Varshamov ound [7], as follows. Lemma 1: For α < 1 def, there exists a class of M n1 hα o1 distriutions satisfying the properties and such that the Hamming distance etween the strings that they map to satisfies, SP i SP j < n1 α/. In other words for any pair of distriutions, their distriutions are different in at least a fraction 1 α/ of the intervals. Using this, we prove the following theorem. Theorem 13: The class of distriutions defined aove is 0.9 distinguishale and contains n/100 distriutions. Applying Lemma 11 to this theorem proves the lower ound. Due to space constraints, we only sketch the function f that maps [k] n to the collection of distriutions in the previous theorem and defer the complete proof to the full version of the paper. For a distriution P in the theorem, let LP [n] e the intervals where P I 4/3n. Note that LP n/, and we expect /3rd of the n symols to appear in these intervals when P is sampled n times. Given a sample in [k] n, compute the fraction of symols lying in LP for each of the n/100 such P s. Output the P that has the highest fraction of elements in its corresponding LP. The proof of distinguishaility uses Chernoff ounds and is along the lines of the proof of lower ound in [19]. REFERENCES [1] G. K. Zipf, The Psychoiology of Language. New York, NY, USA: Houghton-Mifflin, [] N. Merhav, G. Seroussi, and M. J. Weinerger, Optimal prefix codes for sources with two-sided geometric distriutions, IEEE Transactions on Information Theory, vol. 46, no. 1, pp , 000. [3] K. Pearson, Contriutions to the Mathematical Theory of Evolution, Philosophical Transactions of the Royal Society of London. A, vol. 185, pp , [4] J. Kieffer, A unified approach to weak universal source coding, IEEE Transactions on Information Theory, vol. 4, no. 6, pp , Nov [5] L. D. Davisson, Universal noiseless coding, IEEE Transactions on Information Theory, vol. 19, no. 6, pp , Nov [6] L. D. Davisson, R. J. McEliece, M. B. Pursley, and M. S. Wallace, Efficient universal noiseless source codes, IEEE Transactions on Information Theory, vol. 7, no. 3, pp , [7] F. M. J. Willems, Y. M. Shtarkov, and T. J. Tjalkens, The context-tree weighting method: asic properties. IEEE Transactions on Information Theory, vol. 41, no. 3, pp , [8] J. Rissanen, Fisher information and stochastic complexity, IEEE Transactions on Information Theory, vol. 4, no. 1, pp , January [9] W. Szpankowski, On asymptotics of certain recurrences arising in universal coding, Prolems of Information Transmission, vol. 34, no., pp , [10] Q. Xie and A. R. Barron, Asymptotic minimax regret for data compression, gamling and prediction, IEEE Transactions on Information Theory, vol. 46, no., pp , 000. [11] W. Szpankowski and M. J. Weinerger, Minimax redundancy for large alphaets, in ISIT, 010, pp [1] P. Elias, Universal codeword sets and representations of integers, IEEE Transactions on Information Theory, vol. 1, no., pp , Mar [13] B. Y. Ryako, Coding of a source with unknown ut ordered proailities, Prolemy Peredachi Informatsii, vol. 15, no., pp , [14] J. Rissanen, Minimax codes for finite alphaets, Information Theory, IEEE Transactions on, vol. 4, no. 3, pp , [15] L. D. Davisson and A. Leon-Garcia, A source matching approach to finding minimax codes. IEEE Transactions on Information Theory, vol. 6, no., pp , [16] M. Khosravifard, H. Saidi, M. Esmaeili, and T. A. Gulliver, The minimum average code for finite memoryless monotone sources. IEEE Transactions on Information Theory, vol. 53, no. 3, pp , 007. [17] D. Foster, R. Stine, and A. Wyner, Universal codes for finite sequences of integers drawn from a monotone distriution, IEEE Transactions on Information Theory, vol. 48, no. 6, pp , June 00. [18] G. I. Shamir, Universal source coding for monotonic and fast decaying monotonic distriutions, IEEE Transactions on Information Theory, vol. 59, no. 11, pp , 013. [19] J. Acharya, H. Das, and A. Orlitsky, Tight ounds on profile redundancy and distinguishaility, in Neural Information Processing Systems NIPS, 01. [0] L. Birgé, On the risk of histograms for estimating decreasing densities, Annals of Statistics, vol. 15, no. 3, pp , [1] C. Daskalakis, I. Diakonikolas, and R. A. Servedio, Learning k-modal distriutions via testing, in SODA, 01, pp [] M. Woodroofe and J. Sun, Testing uniformity versus a monotone density, The Annals of Statistics, vol. 7, no. 1, pp. pp , [3] O. Goldreich, S. Goldwasser, E. Lehman, and D. Ron, Testing monotonicity, Foundations of Computer Science, IEEE Annual Symposium on, vol. 0, p. 46, [4] T. Batu, R. Kumar, and R. Ruinfeld, in STOC. ACM, pp [5] R. Ruinfeld and R. A. Servedio, Testing monotone high-dimensional distriutions, Random Struct. Algorithms, vol. 34, no. 1, pp. 4 44, 009. [6] J. Acharya, A. Jafarpour, A. Orlitsky, and A. T. Suresh, A competitive test for uniformity of monotone distriutions, in AISTATS, 013, pp [7] R. M. Roth, Introduction to coding theory. Camridge University Press, 006.
WE start with a general discussion. Suppose we have
646 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 43, NO. 2, MARCH 1997 Minimax Redundancy for the Class of Memoryless Sources Qun Xie and Andrew R. Barron, Member, IEEE Abstract Let X n = (X 1 ; 111;Xn)be
More informationDepth versus Breadth in Convolutional Polar Codes
Depth versus Breadth in Convolutional Polar Codes Maxime Tremlay, Benjamin Bourassa and David Poulin,2 Département de physique & Institut quantique, Université de Sherrooke, Sherrooke, Quéec, Canada JK
More informationCoding on Countably Infinite Alphabets
Coding on Countably Infinite Alphabets Non-parametric Information Theory Licence de droits d usage Outline Lossless Coding on infinite alphabets Source Coding Universal Coding Infinite Alphabets Enveloppe
More informationSTATISTICAL CURVATURE AND STOCHASTIC COMPLEXITY
2nd International Symposium on Information Geometry and its Applications December 2-6, 2005, Tokyo Pages 000 000 STATISTICAL CURVATURE AND STOCHASTIC COMPLEXITY JUN-ICHI TAKEUCHI, ANDREW R. BARRON, AND
More informationOptimal Testing for Properties of Distributions
Optimal Testing for Properties of Distributions Jayadev Acharya, Constantinos Daskalakis, Gautam Kamath EECS, MIT {jayadev, costis, g}@mit.edu Abstract Given samples from an unknown discrete distribution
More informationSTRONG NORMALITY AND GENERALIZED COPELAND ERDŐS NUMBERS
#A INTEGERS 6 (206) STRONG NORMALITY AND GENERALIZED COPELAND ERDŐS NUMBERS Elliot Catt School of Mathematical and Physical Sciences, The University of Newcastle, Callaghan, New South Wales, Australia
More information1 Caveats of Parallel Algorithms
CME 323: Distriuted Algorithms and Optimization, Spring 2015 http://stanford.edu/ reza/dao. Instructor: Reza Zadeh, Matroid and Stanford. Lecture 1, 9/26/2015. Scried y Suhas Suresha, Pin Pin, Andreas
More informationVariable-to-Variable Codes with Small Redundancy Rates
Variable-to-Variable Codes with Small Redundancy Rates M. Drmota W. Szpankowski September 25, 2004 This research is supported by NSF, NSA and NIH. Institut f. Diskrete Mathematik und Geometrie, TU Wien,
More informationContext tree models for source coding
Context tree models for source coding Toward Non-parametric Information Theory Licence de droits d usage Outline Lossless Source Coding = density estimation with log-loss Source Coding and Universal Coding
More informationSquare Hellinger Subadditivity for Bayesian Networks and its Applications to Identity Testing. Extended Abstract 1
Proceedings of Machine Learning Research vol 65:1 7, 2017 Square Hellinger Subadditivity for Bayesian Networks and its Applications to Identity Testing Constantinos Daskslakis EECS and CSAIL, MIT Qinxuan
More informationBringing Order to Special Cases of Klee s Measure Problem
Bringing Order to Special Cases of Klee s Measure Prolem Karl Bringmann Max Planck Institute for Informatics Astract. Klee s Measure Prolem (KMP) asks for the volume of the union of n axis-aligned oxes
More informationA One-to-One Code and Its Anti-Redundancy
A One-to-One Code and Its Anti-Redundancy W. Szpankowski Department of Computer Science, Purdue University July 4, 2005 This research is supported by NSF, NSA and NIH. Outline of the Talk. Prefix Codes
More informationTesting Distributions Candidacy Talk
Testing Distributions Candidacy Talk Clément Canonne Columbia University 2015 Introduction Introduction Property testing: what can we say about an object while barely looking at it? P 3 / 20 Introduction
More informationUpper Bounds for Stern s Diatomic Sequence and Related Sequences
Upper Bounds for Stern s Diatomic Sequence and Related Sequences Colin Defant Department of Mathematics University of Florida, U.S.A. cdefant@ufl.edu Sumitted: Jun 18, 01; Accepted: Oct, 016; Pulished:
More informationAsymptotic Minimax Regret for Data Compression, Gambling, and Prediction
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 46, NO. 2, MARCH 2000 431 Asymptotic Minimax Regret for Data Compression, Gambling, Prediction Qun Xie Andrew R. Barron, Member, IEEE Abstract For problems
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science Transmission of Information Spring 2006
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.44 Transmission of Information Spring 2006 Homework 2 Solution name username April 4, 2006 Reading: Chapter
More informationQuantum query complexity of entropy estimation
Quantum query complexity of entropy estimation Xiaodi Wu QuICS, University of Maryland MSR Redmond, July 19th, 2017 J o i n t C e n t e r f o r Quantum Information and Computer Science Outline Motivation
More information#A50 INTEGERS 14 (2014) ON RATS SEQUENCES IN GENERAL BASES
#A50 INTEGERS 14 (014) ON RATS SEQUENCES IN GENERAL BASES Johann Thiel Dept. of Mathematics, New York City College of Technology, Brooklyn, New York jthiel@citytech.cuny.edu Received: 6/11/13, Revised:
More informationInformation Theory in Intelligent Decision Making
Information Theory in Intelligent Decision Making Adaptive Systems and Algorithms Research Groups School of Computer Science University of Hertfordshire, United Kingdom June 7, 2015 Information Theory
More informationMaximum Likelihood Approach for Symmetric Distribution Property Estimation
Maximum Likelihood Approach for Symmetric Distribution Property Estimation Jayadev Acharya, Hirakendu Das, Alon Orlitsky, Ananda Suresh Cornell, Yahoo, UCSD, Google Property estimation p: unknown discrete
More informationA vector version of Witsenhausen s counterexample: Towards convergence of control, communication and computation
A vector version of Witsenhausen s counterexample: Towards convergence of control, communication and computation Pulkit Grover and Anant Sahai Wireless Foundations, Department of EECS University of California
More informationExpansion formula using properties of dot product (analogous to FOIL in algebra): u v 2 u v u v u u 2u v v v u 2 2u v v 2
Least squares: Mathematical theory Below we provide the "vector space" formulation, and solution, of the least squares prolem. While not strictly necessary until we ring in the machinery of matrix algera,
More informationLecture 12: Grover s Algorithm
CPSC 519/619: Quantum Computation John Watrous, University of Calgary Lecture 12: Grover s Algorithm March 7, 2006 We have completed our study of Shor s factoring algorithm. The asic technique ehind Shor
More informationUpper Bounds on the Capacity of Binary Intermittent Communication
Upper Bounds on the Capacity of Binary Intermittent Communication Mostafa Khoshnevisan and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame, Indiana 46556 Email:{mhoshne,
More informationLearning and Testing Structured Distributions
Learning and Testing Structured Distributions Ilias Diakonikolas University of Edinburgh Simons Institute March 2015 This Talk Algorithmic Framework for Distribution Estimation: Leads to fast & robust
More informationEECS 750. Hypothesis Testing with Communication Constraints
EECS 750 Hypothesis Testing with Communication Constraints Name: Dinesh Krithivasan Abstract In this report, we study a modification of the classical statistical problem of bivariate hypothesis testing.
More informationPointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression
Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression I. Kontoyiannis To appear, IEEE Transactions on Information Theory, Jan. 2000 Last revision, November 21, 1999 Abstract
More informationDeterminants of generalized binary band matrices
Determinants of generalized inary and matrices Dmitry Efimov arxiv:17005655v1 [mathra] 18 Fe 017 Department of Mathematics, Komi Science Centre UrD RAS, Syktyvkar, Russia Astract Under inary matrices we
More informationA Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network
A Sufficient Condition for Optimality of Digital versus Analog Relaying in a Sensor Network Chandrashekhar Thejaswi PS Douglas Cochran and Junshan Zhang Department of Electrical Engineering Arizona State
More informationEE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16
EE539R: Problem Set 4 Assigned: 3/08/6, Due: 07/09/6. Cover and Thomas: Problem 3.5 Sets defined by probabilities: Define the set C n (t = {x n : P X n(x n 2 nt } (a We have = P X n(x n P X n(x n 2 nt
More informationA lower bound on compression of unknown alphabets
Theoretical Computer Science 33 005 93 3 wwwelseviercom/locate/tcs A lower bound on compression of unknown alphabets Nikola Jevtić a, Alon Orlitsky a,b,, Narayana P Santhanam a a ECE Department, University
More informationMMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only
MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University
More informationSHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe
SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/40 Acknowledgement Praneeth Boda Himanshu Tyagi Shun Watanabe 3/40 Outline Two-terminal model: Mutual
More informationMinimax Redundancy for Large Alphabets by Analytic Methods
Minimax Redundancy for Large Alphabets by Analytic Methods Wojciech Szpankowski Purdue University W. Lafayette, IN 47907 March 15, 2012 NSF STC Center for Science of Information CISS, Princeton 2012 Joint
More informationSequential prediction with coded side information under logarithmic loss
under logarithmic loss Yanina Shkel Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Maxim Raginsky Department of Electrical and Computer Engineering Coordinated Science
More informationTIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION
TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION Jiangping Wang and Dapeng Wu Department of Electrical and Computer Engineering University of Florida, Gainesville, FL 3611 Correspondence author: Prof.
More informationEstimating properties of distributions. Ronitt Rubinfeld MIT and Tel Aviv University
Estimating properties of distributions Ronitt Rubinfeld MIT and Tel Aviv University Distributions are everywhere What properties do your distributions have? Play the lottery? Is it independent? Is it uniform?
More informationCodes in the Damerau Distance for Deletion and Adjacent Transposition Correction
1 Codes in the Damerau Distance for Deletion and Adjacent Transposition Correction Ryan Garys, Eitan Yaakoi, and Olgica Milenkovic ECE Department, University of Illinois, Urana-Champaign Technion University
More informationSHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe
SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/41 Outline Two-terminal model: Mutual information Operational meaning in: Channel coding: channel
More informationEstimating Symmetric Distribution Properties Maximum Likelihood Strikes Back!
Estimating Symmetric Distribution Properties Maximum Likelihood Strikes Back! Jayadev Acharya Cornell Hirakendu Das Yahoo Alon Orlitsky UCSD Ananda Suresh Google Shannon Channel March 5, 2018 Based on:
More informationAsymptotic redundancy and prolixity
Asymptotic redundancy and prolixity Yuval Dagan, Yuval Filmus, and Shay Moran April 6, 2017 Abstract Gallager (1978) considered the worst-case redundancy of Huffman codes as the maximum probability tends
More informationSpiking problem in monotone regression : penalized residual sum of squares
Spiking prolem in monotone regression : penalized residual sum of squares Jayanta Kumar Pal 12 SAMSI, NC 27606, U.S.A. Astract We consider the estimation of a monotone regression at its end-point, where
More informationSource Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria
Source Coding Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Asymptotic Equipartition Property Optimal Codes (Huffman Coding) Universal
More informationThe variance for partial match retrievals in k-dimensional bucket digital trees
The variance for partial match retrievals in k-dimensional ucket digital trees Michael FUCHS Department of Applied Mathematics National Chiao Tung University January 12, 21 Astract The variance of partial
More informationTrace Reconstruction Revisited
Trace Reconstruction Revisited Andrew McGregor 1, Eric Price 2, and Sofya Vorotnikova 1 1 University of Massachusetts Amherst {mcgregor,svorotni}@cs.umass.edu 2 IBM Almaden Research Center ecprice@mit.edu
More informationQuantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability
Quantifying Computational Security Subject to Source Constraints, Guesswork and Inscrutability Ahmad Beirami, Robert Calderbank, Ken Duffy, Muriel Médard Department of Electrical and Computer Engineering,
More informationStructuring Unreliable Radio Networks
Structuring Unreliale Radio Networks Keren Censor-Hillel Seth Gilert Faian Kuhn Nancy Lynch Calvin Newport March 29, 2011 Astract In this paper we study the prolem of uilding a connected dominating set
More informationAPC486/ELE486: Transmission and Compression of Information. Bounds on the Expected Length of Code Words
APC486/ELE486: Transmission and Compression of Information Bounds on the Expected Length of Code Words Scribe: Kiran Vodrahalli September 8, 204 Notations In these notes, denotes a finite set, called the
More informationTight Bounds on Minimum Maximum Pointwise Redundancy
Tight Bounds on Minimum Maximum Pointwise Redundancy Michael B. Baer vlnks Mountain View, CA 94041-2803, USA Email:.calbear@ 1eee.org Abstract This paper presents new lower and upper bounds for the optimal
More informationBounded Queries, Approximations and the Boolean Hierarchy
Bounded Queries, Approximations and the Boolean Hierarchy Richard Chang Department of Computer Science and Electrical Engineering University of Maryland Baltimore County March 23, 1999 Astract This paper
More informationOptimal Routing in Chord
Optimal Routing in Chord Prasanna Ganesan Gurmeet Singh Manku Astract We propose optimal routing algorithms for Chord [1], a popular topology for routing in peer-to-peer networks. Chord is an undirected
More informationLong non-crossing configurations in the plane
Long non-crossing configurations in the plane Adrian Dumitrescu Csaa D. Tóth July 4, 00 Astract We revisit some maximization prolems for geometric networks design under the non-crossing constraint, first
More informationVariable-Rate Universal Slepian-Wolf Coding with Feedback
Variable-Rate Universal Slepian-Wolf Coding with Feedback Shriram Sarvotham, Dror Baron, and Richard G. Baraniuk Dept. of Electrical and Computer Engineering Rice University, Houston, TX 77005 Abstract
More informationContent Delivery in Erasure Broadcast. Channels with Cache and Feedback
Content Delivery in Erasure Broadcast Channels with Cache and Feedack Asma Ghorel, Mari Koayashi, and Sheng Yang LSS, CentraleSupélec Gif-sur-Yvette, France arxiv:602.04630v [cs.t] 5 Fe 206 {asma.ghorel,
More informationarxiv: v1 [math.nt] 27 May 2018
SETTLIG SOME SUM SUPPOSITIOS TAAY WAKHARE AD CHRISTOPHE VIGAT arxiv:180510569v1 matht] 7 May 018 Astract We solve multiple conjectures y Byszewsi and Ulas aout the sum of ase digits function In order to
More informationMinimax Optimal Bayes Mixtures for Memoryless Sources over Large Alphabets
Proceedings of Machine Learning Research 8: 8, 208 Algorithmic Learning Theory 208 Minimax Optimal Bayes Mixtures for Memoryless Sources over Large Alphabets Elias Jääsaari Helsinki Institute for Information
More informationWorst-Case Bounds for Gaussian Process Models
Worst-Case Bounds for Gaussian Process Models Sham M. Kakade University of Pennsylvania Matthias W. Seeger UC Berkeley Abstract Dean P. Foster University of Pennsylvania We present a competitive analysis
More informationLuis Manuel Santana Gallego 100 Investigation and simulation of the clock skew in modern integrated circuits. Clock Skew Model
Luis Manuel Santana Gallego 100 Appendix 3 Clock Skew Model Xiaohong Jiang and Susumu Horiguchi [JIA-01] 1. Introduction The evolution of VLSI chips toward larger die sizes and faster clock speeds makes
More informationSoft Covering with High Probability
Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information
More informationTHE multiway relay channel is a multicast network model
1 A Multiway Relay Channel with Balanced Sources awrence Ong and Roy Timo arxiv:160509493v1 [csit] 31 May 2016 Astract We consider a joint source-channel coding prolem on a finite-field multiway relay
More informationGenetic Algorithms applied to Problems of Forbidden Configurations
Genetic Algorithms applied to Prolems of Foridden Configurations R.P. Anstee Miguel Raggi Department of Mathematics University of British Columia Vancouver, B.C. Canada V6T Z2 anstee@math.uc.ca mraggi@gmail.com
More informationCSCI 2570 Introduction to Nanocomputing
CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication
More informationGeneralized Sampling and Variance in Counterfactual Regret Minimization
Generalized Sampling and Variance in Counterfactual Regret Minimization Richard Gison and Marc Lanctot and Neil Burch and Duane Szafron and Michael Bowling Department of Computing Science, University of
More informationLearning Methods for Online Prediction Problems. Peter Bartlett Statistics and EECS UC Berkeley
Learning Methods for Online Prediction Problems Peter Bartlett Statistics and EECS UC Berkeley Course Synopsis A finite comparison class: A = {1,..., m}. Converting online to batch. Online convex optimization.
More informationNecessary and Sufficient Conditions for High-Dimensional Salient Feature Subset Recovery
Necessary and Sufficient Conditions for High-Dimensional Salient Feature Subset Recovery Vincent Tan, Matt Johnson, Alan S. Willsky Stochastic Systems Group, Laboratory for Information and Decision Systems,
More informationSUFFIX TREE. SYNONYMS Compact suffix trie
SUFFIX TREE Maxime Crochemore King s College London and Université Paris-Est, http://www.dcs.kcl.ac.uk/staff/mac/ Thierry Lecroq Université de Rouen, http://monge.univ-mlv.fr/~lecroq SYNONYMS Compact suffix
More informationOn the Optimal Performance in Asymmetric Gaussian Wireless Sensor Networks With Fading
436 IEEE TRANSACTIONS ON SIGNA ROCESSING, O. 58, NO. 4, ARI 00 [9] B. Chen and. K. Willett, On the optimality of the likelihood-ratio test for local sensor decision rules in the presence of nonideal channels,
More informationThe Optimal Fix-Free Code for Anti-Uniform Sources
Entropy 2015, 17, 1379-1386; doi:10.3390/e17031379 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article The Optimal Fix-Free Code for Anti-Uniform Sources Ali Zaghian 1, Adel Aghajan
More informationTaming Big Probability Distributions
Taming Big Probability Distributions Ronitt Rubinfeld June 18, 2012 CSAIL, Massachusetts Institute of Technology, Cambridge MA 02139 and the Blavatnik School of Computer Science, Tel Aviv University. Research
More informationCompactness vs Collusion Resistance in Functional Encryption
Compactness vs Collusion Resistance in Functional Encryption Baiyu Li Daniele Micciancio April 10, 2017 Astract We present two general constructions that can e used to comine any two functional encryption
More informationERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA)
ERASMUS UNIVERSITY ROTTERDAM Information concerning the Entrance examination Mathematics level 2 for International Business Administration (IBA) General information Availale time: 2.5 hours (150 minutes).
More informationUnified Maximum Likelihood Estimation of Symmetric Distribution Properties
Unified Maximum Likelihood Estimation of Symmetric Distribution Properties Jayadev Acharya HirakenduDas Alon Orlitsky Ananda Suresh Cornell Yahoo UCSD Google Frontiers in Distribution Testing Workshop
More informationPrequential Plug-In Codes that Achieve Optimal Redundancy Rates even if the Model is Wrong
Prequential Plug-In Codes that Achieve Optimal Redundancy Rates even if the Model is Wrong Peter Grünwald pdg@cwi.nl Wojciech Kotłowski kotlowsk@cwi.nl National Research Institute for Mathematics and Computer
More informationNew Constructions of Sonar Sequences
INTERNATIONAL JOURNAL OF BASIC & APPLIED SCIENCES IJBAS-IJENS VOL.:14 NO.:01 12 New Constructions of Sonar Sequences Diego F. Ruiz 1, Carlos A. Trujillo 1, and Yadira Caicedo 2 1 Department of Mathematics,
More informationCHAPTER 5. Linear Operators, Span, Linear Independence, Basis Sets, and Dimension
A SERIES OF CLASS NOTES TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS LINEAR CLASS NOTES: A COLLECTION OF HANDOUTS FOR REVIEW AND PREVIEW OF LINEAR THEORY
More informationMinimizing a convex separable exponential function subject to linear equality constraint and bounded variables
Minimizing a convex separale exponential function suect to linear equality constraint and ounded variales Stefan M. Stefanov Department of Mathematics Neofit Rilski South-Western University 2700 Blagoevgrad
More informationLattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function
Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function Dinesh Krithivasan and S. Sandeep Pradhan Department of Electrical Engineering and Computer Science,
More informationwhere L(y) = P Y X (y 0) C PPM = log 2 (M) ( For n b 0 the expression reduces to D. Polar Coding (2)
Polar-Coded Pulse Position odulation for the Poisson Channel Delcho Donev Institute for Communications Engineering Technical University of unich Email: delchodonev@tumde Georg Böcherer athematical and
More informationIEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence
IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence 1 Overview We started by stating the two principal laws of large numbers: the strong
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationRefined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities
Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Maxim Raginsky and Igal Sason ISIT 2013, Istanbul, Turkey Capacity-Achieving Channel Codes The set-up DMC
More information1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE Source Coding, Large Deviations, and Approximate Pattern Matching
1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE 2002 Source Coding, Large Deviations, and Approximate Pattern Matching Amir Dembo and Ioannis Kontoyiannis, Member, IEEE Invited Paper
More informationGeneralized Geometric Series, The Ratio Comparison Test and Raabe s Test
Generalized Geometric Series The Ratio Comparison Test and Raae s Test William M. Faucette Decemer 2003 The goal of this paper is to examine the convergence of a type of infinite series in which the summands
More informationQuadratic-backtracking Algorithm for String Reconstruction from Substring Compositions
Quadratic-backtracking Algorithm for String Reconstruction from Substring Compositions Jayadev Acharya UCSD jacharya@ucsdedu Hirakendu Das Yahoo Labs hdas@yahoo-inccom Olgica Milenkovic UIUC milenkov@uiucedu
More informationIN this paper, we consider the estimation of the frequency
Iterative Frequency Estimation y Interpolation on Fourier Coefficients Elias Aoutanios, MIEEE, Bernard Mulgrew, MIEEE Astract The estimation of the frequency of a complex exponential is a prolem that is
More informationEssential Maths 1. Macquarie University MAFC_Essential_Maths Page 1 of These notes were prepared by Anne Cooper and Catriona March.
Essential Maths 1 The information in this document is the minimum assumed knowledge for students undertaking the Macquarie University Masters of Applied Finance, Graduate Diploma of Applied Finance, and
More information(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute
ENEE 739C: Advanced Topics in Signal Processing: Coding Theory Instructor: Alexander Barg Lecture 6 (draft; 9/6/03. Error exponents for Discrete Memoryless Channels http://www.enee.umd.edu/ abarg/enee739c/course.html
More informationJumping Sequences. Steve Butler 1. (joint work with Ron Graham and Nan Zang) University of California Los Angelese
Jumping Sequences Steve Butler 1 (joint work with Ron Graham and Nan Zang) 1 Department of Mathematics University of California Los Angelese www.math.ucla.edu/~butler UCSD Combinatorics Seminar 14 October
More informationMotivation: Can the equations of physics be derived from information-theoretic principles?
10. Physics from Fisher Information. Motivation: Can the equations of physics e derived from information-theoretic principles? I. Fisher Information. Task: To otain a measure of the accuracy of estimated
More informationJoint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 5, MAY 2009 1945 Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE Abstract
More information1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =
Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values
More informationFaster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians. Constantinos Daskalakis, MIT Gautam Kamath, MIT
Faster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians Constantinos Daskalakis, MIT Gautam Kamath, MIT What s a Gaussian Mixture Model (GMM)? Interpretation 1: PDF is a convex
More informationLecture 6 January 15, 2014
Advanced Graph Algorithms Jan-Apr 2014 Lecture 6 January 15, 2014 Lecturer: Saket Sourah Scrie: Prafullkumar P Tale 1 Overview In the last lecture we defined simple tree decomposition and stated that for
More informationEntropy power inequality for a family of discrete random variables
20 IEEE International Symposium on Information Theory Proceedings Entropy power inequality for a family of discrete random variables Naresh Sharma, Smarajit Das and Siddharth Muthurishnan School of Technology
More information1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.
Problem sheet Ex. Verify that the function H(p,..., p n ) = k p k log p k satisfies all 8 axioms on H. Ex. (Not to be handed in). looking at the notes). List as many of the 8 axioms as you can, (without
More informationEE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018
Please submit the solutions on Gradescope. EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 1. Optimal codeword lengths. Although the codeword lengths of an optimal variable length code
More information16.1 Bounding Capacity with Covering Number
ECE598: Information-theoretic methods in high-dimensional statistics Spring 206 Lecture 6: Upper Bounds for Density Estimation Lecturer: Yihong Wu Scribe: Yang Zhang, Apr, 206 So far we have been mostly
More informationNetwork Coding for Computing
Networ Coding for Computing Rathinaumar Appuswamy, Massimo Franceschetti, Nihil Karamchandani, and Kenneth Zeger Abstract The following networ computation problem is considered A set of source nodes in
More informationGaussian Estimation under Attack Uncertainty
Gaussian Estimation under Attack Uncertainty Tara Javidi Yonatan Kaspi Himanshu Tyagi Abstract We consider the estimation of a standard Gaussian random variable under an observation attack where an adversary
More informationON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria
ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA Thomas Adidaume Uge and Stephen Seastian Akpan, Department Of Mathematics/Statistics And Computer
More information