State Variation Mining: On Information Divergence with Message Importance in Big Data
|
|
- James McCarthy
- 6 years ago
- Views:
Transcription
1 State Variation Mining: On Information Divergence with Message Importance in Big Data Rui She, Shanyun Liu, and Pingyi Fan State Key Laboratory on Microwave and Digital Communications Tsinghua National Laboratory for Information Science and Technology Department of Electronic Engineering, Tsinghua University, Beijing, P.R. China s: {sher15, arxiv: v2 [cs.it] 15 Jan 2018 Abstract Information transfer which reveals the state variation of variables can play a vital role in big data analytics and processing. In fact, the measure for information transfer can reflect the system change from the statistics by using the variable distributions, similar to KL divergence and Renyi divergence. Furthermore, in terms of the information transfer in big data, small probability events dominate the importance of the total message to some degree. Therefore, it is significant to design an information transfer measure based on the message importance which emphasizes the small probability events. In this paper, we propose the message importance divergence (MID) and investigate its characteristics and applications on three aspects. First, the message importance transfer capacity based on MID is presented to offer an upper bound for the information transfer with disturbance. Then, we utilize the MID to guide the queue length selection, which is the fundamental problem considered to have higher social or academic value in the caching operation of mobile edge computing. Finally, we extend the MID to the continuous case and discuss the robustness by using it to measuring information distance. I. INTRODUCTION Recently, the amount of data is exploding rapidly and the computing complexity for data processing is also increasing. To some degree, this phenomenon is resulted from more and more mobile devices as well as the growing service of clouds. In the literature, it is favored to process the collected data to dig out the hidden important information. On the one hand, it is necessary to improve the computation technologies for big data processing, such as cloud computing, fog computing and mobile edge computing (MEC). On the other hand, a series of technologies for big data analysis and mining are required, such as neural networks and machine learning, as well as distributed parallel computing, etc. In many scenarios of big data, the small probability events attract more attention than the large probability ones [1]. That is, the rarity of small probability events has higher social or academic value. Message importance is used to reflect such kind of values in applications. How to characterize the message importance variation in the processing of big data is a critical and interesting problem. Here, we shall investigate and measure information transfer process by use of the message importance, focusing on small probability events. Before we proceed, let us review the definition of message importance measure (MIM) briefly first [2]. In a finite alphabet, for a given probability distribution P = {p(x 1 ),p(x 2 ),...,p(x n )}, the MIM with importance coefficient 0 is defined as L(P, ) = log { x i p(x i )e (1 p(xi)) }, (1) which measures the information importance of the distribution. Then, by setting the parameter = 1 and simplifying the form of MIM, it is easy to obtain its fundamental definition as follows. Definition 1. For the discrete probability P ={p(x 1 ), p(x 2 ),...,p(x n )}, the MIM can be given by L(P) = x i p(x i )e p(xi). (2) On account of the Definition 1, it is available to regard the MIM as an element to measure the message importance variation for a dynamic system. Then, new information transfer measures based on the MIM are defined as follows. Definition 2. In terms of two discrete probability Q = {q(x 1 ),q(x 2 ),...,q(x n )} and P = {p(x 1 ),p(x 2 ),...,p(x n )}, the message importance divergence (MID) is defined as (Q P) = x i {q(x i )e q(xi) p(x i )e p(xi) }. (3) Note that the Definition 2 characterizes the information transfer from the statistics. That is, we can make use of MID to measure the end-to-end change for an information transfer process. In fact, there exist a variety of different information measures handling the problem of information transfer process. Shannon entropy and Renyi entropy are applicable to intrinsic dimension estimation [3]. As well, the NMIM and M-I divergence can be used in anomaly detection [4], [5]. Moreover, the directed information [6] and Schreibers transfer entropy [7] are commonly applied to inferring the causality structure and characterizing the information transfer process. In addition, referring to the idea from dynamical system theory, new information transfer measures are proposed to explore and exploit the causality between states in the system control [8]. However, in spite of numerous kinds of information measures, few works focus on how to characterize the information
2 transfer from the perspective of message importance in big data. To this end, the MID different from the above information measures is introduced. We organize the rest of this paper as follows. In Section II, we introduce the message importance transfer capacity measured by the MID to describe the information transfer with disturbance. In Section III, the MID is used to discuss the queue length selection for the data caching in MEC from the viewpoint of queue theory. Moreover, some simulations are presented to validate our theoretical results. In Section IV, we extend the MID to the continuous case to investigate the correlation of message importance in the information transfer process. Finally, we conclude it in Section VI. II. MESSAGE IMPORTANCE TRANSFER CAPACITY BASED ON MESSAGE IMPORTANCE DIVERGENCE In this section, we will introduce the MID to describe the information transfer with additive disturbance. To do so, we define the message importance transfer capacity measured by the MID as follows. Definition 3. Assume that there exists an information transfer process as, {X,p(y x),y}, (4) where the p(y x) denotes a probability distribution matrix describing the information transfer from the variable X following distribution p(x) to the Y following p(y). We define the message importance transfer capacity as C = max p(x) {L(Y) L(Y X)}, (5) where p(y j ) = x i p(x i )p(y j x i ), L(Y) = y j p(y j )e p(yj), L(Y X) = y j x i p(x i,y j )e p(yj xi). In order to have an insight into the applications of message importance transfer capacity, some specific information transfer scenarios are discussed as follows. A. Binary symmetric information transfer Proposition 1. Assume that there exists an information transfer process {X, p(y x), Y}, where the information transfer matrix is [ ] 1 p(y x) =, (6) 1 which indicates that variables X and Y both obey binary distributions. In this case, we have C() = e 1 2 L(), (7) where L() = e +(1 )e (1 ) and 0 < < 1. Proof. Considering a variable X following binary distribution (p,1 p), it is not difficult to see that L(Y X) = e +(1 )e (1 ). (8) Moreover, according to Eq. (5), we have C(p,) = max p { [p+(1 2p)]e [p+(1 2p)] +[(1 p)+(2p 1)]e [(1 p)+(2p 1)]} L(). Then, it is readily seen that C(p, ) p { = (1 2) [1 p (1 2p)]e [p+(1 2p)] [1 (1 p) ε(2p 1)]e [(1 p)+(2p 1)]}. (10) In the light of the monotonically decreasing of C(p,) p for p [0,1], it is apparent that p = 1/2 is the only solution for = 0. Therefore, the proposition can be testified. C(p,) p B. Strongly symmetric information transfer Proposition 2. Assume that there exists a strongly symmetric information transfer matrix as follows 1... p(y x) = , (11) 1... which indicates that variables X and Y both follow K- ary distributions. We have the message importance transfer capacity as (9) C() = e 1 K {(1 )e (1 ) +e }, (12) where the parameter (0,1). Proof. On account of the information transfer matrix and the Eq. (2), we have L(Y X) = e +(1 )e (1 ). (13) Then, similar to the proof of Proposition 1, we can also use Lagrange multiplier method to obtain the message information transfer capacity. In this case, the distribution of Y should satisfy p(y 1 ) = p(y 2 ) =... = p(y K ) = 1/K. In addition, consider that the probability distribution of variablex is {p(x 1 ),p(x 2 ),...,p(x K )}. In the strongly symmetric transfer matrix, if the variable X follows uniform distribution, namely p(x 1 ) = p(x 2 ) =... = p(x K ) = 1/K, we will have K K p(y j ) = p(x i,y j ) = p(x i )p(y j x i ) = 1 K (14) K p(y j x i ) = 1 K, which indicates that Y also follows the uniform distribution. Therefore, it is testified that when the variable X follows uniform distribution which leads to the uniform distribution for variable Y, we will obtain the message importance transfer capacity C().
3 III. APPLICATION IN MOBILE EDGE COMPUTING WITH THE M/M/S/K QUEUE Consider the MEC system that consists of numerous mobile devices, an edge server cluster, and a central cloud far from the local devices. The queue models for mobile devices, the edge server cluster and the central cloud are regarded as a M/M/1 queue, a M/M/s queue, and a M/M/ queue, respectively. In particular, each mobile device can offload a part of or the entire service requests to the corresponding edge server cluster when the communication is disturbed. As well, the overloaded requests from edge servers will be offloaded to the central cloud for processing [10]. In terms of the MID, it can be applied to distinguish the state probability distributions in a queue model such as M/M/s queue. In particular, the state probability is derived from the stationary process, namely a dynamic equilibrium based on birth and death process. Then, we can obtain the steady queue state probability in the M/M/s/k model. By use of Taylor series expansion, the approximate MIM is given by s+k s+k p j e pj = p j [1 p j +O(p 2 j )]. = 1 p 2 0 { s 1 ( aj j! )2 + (as /s!) 2 [1 ρ 2(k+1) } ] 1 ρ 2, (15) where λ and µ are arrival rate and service rate,sis the number of servers, k is the buffer or caching size, the traffic intensity ρ = a/s as well as a = λ/ µ. Note that, in the M/M/s/k queue, the queue state probability p j (j = 0,...,s+k) are given by p 0 = [ s 1 a j ] 1, j! + as s! 1 ρk+1 (16a) 1 ρ p j = aj j! p 0, (0 < j < s), (16b) p j = as s! ρj s p 0, (s j s+k). (16c) Then, referring to Eq. (15), we can obtain the MID to characterize the message importance gap for the queue model M/M/s as follows. Proposition 3. As for the M/M/s model, the information difference between two queue state probability distributions P k+1 and P k which are with buffer size k and k + 1 respectively, can be measured by MID as (P k+1 P k ) = s+k+1 s+k p j e pj p j e pj. 1 = ( (ϕ 1 +ϕ ρ )2 1 ρ (ϕ 1 +ϕ k ϕ2 2 1 ρ 2[ 1 ρ 2(k+1) (ϕ 1 +ϕ 2 1 ρ )2 s 1 1 ρ )2) 1 ρ2(k+2) ( aj j! )2 (ϕ 1 +ϕ 2 1 ρ k+2 1 ρ )2]. (17) where p j and p j are queue state probability in the M/M/s/k and M/M/s/k+1 models respectively, as well as parameter ϕ 1 and ϕ 2 are given by ϕ 1 = s 1 aj /j! and ϕ 2 = a s /s!. Similarly, it is not difficult to derive the MID between the queue state probability distributions with buffer size and k, which is given by (P P k ) 1 = ( (ϕ 1 +ϕ ρ )2 (ϕ 1 + ϕ2 + ϕ2 2 1 ρ 2(k+1) s 1 1 ρ )2) ( aj j! )2 1 ρ 2[ (ϕ 1 +ϕ ρ )2 (ϕ 1 + ϕ2 1 ρ )2]. (18) Especially, when the number of server is one, that is, the queue model is M/M/1/k model, it it readily seen that (s=1) (P P k ) = 2a 1+a 2a(1 ak+1 ) (1+a)(1 a k+2 ), (19) where (s=1) (P P k ) denotes the MID with s = 1. Moreover, for the queue length selection, it is required that the distinction between two distribution P and P k should be small enough, namely, (s=1) (P P k ) ǫ. Then we have ǫ(1+a) a(2+ǫ (2 ǫ)a) k ln 1. (20) lna It is easy to see that ǫ plays a key role in the caching size selection when using finite size caching to imitate the infinite caching working mode. Similar to MID, the KL divergence between the queue state probability distributions with buffer size k+1 and k is given by D(P k P k+1 ) = j = log p j log 1 p j j p j log 1 p j {1+ ϕ 2 (1 ρ)ρ k+1 }, ϕ 1(1 ρ) +ϕ 2 (21) where the parameters p j, p j, ϕ 1 and ϕ 2 are the same as them in Proposition 3. Likewise, we can derive the KL divergence between the queue state distributions with buffer size and k as D(P k P ) = log ϕ 1 1 +ϕ 2 1 ρ. (22) 1 ρ ϕ 1 +ϕ k ρ For the queue length selection with KL divergence, it is readily seen that k ln(1 1 2 ) ǫ 2. (23) lna To validate our derived results in theory, some simulations are presented. The events arrivals are listed in Table I. It is readily seen that they have the same average interarrival time as 1/ λ 0. Besides, the traffic intensity is selected as ρ = 0.9 in all cases.
4 Type of Imfornation measure value TABLE I THE DISTRIBUTIONS OF EVENT INTERARRIVAL TIME Exponential Uniform Normal P(X) X E( λ 0 ) X U(0,2/ λ 0 ) X N( λ 1, Queue length 1 ) λ 2 0 -Sim-Poisson -Sim-Uniform -Sim-Normal D-Sim-Poisson D-Sim-Uniform D-Sim-Normal -Ana-Poisson D-Ana-Poisson Fig. 1. The performance of different information measures between the queue length k and k+1 for the queue model in the case of different arrival events distributions. In Fig. 1, the legends -Sim, -Ana and D-Sim, D-Ana are used to denote the simulation results and the analytical results for MID and KL divergence, respectively. It is illustrated that the convergence of MID is faster than that of KL divergence, which indicates that MID may provide a reasonable lower bound to select the caching size for MEC. In addition, we can see that the Poisson distribution corresponds the worst case for the arrival process among the three discussed cases with respect to the convergence of both MID and KL divergence. IV. MESSAGE IMPORTANCE DIVERGENCE IN CONTINUOUS CASES Similar to the definition 1 and 2, we can extend the two definition to the case with continuous distributions as L(f(x)) = f(x)e f(x) dx, x, (24) (g(x) f(x)) = L(g(x)) L(f(x)) = g(x)e g(x) f(x)e f(x) dx,x, (25) where g(x) and f(x) are two probability distributions with respect to the variable X in a given interval. Moreover, L(f(x)) and (g(x) f(x)) can be called differential message importance measure (DMIM) and differential message importance divergence (DMID). Then, we investigate the correlation of message importance by using DMID, which can also reflect the robustness of DMID. Consider the observation model, P g0 f 0 : f 0 (x) g 0 (x), that denotes an information transfer map for the variable X from the probability distribution f 0 (x) to g 0 (x). By using the similar way in [9], the relationship between f 0 (x) and g 0 (x) can be described as g 0 (x) = f 0 (x)+ǫf0 α (x)u(x), (26) and the constraint condition satisfies ǫf0 α (x)u(x)dx = 0, (27) where ǫ and α are adjustable coefficients. u(x) is a perturbation function of the variable X in the interval. Then, by using the above model, the end-to-end information distance measured by DMID is given as follows. Proposition 4. For two probability distributions g 0 (x) and f 0 (x) whose relationship satisfies the conditions Eq. (26) and Eq. (27), the information distance measured by DMID is given by (g 0 (x) f 0 (x)) { = g 0 (x)e g0(x) f 0 (x)e f0(x)} dx = ǫ i! (i 1)! f0 i+α (x)u(x)dx + ǫ2 2 (28) where ǫ and α denote parameters, u(x) is a function of the variable X in the interval. f0 i 1+2α (x)u 2 (x)dx+o(ǫ 2 ), Proposition 4 describes the perturbation between f 0 (x) and g 0 (x). Furthermore, it is not difficult to obtain the DMID between two different distributions g (u) 1 and g (u) 2 based on the same reference distribution f 0 (x), which is given by (g (u) 1 (x) g(u) 2 (x)) = [L(g (u) 1 (x)) L(f 0(x))] [L(g (u) 2 (x)) L(f 0(x))] = ǫ + ǫ2 2 +o(ǫ 2 ), where i! (i 1)! f0 i+α (x)[u 1 (x) u 2 (x)]dx f0 i 1+2α (x)[u 2 1 (x) u2 2 (x)]dx (29) g (u) 1 (x) = f 0(x)+ǫf α 0 (x)u 1(x), x, (30a) g (u) 2 (x) = f 0(x)+ǫf α 0 (x)u 2(x), x, (30b) in which the ǫ and α are parameters, u 1 (x) and u 2 (x) are functions of the variable X in the interval.
5 Remark 1. It is apparent that when the parameter ǫ is small enough, the DMID of two distributions is convergent with the order of O(ǫ). Actually, this provides a way to apply DMID to measure the correlation of message importance, if the system does not have relatively large change. In addition, by extending the information transfer process described as Eq. (4) to the case with continuous distributions, it is also significant to clarify the effect of the additive disturbance on the message importance transfer capacity as follows. Theorem 1. Assume that there exists an information transfer process between the variable X and Y both following continuous distributoins, denoted by {X, p(y x), Y}, where E[X] = 0, E[X 2 ] = P s, Y = X + Z, and Z denotes independent memoryless additive variable, namely a continuous additive disturbance, whose mean and variance satisfy that E[Z] = µ and E[(Z µ) 2 ] = σ 2, respectively. Then, the message importance transfer capacity is determined by the variances of the variable X and Z as C(P s ) = max p(x) (Y Z) = max {L(Y)} L(Z) p(x) { + } = max p(y)e p(y) dy p(x) ( 1) j j! s.t. E[Y 2 ] = P s +P N, p j+1 (z)dz, (31a) (31b) where P N = µ 2 +σ 2, L( ) is the DMIM operator, and p(y) = p(x)p(y x)dx. Proof. According to Eq. (4), we have { x(x,y) = x z(x,y) = y x. (32) Then, it is not difficult to see that p(x,y) = p(x,z) J( xz x x x y ) = p(x,z) z z xy x y = p(x,z). (33) Moreover, by virtue of the independence of X and Z, we have p(x)p(y x) = p(x)p(z), which indicates that p(y x) = p(z). Then, we have L(Y X) = p(x)p(y x)e p(y x) dxdy x y = p(x)p(z)e p(z) dxdz (34) x z = p(x){ p(z)e p(z) dz}dx = L(Z). x z Consequently, in terms of the Definition 3, it is readily seen that L(Y) L(Y X) can be written as L(Y) L(Z), which testifies Eq. (31a). Furthermore, according to the fact that E[Y 2 ] = E[(X + Z) 2 ] = E[X 2 ] +E[Z 2 ] = P s + P N, we have the constraint condition Eq. (31b). As well, by substituting the definition of MID into Eq. (31a), the Theorem 1 is proved. In practice, the variance can be regarded as the power of signals. Therefore, the message importance transfer capacity in the Theorem 1 may be applied to guide the signal transfer processing with an additive disturbance. V. CONCLUSION In this paper, we investigated the information transfer problem in big data and proposed an information measure, i.e., MID. Furthermore, this information measure has its own dramatic characteristics on paying more attention to the message importance hidden in big data. This makes the information measure as a promising tool for information transfer measure in big data. We presented the message importance transfer capacity measured by the MID which can give an upper bound for the information transfer with disturbance. Furthermore, we employed the MID to discuss the caching size selection in MEC. In addition, the MID was extended to the continuous case to investigate the correlation of message importance in the information transfer process. ACKNOWLEDGMENT We indeed appreciate the support of the China Major State Basic Research Development Program (973 Program) No.2012CB316100(2), and the National Natural Science Foundation of China (NSFC) No REFERENCES [1] Y. Bu, S. Zou, and V. V. Veeravalli, Linear-complexity exponentiallyconsistent tests for universal outlying sequence detection, in Proc IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, June. 2017, pp [2] P. Fan, Y. Dong, J. Lu, and S. Liu, Message importance measure and its application to minority subset detection in big data, in Proc. IEEE Globecom Workshops (GC Wkshps), Washington D.C., USA, Dec. 2016, pp 1 5. [3] K. M. Carter, R. Raich, and A. O. Hero, On local intrinsic dimension estimation and its applications, IEEE Trans. Signal Process., vol. 58, no. 2, pp , Feb [4] S. Liu, R. She, P. Fan, and K. B. Letaief. (2017). Non-parametric message important measure: storage code design and transmission planning for big data. [Online]. Available: [5] R. She, S. Y. Liu, and P. Fan, Amplifying inter-message distance: on information divergence measures in big data, IEEE Access, vol. 5, pp , Nov [6] L. Zhao, Y. H. Kim, H. H. Permuter, and T. Weissman, Universal estimation of directed information, in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Austin, USA, June. 2010, pp [7] T. Schreiber, Measuring information transfer, Physical Review Letters, vol. 85, no. 2, pp , July, [8] S. Sinha and U. Vaidya, Causality preserving information transfer measure for control dynamical system, in Proc. IEEE 55th Conference on Decision and Control (CDC), Las Vegas, USA, Dec. 2016, pp [9] S. Huang, A. Makur, L. Zheng, and G. W. Wornell, An informationtheoretic approach to universal feature selection in high-dimensional inference, in Proc IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, June. 2017, pp
6 [10] L. Liu, Z. Chang, X. Guo, and T. Ristaniemi, Multi-objective optimization for computation offloading in mobile-edge computing, In Proc. IEEE Symposium on Computers and Communications (ISCC), Heraklion, Greece, July. 2017, pp
Minor probability events detection in big data: An integrated approach with Bayesian testing and MIM
Minor probability events detection in big data: An integrated approach with Bayesian testing and MIM Shuo Wan, Jiaxun Lu, Pingyi Fan, and Khaled B. Letaief*, Tsinghua National Laboratory for Information
More informationFocusing on a Probability Element: Parameter Selection of Message Importance Measure in Big Data
Focusing on a Probability Element: Parameter Selection of Message Importance Measure in Big Data arxiv:1701.03234v2 [cs.it] 11 Nov 2018 Rui She, Shanyun Liu, Yunquan Dong, and Pingyi Fan, SeniorMember,
More informationPerformance Modelling of Computer Systems
Performance Modelling of Computer Systems Mirco Tribastone Institut für Informatik Ludwig-Maximilians-Universität München Fundamentals of Queueing Theory Tribastone (IFI LMU) Performance Modelling of Computer
More informationAge of Information Upon Decisions
Age of Information Upon Decisions Yunquan Dong, Zhengchuan Chen, Shanyun Liu, and Pingyi Fan School of Electronic & Information Engineering, Nanjing University of Information Science & Technology, Nanjing,
More informationFeedback Capacity of a Class of Symmetric Finite-State Markov Channels
Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Nevroz Şen, Fady Alajaji and Serdar Yüksel Department of Mathematics and Statistics Queen s University Kingston, ON K7L 3N6, Canada
More informationBIRTH DEATH PROCESSES AND QUEUEING SYSTEMS
BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS Andrea Bobbio Anno Accademico 999-2000 Queueing Systems 2 Notation for Queueing Systems /λ mean time between arrivals S = /µ ρ = λ/µ N mean service time traffic
More informationPerformance Evaluation of Queuing Systems
Performance Evaluation of Queuing Systems Introduction to Queuing Systems System Performance Measures & Little s Law Equilibrium Solution of Birth-Death Processes Analysis of Single-Station Queuing Systems
More informationQueueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "
Queueing Theory I Summary Little s Law Queueing System Notation Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K " Little s Law a(t): the process that counts the number of arrivals
More informationAn Extended Fano s Inequality for the Finite Blocklength Coding
An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.
More informationMath 240 Calculus III
Calculus III Summer 2015, Session II Monday, August 3, 2015 Agenda 1. 2. Introduction The reduction of technique, which applies to second- linear differential equations, allows us to go beyond equations
More informationContinuous Time Processes
page 102 Chapter 7 Continuous Time Processes 7.1 Introduction In a continuous time stochastic process (with discrete state space), a change of state can occur at any time instant. The associated point
More informationCan Feedback Increase the Capacity of the Energy Harvesting Channel?
Can Feedback Increase the Capacity of the Energy Harvesting Channel? Dor Shaviv EE Dept., Stanford University shaviv@stanford.edu Ayfer Özgür EE Dept., Stanford University aozgur@stanford.edu Haim Permuter
More informationThe Least Degraded and the Least Upgraded Channel with respect to a Channel Family
The Least Degraded and the Least Upgraded Channel with respect to a Channel Family Wei Liu, S. Hamed Hassani, and Rüdiger Urbanke School of Computer and Communication Sciences EPFL, Switzerland Email:
More informationIntroduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.
L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission
More informationA Modification of Linfoot s Informational Correlation Coefficient
Austrian Journal of Statistics April 07, Volume 46, 99 05. AJS http://www.ajs.or.at/ doi:0.773/ajs.v46i3-4.675 A Modification of Linfoot s Informational Correlation Coefficient Georgy Shevlyakov Peter
More informationOptimization in Information Theory
Optimization in Information Theory Dawei Shen November 11, 2005 Abstract This tutorial introduces the application of optimization techniques in information theory. We revisit channel capacity problem from
More informationOptimal Sleeping Mechanism for Multiple Servers with MMPP-Based Bursty Traffic Arrival
1 Optimal Sleeping Mechanism for Multiple Servers with MMPP-Based Bursty Traffic Arrival Zhiyuan Jiang, Bhaskar Krishnamachari, Sheng Zhou, arxiv:1711.07912v1 [cs.it] 21 Nov 2017 Zhisheng Niu, Fellow,
More informationModule 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur
Module ntroduction to Digital Communications and nformation Theory Lesson 3 nformation Theoretic Approach to Digital Communications After reading this lesson, you will learn about Scope of nformation Theory
More informationCover Page. The handle holds various files of this Leiden University dissertation
Cover Page The handle http://hdl.handle.net/1887/39637 holds various files of this Leiden University dissertation Author: Smit, Laurens Title: Steady-state analysis of large scale systems : the successive
More informationSoft Covering with High Probability
Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information
More informationStatistical Learning Theory
Statistical Learning Theory Part I : Mathematical Learning Theory (1-8) By Sumio Watanabe, Evaluation : Report Part II : Information Statistical Mechanics (9-15) By Yoshiyuki Kabashima, Evaluation : Report
More informationInformation Theory. Coding and Information Theory. Information Theory Textbooks. Entropy
Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is
More informationQueueing Theory. VK Room: M Last updated: October 17, 2013.
Queueing Theory VK Room: M1.30 knightva@cf.ac.uk www.vincent-knight.com Last updated: October 17, 2013. 1 / 63 Overview Description of Queueing Processes The Single Server Markovian Queue Multi Server
More informationCPSC 531: System Modeling and Simulation. Carey Williamson Department of Computer Science University of Calgary Fall 2017
CPSC 531: System Modeling and Simulation Carey Williamson Department of Computer Science University of Calgary Fall 2017 Motivating Quote for Queueing Models Good things come to those who wait - poet/writer
More informationLink Models for Packet Switching
Link Models for Packet Switching To begin our study of the performance of communications networks, we will study a model of a single link in a message switched network. The important feature of this model
More informationQueueing Theory and Simulation. Introduction
Queueing Theory and Simulation Based on the slides of Dr. Dharma P. Agrawal, University of Cincinnati and Dr. Hiroyuki Ohsaki Graduate School of Information Science & Technology, Osaka University, Japan
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality
More informationAn Outer Bound for the Gaussian. Interference channel with a relay.
An Outer Bound for the Gaussian Interference Channel with a Relay Ivana Marić Stanford University Stanford, CA ivanam@wsl.stanford.edu Ron Dabora Ben-Gurion University Be er-sheva, Israel ron@ee.bgu.ac.il
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationCapacity of the Discrete Memoryless Energy Harvesting Channel with Side Information
204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener
More informationQueuing Approaches to Principal-Agent Communication under Information Overload
Queuing Approaches to Principal-Agent Communication under Information Overload Aseem Sharma, Krishna Jagannathan, Member, IEEE, and Lav R. Varshney, Senior Member, IEEE arxiv:1605.00802v1 [cs.it] 3 May
More informationLatent Tree Approximation in Linear Model
Latent Tree Approximation in Linear Model Navid Tafaghodi Khajavi Dept. of Electrical Engineering, University of Hawaii, Honolulu, HI 96822 Email: navidt@hawaii.edu ariv:1710.01838v1 [cs.it] 5 Oct 2017
More informationBRAVO for QED Queues
1 BRAVO for QED Queues Yoni Nazarathy, The University of Queensland Joint work with Daryl J. Daley, The University of Melbourne, Johan van Leeuwaarden, EURANDOM, Eindhoven University of Technology. Applied
More information2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES
295 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES 16 Queueing Systems with Two Types of Customers In this section, we discuss queueing systems with two types of customers.
More informationStochastic Models in Computer Science A Tutorial
Stochastic Models in Computer Science A Tutorial Dr. Snehanshu Saha Department of Computer Science PESIT BSC, Bengaluru WCI 2015 - August 10 to August 13 1 Introduction 2 Random Variable 3 Introduction
More informationOn the static assignment to parallel servers
On the static assignment to parallel servers Ger Koole Vrije Universiteit Faculty of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The Netherlands Email: koole@cs.vu.nl, Url: www.cs.vu.nl/
More informationEnergy State Amplification in an Energy Harvesting Communication System
Energy State Amplification in an Energy Harvesting Communication System Omur Ozel Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland College Park, MD 20742 omur@umd.edu
More informationOn the Capacity of the Interference Channel with a Relay
On the Capacity of the Interference Channel with a Relay Ivana Marić, Ron Dabora and Andrea Goldsmith Stanford University, Stanford, CA {ivanam,ron,andrea}@wsl.stanford.edu Abstract Capacity gains due
More informationArimoto Channel Coding Converse and Rényi Divergence
Arimoto Channel Coding Converse and Rényi Divergence Yury Polyanskiy and Sergio Verdú Abstract Arimoto proved a non-asymptotic upper bound on the probability of successful decoding achievable by any code
More informationPractical Polar Code Construction Using Generalised Generator Matrices
Practical Polar Code Construction Using Generalised Generator Matrices Berksan Serbetci and Ali E. Pusane Department of Electrical and Electronics Engineering Bogazici University Istanbul, Turkey E-mail:
More information2905 Queueing Theory and Simulation PART IV: SIMULATION
2905 Queueing Theory and Simulation PART IV: SIMULATION 22 Random Numbers A fundamental step in a simulation study is the generation of random numbers, where a random number represents the value of a random
More informationOutline. Finite source queue M/M/c//K Queues with impatience (balking, reneging, jockeying, retrial) Transient behavior Advanced Queue.
Outline Finite source queue M/M/c//K Queues with impatience (balking, reneging, jockeying, retrial) Transient behavior Advanced Queue Batch queue Bulk input queue M [X] /M/1 Bulk service queue M/M [Y]
More informationInformation Measure Estimation and Applications: Boosting the Effective Sample Size from n to n ln n
Information Measure Estimation and Applications: Boosting the Effective Sample Size from n to n ln n Jiantao Jiao (Stanford EE) Joint work with: Kartik Venkat Yanjun Han Tsachy Weissman Stanford EE Tsinghua
More informationTaylor series. Chapter Introduction From geometric series to Taylor polynomials
Chapter 2 Taylor series 2. Introduction The topic of this chapter is find approximations of functions in terms of power series, also called Taylor series. Such series can be described informally as infinite
More informationFigure 10.1: Recording when the event E occurs
10 Poisson Processes Let T R be an interval. A family of random variables {X(t) ; t T} is called a continuous time stochastic process. We often consider T = [0, 1] and T = [0, ). As X(t) is a random variable
More informationIntroduction to queuing theory
Introduction to queuing theory Queu(e)ing theory Queu(e)ing theory is the branch of mathematics devoted to how objects (packets in a network, people in a bank, processes in a CPU etc etc) join and leave
More informationHomework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015
10-704 Homework 1 Due: Thursday 2/5/2015 Instructions: Turn in your homework in class on Thursday 2/5/2015 1. Information Theory Basics and Inequalities C&T 2.47, 2.29 (a) A deck of n cards in order 1,
More informationLecture 22: Variance and Covariance
EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce
More informationEnergy-Efficient Resource Allocation for Multi-User Mobile Edge Computing
Energy-Efficient Resource Allocation for Multi-User Mobile Edge Computing Junfeng Guo, Zhaozhe Song, Ying Cui Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, P. R. China
More informationChapter 2 Review of Classical Information Theory
Chapter 2 Review of Classical Information Theory Abstract This chapter presents a review of the classical information theory which plays a crucial role in this thesis. We introduce the various types of
More informationLatent Variable Models and EM algorithm
Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic
More informationTight Bounds on Minimum Maximum Pointwise Redundancy
Tight Bounds on Minimum Maximum Pointwise Redundancy Michael B. Baer vlnks Mountain View, CA 94041-2803, USA Email:.calbear@ 1eee.org Abstract This paper presents new lower and upper bounds for the optimal
More informationInformation in Biology
Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve
More informationLower Bounds on the Graphical Complexity of Finite-Length LDPC Codes
Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel 2009 IEEE International
More informationQuantifying Stochastic Model Errors via Robust Optimization
Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations
More informationSuperposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels
Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:
More informationAP Calculus Chapter 9: Infinite Series
AP Calculus Chapter 9: Infinite Series 9. Sequences a, a 2, a 3, a 4, a 5,... Sequence: A function whose domain is the set of positive integers n = 2 3 4 a n = a a 2 a 3 a 4 terms of the sequence Begin
More informationConvexity/Concavity of Renyi Entropy and α-mutual Information
Convexity/Concavity of Renyi Entropy and -Mutual Information Siu-Wai Ho Institute for Telecommunications Research University of South Australia Adelaide, SA 5095, Australia Email: siuwai.ho@unisa.edu.au
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationSimple Channel Coding Bounds
ISIT 2009, Seoul, Korea, June 28 - July 3,2009 Simple Channel Coding Bounds Ligong Wang Signal and Information Processing Laboratory wang@isi.ee.ethz.ch Roger Colbeck Institute for Theoretical Physics,
More informationInformation in Biology
Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1 Living
More informationThe Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels
The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels Parastoo Sadeghi National ICT Australia (NICTA) Sydney NSW 252 Australia Email: parastoo@student.unsw.edu.au Predrag
More informationBioinformatics: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Bayesian Interpretation of Probabilities 2 Where (or of what)
More informationStochastic Models of Manufacturing Systems
Stochastic Models of Manufacturing Systems Ivo Adan Organization 2/47 7 lectures (lecture of May 12 is canceled) Studyguide available (with notes, slides, assignments, references), see http://www.win.tue.nl/
More informationChapter 5 Joint Probability Distributions
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 5 Joint Probability Distributions 5 Joint Probability Distributions CHAPTER OUTLINE 5-1 Two
More informationIntroduction to Markov Chains, Queuing Theory, and Network Performance
Introduction to Markov Chains, Queuing Theory, and Network Performance Marceau Coupechoux Telecom ParisTech, departement Informatique et Réseaux marceau.coupechoux@telecom-paristech.fr IT.2403 Modélisation
More informationSlepian-Wolf Code Design via Source-Channel Correspondence
Slepian-Wolf Code Design via Source-Channel Correspondence Jun Chen University of Illinois at Urbana-Champaign Urbana, IL 61801, USA Email: junchen@ifpuiucedu Dake He IBM T J Watson Research Center Yorktown
More informationMultiaccess Problem. How to let distributed users (efficiently) share a single broadcast channel? How to form a queue for distributed users?
Multiaccess Problem How to let distributed users (efficiently) share a single broadcast channel? How to form a queue for distributed users? z The protocols we used to solve this multiaccess problem are
More informationExercises Solutions. Automation IEA, LTH. Chapter 2 Manufacturing and process systems. Chapter 5 Discrete manufacturing problems
Exercises Solutions Note, that we have not formulated the answers for all the review questions. You will find the answers for many questions by reading and reflecting about the text in the book. Chapter
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationLink Models for Circuit Switching
Link Models for Circuit Switching The basis of traffic engineering for telecommunication networks is the Erlang loss function. It basically allows us to determine the amount of telephone traffic that can
More informationOne-Bit LDPC Message Passing Decoding Based on Maximization of Mutual Information
One-Bit LDPC Message Passing Decoding Based on Maximization of Mutual Information ZOU Sheng and Brian M. Kurkoski kurkoski@ice.uec.ac.jp University of Electro-Communications Tokyo, Japan University of
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationComputation of Information Rates from Finite-State Source/Channel Models
Allerton 2002 Computation of Information Rates from Finite-State Source/Channel Models Dieter Arnold arnold@isi.ee.ethz.ch Hans-Andrea Loeliger loeliger@isi.ee.ethz.ch Pascal O. Vontobel vontobel@isi.ee.ethz.ch
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationSTABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL
First published in Journal of Applied Probability 43(1) c 2006 Applied Probability Trust STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL LASSE LESKELÄ, Helsinki
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationSynchronized Queues with Deterministic Arrivals
Synchronized Queues with Deterministic Arrivals Dimitra Pinotsi and Michael A. Zazanis Department of Statistics Athens University of Economics and Business 76 Patission str., Athens 14 34, Greece Abstract
More informationStat 100a, Introduction to Probability.
Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous
More informationLecture 20: Reversible Processes and Queues
Lecture 20: Reversible Processes and Queues 1 Examples of reversible processes 11 Birth-death processes We define two non-negative sequences birth and death rates denoted by {λ n : n N 0 } and {µ n : n
More informationReducing Age-of-Information for Computation-Intensive Messages via Packet Replacement
Reducing Age-of-Information for Computation-Intensive Messages via Packet Replacement arxiv:191.4654v1 [cs.it] 15 Jan 19 Jie Gong, Qiaobin Kuang, Xiang Chen and Xiao Ma School of Data and Computer Science,
More informationSTAT Chapter 5 Continuous Distributions
STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationL 2 -induced Gains of Switched Systems and Classes of Switching Signals
L 2 -induced Gains of Switched Systems and Classes of Switching Signals Kenji Hirata and João P. Hespanha Abstract This paper addresses the L 2-induced gain analysis for switched linear systems. We exploit
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationThe Behavior of a Multichannel Queueing System under Three Queue Disciplines
The Behavior of a Multichannel Queueing System under Three Queue Disciplines John K Karlof John Jenkins November 11, 2002 Abstract In this paper we investigate a multichannel channel queueing system where
More informationDivergence measure of intuitionistic fuzzy sets
Divergence measure of intuitionistic fuzzy sets Fuyuan Xiao a, a School of Computer and Information Science, Southwest University, Chongqing, 400715, China Abstract As a generation of fuzzy sets, the intuitionistic
More informationCitation PHYSICAL REVIEW E (2002), 65(6) RightCopyright 2002 American Physical So
TitleStatistical mechanics of two hard d Author(s) Munakata, T; Hu, G Citation PHYSICAL REVIEW E (2002), 65(6) Issue Date 2002-06 URL http://hdl.handle.net/2433/50310 RightCopyright 2002 American Physical
More informationSTEADY-STATE PROBABILITIES FOR SOME QUEUEING SYSTEMS: Part II
STEADY-STATE PROBABILITIES FOR SOME QUEUEING SYSTEMS: Part II Mahmoud Abdel Moety Hussain Faculty of Computers and Informatics Zagazig University Abstract This paper applies the theoretical results presented
More information6 Solving Queueing Models
6 Solving Queueing Models 6.1 Introduction In this note we look at the solution of systems of queues, starting with simple isolated queues. The benefits of using predefined, easily classified queues will
More informationA New Technique for Link Utilization Estimation
A New Technique for Link Utilization Estimation in Packet Data Networks using SNMP Variables S. Amarnath and Anurag Kumar* Dept. of Electrical Communication Engineering Indian Institute of Science, Bangalore
More informationFundamental Limits of Invisible Flow Fingerprinting
Fundamental Limits of Invisible Flow Fingerprinting Ramin Soltani, Dennis Goeckel, Don Towsley, and Amir Houmansadr Electrical and Computer Engineering Department, University of Massachusetts, Amherst,
More informationChapter 2 Queueing Theory and Simulation
Chapter 2 Queueing Theory and Simulation Based on the slides of Dr. Dharma P. Agrawal, University of Cincinnati and Dr. Hiroyuki Ohsaki Graduate School of Information Science & Technology, Osaka University,
More informationAsymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime
Asymptotic Filtering and Entropy Rate of a Hidden Marov Process in the Rare Transitions Regime Chandra Nair Dept. of Elect. Engg. Stanford University Stanford CA 94305, USA mchandra@stanford.edu Eri Ordentlich
More informationMultiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets
Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,
More informationFirst Order Differential Equations
Chapter 2 First Order Differential Equations Introduction Any first order differential equation can be written as F (x, y, y )=0 by moving all nonzero terms to the left hand side of the equation. Of course,
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More information