State Variation Mining: On Information Divergence with Message Importance in Big Data

Size: px
Start display at page:

Download "State Variation Mining: On Information Divergence with Message Importance in Big Data"

Transcription

1 State Variation Mining: On Information Divergence with Message Importance in Big Data Rui She, Shanyun Liu, and Pingyi Fan State Key Laboratory on Microwave and Digital Communications Tsinghua National Laboratory for Information Science and Technology Department of Electronic Engineering, Tsinghua University, Beijing, P.R. China s: {sher15, arxiv: v2 [cs.it] 15 Jan 2018 Abstract Information transfer which reveals the state variation of variables can play a vital role in big data analytics and processing. In fact, the measure for information transfer can reflect the system change from the statistics by using the variable distributions, similar to KL divergence and Renyi divergence. Furthermore, in terms of the information transfer in big data, small probability events dominate the importance of the total message to some degree. Therefore, it is significant to design an information transfer measure based on the message importance which emphasizes the small probability events. In this paper, we propose the message importance divergence (MID) and investigate its characteristics and applications on three aspects. First, the message importance transfer capacity based on MID is presented to offer an upper bound for the information transfer with disturbance. Then, we utilize the MID to guide the queue length selection, which is the fundamental problem considered to have higher social or academic value in the caching operation of mobile edge computing. Finally, we extend the MID to the continuous case and discuss the robustness by using it to measuring information distance. I. INTRODUCTION Recently, the amount of data is exploding rapidly and the computing complexity for data processing is also increasing. To some degree, this phenomenon is resulted from more and more mobile devices as well as the growing service of clouds. In the literature, it is favored to process the collected data to dig out the hidden important information. On the one hand, it is necessary to improve the computation technologies for big data processing, such as cloud computing, fog computing and mobile edge computing (MEC). On the other hand, a series of technologies for big data analysis and mining are required, such as neural networks and machine learning, as well as distributed parallel computing, etc. In many scenarios of big data, the small probability events attract more attention than the large probability ones [1]. That is, the rarity of small probability events has higher social or academic value. Message importance is used to reflect such kind of values in applications. How to characterize the message importance variation in the processing of big data is a critical and interesting problem. Here, we shall investigate and measure information transfer process by use of the message importance, focusing on small probability events. Before we proceed, let us review the definition of message importance measure (MIM) briefly first [2]. In a finite alphabet, for a given probability distribution P = {p(x 1 ),p(x 2 ),...,p(x n )}, the MIM with importance coefficient 0 is defined as L(P, ) = log { x i p(x i )e (1 p(xi)) }, (1) which measures the information importance of the distribution. Then, by setting the parameter = 1 and simplifying the form of MIM, it is easy to obtain its fundamental definition as follows. Definition 1. For the discrete probability P ={p(x 1 ), p(x 2 ),...,p(x n )}, the MIM can be given by L(P) = x i p(x i )e p(xi). (2) On account of the Definition 1, it is available to regard the MIM as an element to measure the message importance variation for a dynamic system. Then, new information transfer measures based on the MIM are defined as follows. Definition 2. In terms of two discrete probability Q = {q(x 1 ),q(x 2 ),...,q(x n )} and P = {p(x 1 ),p(x 2 ),...,p(x n )}, the message importance divergence (MID) is defined as (Q P) = x i {q(x i )e q(xi) p(x i )e p(xi) }. (3) Note that the Definition 2 characterizes the information transfer from the statistics. That is, we can make use of MID to measure the end-to-end change for an information transfer process. In fact, there exist a variety of different information measures handling the problem of information transfer process. Shannon entropy and Renyi entropy are applicable to intrinsic dimension estimation [3]. As well, the NMIM and M-I divergence can be used in anomaly detection [4], [5]. Moreover, the directed information [6] and Schreibers transfer entropy [7] are commonly applied to inferring the causality structure and characterizing the information transfer process. In addition, referring to the idea from dynamical system theory, new information transfer measures are proposed to explore and exploit the causality between states in the system control [8]. However, in spite of numerous kinds of information measures, few works focus on how to characterize the information

2 transfer from the perspective of message importance in big data. To this end, the MID different from the above information measures is introduced. We organize the rest of this paper as follows. In Section II, we introduce the message importance transfer capacity measured by the MID to describe the information transfer with disturbance. In Section III, the MID is used to discuss the queue length selection for the data caching in MEC from the viewpoint of queue theory. Moreover, some simulations are presented to validate our theoretical results. In Section IV, we extend the MID to the continuous case to investigate the correlation of message importance in the information transfer process. Finally, we conclude it in Section VI. II. MESSAGE IMPORTANCE TRANSFER CAPACITY BASED ON MESSAGE IMPORTANCE DIVERGENCE In this section, we will introduce the MID to describe the information transfer with additive disturbance. To do so, we define the message importance transfer capacity measured by the MID as follows. Definition 3. Assume that there exists an information transfer process as, {X,p(y x),y}, (4) where the p(y x) denotes a probability distribution matrix describing the information transfer from the variable X following distribution p(x) to the Y following p(y). We define the message importance transfer capacity as C = max p(x) {L(Y) L(Y X)}, (5) where p(y j ) = x i p(x i )p(y j x i ), L(Y) = y j p(y j )e p(yj), L(Y X) = y j x i p(x i,y j )e p(yj xi). In order to have an insight into the applications of message importance transfer capacity, some specific information transfer scenarios are discussed as follows. A. Binary symmetric information transfer Proposition 1. Assume that there exists an information transfer process {X, p(y x), Y}, where the information transfer matrix is [ ] 1 p(y x) =, (6) 1 which indicates that variables X and Y both obey binary distributions. In this case, we have C() = e 1 2 L(), (7) where L() = e +(1 )e (1 ) and 0 < < 1. Proof. Considering a variable X following binary distribution (p,1 p), it is not difficult to see that L(Y X) = e +(1 )e (1 ). (8) Moreover, according to Eq. (5), we have C(p,) = max p { [p+(1 2p)]e [p+(1 2p)] +[(1 p)+(2p 1)]e [(1 p)+(2p 1)]} L(). Then, it is readily seen that C(p, ) p { = (1 2) [1 p (1 2p)]e [p+(1 2p)] [1 (1 p) ε(2p 1)]e [(1 p)+(2p 1)]}. (10) In the light of the monotonically decreasing of C(p,) p for p [0,1], it is apparent that p = 1/2 is the only solution for = 0. Therefore, the proposition can be testified. C(p,) p B. Strongly symmetric information transfer Proposition 2. Assume that there exists a strongly symmetric information transfer matrix as follows 1... p(y x) = , (11) 1... which indicates that variables X and Y both follow K- ary distributions. We have the message importance transfer capacity as (9) C() = e 1 K {(1 )e (1 ) +e }, (12) where the parameter (0,1). Proof. On account of the information transfer matrix and the Eq. (2), we have L(Y X) = e +(1 )e (1 ). (13) Then, similar to the proof of Proposition 1, we can also use Lagrange multiplier method to obtain the message information transfer capacity. In this case, the distribution of Y should satisfy p(y 1 ) = p(y 2 ) =... = p(y K ) = 1/K. In addition, consider that the probability distribution of variablex is {p(x 1 ),p(x 2 ),...,p(x K )}. In the strongly symmetric transfer matrix, if the variable X follows uniform distribution, namely p(x 1 ) = p(x 2 ) =... = p(x K ) = 1/K, we will have K K p(y j ) = p(x i,y j ) = p(x i )p(y j x i ) = 1 K (14) K p(y j x i ) = 1 K, which indicates that Y also follows the uniform distribution. Therefore, it is testified that when the variable X follows uniform distribution which leads to the uniform distribution for variable Y, we will obtain the message importance transfer capacity C().

3 III. APPLICATION IN MOBILE EDGE COMPUTING WITH THE M/M/S/K QUEUE Consider the MEC system that consists of numerous mobile devices, an edge server cluster, and a central cloud far from the local devices. The queue models for mobile devices, the edge server cluster and the central cloud are regarded as a M/M/1 queue, a M/M/s queue, and a M/M/ queue, respectively. In particular, each mobile device can offload a part of or the entire service requests to the corresponding edge server cluster when the communication is disturbed. As well, the overloaded requests from edge servers will be offloaded to the central cloud for processing [10]. In terms of the MID, it can be applied to distinguish the state probability distributions in a queue model such as M/M/s queue. In particular, the state probability is derived from the stationary process, namely a dynamic equilibrium based on birth and death process. Then, we can obtain the steady queue state probability in the M/M/s/k model. By use of Taylor series expansion, the approximate MIM is given by s+k s+k p j e pj = p j [1 p j +O(p 2 j )]. = 1 p 2 0 { s 1 ( aj j! )2 + (as /s!) 2 [1 ρ 2(k+1) } ] 1 ρ 2, (15) where λ and µ are arrival rate and service rate,sis the number of servers, k is the buffer or caching size, the traffic intensity ρ = a/s as well as a = λ/ µ. Note that, in the M/M/s/k queue, the queue state probability p j (j = 0,...,s+k) are given by p 0 = [ s 1 a j ] 1, j! + as s! 1 ρk+1 (16a) 1 ρ p j = aj j! p 0, (0 < j < s), (16b) p j = as s! ρj s p 0, (s j s+k). (16c) Then, referring to Eq. (15), we can obtain the MID to characterize the message importance gap for the queue model M/M/s as follows. Proposition 3. As for the M/M/s model, the information difference between two queue state probability distributions P k+1 and P k which are with buffer size k and k + 1 respectively, can be measured by MID as (P k+1 P k ) = s+k+1 s+k p j e pj p j e pj. 1 = ( (ϕ 1 +ϕ ρ )2 1 ρ (ϕ 1 +ϕ k ϕ2 2 1 ρ 2[ 1 ρ 2(k+1) (ϕ 1 +ϕ 2 1 ρ )2 s 1 1 ρ )2) 1 ρ2(k+2) ( aj j! )2 (ϕ 1 +ϕ 2 1 ρ k+2 1 ρ )2]. (17) where p j and p j are queue state probability in the M/M/s/k and M/M/s/k+1 models respectively, as well as parameter ϕ 1 and ϕ 2 are given by ϕ 1 = s 1 aj /j! and ϕ 2 = a s /s!. Similarly, it is not difficult to derive the MID between the queue state probability distributions with buffer size and k, which is given by (P P k ) 1 = ( (ϕ 1 +ϕ ρ )2 (ϕ 1 + ϕ2 + ϕ2 2 1 ρ 2(k+1) s 1 1 ρ )2) ( aj j! )2 1 ρ 2[ (ϕ 1 +ϕ ρ )2 (ϕ 1 + ϕ2 1 ρ )2]. (18) Especially, when the number of server is one, that is, the queue model is M/M/1/k model, it it readily seen that (s=1) (P P k ) = 2a 1+a 2a(1 ak+1 ) (1+a)(1 a k+2 ), (19) where (s=1) (P P k ) denotes the MID with s = 1. Moreover, for the queue length selection, it is required that the distinction between two distribution P and P k should be small enough, namely, (s=1) (P P k ) ǫ. Then we have ǫ(1+a) a(2+ǫ (2 ǫ)a) k ln 1. (20) lna It is easy to see that ǫ plays a key role in the caching size selection when using finite size caching to imitate the infinite caching working mode. Similar to MID, the KL divergence between the queue state probability distributions with buffer size k+1 and k is given by D(P k P k+1 ) = j = log p j log 1 p j j p j log 1 p j {1+ ϕ 2 (1 ρ)ρ k+1 }, ϕ 1(1 ρ) +ϕ 2 (21) where the parameters p j, p j, ϕ 1 and ϕ 2 are the same as them in Proposition 3. Likewise, we can derive the KL divergence between the queue state distributions with buffer size and k as D(P k P ) = log ϕ 1 1 +ϕ 2 1 ρ. (22) 1 ρ ϕ 1 +ϕ k ρ For the queue length selection with KL divergence, it is readily seen that k ln(1 1 2 ) ǫ 2. (23) lna To validate our derived results in theory, some simulations are presented. The events arrivals are listed in Table I. It is readily seen that they have the same average interarrival time as 1/ λ 0. Besides, the traffic intensity is selected as ρ = 0.9 in all cases.

4 Type of Imfornation measure value TABLE I THE DISTRIBUTIONS OF EVENT INTERARRIVAL TIME Exponential Uniform Normal P(X) X E( λ 0 ) X U(0,2/ λ 0 ) X N( λ 1, Queue length 1 ) λ 2 0 -Sim-Poisson -Sim-Uniform -Sim-Normal D-Sim-Poisson D-Sim-Uniform D-Sim-Normal -Ana-Poisson D-Ana-Poisson Fig. 1. The performance of different information measures between the queue length k and k+1 for the queue model in the case of different arrival events distributions. In Fig. 1, the legends -Sim, -Ana and D-Sim, D-Ana are used to denote the simulation results and the analytical results for MID and KL divergence, respectively. It is illustrated that the convergence of MID is faster than that of KL divergence, which indicates that MID may provide a reasonable lower bound to select the caching size for MEC. In addition, we can see that the Poisson distribution corresponds the worst case for the arrival process among the three discussed cases with respect to the convergence of both MID and KL divergence. IV. MESSAGE IMPORTANCE DIVERGENCE IN CONTINUOUS CASES Similar to the definition 1 and 2, we can extend the two definition to the case with continuous distributions as L(f(x)) = f(x)e f(x) dx, x, (24) (g(x) f(x)) = L(g(x)) L(f(x)) = g(x)e g(x) f(x)e f(x) dx,x, (25) where g(x) and f(x) are two probability distributions with respect to the variable X in a given interval. Moreover, L(f(x)) and (g(x) f(x)) can be called differential message importance measure (DMIM) and differential message importance divergence (DMID). Then, we investigate the correlation of message importance by using DMID, which can also reflect the robustness of DMID. Consider the observation model, P g0 f 0 : f 0 (x) g 0 (x), that denotes an information transfer map for the variable X from the probability distribution f 0 (x) to g 0 (x). By using the similar way in [9], the relationship between f 0 (x) and g 0 (x) can be described as g 0 (x) = f 0 (x)+ǫf0 α (x)u(x), (26) and the constraint condition satisfies ǫf0 α (x)u(x)dx = 0, (27) where ǫ and α are adjustable coefficients. u(x) is a perturbation function of the variable X in the interval. Then, by using the above model, the end-to-end information distance measured by DMID is given as follows. Proposition 4. For two probability distributions g 0 (x) and f 0 (x) whose relationship satisfies the conditions Eq. (26) and Eq. (27), the information distance measured by DMID is given by (g 0 (x) f 0 (x)) { = g 0 (x)e g0(x) f 0 (x)e f0(x)} dx = ǫ i! (i 1)! f0 i+α (x)u(x)dx + ǫ2 2 (28) where ǫ and α denote parameters, u(x) is a function of the variable X in the interval. f0 i 1+2α (x)u 2 (x)dx+o(ǫ 2 ), Proposition 4 describes the perturbation between f 0 (x) and g 0 (x). Furthermore, it is not difficult to obtain the DMID between two different distributions g (u) 1 and g (u) 2 based on the same reference distribution f 0 (x), which is given by (g (u) 1 (x) g(u) 2 (x)) = [L(g (u) 1 (x)) L(f 0(x))] [L(g (u) 2 (x)) L(f 0(x))] = ǫ + ǫ2 2 +o(ǫ 2 ), where i! (i 1)! f0 i+α (x)[u 1 (x) u 2 (x)]dx f0 i 1+2α (x)[u 2 1 (x) u2 2 (x)]dx (29) g (u) 1 (x) = f 0(x)+ǫf α 0 (x)u 1(x), x, (30a) g (u) 2 (x) = f 0(x)+ǫf α 0 (x)u 2(x), x, (30b) in which the ǫ and α are parameters, u 1 (x) and u 2 (x) are functions of the variable X in the interval.

5 Remark 1. It is apparent that when the parameter ǫ is small enough, the DMID of two distributions is convergent with the order of O(ǫ). Actually, this provides a way to apply DMID to measure the correlation of message importance, if the system does not have relatively large change. In addition, by extending the information transfer process described as Eq. (4) to the case with continuous distributions, it is also significant to clarify the effect of the additive disturbance on the message importance transfer capacity as follows. Theorem 1. Assume that there exists an information transfer process between the variable X and Y both following continuous distributoins, denoted by {X, p(y x), Y}, where E[X] = 0, E[X 2 ] = P s, Y = X + Z, and Z denotes independent memoryless additive variable, namely a continuous additive disturbance, whose mean and variance satisfy that E[Z] = µ and E[(Z µ) 2 ] = σ 2, respectively. Then, the message importance transfer capacity is determined by the variances of the variable X and Z as C(P s ) = max p(x) (Y Z) = max {L(Y)} L(Z) p(x) { + } = max p(y)e p(y) dy p(x) ( 1) j j! s.t. E[Y 2 ] = P s +P N, p j+1 (z)dz, (31a) (31b) where P N = µ 2 +σ 2, L( ) is the DMIM operator, and p(y) = p(x)p(y x)dx. Proof. According to Eq. (4), we have { x(x,y) = x z(x,y) = y x. (32) Then, it is not difficult to see that p(x,y) = p(x,z) J( xz x x x y ) = p(x,z) z z xy x y = p(x,z). (33) Moreover, by virtue of the independence of X and Z, we have p(x)p(y x) = p(x)p(z), which indicates that p(y x) = p(z). Then, we have L(Y X) = p(x)p(y x)e p(y x) dxdy x y = p(x)p(z)e p(z) dxdz (34) x z = p(x){ p(z)e p(z) dz}dx = L(Z). x z Consequently, in terms of the Definition 3, it is readily seen that L(Y) L(Y X) can be written as L(Y) L(Z), which testifies Eq. (31a). Furthermore, according to the fact that E[Y 2 ] = E[(X + Z) 2 ] = E[X 2 ] +E[Z 2 ] = P s + P N, we have the constraint condition Eq. (31b). As well, by substituting the definition of MID into Eq. (31a), the Theorem 1 is proved. In practice, the variance can be regarded as the power of signals. Therefore, the message importance transfer capacity in the Theorem 1 may be applied to guide the signal transfer processing with an additive disturbance. V. CONCLUSION In this paper, we investigated the information transfer problem in big data and proposed an information measure, i.e., MID. Furthermore, this information measure has its own dramatic characteristics on paying more attention to the message importance hidden in big data. This makes the information measure as a promising tool for information transfer measure in big data. We presented the message importance transfer capacity measured by the MID which can give an upper bound for the information transfer with disturbance. Furthermore, we employed the MID to discuss the caching size selection in MEC. In addition, the MID was extended to the continuous case to investigate the correlation of message importance in the information transfer process. ACKNOWLEDGMENT We indeed appreciate the support of the China Major State Basic Research Development Program (973 Program) No.2012CB316100(2), and the National Natural Science Foundation of China (NSFC) No REFERENCES [1] Y. Bu, S. Zou, and V. V. Veeravalli, Linear-complexity exponentiallyconsistent tests for universal outlying sequence detection, in Proc IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, June. 2017, pp [2] P. Fan, Y. Dong, J. Lu, and S. Liu, Message importance measure and its application to minority subset detection in big data, in Proc. IEEE Globecom Workshops (GC Wkshps), Washington D.C., USA, Dec. 2016, pp 1 5. [3] K. M. Carter, R. Raich, and A. O. Hero, On local intrinsic dimension estimation and its applications, IEEE Trans. Signal Process., vol. 58, no. 2, pp , Feb [4] S. Liu, R. She, P. Fan, and K. B. Letaief. (2017). Non-parametric message important measure: storage code design and transmission planning for big data. [Online]. Available: [5] R. She, S. Y. Liu, and P. Fan, Amplifying inter-message distance: on information divergence measures in big data, IEEE Access, vol. 5, pp , Nov [6] L. Zhao, Y. H. Kim, H. H. Permuter, and T. Weissman, Universal estimation of directed information, in Proc. IEEE Int. Symp. Inf. Theory (ISIT), Austin, USA, June. 2010, pp [7] T. Schreiber, Measuring information transfer, Physical Review Letters, vol. 85, no. 2, pp , July, [8] S. Sinha and U. Vaidya, Causality preserving information transfer measure for control dynamical system, in Proc. IEEE 55th Conference on Decision and Control (CDC), Las Vegas, USA, Dec. 2016, pp [9] S. Huang, A. Makur, L. Zheng, and G. W. Wornell, An informationtheoretic approach to universal feature selection in high-dimensional inference, in Proc IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, June. 2017, pp

6 [10] L. Liu, Z. Chang, X. Guo, and T. Ristaniemi, Multi-objective optimization for computation offloading in mobile-edge computing, In Proc. IEEE Symposium on Computers and Communications (ISCC), Heraklion, Greece, July. 2017, pp

Minor probability events detection in big data: An integrated approach with Bayesian testing and MIM

Minor probability events detection in big data: An integrated approach with Bayesian testing and MIM Minor probability events detection in big data: An integrated approach with Bayesian testing and MIM Shuo Wan, Jiaxun Lu, Pingyi Fan, and Khaled B. Letaief*, Tsinghua National Laboratory for Information

More information

Focusing on a Probability Element: Parameter Selection of Message Importance Measure in Big Data

Focusing on a Probability Element: Parameter Selection of Message Importance Measure in Big Data Focusing on a Probability Element: Parameter Selection of Message Importance Measure in Big Data arxiv:1701.03234v2 [cs.it] 11 Nov 2018 Rui She, Shanyun Liu, Yunquan Dong, and Pingyi Fan, SeniorMember,

More information

Performance Modelling of Computer Systems

Performance Modelling of Computer Systems Performance Modelling of Computer Systems Mirco Tribastone Institut für Informatik Ludwig-Maximilians-Universität München Fundamentals of Queueing Theory Tribastone (IFI LMU) Performance Modelling of Computer

More information

Age of Information Upon Decisions

Age of Information Upon Decisions Age of Information Upon Decisions Yunquan Dong, Zhengchuan Chen, Shanyun Liu, and Pingyi Fan School of Electronic & Information Engineering, Nanjing University of Information Science & Technology, Nanjing,

More information

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Nevroz Şen, Fady Alajaji and Serdar Yüksel Department of Mathematics and Statistics Queen s University Kingston, ON K7L 3N6, Canada

More information

BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS

BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS Andrea Bobbio Anno Accademico 999-2000 Queueing Systems 2 Notation for Queueing Systems /λ mean time between arrivals S = /µ ρ = λ/µ N mean service time traffic

More information

Performance Evaluation of Queuing Systems

Performance Evaluation of Queuing Systems Performance Evaluation of Queuing Systems Introduction to Queuing Systems System Performance Measures & Little s Law Equilibrium Solution of Birth-Death Processes Analysis of Single-Station Queuing Systems

More information

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems  M/M/1  M/M/m  M/M/1/K Queueing Theory I Summary Little s Law Queueing System Notation Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K " Little s Law a(t): the process that counts the number of arrivals

More information

An Extended Fano s Inequality for the Finite Blocklength Coding

An Extended Fano s Inequality for the Finite Blocklength Coding An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.

More information

Math 240 Calculus III

Math 240 Calculus III Calculus III Summer 2015, Session II Monday, August 3, 2015 Agenda 1. 2. Introduction The reduction of technique, which applies to second- linear differential equations, allows us to go beyond equations

More information

Continuous Time Processes

Continuous Time Processes page 102 Chapter 7 Continuous Time Processes 7.1 Introduction In a continuous time stochastic process (with discrete state space), a change of state can occur at any time instant. The associated point

More information

Can Feedback Increase the Capacity of the Energy Harvesting Channel?

Can Feedback Increase the Capacity of the Energy Harvesting Channel? Can Feedback Increase the Capacity of the Energy Harvesting Channel? Dor Shaviv EE Dept., Stanford University shaviv@stanford.edu Ayfer Özgür EE Dept., Stanford University aozgur@stanford.edu Haim Permuter

More information

The Least Degraded and the Least Upgraded Channel with respect to a Channel Family

The Least Degraded and the Least Upgraded Channel with respect to a Channel Family The Least Degraded and the Least Upgraded Channel with respect to a Channel Family Wei Liu, S. Hamed Hassani, and Rüdiger Urbanke School of Computer and Communication Sciences EPFL, Switzerland Email:

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

A Modification of Linfoot s Informational Correlation Coefficient

A Modification of Linfoot s Informational Correlation Coefficient Austrian Journal of Statistics April 07, Volume 46, 99 05. AJS http://www.ajs.or.at/ doi:0.773/ajs.v46i3-4.675 A Modification of Linfoot s Informational Correlation Coefficient Georgy Shevlyakov Peter

More information

Optimization in Information Theory

Optimization in Information Theory Optimization in Information Theory Dawei Shen November 11, 2005 Abstract This tutorial introduces the application of optimization techniques in information theory. We revisit channel capacity problem from

More information

Optimal Sleeping Mechanism for Multiple Servers with MMPP-Based Bursty Traffic Arrival

Optimal Sleeping Mechanism for Multiple Servers with MMPP-Based Bursty Traffic Arrival 1 Optimal Sleeping Mechanism for Multiple Servers with MMPP-Based Bursty Traffic Arrival Zhiyuan Jiang, Bhaskar Krishnamachari, Sheng Zhou, arxiv:1711.07912v1 [cs.it] 21 Nov 2017 Zhisheng Niu, Fellow,

More information

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur Module ntroduction to Digital Communications and nformation Theory Lesson 3 nformation Theoretic Approach to Digital Communications After reading this lesson, you will learn about Scope of nformation Theory

More information

Cover Page. The handle holds various files of this Leiden University dissertation

Cover Page. The handle  holds various files of this Leiden University dissertation Cover Page The handle http://hdl.handle.net/1887/39637 holds various files of this Leiden University dissertation Author: Smit, Laurens Title: Steady-state analysis of large scale systems : the successive

More information

Soft Covering with High Probability

Soft Covering with High Probability Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information

More information

Statistical Learning Theory

Statistical Learning Theory Statistical Learning Theory Part I : Mathematical Learning Theory (1-8) By Sumio Watanabe, Evaluation : Report Part II : Information Statistical Mechanics (9-15) By Yoshiyuki Kabashima, Evaluation : Report

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Queueing Theory. VK Room: M Last updated: October 17, 2013.

Queueing Theory. VK Room: M Last updated: October 17, 2013. Queueing Theory VK Room: M1.30 knightva@cf.ac.uk www.vincent-knight.com Last updated: October 17, 2013. 1 / 63 Overview Description of Queueing Processes The Single Server Markovian Queue Multi Server

More information

CPSC 531: System Modeling and Simulation. Carey Williamson Department of Computer Science University of Calgary Fall 2017

CPSC 531: System Modeling and Simulation. Carey Williamson Department of Computer Science University of Calgary Fall 2017 CPSC 531: System Modeling and Simulation Carey Williamson Department of Computer Science University of Calgary Fall 2017 Motivating Quote for Queueing Models Good things come to those who wait - poet/writer

More information

Link Models for Packet Switching

Link Models for Packet Switching Link Models for Packet Switching To begin our study of the performance of communications networks, we will study a model of a single link in a message switched network. The important feature of this model

More information

Queueing Theory and Simulation. Introduction

Queueing Theory and Simulation. Introduction Queueing Theory and Simulation Based on the slides of Dr. Dharma P. Agrawal, University of Cincinnati and Dr. Hiroyuki Ohsaki Graduate School of Information Science & Technology, Osaka University, Japan

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality

More information

An Outer Bound for the Gaussian. Interference channel with a relay.

An Outer Bound for the Gaussian. Interference channel with a relay. An Outer Bound for the Gaussian Interference Channel with a Relay Ivana Marić Stanford University Stanford, CA ivanam@wsl.stanford.edu Ron Dabora Ben-Gurion University Be er-sheva, Israel ron@ee.bgu.ac.il

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information 204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener

More information

Queuing Approaches to Principal-Agent Communication under Information Overload

Queuing Approaches to Principal-Agent Communication under Information Overload Queuing Approaches to Principal-Agent Communication under Information Overload Aseem Sharma, Krishna Jagannathan, Member, IEEE, and Lav R. Varshney, Senior Member, IEEE arxiv:1605.00802v1 [cs.it] 3 May

More information

Latent Tree Approximation in Linear Model

Latent Tree Approximation in Linear Model Latent Tree Approximation in Linear Model Navid Tafaghodi Khajavi Dept. of Electrical Engineering, University of Hawaii, Honolulu, HI 96822 Email: navidt@hawaii.edu ariv:1710.01838v1 [cs.it] 5 Oct 2017

More information

BRAVO for QED Queues

BRAVO for QED Queues 1 BRAVO for QED Queues Yoni Nazarathy, The University of Queensland Joint work with Daryl J. Daley, The University of Melbourne, Johan van Leeuwaarden, EURANDOM, Eindhoven University of Technology. Applied

More information

2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES

2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES 295 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES 16 Queueing Systems with Two Types of Customers In this section, we discuss queueing systems with two types of customers.

More information

Stochastic Models in Computer Science A Tutorial

Stochastic Models in Computer Science A Tutorial Stochastic Models in Computer Science A Tutorial Dr. Snehanshu Saha Department of Computer Science PESIT BSC, Bengaluru WCI 2015 - August 10 to August 13 1 Introduction 2 Random Variable 3 Introduction

More information

On the static assignment to parallel servers

On the static assignment to parallel servers On the static assignment to parallel servers Ger Koole Vrije Universiteit Faculty of Mathematics and Computer Science De Boelelaan 1081a, 1081 HV Amsterdam The Netherlands Email: koole@cs.vu.nl, Url: www.cs.vu.nl/

More information

Energy State Amplification in an Energy Harvesting Communication System

Energy State Amplification in an Energy Harvesting Communication System Energy State Amplification in an Energy Harvesting Communication System Omur Ozel Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland College Park, MD 20742 omur@umd.edu

More information

On the Capacity of the Interference Channel with a Relay

On the Capacity of the Interference Channel with a Relay On the Capacity of the Interference Channel with a Relay Ivana Marić, Ron Dabora and Andrea Goldsmith Stanford University, Stanford, CA {ivanam,ron,andrea}@wsl.stanford.edu Abstract Capacity gains due

More information

Arimoto Channel Coding Converse and Rényi Divergence

Arimoto Channel Coding Converse and Rényi Divergence Arimoto Channel Coding Converse and Rényi Divergence Yury Polyanskiy and Sergio Verdú Abstract Arimoto proved a non-asymptotic upper bound on the probability of successful decoding achievable by any code

More information

Practical Polar Code Construction Using Generalised Generator Matrices

Practical Polar Code Construction Using Generalised Generator Matrices Practical Polar Code Construction Using Generalised Generator Matrices Berksan Serbetci and Ali E. Pusane Department of Electrical and Electronics Engineering Bogazici University Istanbul, Turkey E-mail:

More information

2905 Queueing Theory and Simulation PART IV: SIMULATION

2905 Queueing Theory and Simulation PART IV: SIMULATION 2905 Queueing Theory and Simulation PART IV: SIMULATION 22 Random Numbers A fundamental step in a simulation study is the generation of random numbers, where a random number represents the value of a random

More information

Outline. Finite source queue M/M/c//K Queues with impatience (balking, reneging, jockeying, retrial) Transient behavior Advanced Queue.

Outline. Finite source queue M/M/c//K Queues with impatience (balking, reneging, jockeying, retrial) Transient behavior Advanced Queue. Outline Finite source queue M/M/c//K Queues with impatience (balking, reneging, jockeying, retrial) Transient behavior Advanced Queue Batch queue Bulk input queue M [X] /M/1 Bulk service queue M/M [Y]

More information

Information Measure Estimation and Applications: Boosting the Effective Sample Size from n to n ln n

Information Measure Estimation and Applications: Boosting the Effective Sample Size from n to n ln n Information Measure Estimation and Applications: Boosting the Effective Sample Size from n to n ln n Jiantao Jiao (Stanford EE) Joint work with: Kartik Venkat Yanjun Han Tsachy Weissman Stanford EE Tsinghua

More information

Taylor series. Chapter Introduction From geometric series to Taylor polynomials

Taylor series. Chapter Introduction From geometric series to Taylor polynomials Chapter 2 Taylor series 2. Introduction The topic of this chapter is find approximations of functions in terms of power series, also called Taylor series. Such series can be described informally as infinite

More information

Figure 10.1: Recording when the event E occurs

Figure 10.1: Recording when the event E occurs 10 Poisson Processes Let T R be an interval. A family of random variables {X(t) ; t T} is called a continuous time stochastic process. We often consider T = [0, 1] and T = [0, ). As X(t) is a random variable

More information

Introduction to queuing theory

Introduction to queuing theory Introduction to queuing theory Queu(e)ing theory Queu(e)ing theory is the branch of mathematics devoted to how objects (packets in a network, people in a bank, processes in a CPU etc etc) join and leave

More information

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015 10-704 Homework 1 Due: Thursday 2/5/2015 Instructions: Turn in your homework in class on Thursday 2/5/2015 1. Information Theory Basics and Inequalities C&T 2.47, 2.29 (a) A deck of n cards in order 1,

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Energy-Efficient Resource Allocation for Multi-User Mobile Edge Computing

Energy-Efficient Resource Allocation for Multi-User Mobile Edge Computing Energy-Efficient Resource Allocation for Multi-User Mobile Edge Computing Junfeng Guo, Zhaozhe Song, Ying Cui Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, P. R. China

More information

Chapter 2 Review of Classical Information Theory

Chapter 2 Review of Classical Information Theory Chapter 2 Review of Classical Information Theory Abstract This chapter presents a review of the classical information theory which plays a crucial role in this thesis. We introduce the various types of

More information

Latent Variable Models and EM algorithm

Latent Variable Models and EM algorithm Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic

More information

Tight Bounds on Minimum Maximum Pointwise Redundancy

Tight Bounds on Minimum Maximum Pointwise Redundancy Tight Bounds on Minimum Maximum Pointwise Redundancy Michael B. Baer vlnks Mountain View, CA 94041-2803, USA Email:.calbear@ 1eee.org Abstract This paper presents new lower and upper bounds for the optimal

More information

Information in Biology

Information in Biology Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve

More information

Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes

Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel 2009 IEEE International

More information

Quantifying Stochastic Model Errors via Robust Optimization

Quantifying Stochastic Model Errors via Robust Optimization Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations

More information

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:

More information

AP Calculus Chapter 9: Infinite Series

AP Calculus Chapter 9: Infinite Series AP Calculus Chapter 9: Infinite Series 9. Sequences a, a 2, a 3, a 4, a 5,... Sequence: A function whose domain is the set of positive integers n = 2 3 4 a n = a a 2 a 3 a 4 terms of the sequence Begin

More information

Convexity/Concavity of Renyi Entropy and α-mutual Information

Convexity/Concavity of Renyi Entropy and α-mutual Information Convexity/Concavity of Renyi Entropy and -Mutual Information Siu-Wai Ho Institute for Telecommunications Research University of South Australia Adelaide, SA 5095, Australia Email: siuwai.ho@unisa.edu.au

More information

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009. NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total

More information

Simple Channel Coding Bounds

Simple Channel Coding Bounds ISIT 2009, Seoul, Korea, June 28 - July 3,2009 Simple Channel Coding Bounds Ligong Wang Signal and Information Processing Laboratory wang@isi.ee.ethz.ch Roger Colbeck Institute for Theoretical Physics,

More information

Information in Biology

Information in Biology Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1 Living

More information

The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels

The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels Parastoo Sadeghi National ICT Australia (NICTA) Sydney NSW 252 Australia Email: parastoo@student.unsw.edu.au Predrag

More information

Bioinformatics: Biology X

Bioinformatics: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Bayesian Interpretation of Probabilities 2 Where (or of what)

More information

Stochastic Models of Manufacturing Systems

Stochastic Models of Manufacturing Systems Stochastic Models of Manufacturing Systems Ivo Adan Organization 2/47 7 lectures (lecture of May 12 is canceled) Studyguide available (with notes, slides, assignments, references), see http://www.win.tue.nl/

More information

Chapter 5 Joint Probability Distributions

Chapter 5 Joint Probability Distributions Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 5 Joint Probability Distributions 5 Joint Probability Distributions CHAPTER OUTLINE 5-1 Two

More information

Introduction to Markov Chains, Queuing Theory, and Network Performance

Introduction to Markov Chains, Queuing Theory, and Network Performance Introduction to Markov Chains, Queuing Theory, and Network Performance Marceau Coupechoux Telecom ParisTech, departement Informatique et Réseaux marceau.coupechoux@telecom-paristech.fr IT.2403 Modélisation

More information

Slepian-Wolf Code Design via Source-Channel Correspondence

Slepian-Wolf Code Design via Source-Channel Correspondence Slepian-Wolf Code Design via Source-Channel Correspondence Jun Chen University of Illinois at Urbana-Champaign Urbana, IL 61801, USA Email: junchen@ifpuiucedu Dake He IBM T J Watson Research Center Yorktown

More information

Multiaccess Problem. How to let distributed users (efficiently) share a single broadcast channel? How to form a queue for distributed users?

Multiaccess Problem. How to let distributed users (efficiently) share a single broadcast channel? How to form a queue for distributed users? Multiaccess Problem How to let distributed users (efficiently) share a single broadcast channel? How to form a queue for distributed users? z The protocols we used to solve this multiaccess problem are

More information

Exercises Solutions. Automation IEA, LTH. Chapter 2 Manufacturing and process systems. Chapter 5 Discrete manufacturing problems

Exercises Solutions. Automation IEA, LTH. Chapter 2 Manufacturing and process systems. Chapter 5 Discrete manufacturing problems Exercises Solutions Note, that we have not formulated the answers for all the review questions. You will find the answers for many questions by reading and reflecting about the text in the book. Chapter

More information

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) = Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of

More information

Link Models for Circuit Switching

Link Models for Circuit Switching Link Models for Circuit Switching The basis of traffic engineering for telecommunication networks is the Erlang loss function. It basically allows us to determine the amount of telephone traffic that can

More information

One-Bit LDPC Message Passing Decoding Based on Maximization of Mutual Information

One-Bit LDPC Message Passing Decoding Based on Maximization of Mutual Information One-Bit LDPC Message Passing Decoding Based on Maximization of Mutual Information ZOU Sheng and Brian M. Kurkoski kurkoski@ice.uec.ac.jp University of Electro-Communications Tokyo, Japan University of

More information

THE QUEEN S UNIVERSITY OF BELFAST

THE QUEEN S UNIVERSITY OF BELFAST THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M

More information

Computation of Information Rates from Finite-State Source/Channel Models

Computation of Information Rates from Finite-State Source/Channel Models Allerton 2002 Computation of Information Rates from Finite-State Source/Channel Models Dieter Arnold arnold@isi.ee.ethz.ch Hans-Andrea Loeliger loeliger@isi.ee.ethz.ch Pascal O. Vontobel vontobel@isi.ee.ethz.ch

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL First published in Journal of Applied Probability 43(1) c 2006 Applied Probability Trust STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL LASSE LESKELÄ, Helsinki

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Synchronized Queues with Deterministic Arrivals

Synchronized Queues with Deterministic Arrivals Synchronized Queues with Deterministic Arrivals Dimitra Pinotsi and Michael A. Zazanis Department of Statistics Athens University of Economics and Business 76 Patission str., Athens 14 34, Greece Abstract

More information

Stat 100a, Introduction to Probability.

Stat 100a, Introduction to Probability. Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous

More information

Lecture 20: Reversible Processes and Queues

Lecture 20: Reversible Processes and Queues Lecture 20: Reversible Processes and Queues 1 Examples of reversible processes 11 Birth-death processes We define two non-negative sequences birth and death rates denoted by {λ n : n N 0 } and {µ n : n

More information

Reducing Age-of-Information for Computation-Intensive Messages via Packet Replacement

Reducing Age-of-Information for Computation-Intensive Messages via Packet Replacement Reducing Age-of-Information for Computation-Intensive Messages via Packet Replacement arxiv:191.4654v1 [cs.it] 15 Jan 19 Jie Gong, Qiaobin Kuang, Xiang Chen and Xiao Ma School of Data and Computer Science,

More information

STAT Chapter 5 Continuous Distributions

STAT Chapter 5 Continuous Distributions STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

L 2 -induced Gains of Switched Systems and Classes of Switching Signals

L 2 -induced Gains of Switched Systems and Classes of Switching Signals L 2 -induced Gains of Switched Systems and Classes of Switching Signals Kenji Hirata and João P. Hespanha Abstract This paper addresses the L 2-induced gain analysis for switched linear systems. We exploit

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017) UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable

More information

The Behavior of a Multichannel Queueing System under Three Queue Disciplines

The Behavior of a Multichannel Queueing System under Three Queue Disciplines The Behavior of a Multichannel Queueing System under Three Queue Disciplines John K Karlof John Jenkins November 11, 2002 Abstract In this paper we investigate a multichannel channel queueing system where

More information

Divergence measure of intuitionistic fuzzy sets

Divergence measure of intuitionistic fuzzy sets Divergence measure of intuitionistic fuzzy sets Fuyuan Xiao a, a School of Computer and Information Science, Southwest University, Chongqing, 400715, China Abstract As a generation of fuzzy sets, the intuitionistic

More information

Citation PHYSICAL REVIEW E (2002), 65(6) RightCopyright 2002 American Physical So

Citation PHYSICAL REVIEW E (2002), 65(6)   RightCopyright 2002 American Physical So TitleStatistical mechanics of two hard d Author(s) Munakata, T; Hu, G Citation PHYSICAL REVIEW E (2002), 65(6) Issue Date 2002-06 URL http://hdl.handle.net/2433/50310 RightCopyright 2002 American Physical

More information

STEADY-STATE PROBABILITIES FOR SOME QUEUEING SYSTEMS: Part II

STEADY-STATE PROBABILITIES FOR SOME QUEUEING SYSTEMS: Part II STEADY-STATE PROBABILITIES FOR SOME QUEUEING SYSTEMS: Part II Mahmoud Abdel Moety Hussain Faculty of Computers and Informatics Zagazig University Abstract This paper applies the theoretical results presented

More information

6 Solving Queueing Models

6 Solving Queueing Models 6 Solving Queueing Models 6.1 Introduction In this note we look at the solution of systems of queues, starting with simple isolated queues. The benefits of using predefined, easily classified queues will

More information

A New Technique for Link Utilization Estimation

A New Technique for Link Utilization Estimation A New Technique for Link Utilization Estimation in Packet Data Networks using SNMP Variables S. Amarnath and Anurag Kumar* Dept. of Electrical Communication Engineering Indian Institute of Science, Bangalore

More information

Fundamental Limits of Invisible Flow Fingerprinting

Fundamental Limits of Invisible Flow Fingerprinting Fundamental Limits of Invisible Flow Fingerprinting Ramin Soltani, Dennis Goeckel, Don Towsley, and Amir Houmansadr Electrical and Computer Engineering Department, University of Massachusetts, Amherst,

More information

Chapter 2 Queueing Theory and Simulation

Chapter 2 Queueing Theory and Simulation Chapter 2 Queueing Theory and Simulation Based on the slides of Dr. Dharma P. Agrawal, University of Cincinnati and Dr. Hiroyuki Ohsaki Graduate School of Information Science & Technology, Osaka University,

More information

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime Asymptotic Filtering and Entropy Rate of a Hidden Marov Process in the Rare Transitions Regime Chandra Nair Dept. of Elect. Engg. Stanford University Stanford CA 94305, USA mchandra@stanford.edu Eri Ordentlich

More information

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,

More information

First Order Differential Equations

First Order Differential Equations Chapter 2 First Order Differential Equations Introduction Any first order differential equation can be written as F (x, y, y )=0 by moving all nonzero terms to the left hand side of the equation. Of course,

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information