A General Formula for Compound Channel Capacity
|
|
- Aldous Barnett
- 5 years ago
- Views:
Transcription
1 A General Formula for Compound Channel Capacity Sergey Loyka, Charalambos D. Charalambous University of Ottawa, University of Cyprus ETH Zurich (May 2015), ISIT-15 1/32
2 Outline 1 Introduction 2 Channel Model 3 Definitions 4 Capacity for a given channel state 5 General Formula for Compound Channel Capacity 6 Uniform Compound Channel 7 Examples & mistakes 8 Conclusion 2/32
3 Introduction Impact of channel state information (CSI) on capacity/system design Real world: measurement/modeling uncertainty Wireless channels dynamic nature estimation error feedback limitations Uncertain CSI: deterministic vs. stochastic models 3/32
4 Introduction Impact of channel state information (CSI) on capacity/system design Real world: measurement/modeling uncertainty Wireless channels dynamic nature estimation error feedback limitations Uncertain CSI: deterministic vs. stochastic models 4/32
5 Introduction Impact of channel state information (CSI) on capacity/system design Real world: measurement/modeling uncertainty Wireless channels dynamic nature estimation error feedback limitations Uncertain CSI: deterministic vs. stochastic models 5/32
6 Introduction Impact of channel state information (CSI) on capacity/system design Real world: measurement/modeling uncertainty Wireless channels dynamic nature estimation error feedback limitations Uncertain CSI: deterministic vs. stochastic models 6/32
7 Introduction Impact of channel uncertainty: extensive studies since 1950s 12 Several approaches compound channel model mixed/composite channel arbitrary varying channel CDI instead of CSI 1 A. Lapidoth and P. Narayan, Reliable Communication Under Channel Uncertainty, IEEE Trans. Inform. Theory, vol. 44, No. 6, Oct E. Biglieri, J. Proakis, and S. Shamai, Fading Channels: Information-Theoretic and Communications Aspects, IEEE Trans. Inform. Theory, vol. 44, No. 6, Oct /32
8 Introduction Impact of channel uncertainty: extensive studies since 1950s 12 Several approaches compound channel model mixed/composite channel arbitrary varying channel CDI instead of CSI 1 A. Lapidoth and P. Narayan, Reliable Communication Under Channel Uncertainty, IEEE Trans. Inform. Theory, vol. 44, No. 6, Oct E. Biglieri, J. Proakis, and S. Shamai, Fading Channels: Information-Theoretic and Communications Aspects, IEEE Trans. Inform. Theory, vol. 44, No. 6, Oct /32
9 Introduction Compound channel/capacity Prior work: information-stable channels (e.g. stationary ergodic) Many/all channels are non-stationary, non-ergodic modulation-induced channels quasi-static fading channels how to measure ergodicity??? This work: information-unstable channels no ergodicity/stationarity is required via information density/spectrum 34 3 S. Verdu, T.S. Han, A General Formula for Channel Capacity, IEEE Trans. Info. Theory, vol. 40, no. 4, July T. S. Han, Information-Spectrum Method in Information Theory, New York: Springer, /32
10 Introduction Compound channel/capacity Prior work: information-stable channels (e.g. stationary ergodic) Many/all channels are non-stationary, non-ergodic modulation-induced channels quasi-static fading channels how to measure ergodicity??? This work: information-unstable channels no ergodicity/stationarity is required via information density/spectrum 34 3 S. Verdu, T.S. Han, A General Formula for Channel Capacity, IEEE Trans. Info. Theory, vol. 40, no. 4, July T. S. Han, Information-Spectrum Method in Information Theory, New York: Springer, /32
11 Introduction Compound channel/capacity Prior work: information-stable channels (e.g. stationary ergodic) Many/all channels are non-stationary, non-ergodic modulation-induced channels quasi-static fading channels how to measure ergodicity??? This work: information-unstable channels no ergodicity/stationarity is required via information density/spectrum 34 3 S. Verdu, T.S. Han, A General Formula for Channel Capacity, IEEE Trans. Info. Theory, vol. 40, no. 4, July T. S. Han, Information-Spectrum Method in Information Theory, New York: Springer, /32
12 Introduction Compound channel/capacity Prior work: information-stable channels (e.g. stationary ergodic) Many/all channels are non-stationary, non-ergodic modulation-induced channels quasi-static fading channels how to measure ergodicity??? This work: information-unstable channels no ergodicity/stationarity is required via information density/spectrum 34 3 S. Verdu, T.S. Han, A General Formula for Channel Capacity, IEEE Trans. Info. Theory, vol. 40, no. 4, July T. S. Han, Information-Spectrum Method in Information Theory, New York: Springer, /32
13 Compound Channel Model I a channel is selected from a given uncertainty set stays fixed during entire transmission design a single code good for any channel in the set largest achievable rate = compound capacity No CSI at Tx, complete CSI at Rx 13/32
14 Compound Channel Model II X n = {X 1...X n } - (random) sequence of n input symbols Y n - corresponding output sequence; s S - channel state (which may also be a sequence) p(x n ), p s (y n ) - input/output distributions under state s. p s (y n x n ) - channel transition probability; information density ( unexpected mutual information): i(x n ;y n,s) = ln p s(x n,y n ) p(x n )p s (y n ) = i(xn ;y n s) (1) 14/32
15 Compound Channel Model III If information-stable: If information-unstable: 1 n i(xn ;y n s) I(X;Y s) as n (2) 1 n i(xn ;y n s) RV (3) 15/32
16 Definitions I (n, r n, ε ns )-code: n - the block length ε ns - the error probability (for channel state s) r n = lnm n /n - the code rate M n - the number of codewords compound error probability: achievable rate R: (n, r n, ε ns )-code such that compound channel capacity C: ε n = supε ns (4) s S lim supε n = 0, liminf r n R (5) n n C = sup{r : R is achievable} (6) R 16/32
17 Capacity for a given channel state Very general - information-unstable channels Given channel state s, the capacity is [Verdu-Han 94] 5 C(s) = supi(x;y s) (7) p(x) where I(X; Y s) is the inf-information rate I(X;Y s) = sup R { R : lim n Pr{ n 1 i(x n ;Y n s) R } = 0 } (8) Proof: via Feinstein and Verdu-Han Lemmas 5 S. Verdu, T.S. Han, A General Formula for Channel Capacity, IEEE Trans. Info. Theory, vol. 40, no. 4, July /32
18 Capacity for a given channel state Lemma (Feinstein) For arbitrary input X n, any r n and a given channel state s, there exists a code satisfying the following inequality, for any γ > 0. Lemma (Verdu-Han 94) ε ns Pr { n 1 i(x n ;Y n s) r n +γ } +e γn (9) Every (n,r n,ε ns )-code satisfies the following inequality, ε ns Pr { n 1 i(x n ;Y n s) r n γ } e γn (10) for any γ > 0, where X n is uniformly distributed over all codewords and Y n is the corresponding channel output under channel state s. 18/32
19 Capacity for a given channel state Lemma (Feinstein) For arbitrary input X n, any r n and a given channel state s, there exists a code satisfying the following inequality, for any γ > 0. Lemma (Verdu-Han 94) ε ns Pr { n 1 i(x n ;Y n s) r n +γ } +e γn (9) Every (n,r n,ε ns )-code satisfies the following inequality, ε ns Pr { n 1 i(x n ;Y n s) r n γ } e γn (10) for any γ > 0, where X n is uniformly distributed over all codewords and Y n is the corresponding channel output under channel state s. 19/32
20 General Formula for Compound Channel Capacity Theorem Consider a general compound channel; the Rx knows s S, but not the Tx; the Tx knows the (arbitrary) S. The compound channel capacity is C c = supi(x;y) (11) p(x) where I(X;Y) = sup R {R Ω} is the compound inf-information rate, Ω = { R : lim sup n s S { } } 1 Pr n i(xn ;Y n s) R = 0 (12) Proof: via Feinstein and Verdu-Han Lemmas extended to compound setting 20/32
21 General Formula for Compound Channel Capacity Theorem Consider a general compound channel; the Rx knows s S, but not the Tx; the Tx knows the (arbitrary) S. The compound channel capacity is C c = supi(x;y) (11) p(x) where I(X;Y) = sup R {R Ω} is the compound inf-information rate, Ω = { R : lim sup n s S { } } 1 Pr n i(xn ;Y n s) R = 0 (12) Proof: via Feinstein and Verdu-Han Lemmas extended to compound setting 21/32
22 Compound Lemmas Lemma (compound Feinstein) For any X n, S and r n, there exists a (n,r n,ε n )-code (where the codewords are independent of channel state s), such that, for any γ > 0, ε n suppr { n 1 i(x n ;Y n s) r n +γ } +e γn (13) s S Lemma (compound Verdu-Han) Every (n,r n,ε n )-code satisfies ε n suppr { n 1 i(x n ;Y n s) r n γ } e γn (14) s S for any γ > 0, where X n is uniformly distributed over all codewords and Y n is the corresponding channel output under channel state s. 22/32
23 Compound Lemmas Lemma (compound Feinstein) For any X n, S and r n, there exists a (n,r n,ε n )-code (where the codewords are independent of channel state s), such that, for any γ > 0, ε n suppr { n 1 i(x n ;Y n s) r n +γ } +e γn (13) s S Lemma (compound Verdu-Han) Every (n,r n,ε n )-code satisfies ε n suppr { n 1 i(x n ;Y n s) r n γ } e γn (14) s S for any γ > 0, where X n is uniformly distributed over all codewords and Y n is the corresponding channel output under channel state s. 23/32
24 Properties Remark I(X,Y) is an extension of I(X,Y s) to the compound channel setting, not inf s I(X,Y s), in the general case. The relationship between I(X,Y) and inf s I(X,Y s) is established below. Proposition The following inequality holds for a general compound channel I(X,Y) I(X,Y) = infi(x,y s) (15) s Strict inequality can be shown by examples. 24/32
25 Properties Remark I(X,Y) is an extension of I(X,Y s) to the compound channel setting, not inf s I(X,Y s), in the general case. The relationship between I(X,Y) and inf s I(X,Y s) is established below. Proposition The following inequality holds for a general compound channel I(X,Y) I(X,Y) = infi(x,y s) (15) s Strict inequality can be shown by examples. 25/32
26 Uniform Compound Channel When does the equality hold? Definition A compound channel is uniform if Pr { n 1 i(x n ;Y n s) I(X,Y) γ } 0 γ > 0 (16) uniformly in s S as n. Proposition The following equality holds for a uniform compound channel I(X,Y) = I(X,Y) = infi(x,y s) (17) s 26/32
27 Uniform Compound Channel When does the equality hold? Definition A compound channel is uniform if Pr { n 1 i(x n ;Y n s) I(X,Y) γ } 0 γ > 0 (16) uniformly in s S as n. Proposition The following equality holds for a uniform compound channel I(X,Y) = I(X,Y) = infi(x,y s) (17) s 27/32
28 Uniform Compound Channel Theorem The capacity of a uniform compound channel: C c = sup inf I(X;Y s) (18) p(x) s S Does not hold for non-uniform channels, contrary to what Theorem in [Han 03] 6 claims. 6 T. S. Han, Information-Spectrum Method in Information Theory, New York: Springer, /32
29 Examples & mistakes Binary non-stationary channel with memory: p s (y n x n ) = p s (y n ) if n s (19) = BSC(0) if n > s (20) Models the noise coherence time τ = s S = {1,2,...}. It follows that I(X;Y) = 0 < I(X;Y) = infi(x;y s) = ln2 s (21) C c = supi(x;y) = 0 < ln2 = sup inf I(X;Y s) p(x) p(x) s S (22) Mistake: Theorem in [Han 03] claims C c = ln2. 29/32
30 Examples & mistakes Mistake: Theorem in [Han 03] claims C c = ln2. Reason: improper error probability definition for compound channel: In general, In our example, lim ε ns = 0 s lim n sup s 0 = sup s sup n s lim ε ns lim n sup n s lim ε ns < lim n sup n s ε ns = 0 (23) ε ns (24) ε ns = 1 (25) 30/32
31 Examples & mistakes Binary compound channel with additive noise: Y k = X k +Z ks, Z n s = {w 1,w 2,...w s,0,0...0} (26) and w 1...w s are i.i.d. equiprobable. Its capacity Why C c = 0? C c = supi(x;y) = 0 < ln2 = sup p(x) p(x) inf I(X;Y s) (27) For any n, does not matter how large, there are always channel states s n for which the channel is BSC(1/2), i.e. useless. The standard sup inf expression falls short of the channel capacity because this compound channel is not uniform. s S 31/32
32 Conclusion CSI uncertainty - compound channel Real world - information-unstable channels General formula for compound channel capacity Uniform compound channels Examples & mistakes 32/32
The Poisson Channel with Side Information
The Poisson Channel with Side Information Shraga Bross School of Enginerring Bar-Ilan University, Israel brosss@macs.biu.ac.il Amos Lapidoth Ligong Wang Signal and Information Processing Laboratory ETH
More informationDispersion of the Gilbert-Elliott Channel
Dispersion of the Gilbert-Elliott Channel Yury Polyanskiy Email: ypolyans@princeton.edu H. Vincent Poor Email: poor@princeton.edu Sergio Verdú Email: verdu@princeton.edu Abstract Channel dispersion plays
More informationChapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University
Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission
More informationVariable Rate Channel Capacity. Jie Ren 2013/4/26
Variable Rate Channel Capacity Jie Ren 2013/4/26 Reference This is a introduc?on of Sergio Verdu and Shlomo Shamai s paper. Sergio Verdu and Shlomo Shamai, Variable- Rate Channel Capacity, IEEE Transac?ons
More informationOn the Sufficiency of Power Control for a Class of Channels with Feedback
On the Sufficiency of Power Control for a Class of Channels with Feedback Desmond S. Lun and Muriel Médard Laboratory for Information and Decision Systems Massachusetts Institute of Technology Cambridge,
More informationWE study the capacity of peak-power limited, single-antenna,
1158 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 3, MARCH 2010 Gaussian Fading Is the Worst Fading Tobias Koch, Member, IEEE, and Amos Lapidoth, Fellow, IEEE Abstract The capacity of peak-power
More informationDiscrete Memoryless Channels with Memoryless Output Sequences
Discrete Memoryless Channels with Memoryless utput Sequences Marcelo S Pinho Department of Electronic Engineering Instituto Tecnologico de Aeronautica Sao Jose dos Campos, SP 12228-900, Brazil Email: mpinho@ieeeorg
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationLecture 4 Noisy Channel Coding
Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem
More informationEE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15
EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 1. Cascade of Binary Symmetric Channels The conditional probability distribution py x for each of the BSCs may be expressed by the transition probability
More informationOn the Distribution of Mutual Information
On the Distribution of Mutual Information J. Nicholas Laneman Dept. of Electrical Engineering University of Notre Dame Notre Dame IN 46556 Email: jnl@nd.edu Abstract In the early years of information theory
More informationA Formula for the Capacity of the General Gel fand-pinsker Channel
A Formula for the Capacity of the General Gel fand-pinsker Channel Vincent Y. F. Tan Institute for Infocomm Research (I 2 R, A*STAR, Email: tanyfv@i2r.a-star.edu.sg ECE Dept., National University of Singapore
More informationAn Extended Fano s Inequality for the Finite Blocklength Coding
An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.
More informationFeedback Capacity of a Class of Symmetric Finite-State Markov Channels
Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Nevroz Şen, Fady Alajaji and Serdar Yüksel Department of Mathematics and Statistics Queen s University Kingston, ON K7L 3N6, Canada
More informationDiversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007
Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT ECE 559 Presentation Hoa Pham Dec 3, 2007 Introduction MIMO systems provide two types of gains Diversity Gain: each path from a transmitter
More informationCapacity Definitions for General Channels with Receiver Side Information
Capacity Definitions for General Channels with Receiver Side Information Michelle Effros, Senior Member, IEEE, Andrea Goldsmith, Fellow, IEEE, and Yifan Liang, Student Member, IEEE arxiv:0804.4239v [cs.it]
More informationFeedback Capacity of the Compound Channel
Feedback Capacity of the Compound Channel The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher Shrader,
More informationCapacity of a channel Shannon s second theorem. Information Theory 1/33
Capacity of a channel Shannon s second theorem Information Theory 1/33 Outline 1. Memoryless channels, examples ; 2. Capacity ; 3. Symmetric channels ; 4. Channel Coding ; 5. Shannon s second theorem,
More informationGeneralized Writing on Dirty Paper
Generalized Writing on Dirty Paper Aaron S. Cohen acohen@mit.edu MIT, 36-689 77 Massachusetts Ave. Cambridge, MA 02139-4307 Amos Lapidoth lapidoth@isi.ee.ethz.ch ETF E107 ETH-Zentrum CH-8092 Zürich, Switzerland
More informationCan Feedback Increase the Capacity of the Energy Harvesting Channel?
Can Feedback Increase the Capacity of the Energy Harvesting Channel? Dor Shaviv EE Dept., Stanford University shaviv@stanford.edu Ayfer Özgür EE Dept., Stanford University aozgur@stanford.edu Haim Permuter
More informationComputation of Information Rates from Finite-State Source/Channel Models
Allerton 2002 Computation of Information Rates from Finite-State Source/Channel Models Dieter Arnold arnold@isi.ee.ethz.ch Hans-Andrea Loeliger loeliger@isi.ee.ethz.ch Pascal O. Vontobel vontobel@isi.ee.ethz.ch
More informationTowards control over fading channels
Towards control over fading channels Paolo Minero, Massimo Franceschetti Advanced Network Science University of California San Diego, CA, USA mail: {minero,massimo}@ucsd.edu Invited Paper) Subhrakanti
More informationRevision of Lecture 4
Revision of Lecture 4 We have completed studying digital sources from information theory viewpoint We have learnt all fundamental principles for source coding, provided by information theory Practical
More informationThroughput analysis for MIMO systems in the high SNR regime
ISIT 2006, Seattle, USA, July 9-4, 2006 Throughput analysis for MIMO systems in the high SNR regime Narayan Prasad and Mahesh K. Varanasi Electrical & Computer Engineering Department University of Colorado,
More informationPerformance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)
Performance-based Security for Encoding of Information Signals FA9550-15-1-0180 (2015-2018) Paul Cuff (Princeton University) Contributors Two students finished PhD Tiance Wang (Goldman Sachs) Eva Song
More informationSoft Covering with High Probability
Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information
More informationMultiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets
Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,
More informationThe Method of Types and Its Application to Information Hiding
The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,
More informationCapacity-achieving Feedback Scheme for Flat Fading Channels with Channel State Information
Capacity-achieving Feedback Scheme for Flat Fading Channels with Channel State Information Jialing Liu liujl@iastate.edu Sekhar Tatikonda sekhar.tatikonda@yale.edu Nicola Elia nelia@iastate.edu Dept. of
More informationLecture 2. Capacity of the Gaussian channel
Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN
More informationEquivalence for Networks with Adversarial State
Equivalence for Networks with Adversarial State Oliver Kosut Department of Electrical, Computer and Energy Engineering Arizona State University Tempe, AZ 85287 Email: okosut@asu.edu Jörg Kliewer Department
More informationUniversal Anytime Codes: An approach to uncertain channels in control
Universal Anytime Codes: An approach to uncertain channels in control paper by Stark Draper and Anant Sahai presented by Sekhar Tatikonda Wireless Foundations Department of Electrical Engineering and Computer
More informationVariable Length Codes for Degraded Broadcast Channels
Variable Length Codes for Degraded Broadcast Channels Stéphane Musy School of Computer and Communication Sciences, EPFL CH-1015 Lausanne, Switzerland Email: stephane.musy@ep.ch Abstract This paper investigates
More informationThe Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels
The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels Parastoo Sadeghi National ICT Australia (NICTA) Sydney NSW 252 Australia Email: parastoo@student.unsw.edu.au Predrag
More informationBinary Transmissions over Additive Gaussian Noise: A Closed-Form Expression for the Channel Capacity 1
5 Conference on Information Sciences and Systems, The Johns Hopkins University, March 6 8, 5 inary Transmissions over Additive Gaussian Noise: A Closed-Form Expression for the Channel Capacity Ahmed O.
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationPolar codes for the m-user MAC and matroids
Research Collection Conference Paper Polar codes for the m-user MAC and matroids Author(s): Abbe, Emmanuel; Telatar, Emre Publication Date: 2010 Permanent Link: https://doi.org/10.3929/ethz-a-005997169
More informationCapacity of the Discrete Memoryless Energy Harvesting Channel with Side Information
204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener
More informationOn the Secrecy Capacity of Fading Channels
On the Secrecy Capacity of Fading Channels arxiv:cs/63v [cs.it] 7 Oct 26 Praveen Kumar Gopala, Lifeng Lai and Hesham El Gamal Department of Electrical and Computer Engineering The Ohio State University
More informationRefined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities
Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Maxim Raginsky and Igal Sason ISIT 2013, Istanbul, Turkey Capacity-Achieving Channel Codes The set-up DMC
More informationAchievable Rates for Probabilistic Shaping
Achievable Rates for Probabilistic Shaping Georg Böcherer Mathematical and Algorithmic Sciences Lab Huawei Technologies France S.A.S.U. georg.boecherer@ieee.org May 3, 08 arxiv:707.034v5 cs.it May 08 For
More informationTwo Applications of the Gaussian Poincaré Inequality in the Shannon Theory
Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Vincent Y. F. Tan (Joint work with Silas L. Fong) National University of Singapore (NUS) 2016 International Zurich Seminar on
More informationOn the Shamai-Laroia Approximation for the Information Rate of the ISI Channel
On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel Yair Carmon and Shlomo Shamai (Shitz) Department of Electrical Engineering, Technion - Israel Institute of Technology 2014
More informationCapacity Pre-Log of SIMO Correlated Block-Fading Channels
Capacity Pre-Log of SIMO Correlated Block-Fading Channels Wei Yang, Giuseppe Durisi, Veniamin I. Morgenshtern, Erwin Riegler 3 Chalmers University of Technology, 496 Gothenburg, Sweden ETH Zurich, 809
More informationA Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback
A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback Vincent Y. F. Tan (NUS) Joint work with Silas L. Fong (Toronto) 2017 Information Theory Workshop, Kaohsiung,
More informationLecture 6 I. CHANNEL CODING. X n (m) P Y X
6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder
More informationApproximate Capacity of Fast Fading Interference Channels with no CSIT
Approximate Capacity of Fast Fading Interference Channels with no CSIT Joyson Sebastian, Can Karakus, Suhas Diggavi Abstract We develop a characterization of fading models, which assigns a number called
More informationInformation Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST
Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it
More informationFinite-Blocklength Bounds on the Maximum Coding Rate of Rician Fading Channels with Applications to Pilot-Assisted Transmission
Finite-Blocklength Bounds on the Maximum Coding Rate of Rician Fading Channels with Applications to Pilot-Assisted Transmission Johan Östman, Giuseppe Durisi, and Erik G. Ström Dept. of Signals and Systems,
More informationProblem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.
(b) We note that the above capacity is the same to the capacity of the binary symmetric channel. Indeed, if we considerthe grouping of the output symbols into a = {y 1,y 2 } and b = {y 3,y 4 } we get a
More informationA Mathematical Framework for Robust Control Over Uncertain Communication Channels
A Mathematical Framework for Robust Control Over Uncertain Communication Channels Charalambos D. Charalambous and Alireza Farhadi Abstract In this paper, a mathematical framework for studying robust control
More informationInteractions of Information Theory and Estimation in Single- and Multi-user Communications
Interactions of Information Theory and Estimation in Single- and Multi-user Communications Dongning Guo Department of Electrical Engineering Princeton University March 8, 2004 p 1 Dongning Guo Communications
More informationOn the Capacity of Free-Space Optical Intensity Channels
On the Capacity of Free-Space Optical Intensity Channels Amos Lapidoth TH Zurich Zurich, Switzerl mail: lapidoth@isi.ee.ethz.ch Stefan M. Moser National Chiao Tung University NCTU Hsinchu, Taiwan mail:
More informationThe Robustness of Dirty Paper Coding and The Binary Dirty Multiple Access Channel with Common Interference
The and The Binary Dirty Multiple Access Channel with Common Interference Dept. EE - Systems, Tel Aviv University, Tel Aviv, Israel April 25th, 2010 M.Sc. Presentation The B/G Model Compound CSI Smart
More informationLecture 4 Channel Coding
Capacity and the Weak Converse Lecture 4 Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 15, 2014 1 / 16 I-Hsiang Wang NIT Lecture 4 Capacity
More informationOn the Optimality of Treating Interference as Noise in Competitive Scenarios
On the Optimality of Treating Interference as Noise in Competitive Scenarios A. DYTSO, D. TUNINETTI, N. DEVROYE WORK PARTIALLY FUNDED BY NSF UNDER AWARD 1017436 OUTLINE CHANNEL MODEL AND PAST WORK ADVANTAGES
More informationMultiuser Capacity in Block Fading Channel
Multiuser Capacity in Block Fading Channel April 2003 1 Introduction and Model We use a block-fading model, with coherence interval T where M independent users simultaneously transmit to a single receiver
More informationChapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye
Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for
More informationShannon s noisy-channel theorem
Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for
More informationJoint FEC Encoder and Linear Precoder Design for MIMO Systems with Antenna Correlation
Joint FEC Encoder and Linear Precoder Design for MIMO Systems with Antenna Correlation Chongbin Xu, Peng Wang, Zhonghao Zhang, and Li Ping City University of Hong Kong 1 Outline Background Mutual Information
More informationLECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem
LECTURE 15 Last time: Feedback channel: setting up the problem Perfect feedback Feedback capacity Data compression Lecture outline Joint source and channel coding theorem Converse Robustness Brain teaser
More informationChannel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter
Channel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter Bradford D. Boyle, John MacLaren Walsh, and Steven Weber Modeling & Analysis of Networks Laboratory
More informationTHIS paper constitutes a systematic study of the capacity
4 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 1, JANUARY 2002 Capacities of TimeVarying MultipleAccess Channels With Side Information Arnab Das Prakash Narayan, Fellow, IEEE Abstract We determine
More information3052 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 7, JULY /$ IEEE
3052 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 52, NO 7, JULY 2006 Capacity of Nearly Decomposable Markovian Fading Channels Under Asymmetric Receiver Sender Side Information Muriel Médard, Senior Member,
More informationInformation Theoretic Limits of Randomness Generation
Information Theoretic Limits of Randomness Generation Abbas El Gamal Stanford University Shannon Centennial, University of Michigan, September 2016 Information theory The fundamental problem of communication
More informationCommon Information. Abbas El Gamal. Stanford University. Viterbi Lecture, USC, April 2014
Common Information Abbas El Gamal Stanford University Viterbi Lecture, USC, April 2014 Andrew Viterbi s Fabulous Formula, IEEE Spectrum, 2010 El Gamal (Stanford University) Disclaimer Viterbi Lecture 2
More informationLECTURE 13. Last time: Lecture outline
LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to
More informationNUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS
NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS Justin Dauwels Dept. of Information Technology and Electrical Engineering ETH, CH-8092 Zürich, Switzerland dauwels@isi.ee.ethz.ch
More informationOptimal Power Control for LDPC Codes in Block-Fading Channels
IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 59, NO. 7, JULY 2011 1759 Optimal Power Control for LDPC Codes in Block-Fading Channels Gottfried Lechner, Khoa D. Nguyen, Albert Guillén i Fàbregas, and Lars
More informationLECTURE 3. Last time:
LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate
More informationOptimal Power Control in Decentralized Gaussian Multiple Access Channels
1 Optimal Power Control in Decentralized Gaussian Multiple Access Channels Kamal Singh Department of Electrical Engineering Indian Institute of Technology Bombay. arxiv:1711.08272v1 [eess.sp] 21 Nov 2017
More informationA Summary of Multiple Access Channels
A Summary of Multiple Access Channels Wenyi Zhang February 24, 2003 Abstract In this summary we attempt to present a brief overview of the classical results on the information-theoretic aspects of multiple
More information(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute
ENEE 739C: Advanced Topics in Signal Processing: Coding Theory Instructor: Alexander Barg Lecture 6 (draft; 9/6/03. Error exponents for Discrete Memoryless Channels http://www.enee.umd.edu/ abarg/enee739c/course.html
More informationWideband Fading Channel Capacity with Training and Partial Feedback
Wideband Fading Channel Capacity with Training and Partial Feedback Manish Agarwal, Michael L. Honig ECE Department, Northwestern University 145 Sheridan Road, Evanston, IL 6008 USA {m-agarwal,mh}@northwestern.edu
More informationSource-Channel Coding Theorems for the Multiple-Access Relay Channel
Source-Channel Coding Theorems for the Multiple-Access Relay Channel Yonathan Murin, Ron Dabora, and Deniz Gündüz Abstract We study reliable transmission of arbitrarily correlated sources over multiple-access
More informationOn the Capacity of the Interference Channel with a Relay
On the Capacity of the Interference Channel with a Relay Ivana Marić, Ron Dabora and Andrea Goldsmith Stanford University, Stanford, CA {ivanam,ron,andrea}@wsl.stanford.edu Abstract Capacity gains due
More informationGraph-based Codes for Quantize-Map-and-Forward Relaying
20 IEEE Information Theory Workshop Graph-based Codes for Quantize-Map-and-Forward Relaying Ayan Sengupta, Siddhartha Brahma, Ayfer Özgür, Christina Fragouli and Suhas Diggavi EPFL, Switzerland, UCLA,
More informationSHANNON [34] showed that the capacity of a memoryless. A Coding Theorem for a Class of Stationary Channels With Feedback Young-Han Kim, Member, IEEE
1488 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 54, NO 4, APRIL 2008 A Coding Theorem for a Cls of Stationary Channels With Feedback Young-Han Kim, Member, IEEE Abstract A coding theorem is proved for
More informationPrinciples of Communications
Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @
More informationAn Achievable Error Exponent for the Mismatched Multiple-Access Channel
An Achievable Error Exponent for the Mismatched Multiple-Access Channel Jonathan Scarlett University of Cambridge jms265@camacuk Albert Guillén i Fàbregas ICREA & Universitat Pompeu Fabra University of
More informationOn Comparability of Multiple Antenna Channels
On Comparability of Multiple Antenna Channels Majid Fozunbal, Steven W. McLaughlin, and Ronald W. Schafer School of Electrical and Computer Engineering Georgia Institute of Technology Atlanta, GA 30332-0250
More informationChannel Dispersion and Moderate Deviations Limits for Memoryless Channels
Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Yury Polyanskiy and Sergio Verdú Abstract Recently, Altug and Wagner ] posed a question regarding the optimal behavior of the probability
More informationTitle. Author(s)Tsai, Shang-Ho. Issue Date Doc URL. Type. Note. File Information. Equal Gain Beamforming in Rayleigh Fading Channels
Title Equal Gain Beamforming in Rayleigh Fading Channels Author(s)Tsai, Shang-Ho Proceedings : APSIPA ASC 29 : Asia-Pacific Signal Citationand Conference: 688-691 Issue Date 29-1-4 Doc URL http://hdl.handle.net/2115/39789
More informationOn the Low-SNR Capacity of Phase-Shift Keying with Hard-Decision Detection
On the Low-SNR Capacity of Phase-Shift Keying with Hard-Decision Detection ustafa Cenk Gursoy Department of Electrical Engineering University of Nebraska-Lincoln, Lincoln, NE 68588 Email: gursoy@engr.unl.edu
More informationOn Two-user Fading Gaussian Broadcast Channels. with Perfect Channel State Information at the Receivers. Daniela Tuninetti
DIMACS Workshop on Network Information Theory - March 2003 Daniela Tuninetti 1 On Two-user Fading Gaussian Broadcast Channels with Perfect Channel State Information at the Receivers Daniela Tuninetti Mobile
More informationII. THE TWO-WAY TWO-RELAY CHANNEL
An Achievable Rate Region for the Two-Way Two-Relay Channel Jonathan Ponniah Liang-Liang Xie Department of Electrical Computer Engineering, University of Waterloo, Canada Abstract We propose an achievable
More informationCapacity of AWGN channels
Chapter 3 Capacity of AWGN channels In this chapter we prove that the capacity of an AWGN channel with bandwidth W and signal-tonoise ratio SNR is W log 2 (1+SNR) bits per second (b/s). The proof that
More informationStrong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach
Strong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach Silas Fong (Joint work with Vincent Tan) Department of Electrical & Computer Engineering National University
More informationthe station can send one of t code packets, transmitting one binary symbol for the time unit. Let > 0 be a xed parameter. Put p = =M, 0 < p < 1. Suppo
Superimposed Codes and ALOHA system Arkady G. D'yachkov, Vyacheslav V. Rykov Moscow State University, Faculty of Mechanics and Mathematics, Department of Probability Theory, Moscow, 119899, Russia, Email:
More informationCapacity Pre-log of Noncoherent SIMO Channels via Hironaka s Theorem
Capacity Pre-log of Noncoherent SIMO Channels via Hironaka s Theorem Veniamin I. Morgenshtern 22. May 2012 Joint work with E. Riegler, W. Yang, G. Durisi, S. Lin, B. Sturmfels, and H. Bőlcskei SISO Fading
More informationA Systematic Approach for Interference Alignment in CSIT-less Relay-Aided X-Networks
A Systematic Approach for Interference Alignment in CSIT-less Relay-Aided X-Networks Daniel Frank, Karlheinz Ochs, Aydin Sezgin Chair of Communication Systems RUB, Germany Email: {danielfrank, karlheinzochs,
More informationLECTURE 10. Last time: Lecture outline
LECTURE 10 Joint AEP Coding Theorem Last time: Error Exponents Lecture outline Strong Coding Theorem Reading: Gallager, Chapter 5. Review Joint AEP A ( ɛ n) (X) A ( ɛ n) (Y ) vs. A ( ɛ n) (X, Y ) 2 nh(x)
More informationApproximately achieving the feedback interference channel capacity with point-to-point codes
Approximately achieving the feedback interference channel capacity with point-to-point codes Joyson Sebastian*, Can Karakus*, Suhas Diggavi* Abstract Superposition codes with rate-splitting have been used
More informationAnalytical Bounds on Maximum-Likelihood Decoded Linear Codes: An Overview
Analytical Bounds on Maximum-Likelihood Decoded Linear Codes: An Overview Igal Sason Department of Electrical Engineering, Technion Haifa 32000, Israel Sason@ee.technion.ac.il December 21, 2004 Background
More informationCapacity and Reliability Function for Small Peak Signal Constraints
828 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 4, APRIL 2002 Capacity and Reliability Function for Small Peak Signal Constraints Bruce Hajek, Fellow, IEEE, and Vijay G. Subramanian, Member,
More informationSecond-Order Asymptotics in Information Theory
Second-Order Asymptotics in Information Theory Vincent Y. F. Tan (vtan@nus.edu.sg) Dept. of ECE and Dept. of Mathematics National University of Singapore (NUS) National Taiwan University November 2015
More informationUpper Bounds on the Capacity of Binary Intermittent Communication
Upper Bounds on the Capacity of Binary Intermittent Communication Mostafa Khoshnevisan and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame, Indiana 46556 Email:{mhoshne,
More informationAchieving Shannon Capacity Region as Secrecy Rate Region in a Multiple Access Wiretap Channel
Achieving Shannon Capacity Region as Secrecy Rate Region in a Multiple Access Wiretap Channel Shahid Mehraj Shah and Vinod Sharma Department of Electrical Communication Engineering, Indian Institute of
More informationA tool oriented approach to network capacity. Ralf Koetter Michelle Effros Muriel Medard
A tool oriented approach to network capacity Ralf Koetter Michelle Effros Muriel Medard ralf.koetter@tum.de effros@caltech.edu medard@mit.edu The main theorem of NC [Ahlswede, Cai, Li Young, 2001] Links
More informationMultiple-Input Multiple-Output Systems
Multiple-Input Multiple-Output Systems What is the best way to use antenna arrays? MIMO! This is a totally new approach ( paradigm ) to wireless communications, which has been discovered in 95-96. Performance
More information