Cover s Open Problem: The Capacity of the Relay Channel

Size: px
Start display at page:

Download "Cover s Open Problem: The Capacity of the Relay Channel"

Transcription

1 Cover s Open Problem: The Capacity of the Relay Channel Ayfer Özgür Stanford University Advanced Networks Colloquia Series University of Maryland, March 2017 Joint work with Xiugang Wu and Leighton Pate Barnes. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 1 / 26

2 Father of the Information Age Claude Shannon ( ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 2 / 26

3 The Bell System Technical Journal Vol. XXVII J Illy, 1948 No.3 A Mathematical Theory of Communication By c. E. SHANNON IXTRODUCTION HE recent development of various methods of modulation such as rem T and PPM which exchange bandwidth for signal-to-noise ratio has intensified the interest in a general theory of communication. A basis for such a theory is contained in the important papers of Nyquist! and Hartley" on this subject. In the present paper we will extend the theory to include a number of new factors, in particular the effect of noise in the channel, and Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 3 / 26

4 A method is developed for representing any communication system geometrically... Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 4 / 26

5 AWGN Channel P Transmitter Receiver Capacity ( C = log 1 + P ) N Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 5 / 26

6 Converse: Sphere Packing p nn X n (1) X n (2) p nn 0 p n(p + N) Y sphere X n (3) p nn noise sphere ( n(p ) Sphere + N) # of X n. 2 n log 2πe(P+N) 2 ( nn ) = Sphere 2 n 2 log 2πeN = 2 n 2 log(1+ P N ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 6 / 26

7 Achievability: Geometric Random Coding Pr( false X n Lens ) ( np ) 2 nr Sphere. = 2 n 2 2 n 2 PN log 2πe P+N 2nR log 2πeP = 2 n 2 log(1+ P N ) 2 nr Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 7 / 26

8 The story goes... Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 8 / 26

9 Cover s Open Problem Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 9 / 26

10 Gaussian case W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Achievability: C 0 =. C( ) = 1 2 log ( 1 + 2P N Destination Y n Cutset-bound (Cover and El Gamal 79): C0 1 ( 2 log 1 + 2P ) 12 ( N log 1 + P ). N ) Potentially, C 0 0 as P/N 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

11 Main Result W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Destination Y n Theorem C 0 = Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

12 Upper Bound on the Capacity Cut-set bound C-F Old bound New bound SNR = 15 db C 0 (bit/channel use) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

13 Cutset Bound W 1 N(0,N) W 2 N(0,N) Z n C 0 Relay I n = f(z n ) Source X n Destination Y n nr I (X n ; Y n, I n ) + nɛ n = I (X n ; Y n ) + I (X n ; I n Y n ) + nɛ n = I (X n ; Y n ) + H(I n Y n ) H(I }{{} n Y n, X n ) +nɛ }{{} n nc 0 0 n(i (X ; Y ) + C 0 + ɛ n ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

14 Cutset Bound W 1 N(0,N) W 2 N(0,N) Z n C 0 Relay I n = f(z n ) Source X n Destination Y n nr I (X n ; Y n, I n ) + nɛ n = I (X n ; Y n ) + I (X n ; I n Y n ) + nɛ n = I (X n ; Y n ) + H(I n Y n ) H(I }{{} n X n ) +nɛ }{{} n nc 0 0 n(i (X ; Y ) + C 0 + ɛ n ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

15 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

16 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin If H(I n X n ) = 0, Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

17 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin If H(I n X n ) = 0, then H(I n Y n ) = 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

18 W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Destination Y n R I (X n ; Y n ) + H(I n Y n ) H(I }{{} n X n ) +nɛ }{{} n? = n log sin θ n Goal: I n = f (Z n ) Z n X n n Y }{{} H(I n X n )= n log sin θ n } {{ } H(I n Y n )? Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

19 H(I n X n ) 0 Typical set of Z n p nn X n Multiple bins # of bins =? P(each bin) =? Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

20 From n- to nb- Dimensional Space X, Y, Z, I : B-length i.i.d. from {(X n (b), Y n (b), Z n (b), I n (b))} B b=1. If H(I n X n ) = n log sin θ n, then for any typical (x, i) p(i x). = 2 nb log sin θn, P(Z A(i) x) =. nb log sin θn 2 p nbnx ith bin A x (i) Pr. =2 nb log sin n A x (i). =2 nb( 1 2 log 2 en sin2 n) Typical set of Z/Y Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

21 Isoperimetric Inequalities Isoperimetric Inequality in the Plane (Steiner 1838) Among all closed curves in the plane with a given enclosed area, the circle has the smallest perimeter. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

22 Isoperimetric Inequalities Isoperimetric Inequality on the Sphere (Levy 1919) Among all sets on the sphere with a given volume, the spherical cap has the smallest boundary, or the smallest volume of ω-neighborhood for any ω > 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

23 Blowing-up Lemma A x (i). =2 nb( 1 2 log 2 en sin2 n) Isoperimetric Inequality on the Sphere + Measure Concentration: P(Z blow-up of A x (i) x) 1. P(Y blow-up of A x (i) x) 1. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

24 Geometry of Typical Sets n-dimensional space: Almost all (X n, Y n, Z n, I n ) Z n! I n X n p nn Y n nb-dimensional space: Almost all (x, y, i) p nbn z! i x 2 n y Information Inequality: (Wu and Ozgur, 2015) H(I n Y n ) n(2 log sin θ n + 2 log sin θ n ln 2 log e). Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

25 A new approach A x (i) A x (i) Y Y Control the intersection of a sphere drawn around a randomly chosen Y and A x (i). Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

26 Easy if A x (i) is a spherical cap Cap(z 0, n) z 0 Cap(z 0, n) \ Cap(y 0,! n) p nbn x n! n y 0 Cap(y 0,! n) Cap(z 0, θ n ) Cap(Y, ω n ). = 2 nb( 1 2 log 2πeN(sin2 θ n cos 2 ω n)) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

27 Strengthening of the Isoperimetric Inequality Strengthening of the Isoperimetric Inequality: Among all sets on the sphere with a given volume, the spherical cap has minimal intersection volume at distance ω for almost all points on the sphere for any ω > π/2 θ. Proof: builds on the Riesz rearrangement inequality. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

28 A Packing Argument Given Y, # of I ( ) Sphere Y, nbn4sin 2 ωn 2 2 nb( 1 2 log 2πeN(sin2 θ n cos 2 ω n)) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

29 Summary Solved an problem posed by Cover and named The Capacity of Relay Channel in Open Problems in Communication and Computation, Springer-Verlag, Developed a converse technique that significantly deviates from standard converse techniques based on single-letterization and has some new ingredients: Typicality Measure Concentration Isoperimetric Inequality Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26

Improving on the Cutset Bound via a Geometric Analysis of Typical Sets

Improving on the Cutset Bound via a Geometric Analysis of Typical Sets Imroving on the Cutset Bound via a Geometric Analysis of Tyical Sets Ayfer Özgür Stanford University CUHK, May 28, 2016 Joint work with Xiugang Wu (Stanford). Ayfer Özgür (Stanford) March 16 1 / 33 Gaussian

More information

3F1 Information Theory, Lecture 1

3F1 Information Theory, Lecture 1 3F1 Information Theory, Lecture 1 Jossy Sayir Department of Engineering Michaelmas 2013, 22 November 2013 Organisation History Entropy Mutual Information 2 / 18 Course Organisation 4 lectures Course material:

More information

Lecture 18: Gaussian Channel

Lecture 18: Gaussian Channel Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400

More information

Steiner s formula and large deviations theory

Steiner s formula and large deviations theory Steiner s formula and large deviations theory Venkat Anantharam EECS Department University of California, Berkeley May 19, 2015 Simons Conference on Networks and Stochastic Geometry Blanton Museum of Art

More information

MMSE estimation and lattice encoding/decoding for linear Gaussian channels. Todd P. Coleman /22/02

MMSE estimation and lattice encoding/decoding for linear Gaussian channels. Todd P. Coleman /22/02 MMSE estimation and lattice encoding/decoding for linear Gaussian channels Todd P. Coleman 6.454 9/22/02 Background: the AWGN Channel Y = X + N where N N ( 0, σ 2 N ), 1 n ni=1 X 2 i P X. Shannon: capacity

More information

ECE Information theory Final (Fall 2008)

ECE Information theory Final (Fall 2008) ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1

More information

On the Rate-Limited Gelfand-Pinsker Problem

On the Rate-Limited Gelfand-Pinsker Problem On the Rate-Limited Gelfand-Pinsker Problem Ravi Tandon Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 74 ravit@umd.edu ulukus@umd.edu Abstract

More information

Lecture 2. Capacity of the Gaussian channel

Lecture 2. Capacity of the Gaussian channel Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN

More information

Approximately achieving Gaussian relay. network capacity with lattice-based QMF codes

Approximately achieving Gaussian relay. network capacity with lattice-based QMF codes Approximately achieving Gaussian relay 1 network capacity with lattice-based QMF codes Ayfer Özgür and Suhas Diggavi Abstract In [1], a new relaying strategy, quantize-map-and-forward QMF scheme, has been

More information

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results Information Theory Lecture 10 Network Information Theory (CT15); a focus on channel capacity results The (two-user) multiple access channel (15.3) The (two-user) broadcast channel (15.6) The relay channel

More information

Interactive Decoding of a Broadcast Message

Interactive Decoding of a Broadcast Message In Proc. Allerton Conf. Commun., Contr., Computing, (Illinois), Oct. 2003 Interactive Decoding of a Broadcast Message Stark C. Draper Brendan J. Frey Frank R. Kschischang University of Toronto Toronto,

More information

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Vincent Y. F. Tan (Joint work with Silas L. Fong) National University of Singapore (NUS) 2016 International Zurich Seminar on

More information

Lecture 15: Thu Feb 28, 2019

Lecture 15: Thu Feb 28, 2019 Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:

More information

Leech Constellations of Construction-A Lattices

Leech Constellations of Construction-A Lattices Leech Constellations of Construction-A Lattices Joseph J. Boutros Talk at Nokia Bell Labs, Stuttgart Texas A&M University at Qatar In collaboration with Nicola di Pietro. March 7, 2017 Thanks Many Thanks

More information

Lecture 4 Capacity of Wireless Channels

Lecture 4 Capacity of Wireless Channels Lecture 4 Capacity of Wireless Channels I-Hsiang Wang ihwang@ntu.edu.tw 3/0, 014 What we have learned So far: looked at specific schemes and techniques Lecture : point-to-point wireless channel - Diversity:

More information

II. THE TWO-WAY TWO-RELAY CHANNEL

II. THE TWO-WAY TWO-RELAY CHANNEL An Achievable Rate Region for the Two-Way Two-Relay Channel Jonathan Ponniah Liang-Liang Xie Department of Electrical Computer Engineering, University of Waterloo, Canada Abstract We propose an achievable

More information

Channel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter

Channel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter Channel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter Bradford D. Boyle, John MacLaren Walsh, and Steven Weber Modeling & Analysis of Networks Laboratory

More information

On the Capacity Region of the Gaussian Z-channel

On the Capacity Region of the Gaussian Z-channel On the Capacity Region of the Gaussian Z-channel Nan Liu Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 74 nkancy@eng.umd.edu ulukus@eng.umd.edu

More information

On Multiple User Channels with State Information at the Transmitters

On Multiple User Channels with State Information at the Transmitters On Multiple User Channels with State Information at the Transmitters Styrmir Sigurjónsson and Young-Han Kim* Information Systems Laboratory Stanford University Stanford, CA 94305, USA Email: {styrmir,yhk}@stanford.edu

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

Cooperative Communication with Feedback via Stochastic Approximation

Cooperative Communication with Feedback via Stochastic Approximation Cooperative Communication with Feedback via Stochastic Approximation Utsaw Kumar J Nicholas Laneman and Vijay Gupta Department of Electrical Engineering University of Notre Dame Email: {ukumar jnl vgupta}@ndedu

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

Appendix B Information theory from first principles

Appendix B Information theory from first principles Appendix B Information theory from first principles This appendix discusses the information theory behind the capacity expressions used in the book. Section 8.3.4 is the only part of the book that supposes

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

Dirty Paper Writing and Watermarking Applications

Dirty Paper Writing and Watermarking Applications Dirty Paper Writing and Watermarking Applications G.RaviKiran February 10, 2003 1 Introduction Following an underlying theme in Communication is the duo of Dirty Paper Writing and Watermarking. In 1983

More information

The Shannon s basic inequalities refer to the following fundamental properties of entropy function:

The Shannon s basic inequalities refer to the following fundamental properties of entropy function: COMMUNICATIONS IN INFORMATION AND SYSTEMS c 2003 International Press Vol. 3, No. 1, pp. 47-60, June 2003 004 ON A NEW NON-SHANNON TYPE INFORMATION INEQUALITY ZHEN ZHANG Abstract. Recently, K. Makarychev,

More information

Multicoding Schemes for Interference Channels

Multicoding Schemes for Interference Channels Multicoding Schemes for Interference Channels 1 Ritesh Kolte, Ayfer Özgür, Haim Permuter Abstract arxiv:1502.04273v1 [cs.it] 15 Feb 2015 The best known inner bound for the 2-user discrete memoryless interference

More information

Fat Struts: Constructions and a Bound

Fat Struts: Constructions and a Bound Fat Struts: Constructions and a Bound N. J. A. Sloane, Vinay A. Vaishampayan, Sueli I. R. Costa AT&T Shannon Labs, Florham Park, NJ 0793, USA. Email: {njas,vinay}@research.att.com University of Campinas,

More information

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:

More information

National University of Singapore Department of Electrical & Computer Engineering. Examination for

National University of Singapore Department of Electrical & Computer Engineering. Examination for National University of Singapore Department of Electrical & Computer Engineering Examination for EE5139R Information Theory for Communication Systems (Semester I, 2014/15) November/December 2014 Time Allowed:

More information

Shannon s noisy-channel theorem

Shannon s noisy-channel theorem Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for

More information

An Achievable Rate for the Multiple Level Relay Channel

An Achievable Rate for the Multiple Level Relay Channel An Achievable Rate for the Multiple Level Relay Channel Liang-Liang Xie and P. R. Kumar Department of Electrical and Computer Engineering, and Coordinated Science Laboratory University of Illinois, Urbana-Champaign

More information

Lattices and Lattice Codes

Lattices and Lattice Codes Lattices and Lattice Codes Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology Hyderabad

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

Revision of Lecture 5

Revision of Lecture 5 Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information

More information

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic

More information

Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007

Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007 Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT ECE 559 Presentation Hoa Pham Dec 3, 2007 Introduction MIMO systems provide two types of gains Diversity Gain: each path from a transmitter

More information

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,

More information

Some Goodness Properties of LDA Lattices

Some Goodness Properties of LDA Lattices Some Goodness Properties of LDA Lattices Shashank Vatedka and Navin Kashyap {shashank,nkashyap}@eceiiscernetin Department of ECE Indian Institute of Science Bangalore, India Information Theory Workshop

More information

Covert Communication with Channel-State Information at the Transmitter

Covert Communication with Channel-State Information at the Transmitter Covert Communication with Channel-State Information at the Transmitter Si-Hyeon Lee Joint Work with Ligong Wang, Ashish Khisti, and Gregory W. Wornell July 27, 2017 1 / 21 Covert Communication Transmitter

More information

The Gallager Converse

The Gallager Converse The Gallager Converse Abbas El Gamal Director, Information Systems Laboratory Department of Electrical Engineering Stanford University Gallager s 75th Birthday 1 Information Theoretic Limits Establishing

More information

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING Yichuan Hu (), Javier Garcia-Frias () () Dept. of Elec. and Comp. Engineering University of Delaware Newark, DE

More information

Lecture 6 Channel Coding over Continuous Channels

Lecture 6 Channel Coding over Continuous Channels Lecture 6 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 9, 015 1 / 59 I-Hsiang Wang IT Lecture 6 We have

More information

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR 1 Stefano Rini, Daniela Tuninetti and Natasha Devroye srini2, danielat, devroye @ece.uic.edu University of Illinois at Chicago Abstract

More information

A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels

A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels Mehdi Mohseni Department of Electrical Engineering Stanford University Stanford, CA 94305, USA Email: mmohseni@stanford.edu

More information

The Method of Types and Its Application to Information Hiding

The Method of Types and Its Application to Information Hiding The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,

More information

Optimal Power Control in Decentralized Gaussian Multiple Access Channels

Optimal Power Control in Decentralized Gaussian Multiple Access Channels 1 Optimal Power Control in Decentralized Gaussian Multiple Access Channels Kamal Singh Department of Electrical Engineering Indian Institute of Technology Bombay. arxiv:1711.08272v1 [eess.sp] 21 Nov 2017

More information

Lecture 4. Capacity of Fading Channels

Lecture 4. Capacity of Fading Channels 1 Lecture 4. Capacity of Fading Channels Capacity of AWGN Channels Capacity of Fading Channels Ergodic Capacity Outage Capacity Shannon and Information Theory Claude Elwood Shannon (April 3, 1916 February

More information

IET Commun., 2009, Vol. 3, Iss. 4, pp doi: /iet-com & The Institution of Engineering and Technology 2009

IET Commun., 2009, Vol. 3, Iss. 4, pp doi: /iet-com & The Institution of Engineering and Technology 2009 Published in IET Communications Received on 10th March 2008 Revised on 10th July 2008 ISSN 1751-8628 Comprehensive partial decoding approach for two-level relay networks L. Ghabeli M.R. Aref Information

More information

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless

More information

Lecture 3: Channel Capacity

Lecture 3: Channel Capacity Lecture 3: Channel Capacity 1 Definitions Channel capacity is a measure of maximum information per channel usage one can get through a channel. This one of the fundamental concepts in information theory.

More information

The Role of Directed Information in Network Capacity

The Role of Directed Information in Network Capacity The Role of Directed Information in Network Capacity Sudeep Kamath 1 and Young-Han Kim 2 1 Department of Electrical Engineering Princeton University 2 Department of Electrical and Computer Engineering

More information

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the

More information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information 204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener

More information

On the Duality between Multiple-Access Codes and Computation Codes

On the Duality between Multiple-Access Codes and Computation Codes On the Duality between Multiple-Access Codes and Computation Codes Jingge Zhu University of California, Berkeley jingge.zhu@berkeley.edu Sung Hoon Lim KIOST shlim@kiost.ac.kr Michael Gastpar EPFL michael.gastpar@epfl.ch

More information

PreCalculus: Chapter 9 Test Review

PreCalculus: Chapter 9 Test Review Name: Class: Date: ID: A PreCalculus: Chapter 9 Test Review Short Answer 1. Plot the point given in polar coordinates. 3. Plot the point given in polar coordinates. (-4, -225 ) 2. Plot the point given

More information

LECTURE 13. Last time: Lecture outline

LECTURE 13. Last time: Lecture outline LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to

More information

Capacity of AWGN channels

Capacity of AWGN channels Chapter 3 Capacity of AWGN channels In this chapter we prove that the capacity of an AWGN channel with bandwidth W and signal-tonoise ratio SNR is W log 2 (1+SNR) bits per second (b/s). The proof that

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017 U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 3, 207 Scribed by Keyhan Vakil Lecture 3 In which we complete the study of Independent Set and Max Cut in G n,p random graphs.

More information

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm 1. Feedback does not increase the capacity. Consider a channel with feedback. We assume that all the recieved outputs are sent back immediately

More information

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 6 I. CHANNEL CODING. X n (m) P Y X 6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder

More information

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions 392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6 Problem Set 2 Solutions Problem 2.1 (Cartesian-product constellations) (a) Show that if A is a K-fold Cartesian

More information

A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback

A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback Vincent Y. F. Tan (NUS) Joint work with Silas L. Fong (Toronto) 2017 Information Theory Workshop, Kaohsiung,

More information

Covering a sphere with caps: Rogers bound revisited

Covering a sphere with caps: Rogers bound revisited Covering a sphere with caps: Rogers bound revisited Ilya Dumer Abstract We consider coverings of a sphere S n r of radius r with the balls of radius one in an n-dimensional Euclidean space R n. Our goal

More information

Channel Dispersion and Moderate Deviations Limits for Memoryless Channels

Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Yury Polyanskiy and Sergio Verdú Abstract Recently, Altug and Wagner ] posed a question regarding the optimal behavior of the probability

More information

Wireless Communications Lecture 10

Wireless Communications Lecture 10 Wireless Communications Lecture 1 [SNR per symbol and SNR per bit] SNR = P R N B = E s N BT s = E b N BT b For BPSK: T b = T s, E b = E s, and T s = 1/B. Raised cosine pulse shaper for other pulses. T

More information

The Microcanonical Approach. (a) The volume of accessible phase space for a given total energy is proportional to. dq 1 dq 2 dq N dp 1 dp 2 dp N,

The Microcanonical Approach. (a) The volume of accessible phase space for a given total energy is proportional to. dq 1 dq 2 dq N dp 1 dp 2 dp N, 8333: Statistical Mechanics I Problem Set # 6 Solutions Fall 003 Classical Harmonic Oscillators: The Microcanonical Approach a The volume of accessible phase space for a given total energy is proportional

More information

Lecture 4 Capacity of Wireless Channels

Lecture 4 Capacity of Wireless Channels Lecture 4 Capacity of Wireless Channels I-Hsiang Wang ihwang@ntu.edu.tw 3/0, 014 What we have learned So far: looked at specific schemes and techniques Lecture : point-to-point wireless channel - Diversity:

More information

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture) ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling

More information

Energy State Amplification in an Energy Harvesting Communication System

Energy State Amplification in an Energy Harvesting Communication System Energy State Amplification in an Energy Harvesting Communication System Omur Ozel Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland College Park, MD 20742 omur@umd.edu

More information

Towards control over fading channels

Towards control over fading channels Towards control over fading channels Paolo Minero, Massimo Franceschetti Advanced Network Science University of California San Diego, CA, USA mail: {minero,massimo}@ucsd.edu Invited Paper) Subhrakanti

More information

Hierarchical Cooperation Achieves Optimal Capacity Scaling in Ad Hoc Networks

Hierarchical Cooperation Achieves Optimal Capacity Scaling in Ad Hoc Networks Hierarchical Cooperation Achieves Optimal Capacity Scaling in Ad Hoc Networks Ayfer Özgür, Olivier Lévêque Faculté Informatique et Communications Ecole Polytechnique Fédérale de Lausanne 05 Lausanne, Switzerland

More information

Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel

Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel Ahmad Abu Al Haija ECE Department, McGill University, Montreal, QC, Canada Email: ahmad.abualhaija@mail.mcgill.ca

More information

Space-Time Diversity Enhancements Using Collaborative Communications

Space-Time Diversity Enhancements Using Collaborative Communications IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. XX, NO. YY, MONTH 2005 1 Space-Time Diversity Enhancements Using Collaborative Communications Patrick Mitran, Student Member, IEEE, Hideki Ochiai, Member,

More information

COMBINING DECODED-AND-FORWARDED SIGNALS IN GAUSSIAN COOPERATIVE CHANNELS

COMBINING DECODED-AND-FORWARDED SIGNALS IN GAUSSIAN COOPERATIVE CHANNELS COMBINING DECODED-AND-FORWARDED SIGNALS IN GAUSSIAN COOPERATIVE CHANNELS Brice Djeumou, Samson Lasaulce, Andrew Klein To cite this version: Brice Djeumou, Samson Lasaulce, Andrew Klein. COMBINING DECODED-AND-FORWARDED

More information

Structured signal recovery from non-linear and heavy-tailed measurements

Structured signal recovery from non-linear and heavy-tailed measurements Structured signal recovery from non-linear and heavy-tailed measurements Larry Goldstein* Stanislav Minsker* Xiaohan Wei # *Department of Mathematics # Department of Electrical Engineering UniversityofSouthern

More information

Principles of Communications

Principles of Communications Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @

More information

On Two-user Fading Gaussian Broadcast Channels. with Perfect Channel State Information at the Receivers. Daniela Tuninetti

On Two-user Fading Gaussian Broadcast Channels. with Perfect Channel State Information at the Receivers. Daniela Tuninetti DIMACS Workshop on Network Information Theory - March 2003 Daniela Tuninetti 1 On Two-user Fading Gaussian Broadcast Channels with Perfect Channel State Information at the Receivers Daniela Tuninetti Mobile

More information

Lecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1]

Lecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1] ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture 7: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 3, 06 [Ed. Apr ] In last lecture, we studied the minimax

More information

Solutions to Homework Set #4 Differential Entropy and Gaussian Channel

Solutions to Homework Set #4 Differential Entropy and Gaussian Channel Solutions to Homework Set #4 Differential Entropy and Gaussian Channel 1. Differential entropy. Evaluate the differential entropy h(x = f lnf for the following: (a Find the entropy of the exponential density

More information

LECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs

LECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs LECTURE 18 Last time: White Gaussian noise Bandlimited WGN Additive White Gaussian Noise (AWGN) channel Capacity of AWGN channel Application: DS-CDMA systems Spreading Coding theorem Lecture outline Gaussian

More information

Gaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26

Gaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Gaussian channel Information theory 2013, lecture 6 Jens Sjölund 8 May 2013 Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Outline 1 Definitions 2 The coding theorem for Gaussian channel 3 Bandlimited

More information

Block 2: Introduction to Information Theory

Block 2: Introduction to Information Theory Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation

More information

On Capacity Under Received-Signal Constraints

On Capacity Under Received-Signal Constraints On Capacity Under Received-Signal Constraints Michael Gastpar Dept. of EECS, University of California, Berkeley, CA 9470-770 gastpar@berkeley.edu Abstract In a world where different systems have to share

More information

Harmonic Analysis Homework 5

Harmonic Analysis Homework 5 Harmonic Analysis Homework 5 Bruno Poggi Department of Mathematics, University of Minnesota November 4, 6 Notation Throughout, B, r is the ball of radius r with center in the understood metric space usually

More information

On the Secrecy Capacity of Fading Channels

On the Secrecy Capacity of Fading Channels On the Secrecy Capacity of Fading Channels arxiv:cs/63v [cs.it] 7 Oct 26 Praveen Kumar Gopala, Lifeng Lai and Hesham El Gamal Department of Electrical and Computer Engineering The Ohio State University

More information

Free-Space Optical Communications: Capacity Bounds, Approximations, and a New Sphere-Packing Perspective

Free-Space Optical Communications: Capacity Bounds, Approximations, and a New Sphere-Packing Perspective Free-Space Optical Communications: Capacity Bounds, pproximations, and a New Sphere-Packing Perspective nas Chaaban, Jean-Marie Morvan, and Mohammad Slim louini bstract The capacity of the intensity-modulation

More information

Lecture 14 February 28

Lecture 14 February 28 EE/Stats 376A: Information Theory Winter 07 Lecture 4 February 8 Lecturer: David Tse Scribe: Sagnik M, Vivek B 4 Outline Gaussian channel and capacity Information measures for continuous random variables

More information

Blind phase/frequency synchronization with low-precision ADC: a Bayesian approach

Blind phase/frequency synchronization with low-precision ADC: a Bayesian approach Blind phase/frequency synchronization with low-precision ADC: a Bayesian approach Aseem Wadhwa, Upamanyu Madhow Department of ECE, UCSB 1/26 Modern Communication Receiver Architecture Analog Digital TX

More information

Capacity Bounds for. the Gaussian Interference Channel

Capacity Bounds for. the Gaussian Interference Channel Capacity Bounds for the Gaussian Interference Channel Abolfazl S. Motahari, Student Member, IEEE, and Amir K. Khandani, Member, IEEE Coding & Signal Transmission Laboratory www.cst.uwaterloo.ca {abolfazl,khandani}@cst.uwaterloo.ca

More information

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources Wei Kang Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College

More information

Advanced Calculus II Unit 7.3: 7.3.1a, 7.3.3a, 7.3.6b, 7.3.6f, 7.3.6h Unit 7.4: 7.4.1b, 7.4.1c, 7.4.2b, 7.4.3, 7.4.6, 7.4.7

Advanced Calculus II Unit 7.3: 7.3.1a, 7.3.3a, 7.3.6b, 7.3.6f, 7.3.6h Unit 7.4: 7.4.1b, 7.4.1c, 7.4.2b, 7.4.3, 7.4.6, 7.4.7 Advanced Calculus II Unit 73: 73a, 733a, 736b, 736f, 736h Unit 74: 74b, 74c, 74b, 743, 746, 747 Megan Bryant October 9, 03 73a Prove the following: If lim p a = A, for some p >, then a converges absolutely

More information

Secure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel

Secure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel Secure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel Pritam Mukherjee Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 074 pritamm@umd.edu

More information

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

18.2 Continuous Alphabet (discrete-time, memoryless) Channel 0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not

More information

Lecture 15: Conditional and Joint Typicaility

Lecture 15: Conditional and Joint Typicaility EE376A Information Theory Lecture 1-02/26/2015 Lecture 15: Conditional and Joint Typicaility Lecturer: Kartik Venkat Scribe: Max Zimet, Brian Wai, Sepehr Nezami 1 Notation We always write a sequence of

More information

EXPURGATED GAUSSIAN FINGERPRINTING CODES. Pierre Moulin and Negar Kiyavash

EXPURGATED GAUSSIAN FINGERPRINTING CODES. Pierre Moulin and Negar Kiyavash EXPURGATED GAUSSIAN FINGERPRINTING CODES Pierre Moulin and Negar iyavash Beckman Inst, Coord Sci Lab and ECE Department University of Illinois at Urbana-Champaign, USA ABSTRACT This paper analyzes the

More information

6.02 Fall 2011 Lecture #9

6.02 Fall 2011 Lecture #9 6.02 Fall 2011 Lecture #9 Claude E. Shannon Mutual information Channel capacity Transmission at rates up to channel capacity, and with asymptotically zero error 6.02 Fall 2011 Lecture 9, Slide #1 First

More information

Binary Compressive Sensing via Analog. Fountain Coding

Binary Compressive Sensing via Analog. Fountain Coding Binary Compressive Sensing via Analog 1 Fountain Coding Mahyar Shirvanimoghaddam, Member, IEEE, Yonghui Li, Senior Member, IEEE, Branka Vucetic, Fellow, IEEE, and Jinhong Yuan, Senior Member, IEEE, arxiv:1508.03401v1

More information

Problem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.

Problem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly. (b) We note that the above capacity is the same to the capacity of the binary symmetric channel. Indeed, if we considerthe grouping of the output symbols into a = {y 1,y 2 } and b = {y 3,y 4 } we get a

More information

DECOUPLING LECTURE 6

DECOUPLING LECTURE 6 18.118 DECOUPLING LECTURE 6 INSTRUCTOR: LARRY GUTH TRANSCRIBED BY DONGHAO WANG We begin by recalling basic settings of multi-linear restriction problem. Suppose Σ i,, Σ n are some C 2 hyper-surfaces in

More information