Cover s Open Problem: The Capacity of the Relay Channel
|
|
- Juliet O’Connor’
- 6 years ago
- Views:
Transcription
1 Cover s Open Problem: The Capacity of the Relay Channel Ayfer Özgür Stanford University Advanced Networks Colloquia Series University of Maryland, March 2017 Joint work with Xiugang Wu and Leighton Pate Barnes. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 1 / 26
2 Father of the Information Age Claude Shannon ( ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 2 / 26
3 The Bell System Technical Journal Vol. XXVII J Illy, 1948 No.3 A Mathematical Theory of Communication By c. E. SHANNON IXTRODUCTION HE recent development of various methods of modulation such as rem T and PPM which exchange bandwidth for signal-to-noise ratio has intensified the interest in a general theory of communication. A basis for such a theory is contained in the important papers of Nyquist! and Hartley" on this subject. In the present paper we will extend the theory to include a number of new factors, in particular the effect of noise in the channel, and Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 3 / 26
4 A method is developed for representing any communication system geometrically... Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 4 / 26
5 AWGN Channel P Transmitter Receiver Capacity ( C = log 1 + P ) N Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 5 / 26
6 Converse: Sphere Packing p nn X n (1) X n (2) p nn 0 p n(p + N) Y sphere X n (3) p nn noise sphere ( n(p ) Sphere + N) # of X n. 2 n log 2πe(P+N) 2 ( nn ) = Sphere 2 n 2 log 2πeN = 2 n 2 log(1+ P N ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 6 / 26
7 Achievability: Geometric Random Coding Pr( false X n Lens ) ( np ) 2 nr Sphere. = 2 n 2 2 n 2 PN log 2πe P+N 2nR log 2πeP = 2 n 2 log(1+ P N ) 2 nr Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 7 / 26
8 The story goes... Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 8 / 26
9 Cover s Open Problem Ayfer Özgür (Stanford) The Capacity of the Relay Channel March 17 9 / 26
10 Gaussian case W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Achievability: C 0 =. C( ) = 1 2 log ( 1 + 2P N Destination Y n Cutset-bound (Cover and El Gamal 79): C0 1 ( 2 log 1 + 2P ) 12 ( N log 1 + P ). N ) Potentially, C 0 0 as P/N 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
11 Main Result W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Destination Y n Theorem C 0 = Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
12 Upper Bound on the Capacity Cut-set bound C-F Old bound New bound SNR = 15 db C 0 (bit/channel use) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
13 Cutset Bound W 1 N(0,N) W 2 N(0,N) Z n C 0 Relay I n = f(z n ) Source X n Destination Y n nr I (X n ; Y n, I n ) + nɛ n = I (X n ; Y n ) + I (X n ; I n Y n ) + nɛ n = I (X n ; Y n ) + H(I n Y n ) H(I }{{} n Y n, X n ) +nɛ }{{} n nc 0 0 n(i (X ; Y ) + C 0 + ɛ n ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
14 Cutset Bound W 1 N(0,N) W 2 N(0,N) Z n C 0 Relay I n = f(z n ) Source X n Destination Y n nr I (X n ; Y n, I n ) + nɛ n = I (X n ; Y n ) + I (X n ; I n Y n ) + nɛ n = I (X n ; Y n ) + H(I n Y n ) H(I }{{} n X n ) +nɛ }{{} n nc 0 0 n(i (X ; Y ) + C 0 + ɛ n ) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
15 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
16 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin If H(I n X n ) = 0, Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
17 Z n X n p nn Y n Typical set of Z n /Y n I n -th bin If H(I n X n ) = 0, then H(I n Y n ) = 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
18 W 1 N(0,N) W 2 N(0,N) Z n Relay C 0 I n = f(z n ) Source X n Destination Y n R I (X n ; Y n ) + H(I n Y n ) H(I }{{} n X n ) +nɛ }{{} n? = n log sin θ n Goal: I n = f (Z n ) Z n X n n Y }{{} H(I n X n )= n log sin θ n } {{ } H(I n Y n )? Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
19 H(I n X n ) 0 Typical set of Z n p nn X n Multiple bins # of bins =? P(each bin) =? Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
20 From n- to nb- Dimensional Space X, Y, Z, I : B-length i.i.d. from {(X n (b), Y n (b), Z n (b), I n (b))} B b=1. If H(I n X n ) = n log sin θ n, then for any typical (x, i) p(i x). = 2 nb log sin θn, P(Z A(i) x) =. nb log sin θn 2 p nbnx ith bin A x (i) Pr. =2 nb log sin n A x (i). =2 nb( 1 2 log 2 en sin2 n) Typical set of Z/Y Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
21 Isoperimetric Inequalities Isoperimetric Inequality in the Plane (Steiner 1838) Among all closed curves in the plane with a given enclosed area, the circle has the smallest perimeter. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
22 Isoperimetric Inequalities Isoperimetric Inequality on the Sphere (Levy 1919) Among all sets on the sphere with a given volume, the spherical cap has the smallest boundary, or the smallest volume of ω-neighborhood for any ω > 0. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
23 Blowing-up Lemma A x (i). =2 nb( 1 2 log 2 en sin2 n) Isoperimetric Inequality on the Sphere + Measure Concentration: P(Z blow-up of A x (i) x) 1. P(Y blow-up of A x (i) x) 1. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
24 Geometry of Typical Sets n-dimensional space: Almost all (X n, Y n, Z n, I n ) Z n! I n X n p nn Y n nb-dimensional space: Almost all (x, y, i) p nbn z! i x 2 n y Information Inequality: (Wu and Ozgur, 2015) H(I n Y n ) n(2 log sin θ n + 2 log sin θ n ln 2 log e). Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
25 A new approach A x (i) A x (i) Y Y Control the intersection of a sphere drawn around a randomly chosen Y and A x (i). Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
26 Easy if A x (i) is a spherical cap Cap(z 0, n) z 0 Cap(z 0, n) \ Cap(y 0,! n) p nbn x n! n y 0 Cap(y 0,! n) Cap(z 0, θ n ) Cap(Y, ω n ). = 2 nb( 1 2 log 2πeN(sin2 θ n cos 2 ω n)) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
27 Strengthening of the Isoperimetric Inequality Strengthening of the Isoperimetric Inequality: Among all sets on the sphere with a given volume, the spherical cap has minimal intersection volume at distance ω for almost all points on the sphere for any ω > π/2 θ. Proof: builds on the Riesz rearrangement inequality. Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
28 A Packing Argument Given Y, # of I ( ) Sphere Y, nbn4sin 2 ωn 2 2 nb( 1 2 log 2πeN(sin2 θ n cos 2 ω n)) Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
29 Summary Solved an problem posed by Cover and named The Capacity of Relay Channel in Open Problems in Communication and Computation, Springer-Verlag, Developed a converse technique that significantly deviates from standard converse techniques based on single-letterization and has some new ingredients: Typicality Measure Concentration Isoperimetric Inequality Ayfer Özgür (Stanford) The Capacity of the Relay Channel March / 26
Improving on the Cutset Bound via a Geometric Analysis of Typical Sets
Imroving on the Cutset Bound via a Geometric Analysis of Tyical Sets Ayfer Özgür Stanford University CUHK, May 28, 2016 Joint work with Xiugang Wu (Stanford). Ayfer Özgür (Stanford) March 16 1 / 33 Gaussian
More information3F1 Information Theory, Lecture 1
3F1 Information Theory, Lecture 1 Jossy Sayir Department of Engineering Michaelmas 2013, 22 November 2013 Organisation History Entropy Mutual Information 2 / 18 Course Organisation 4 lectures Course material:
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationSteiner s formula and large deviations theory
Steiner s formula and large deviations theory Venkat Anantharam EECS Department University of California, Berkeley May 19, 2015 Simons Conference on Networks and Stochastic Geometry Blanton Museum of Art
More informationMMSE estimation and lattice encoding/decoding for linear Gaussian channels. Todd P. Coleman /22/02
MMSE estimation and lattice encoding/decoding for linear Gaussian channels Todd P. Coleman 6.454 9/22/02 Background: the AWGN Channel Y = X + N where N N ( 0, σ 2 N ), 1 n ni=1 X 2 i P X. Shannon: capacity
More informationECE Information theory Final (Fall 2008)
ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1
More informationOn the Rate-Limited Gelfand-Pinsker Problem
On the Rate-Limited Gelfand-Pinsker Problem Ravi Tandon Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 74 ravit@umd.edu ulukus@umd.edu Abstract
More informationLecture 2. Capacity of the Gaussian channel
Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN
More informationApproximately achieving Gaussian relay. network capacity with lattice-based QMF codes
Approximately achieving Gaussian relay 1 network capacity with lattice-based QMF codes Ayfer Özgür and Suhas Diggavi Abstract In [1], a new relaying strategy, quantize-map-and-forward QMF scheme, has been
More informationInformation Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results
Information Theory Lecture 10 Network Information Theory (CT15); a focus on channel capacity results The (two-user) multiple access channel (15.3) The (two-user) broadcast channel (15.6) The relay channel
More informationInteractive Decoding of a Broadcast Message
In Proc. Allerton Conf. Commun., Contr., Computing, (Illinois), Oct. 2003 Interactive Decoding of a Broadcast Message Stark C. Draper Brendan J. Frey Frank R. Kschischang University of Toronto Toronto,
More informationTwo Applications of the Gaussian Poincaré Inequality in the Shannon Theory
Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Vincent Y. F. Tan (Joint work with Silas L. Fong) National University of Singapore (NUS) 2016 International Zurich Seminar on
More informationLecture 15: Thu Feb 28, 2019
Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:
More informationLeech Constellations of Construction-A Lattices
Leech Constellations of Construction-A Lattices Joseph J. Boutros Talk at Nokia Bell Labs, Stuttgart Texas A&M University at Qatar In collaboration with Nicola di Pietro. March 7, 2017 Thanks Many Thanks
More informationLecture 4 Capacity of Wireless Channels
Lecture 4 Capacity of Wireless Channels I-Hsiang Wang ihwang@ntu.edu.tw 3/0, 014 What we have learned So far: looked at specific schemes and techniques Lecture : point-to-point wireless channel - Diversity:
More informationII. THE TWO-WAY TWO-RELAY CHANNEL
An Achievable Rate Region for the Two-Way Two-Relay Channel Jonathan Ponniah Liang-Liang Xie Department of Electrical Computer Engineering, University of Waterloo, Canada Abstract We propose an achievable
More informationChannel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter
Channel Dependent Adaptive Modulation and Coding Without Channel State Information at the Transmitter Bradford D. Boyle, John MacLaren Walsh, and Steven Weber Modeling & Analysis of Networks Laboratory
More informationOn the Capacity Region of the Gaussian Z-channel
On the Capacity Region of the Gaussian Z-channel Nan Liu Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 74 nkancy@eng.umd.edu ulukus@eng.umd.edu
More informationOn Multiple User Channels with State Information at the Transmitters
On Multiple User Channels with State Information at the Transmitters Styrmir Sigurjónsson and Young-Han Kim* Information Systems Laboratory Stanford University Stanford, CA 94305, USA Email: {styrmir,yhk}@stanford.edu
More informationLecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157
Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x
More informationCooperative Communication with Feedback via Stochastic Approximation
Cooperative Communication with Feedback via Stochastic Approximation Utsaw Kumar J Nicholas Laneman and Vijay Gupta Department of Electrical Engineering University of Notre Dame Email: {ukumar jnl vgupta}@ndedu
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationAppendix B Information theory from first principles
Appendix B Information theory from first principles This appendix discusses the information theory behind the capacity expressions used in the book. Section 8.3.4 is the only part of the book that supposes
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationDirty Paper Writing and Watermarking Applications
Dirty Paper Writing and Watermarking Applications G.RaviKiran February 10, 2003 1 Introduction Following an underlying theme in Communication is the duo of Dirty Paper Writing and Watermarking. In 1983
More informationThe Shannon s basic inequalities refer to the following fundamental properties of entropy function:
COMMUNICATIONS IN INFORMATION AND SYSTEMS c 2003 International Press Vol. 3, No. 1, pp. 47-60, June 2003 004 ON A NEW NON-SHANNON TYPE INFORMATION INEQUALITY ZHEN ZHANG Abstract. Recently, K. Makarychev,
More informationMulticoding Schemes for Interference Channels
Multicoding Schemes for Interference Channels 1 Ritesh Kolte, Ayfer Özgür, Haim Permuter Abstract arxiv:1502.04273v1 [cs.it] 15 Feb 2015 The best known inner bound for the 2-user discrete memoryless interference
More informationFat Struts: Constructions and a Bound
Fat Struts: Constructions and a Bound N. J. A. Sloane, Vinay A. Vaishampayan, Sueli I. R. Costa AT&T Shannon Labs, Florham Park, NJ 0793, USA. Email: {njas,vinay}@research.att.com University of Campinas,
More informationSuperposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels
Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:
More informationNational University of Singapore Department of Electrical & Computer Engineering. Examination for
National University of Singapore Department of Electrical & Computer Engineering Examination for EE5139R Information Theory for Communication Systems (Semester I, 2014/15) November/December 2014 Time Allowed:
More informationShannon s noisy-channel theorem
Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for
More informationAn Achievable Rate for the Multiple Level Relay Channel
An Achievable Rate for the Multiple Level Relay Channel Liang-Liang Xie and P. R. Kumar Department of Electrical and Computer Engineering, and Coordinated Science Laboratory University of Illinois, Urbana-Champaign
More informationLattices and Lattice Codes
Lattices and Lattice Codes Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology Hyderabad
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationNotes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel
Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic
More informationDiversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007
Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT ECE 559 Presentation Hoa Pham Dec 3, 2007 Introduction MIMO systems provide two types of gains Diversity Gain: each path from a transmitter
More informationMultiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets
Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,
More informationSome Goodness Properties of LDA Lattices
Some Goodness Properties of LDA Lattices Shashank Vatedka and Navin Kashyap {shashank,nkashyap}@eceiiscernetin Department of ECE Indian Institute of Science Bangalore, India Information Theory Workshop
More informationCovert Communication with Channel-State Information at the Transmitter
Covert Communication with Channel-State Information at the Transmitter Si-Hyeon Lee Joint Work with Ligong Wang, Ashish Khisti, and Gregory W. Wornell July 27, 2017 1 / 21 Covert Communication Transmitter
More informationThe Gallager Converse
The Gallager Converse Abbas El Gamal Director, Information Systems Laboratory Department of Electrical Engineering Stanford University Gallager s 75th Birthday 1 Information Theoretic Limits Establishing
More informationMMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING
MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING Yichuan Hu (), Javier Garcia-Frias () () Dept. of Elec. and Comp. Engineering University of Delaware Newark, DE
More informationLecture 6 Channel Coding over Continuous Channels
Lecture 6 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 9, 015 1 / 59 I-Hsiang Wang IT Lecture 6 We have
More informationThe Capacity Region of the Gaussian Cognitive Radio Channels at High SNR
The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR 1 Stefano Rini, Daniela Tuninetti and Natasha Devroye srini2, danielat, devroye @ece.uic.edu University of Illinois at Chicago Abstract
More informationA Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels
A Proof of the Converse for the Capacity of Gaussian MIMO Broadcast Channels Mehdi Mohseni Department of Electrical Engineering Stanford University Stanford, CA 94305, USA Email: mmohseni@stanford.edu
More informationThe Method of Types and Its Application to Information Hiding
The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,
More informationOptimal Power Control in Decentralized Gaussian Multiple Access Channels
1 Optimal Power Control in Decentralized Gaussian Multiple Access Channels Kamal Singh Department of Electrical Engineering Indian Institute of Technology Bombay. arxiv:1711.08272v1 [eess.sp] 21 Nov 2017
More informationLecture 4. Capacity of Fading Channels
1 Lecture 4. Capacity of Fading Channels Capacity of AWGN Channels Capacity of Fading Channels Ergodic Capacity Outage Capacity Shannon and Information Theory Claude Elwood Shannon (April 3, 1916 February
More informationIET Commun., 2009, Vol. 3, Iss. 4, pp doi: /iet-com & The Institution of Engineering and Technology 2009
Published in IET Communications Received on 10th March 2008 Revised on 10th July 2008 ISSN 1751-8628 Comprehensive partial decoding approach for two-level relay networks L. Ghabeli M.R. Aref Information
More informationChannel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.
Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless
More informationLecture 3: Channel Capacity
Lecture 3: Channel Capacity 1 Definitions Channel capacity is a measure of maximum information per channel usage one can get through a channel. This one of the fundamental concepts in information theory.
More informationThe Role of Directed Information in Network Capacity
The Role of Directed Information in Network Capacity Sudeep Kamath 1 and Young-Han Kim 2 1 Department of Electrical Engineering Princeton University 2 Department of Electrical and Computer Engineering
More information3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions
Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the
More informationCapacity of the Discrete Memoryless Energy Harvesting Channel with Side Information
204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener
More informationOn the Duality between Multiple-Access Codes and Computation Codes
On the Duality between Multiple-Access Codes and Computation Codes Jingge Zhu University of California, Berkeley jingge.zhu@berkeley.edu Sung Hoon Lim KIOST shlim@kiost.ac.kr Michael Gastpar EPFL michael.gastpar@epfl.ch
More informationPreCalculus: Chapter 9 Test Review
Name: Class: Date: ID: A PreCalculus: Chapter 9 Test Review Short Answer 1. Plot the point given in polar coordinates. 3. Plot the point given in polar coordinates. (-4, -225 ) 2. Plot the point given
More informationLECTURE 13. Last time: Lecture outline
LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to
More informationCapacity of AWGN channels
Chapter 3 Capacity of AWGN channels In this chapter we prove that the capacity of an AWGN channel with bandwidth W and signal-tonoise ratio SNR is W log 2 (1+SNR) bits per second (b/s). The proof that
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017
U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 3, 207 Scribed by Keyhan Vakil Lecture 3 In which we complete the study of Independent Set and Max Cut in G n,p random graphs.
More informationEE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm
EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm 1. Feedback does not increase the capacity. Consider a channel with feedback. We assume that all the recieved outputs are sent back immediately
More informationLecture 6 I. CHANNEL CODING. X n (m) P Y X
6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder
More information392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions
392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6 Problem Set 2 Solutions Problem 2.1 (Cartesian-product constellations) (a) Show that if A is a K-fold Cartesian
More informationA Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback
A Tight Upper Bound on the Second-Order Coding Rate of Parallel Gaussian Channels with Feedback Vincent Y. F. Tan (NUS) Joint work with Silas L. Fong (Toronto) 2017 Information Theory Workshop, Kaohsiung,
More informationCovering a sphere with caps: Rogers bound revisited
Covering a sphere with caps: Rogers bound revisited Ilya Dumer Abstract We consider coverings of a sphere S n r of radius r with the balls of radius one in an n-dimensional Euclidean space R n. Our goal
More informationChannel Dispersion and Moderate Deviations Limits for Memoryless Channels
Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Yury Polyanskiy and Sergio Verdú Abstract Recently, Altug and Wagner ] posed a question regarding the optimal behavior of the probability
More informationWireless Communications Lecture 10
Wireless Communications Lecture 1 [SNR per symbol and SNR per bit] SNR = P R N B = E s N BT s = E b N BT b For BPSK: T b = T s, E b = E s, and T s = 1/B. Raised cosine pulse shaper for other pulses. T
More informationThe Microcanonical Approach. (a) The volume of accessible phase space for a given total energy is proportional to. dq 1 dq 2 dq N dp 1 dp 2 dp N,
8333: Statistical Mechanics I Problem Set # 6 Solutions Fall 003 Classical Harmonic Oscillators: The Microcanonical Approach a The volume of accessible phase space for a given total energy is proportional
More informationLecture 4 Capacity of Wireless Channels
Lecture 4 Capacity of Wireless Channels I-Hsiang Wang ihwang@ntu.edu.tw 3/0, 014 What we have learned So far: looked at specific schemes and techniques Lecture : point-to-point wireless channel - Diversity:
More informationECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)
ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling
More informationEnergy State Amplification in an Energy Harvesting Communication System
Energy State Amplification in an Energy Harvesting Communication System Omur Ozel Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland College Park, MD 20742 omur@umd.edu
More informationTowards control over fading channels
Towards control over fading channels Paolo Minero, Massimo Franceschetti Advanced Network Science University of California San Diego, CA, USA mail: {minero,massimo}@ucsd.edu Invited Paper) Subhrakanti
More informationHierarchical Cooperation Achieves Optimal Capacity Scaling in Ad Hoc Networks
Hierarchical Cooperation Achieves Optimal Capacity Scaling in Ad Hoc Networks Ayfer Özgür, Olivier Lévêque Faculté Informatique et Communications Ecole Polytechnique Fédérale de Lausanne 05 Lausanne, Switzerland
More informationEfficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel
Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel Ahmad Abu Al Haija ECE Department, McGill University, Montreal, QC, Canada Email: ahmad.abualhaija@mail.mcgill.ca
More informationSpace-Time Diversity Enhancements Using Collaborative Communications
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. XX, NO. YY, MONTH 2005 1 Space-Time Diversity Enhancements Using Collaborative Communications Patrick Mitran, Student Member, IEEE, Hideki Ochiai, Member,
More informationCOMBINING DECODED-AND-FORWARDED SIGNALS IN GAUSSIAN COOPERATIVE CHANNELS
COMBINING DECODED-AND-FORWARDED SIGNALS IN GAUSSIAN COOPERATIVE CHANNELS Brice Djeumou, Samson Lasaulce, Andrew Klein To cite this version: Brice Djeumou, Samson Lasaulce, Andrew Klein. COMBINING DECODED-AND-FORWARDED
More informationStructured signal recovery from non-linear and heavy-tailed measurements
Structured signal recovery from non-linear and heavy-tailed measurements Larry Goldstein* Stanislav Minsker* Xiaohan Wei # *Department of Mathematics # Department of Electrical Engineering UniversityofSouthern
More informationPrinciples of Communications
Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @
More informationOn Two-user Fading Gaussian Broadcast Channels. with Perfect Channel State Information at the Receivers. Daniela Tuninetti
DIMACS Workshop on Network Information Theory - March 2003 Daniela Tuninetti 1 On Two-user Fading Gaussian Broadcast Channels with Perfect Channel State Information at the Receivers Daniela Tuninetti Mobile
More informationLecture 17: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 31, 2016 [Ed. Apr 1]
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture 7: Density Estimation Lecturer: Yihong Wu Scribe: Jiaqi Mu, Mar 3, 06 [Ed. Apr ] In last lecture, we studied the minimax
More informationSolutions to Homework Set #4 Differential Entropy and Gaussian Channel
Solutions to Homework Set #4 Differential Entropy and Gaussian Channel 1. Differential entropy. Evaluate the differential entropy h(x = f lnf for the following: (a Find the entropy of the exponential density
More informationLECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs
LECTURE 18 Last time: White Gaussian noise Bandlimited WGN Additive White Gaussian Noise (AWGN) channel Capacity of AWGN channel Application: DS-CDMA systems Spreading Coding theorem Lecture outline Gaussian
More informationGaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26
Gaussian channel Information theory 2013, lecture 6 Jens Sjölund 8 May 2013 Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Outline 1 Definitions 2 The coding theorem for Gaussian channel 3 Bandlimited
More informationBlock 2: Introduction to Information Theory
Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation
More informationOn Capacity Under Received-Signal Constraints
On Capacity Under Received-Signal Constraints Michael Gastpar Dept. of EECS, University of California, Berkeley, CA 9470-770 gastpar@berkeley.edu Abstract In a world where different systems have to share
More informationHarmonic Analysis Homework 5
Harmonic Analysis Homework 5 Bruno Poggi Department of Mathematics, University of Minnesota November 4, 6 Notation Throughout, B, r is the ball of radius r with center in the understood metric space usually
More informationOn the Secrecy Capacity of Fading Channels
On the Secrecy Capacity of Fading Channels arxiv:cs/63v [cs.it] 7 Oct 26 Praveen Kumar Gopala, Lifeng Lai and Hesham El Gamal Department of Electrical and Computer Engineering The Ohio State University
More informationFree-Space Optical Communications: Capacity Bounds, Approximations, and a New Sphere-Packing Perspective
Free-Space Optical Communications: Capacity Bounds, pproximations, and a New Sphere-Packing Perspective nas Chaaban, Jean-Marie Morvan, and Mohammad Slim louini bstract The capacity of the intensity-modulation
More informationLecture 14 February 28
EE/Stats 376A: Information Theory Winter 07 Lecture 4 February 8 Lecturer: David Tse Scribe: Sagnik M, Vivek B 4 Outline Gaussian channel and capacity Information measures for continuous random variables
More informationBlind phase/frequency synchronization with low-precision ADC: a Bayesian approach
Blind phase/frequency synchronization with low-precision ADC: a Bayesian approach Aseem Wadhwa, Upamanyu Madhow Department of ECE, UCSB 1/26 Modern Communication Receiver Architecture Analog Digital TX
More informationCapacity Bounds for. the Gaussian Interference Channel
Capacity Bounds for the Gaussian Interference Channel Abolfazl S. Motahari, Student Member, IEEE, and Amir K. Khandani, Member, IEEE Coding & Signal Transmission Laboratory www.cst.uwaterloo.ca {abolfazl,khandani}@cst.uwaterloo.ca
More informationA Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources
A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources Wei Kang Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College
More informationAdvanced Calculus II Unit 7.3: 7.3.1a, 7.3.3a, 7.3.6b, 7.3.6f, 7.3.6h Unit 7.4: 7.4.1b, 7.4.1c, 7.4.2b, 7.4.3, 7.4.6, 7.4.7
Advanced Calculus II Unit 73: 73a, 733a, 736b, 736f, 736h Unit 74: 74b, 74c, 74b, 743, 746, 747 Megan Bryant October 9, 03 73a Prove the following: If lim p a = A, for some p >, then a converges absolutely
More informationSecure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel
Secure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel Pritam Mukherjee Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College Park, MD 074 pritamm@umd.edu
More information18.2 Continuous Alphabet (discrete-time, memoryless) Channel
0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not
More informationLecture 15: Conditional and Joint Typicaility
EE376A Information Theory Lecture 1-02/26/2015 Lecture 15: Conditional and Joint Typicaility Lecturer: Kartik Venkat Scribe: Max Zimet, Brian Wai, Sepehr Nezami 1 Notation We always write a sequence of
More informationEXPURGATED GAUSSIAN FINGERPRINTING CODES. Pierre Moulin and Negar Kiyavash
EXPURGATED GAUSSIAN FINGERPRINTING CODES Pierre Moulin and Negar iyavash Beckman Inst, Coord Sci Lab and ECE Department University of Illinois at Urbana-Champaign, USA ABSTRACT This paper analyzes the
More information6.02 Fall 2011 Lecture #9
6.02 Fall 2011 Lecture #9 Claude E. Shannon Mutual information Channel capacity Transmission at rates up to channel capacity, and with asymptotically zero error 6.02 Fall 2011 Lecture 9, Slide #1 First
More informationBinary Compressive Sensing via Analog. Fountain Coding
Binary Compressive Sensing via Analog 1 Fountain Coding Mahyar Shirvanimoghaddam, Member, IEEE, Yonghui Li, Senior Member, IEEE, Branka Vucetic, Fellow, IEEE, and Jinhong Yuan, Senior Member, IEEE, arxiv:1508.03401v1
More informationProblem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.
(b) We note that the above capacity is the same to the capacity of the binary symmetric channel. Indeed, if we considerthe grouping of the output symbols into a = {y 1,y 2 } and b = {y 3,y 4 } we get a
More informationDECOUPLING LECTURE 6
18.118 DECOUPLING LECTURE 6 INSTRUCTOR: LARRY GUTH TRANSCRIBED BY DONGHAO WANG We begin by recalling basic settings of multi-linear restriction problem. Suppose Σ i,, Σ n are some C 2 hyper-surfaces in
More information