Revision of Lecture 4
|
|
- Joshua Simon
- 6 years ago
- Views:
Transcription
1 Revision of Lecture 4 We have completed studying digital sources from information theory viewpoint We have learnt all fundamental principles for source coding, provided by information theory Practical source coding is guided by information theory, with practical constraints, such as performance and processing complexity/delay trade off When you come to practical source coding part, you can smile as you should know everything Information theory does far more, in fact it provides guiding principle for everything in communications For example, what happens to information transmitted through channel? source (Tx) channel destination (Rx) 55
2 Information Across Channels For conveying information across a transmission channel, a suitable model is: AWGN input channel output The channel itself introduces amplitude and phase distortion, is potentially timevarying, and has a limited bandwidth B Error-free reception of symbols is additionally impeded by additive white Gaussian noise (AWGN); it severeness is described by the signal-to-noise ratio (SNR) Therefore, dependent on the above parameters, we are interested in determining the maximum possible error-free information transmission (channel capacity C) We will see that C depends on B and SNR 56
3 Characteristics of Channel The channel can be described by its impulse response h(t) or equivalently its frequency response H(jω) = A(ω) e jφ(ω) with amplitude response A(ω) and phase response Φ(ω); h(t) and H(jω) are Fourier pair Ideal channel (pure delay): h(t) = δ(t T) A(ω) =, Φ(ω) = ωt A(ω) Φ(ω) B ω B ω T Flat magnitude and linear phase ( = constant group delay G(ω) = Φ(ω)/ ω) The only impairment caused by an ideal channel is AWGN Non-ideal channel: channel is dispersive, causing intersymbol interference 57
4 Additive White Gaussian Noise Noise is uncorrelated with the signal 3 2 time series Gaussian noise has a bell shaped probability density function (normally distributed) p(x) = 2πσ e (x µ)/2σ2 x[n] time index n 3 histogram.4 probability density function with mean µ and variance σ 2 White noise has zero mean, and channel noise is usually modelled as an AWGN counts sample value x p(x) sample value x 58
5 White Noise White noise is characterised by a flat power spectral density function, N(ω), or equivalently, its impulse-shaped auto-correlation function, R(τ) N(ω) R( τ) N 2 N 2 ω τ N(ω) and R(τ) are a Fourier pair: R(τ) = 2π Z N(ω)e jωτ dω N(ω) = Z R(τ)e jωτ dτ Two-side spectrum is usually used for convenience, and N is the noise power 59
6 Physic Basis of Channel Transmitted signal is amplified to required power level and launched from transmit antenna to channel Signal power is attenuated, as it travels in distance path loss Copies of signal arrive at receiver with different attenuation and delays, which may cause dispersive and fading (power level fluctuates rapidly) effects Received signal at receiver is very weak, and needs to be amplified to required power level in order to detect digital information contained While receiver amplifier amplifies receive signal, it also introduces thermal noise AWGN in our channel model in fact models this noise How much noise introduced by amplifier is specified by its noise figure Depending on communication carrier frequency, channel bandwidth and actual communication conditions, channel may be modelled as: AWGN channel, i.e. channel is nondispersive or memoryless Or dispersive channel, i.e. channel has memory 6
7 Binary Symmetric Channel (BSC) BSC is the simplest model for information transmission via a discrete channel (channel is ideal, no amplitude and phase distortion, only distortion is due to AWGN): P(X ) P(X ) X X P(Y X ) P(Y X ) P(Y X ) P(Y X ) Y Y P(X i ): probability of occurrence of X i at source output, a priori probability P(Y j X i ): conditional probability of Rx Y j given Tx X i Symmetric refers to P(Y X ) = P(Y X ) = p e, p e being channel error probability The joint probability P(Y j, X i ) (Tx X i and Rx Y j ) is linked with the conditional probabilities P(Y j X i ) by Bayes rule: P(Y j,x i ) = P(X i ) P(Y j X i ) = P(Y j ) P(X i Y j ) = P(X i,y j ) 6
8 Binary Symmetric Channel Example Consider a BSC: P(X = ) =.7 P(X = ) =.3 p e =.2 TX p e =.98 p e =.2 p e =.98 RX This has a non-equiprobable source with P(X = ) =.7 and P(X = ) =.3: on average, 7% of transmitted bits are and 3% are Channel s error probability p e =.2: on average, bit error rate is 2% 62
9 Binary Symmetric Channel Example (continue) Probability of correct reception: Pcorrect = P(Y =, X = ) + P(Y =, X = ) =.98, as P(Y =, X = ) = P(X = ) P(Y = X = ) =.7.98 =.686 P(Y =, X = ) = P(X = ) P(Y = X = ) =.3.98 =.294 Probability of erroneous reception: Perror = P(Y =, X = ) + P(Y =, X = ) =.2, as P(Y =, X = ) = P(X = ) P(Y = X = ) =.3.2 =.6 P(Y =, X = ) = P(X = ) P(Y = X = ) =.7.2 =.4 Total probability of receiving a (or a ): P(Y = ) = P(X = ) P(Y = X = ) + P(X = ) P(Y = X = ) = =.692 P(Y = ) = P(X = ) P(Y = X = ) + P(X = ) P(Y = X = ) = =.38 63
10 A Close Look at Channel For above (binary symmetric channel) example P(X )=.3 P(Y )=.38 P(X )=.7 P(Y )=.692 Something happens in channel to information How to describe this? Y i and X j are connected They have something in common or mutual P(X ) i X i P(Y X ) i i Y i P(Y ) i P(X ) j X j P(Y X ) j i 64
11 Mutual Information Definition of mutual information of X i and Y j : I(X i,y j ) = log 2 P(X i Y j ) P(X i ) (bits) Perfect, noiseless channel: Y i = X i, i.e. P(X i Y i ) = and I(X i, Y i ) = log 2 P(X i ) This is the information of X i, hence no information is lost in the channel Extremely noisy channel with error probability.5 Y i is independent of X i, hence P(X i Y i ) = P(X i, Y i ) = P(X i) P(Y i ) P(Y i ) P(Y i ) Therefore I(X i, Y i ) = log 2 =, meaning all information is lost in the channel In general, I(X i ) > I(X i, Y i ), some information is lost in the channel 65
12 Mutual Information Example Consider the earlier BSC example: P(X ) =.7 X p e =.98 p e =.2 p e =.2 Y P(X Y ) =.993 P(X Y ) =.87 P(X Y ) =.455 P(X ) =.3 X p e =.98 Y P(X Y ) =.9545 Here, the mutual information results in: I(X, Y ) =.52 bits I(X, Y ) =.67 bits (source info: I(X ) =.55 bits I(X ) =.737 bits) (destin info: I(Y ) =.53 bits I(Y ) =.699 bits) I(X, Y ) = 5.3 bits I(X, Y ) = bits Destination info contents are more balanced then source info contents The negative quantities represent mis-information 66
13 Average Mutual Information Based on received symbols Y j given transmitted symbols X i through a BSC, average mutual information is defined as: I(X,Y ) = i = i P(X i, Y j ) I(X i, Y j ) j j P(X i, Y j ) log 2 P(X i Y j ) P(X i ) (bits/symbol) This gives the average amount of source information acquired per received symbol by the receiver, and should be distinguished form the average source information (entropy H(X)) Note that due to Bayes: P(X i Y j ) P(X i ) = P(X i, Y j ) P(X i ) P(Y j ) = P(Y j X i ) P(Y j ) 67
14 Imperfect Channel: Information Loss Consider re-arranging the mutual information between transmitted symbol X i and received symbol Y j : I(X i,y j ) = log 2 P(X i Y j ) P(X i ) = I(X i ) I(X i Y j ) = log 2 P(X i ) log 2 P(X i Y j ) I(X i,y j ) is the amount of information conveyed to receiver when transmitting X i and receiving Y j, I(X i ) is the source information of X i, and I(X i Y j ) can be regarded as the information loss due to the channel Therefore, I(X i ) I(X }{{} i,y j ) }{{} Source Inf. Inf. conveyed to rec. = I(X i Y j ) }{{} Inf. loss I(X i Y j ) I(X i ), see for example the previous cases of p e = and p e =.5 68
15 Imperfect Channel: Average Mutual Information Average mutual information is given by: I(X, Y ) = X X P(X i Y j ) P(X i, Y j ) log 2 (bits/symbol) P(X i j i ) But this average conveyed information I(X, Y ) = X X P(X i, Y j ) log 2 P(X i j i ) X X P(X i, Y j ) log 2 P(X i j i Y j ) = X P(X i, Y j ) A log 2 P(X i j i ) X! X P(Y j ) P(X i Y j ) log 2 P(X j i i Y j ) = X i P(X i ) log 2 P(X i ) X j I(X, Y ) {z } av. converyed information A similar re-arrangement leads to: I(X, Y ) {z } av. converyed information P(Y j ) I(X Y j ) = H(X) H(X Y ) = H(X) {z } H(X Y ) {z } av. source information av. information lost = H(Y ) H(Y X) {z } {z } destination entropy error entropy 69
16 Summary General consideration for transferring information across channels Channel characteristics or channel model Binary symmetric channel assumptions Mutual information between channel input X(k) {X i } and channel output Y (k) {Y i } characterises how information is transferring across channel Average mutual information I(X,Y ) I(X,Y ) }{{} av. conveyed information I(X,Y ) }{{} av. conveyed information = H(X) H(X Y ) }{{}}{{} source entropy av. information lost = H(Y ) H(Y X) }{{}}{{} destination entropy error entropy 7
Revision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationChannel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.
Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless
More informationEC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY
EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase
More informationCommunication Theory II
Communication Theory II Lecture 15: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 29 th, 2015 1 Example: Channel Capacity of BSC o Let then: o For
More informationPrinciples of Communications
Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @
More informationLecture 2. Capacity of the Gaussian channel
Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN
More informationECE Information theory Final
ECE 776 - Information theory Final Q1 (1 point) We would like to compress a Gaussian source with zero mean and variance 1 We consider two strategies In the first, we quantize with a step size so that the
More informationMAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)
MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)
More informationMAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A
MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK DEPARTMENT: ECE SEMESTER: IV SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A 1. What is binary symmetric channel (AUC DEC
More informationModule 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur
Module ntroduction to Digital Communications and nformation Theory Lesson 3 nformation Theoretic Approach to Digital Communications After reading this lesson, you will learn about Scope of nformation Theory
More informationInformation Theory - Entropy. Figure 3
Concept of Information Information Theory - Entropy Figure 3 A typical binary coded digital communication system is shown in Figure 3. What is involved in the transmission of information? - The system
More informationMultiple-Input Multiple-Output Systems
Multiple-Input Multiple-Output Systems What is the best way to use antenna arrays? MIMO! This is a totally new approach ( paradigm ) to wireless communications, which has been discovered in 95-96. Performance
More informationOne Lesson of Information Theory
Institut für One Lesson of Information Theory Prof. Dr.-Ing. Volker Kühn Institute of Communications Engineering University of Rostock, Germany Email: volker.kuehn@uni-rostock.de http://www.int.uni-rostock.de/
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationInformation Sources. Professor A. Manikas. Imperial College London. EE303 - Communication Systems An Overview of Fundamentals
Information Sources Professor A. Manikas Imperial College London EE303 - Communication Systems An Overview of Fundamentals Prof. A. Manikas (Imperial College) EE303: Information Sources 24 Oct. 2011 1
More informationChapter I: Fundamental Information Theory
ECE-S622/T62 Notes Chapter I: Fundamental Information Theory Ruifeng Zhang Dept. of Electrical & Computer Eng. Drexel University. Information Source Information is the outcome of some physical processes.
More informationDigital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10
Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,
More informationKevin Buckley a i. communication. information source. modulator. encoder. channel. encoder. information. demodulator decoder. C k.
Kevin Buckley - -4 ECE877 Information Theory & Coding for Digital Communications Villanova University ECE Department Prof. Kevin M. Buckley Lecture Set Review of Digital Communications, Introduction to
More informationCS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability
More informationCS6956: Wireless and Mobile Networks Lecture Notes: 2/4/2015
CS6956: Wireless and Mobile Networks Lecture Notes: 2/4/2015 [Most of the material for this lecture has been taken from the Wireless Communications & Networks book by Stallings (2 nd edition).] Effective
More information16.36 Communication Systems Engineering
MIT OpenCourseWare http://ocw.mit.edu 16.36 Communication Systems Engineering Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.36: Communication
More informationNoisy channel communication
Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University
More informationClassical Information Theory Notes from the lectures by prof Suhov Trieste - june 2006
Classical Information Theory Notes from the lectures by prof Suhov Trieste - june 2006 Fabio Grazioso... July 3, 2006 1 2 Contents 1 Lecture 1, Entropy 4 1.1 Random variable...............................
More informationBasic Principles of Video Coding
Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion
More informationLecture 4. Capacity of Fading Channels
1 Lecture 4. Capacity of Fading Channels Capacity of AWGN Channels Capacity of Fading Channels Ergodic Capacity Outage Capacity Shannon and Information Theory Claude Elwood Shannon (April 3, 1916 February
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationEstimation of the Capacity of Multipath Infrared Channels
Estimation of the Capacity of Multipath Infrared Channels Jeffrey B. Carruthers Department of Electrical and Computer Engineering Boston University jbc@bu.edu Sachin Padma Department of Electrical and
More informationECE Information theory Final (Fall 2008)
ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1
More informationInformation Theory. Coding and Information Theory. Information Theory Textbooks. Entropy
Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is
More information3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions
Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the
More informationLecture 7: Wireless Channels and Diversity Advanced Digital Communications (EQ2410) 1
Wireless : Wireless Advanced Digital Communications (EQ2410) 1 Thursday, Feb. 11, 2016 10:00-12:00, B24 1 Textbook: U. Madhow, Fundamentals of Digital Communications, 2008 1 / 15 Wireless Lecture 1-6 Equalization
More informationEntropies & Information Theory
Entropies & Information Theory LECTURE I Nilanjana Datta University of Cambridge,U.K. See lecture notes on: http://www.qi.damtp.cam.ac.uk/node/223 Quantum Information Theory Born out of Classical Information
More informationCh. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited
Ch. 8 Math Preliminaries for Lossy Coding 8.4 Info Theory Revisited 1 Info Theory Goals for Lossy Coding Again just as for the lossless case Info Theory provides: Basis for Algorithms & Bounds on Performance
More informationPhysical Layer and Coding
Physical Layer and Coding Muriel Médard Professor EECS Overview A variety of physical media: copper, free space, optical fiber Unified way of addressing signals at the input and the output of these media:
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationEs e j4φ +4N n. 16 KE s /N 0. σ 2ˆφ4 1 γ s. p(φ e )= exp 1 ( 2πσ φ b cos N 2 φ e 0
Problem 6.15 : he received signal-plus-noise vector at the output of the matched filter may be represented as (see (5-2-63) for example) : r n = E s e j(θn φ) + N n where θ n =0,π/2,π,3π/2 for QPSK, and
More informationBASICS OF DETECTION AND ESTIMATION THEORY
BASICS OF DETECTION AND ESTIMATION THEORY 83050E/158 In this chapter we discuss how the transmitted symbols are detected optimally from a noisy received signal (observation). Based on these results, optimal
More informationThe Effect upon Channel Capacity in Wireless Communications of Perfect and Imperfect Knowledge of the Channel
The Effect upon Channel Capacity in Wireless Communications of Perfect and Imperfect Knowledge of the Channel Muriel Medard, Trans. on IT 000 Reading Group Discussion March 0, 008 Intro./ Overview Time
More informationRandom signals II. ÚPGM FIT VUT Brno,
Random signals II. Jan Černocký ÚPGM FIT VUT Brno, cernocky@fit.vutbr.cz 1 Temporal estimate of autocorrelation coefficients for ergodic discrete-time random process. ˆR[k] = 1 N N 1 n=0 x[n]x[n + k],
More informationBasic information theory
Basic information theory Communication system performance is limited by Available signal power Background noise Bandwidth limits. Can we postulate an ideal system based on physical principles, against
More informationMultiuser Capacity Analysis of WDM in Nonlinear Fiber Optics
Multiuser Capacity Analysis of WDM in Nonlinear Fiber Optics Mohammad H. Taghavi N., George C. Papen, and Paul H. Siegel Dept. of ECE, UCSD, La Jolla, CA 92093 Email: {mtaghavi, gpapen, psiegel}@ucsd.edu
More informationRevision of Lecture 4
Revision of Lecture 4 We have discussed all basic components of MODEM Pulse shaping Tx/Rx filter pair Modulator/demodulator Bits map symbols Discussions assume ideal channel, and for dispersive channel
More informationChapter 4: Continuous channel and its capacity
meghdadi@ensil.unilim.fr Reference : Elements of Information Theory by Cover and Thomas Continuous random variable Gaussian multivariate random variable AWGN Band limited channel Parallel channels Flat
More informationSignal Design for Band-Limited Channels
Wireless Information Transmission System Lab. Signal Design for Band-Limited Channels Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal
More informationa) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics.
Digital Modulation and Coding Tutorial-1 1. Consider the signal set shown below in Fig.1 a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics. b) What is the minimum Euclidean
More informationThis examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of
More informationDigital Baseband Systems. Reference: Digital Communications John G. Proakis
Digital Baseband Systems Reference: Digital Communications John G. Proais Baseband Pulse Transmission Baseband digital signals - signals whose spectrum extend down to or near zero frequency. Model of the
More informationLecture 2. Fading Channel
1 Lecture 2. Fading Channel Characteristics of Fading Channels Modeling of Fading Channels Discrete-time Input/Output Model 2 Radio Propagation in Free Space Speed: c = 299,792,458 m/s Isotropic Received
More informationLecture 14 February 28
EE/Stats 376A: Information Theory Winter 07 Lecture 4 February 8 Lecturer: David Tse Scribe: Sagnik M, Vivek B 4 Outline Gaussian channel and capacity Information measures for continuous random variables
More informationCh. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory
Ch. 8 Math Preliminaries for Lossy Coding 8.5 Rate-Distortion Theory 1 Introduction Theory provide insight into the trade between Rate & Distortion This theory is needed to answer: What do typical R-D
More informationNotes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel
Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic
More informationCS 630 Basic Probability and Information Theory. Tim Campbell
CS 630 Basic Probability and Information Theory Tim Campbell 21 January 2003 Probability Theory Probability Theory is the study of how best to predict outcomes of events. An experiment (or trial or event)
More informationEE5713 : Advanced Digital Communications
EE5713 : Advanced Digital Communications Week 12, 13: Inter Symbol Interference (ISI) Nyquist Criteria for ISI Pulse Shaping and Raised-Cosine Filter Eye Pattern Equalization (On Board) 20-May-15 Muhammad
More informationGambling and Information Theory
Gambling and Information Theory Giulio Bertoli UNIVERSITEIT VAN AMSTERDAM December 17, 2014 Overview Introduction Kelly Gambling Horse Races and Mutual Information Some Facts Shannon (1948): definitions/concepts
More informationELEMENT OF INFORMATION THEORY
History Table of Content ELEMENT OF INFORMATION THEORY O. Le Meur olemeur@irisa.fr Univ. of Rennes 1 http://www.irisa.fr/temics/staff/lemeur/ October 2010 1 History Table of Content VERSION: 2009-2010:
More informationSquare Root Raised Cosine Filter
Wireless Information Transmission System Lab. Square Root Raised Cosine Filter Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal design
More informationCapacity of the Discrete Memoryless Energy Harvesting Channel with Side Information
204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener
More informationLecture 8: Channel Capacity, Continuous Random Variables
EE376A/STATS376A Information Theory Lecture 8-02/0/208 Lecture 8: Channel Capacity, Continuous Random Variables Lecturer: Tsachy Weissman Scribe: Augustine Chemparathy, Adithya Ganesh, Philip Hwang Channel
More informationNote that the new channel is noisier than the original two : and H(A I +A2-2A1A2) > H(A2) (why?). min(c,, C2 ) = min(1 - H(a t ), 1 - H(A 2 )).
l I ~-16 / (a) (5 points) What is the capacity Cr of the channel X -> Y? What is C of the channel Y - Z? (b) (5 points) What is the capacity C 3 of the cascaded channel X -3 Z? (c) (5 points) A ow let.
More informationCoding theory: Applications
INF 244 a) Textbook: Lin and Costello b) Lectures (Tu+Th 12.15-14) covering roughly Chapters 1,9-12, and 14-18 c) Weekly exercises: For your convenience d) Mandatory problem: Programming project (counts
More informationthat efficiently utilizes the total available channel bandwidth W.
Signal Design for Band-Limited Channels Wireless Information Transmission System Lab. Institute of Communications Engineering g National Sun Yat-sen University Introduction We consider the problem of signal
More informationX 1 : X Table 1: Y = X X 2
ECE 534: Elements of Information Theory, Fall 200 Homework 3 Solutions (ALL DUE to Kenneth S. Palacio Baus) December, 200. Problem 5.20. Multiple access (a) Find the capacity region for the multiple-access
More informationEE4304 C-term 2007: Lecture 17 Supplemental Slides
EE434 C-term 27: Lecture 17 Supplemental Slides D. Richard Brown III Worcester Polytechnic Institute, Department of Electrical and Computer Engineering February 5, 27 Geometric Representation: Optimal
More informationComputer Vision & Digital Image Processing
Computer Vision & Digital Image Processing Image Restoration and Reconstruction I Dr. D. J. Jackson Lecture 11-1 Image restoration Restoration is an objective process that attempts to recover an image
More informationBlock 2: Introduction to Information Theory
Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation
More informationSimultaneous SDR Optimality via a Joint Matrix Decomp.
Simultaneous SDR Optimality via a Joint Matrix Decomposition Joint work with: Yuval Kochman, MIT Uri Erez, Tel Aviv Uni. May 26, 2011 Model: Source Multicasting over MIMO Channels z 1 H 1 y 1 Rx1 ŝ 1 s
More informationLecture 22: Final Review
Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information
More informationNoisy-Channel Coding
Copyright Cambridge University Press 2003. On-screen viewing permitted. Printing not permitted. http://www.cambridge.org/05264298 Part II Noisy-Channel Coding Copyright Cambridge University Press 2003.
More informationECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)
ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling
More informationMODULATION AND CODING FOR QUANTIZED CHANNELS. Xiaoying Shao and Harm S. Cronie
MODULATION AND CODING FOR QUANTIZED CHANNELS Xiaoying Shao and Harm S. Cronie x.shao@ewi.utwente.nl, h.s.cronie@ewi.utwente.nl University of Twente, Faculty of EEMCS, Signals and Systems Group P.O. box
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Department of Mechanical Engineering 6.050J/2.0J Information and Entropy Spring 2005 Issued: March 7, 2005
More information(Classical) Information Theory III: Noisy channel coding
(Classical) Information Theory III: Noisy channel coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract What is the best possible way
More informationLecture 11: Information theory THURSDAY, FEBRUARY 21, 2019
Lecture 11: Information theory DANIEL WELLER THURSDAY, FEBRUARY 21, 2019 Agenda Information and probability Entropy and coding Mutual information and capacity Both images contain the same fraction of black
More informationThis examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 6 December 2006 This examination consists of
More informationCapacity of multiple-input multiple-output (MIMO) systems in wireless communications
15/11/02 Capacity of multiple-input multiple-output (MIMO) systems in wireless communications Bengt Holter Department of Telecommunications Norwegian University of Science and Technology 1 Outline 15/11/02
More informationOn the Capacity of the Two-Hop Half-Duplex Relay Channel
On the Capacity of the Two-Hop Half-Duplex Relay Channel Nikola Zlatanov, Vahid Jamali, and Robert Schober University of British Columbia, Vancouver, Canada, and Friedrich-Alexander-University Erlangen-Nürnberg,
More informationNeural coding Ecological approach to sensory coding: efficient adaptation to the natural environment
Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment Jean-Pierre Nadal CNRS & EHESS Laboratoire de Physique Statistique (LPS, UMR 8550 CNRS - ENS UPMC Univ.
More informationProblem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.
(b) We note that the above capacity is the same to the capacity of the binary symmetric channel. Indeed, if we considerthe grouping of the output symbols into a = {y 1,y 2 } and b = {y 3,y 4 } we get a
More informationShannon meets Wiener II: On MMSE estimation in successive decoding schemes
Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises
More informationThe Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels
The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels Parastoo Sadeghi National ICT Australia (NICTA) Sydney NSW 252 Australia Email: parastoo@student.unsw.edu.au Predrag
More informationA Brief Introduction to Shannon s Information Theory
A Brief Introduction to Shannon s Information Theory Ricky Xiaofeng Chen arxiv:1612.09316v2 [cs.it] 11 Dec 2018 Abstract This article consists of a brief introduction to the Shannon information theory.
More informationITCT Lecture IV.3: Markov Processes and Sources with Memory
ITCT Lecture IV.3: Markov Processes and Sources with Memory 4. Markov Processes Thus far, we have been occupied with memoryless sources and channels. We must now turn our attention to sources with memory.
More informationLecture 2: August 31
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy
More informationEnergy State Amplification in an Energy Harvesting Communication System
Energy State Amplification in an Energy Harvesting Communication System Omur Ozel Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland College Park, MD 20742 omur@umd.edu
More informationA General Formula for Compound Channel Capacity
A General Formula for Compound Channel Capacity Sergey Loyka, Charalambos D. Charalambous University of Ottawa, University of Cyprus ETH Zurich (May 2015), ISIT-15 1/32 Outline 1 Introduction 2 Channel
More informationInformation Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results
Information Theory Lecture 10 Network Information Theory (CT15); a focus on channel capacity results The (two-user) multiple access channel (15.3) The (two-user) broadcast channel (15.6) The relay channel
More informationELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS
EC 32 (CR) Total No. of Questions :09] [Total No. of Pages : 02 III/IV B.Tech. DEGREE EXAMINATIONS, APRIL/MAY- 207 Second Semester ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS Time: Three Hours
More information16.584: Random (Stochastic) Processes
1 16.584: Random (Stochastic) Processes X(t): X : RV : Continuous function of the independent variable t (time, space etc.) Random process : Collection of X(t, ζ) : Indexed on another independent variable
More informationAnalog Electronics 2 ICS905
Analog Electronics 2 ICS905 G. Rodriguez-Guisantes Dépt. COMELEC http://perso.telecom-paristech.fr/ rodrigez/ens/cycle_master/ November 2016 2/ 67 Schedule Radio channel characteristics ; Analysis and
More informationInformation and Entropy
Information and Entropy Shannon s Separation Principle Source Coding Principles Entropy Variable Length Codes Huffman Codes Joint Sources Arithmetic Codes Adaptive Codes Thomas Wiegand: Digital Image Communication
More informationECE598: Information-theoretic methods in high-dimensional statistics Spring 2016
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma
More informationECE6604 PERSONAL & MOBILE COMMUNICATIONS. Week 3. Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process
1 ECE6604 PERSONAL & MOBILE COMMUNICATIONS Week 3 Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process 2 Multipath-Fading Mechanism local scatterers mobile subscriber base station
More informationGaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26
Gaussian channel Information theory 2013, lecture 6 Jens Sjölund 8 May 2013 Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Outline 1 Definitions 2 The coding theorem for Gaussian channel 3 Bandlimited
More informationInformation Theory CHAPTER. 5.1 Introduction. 5.2 Entropy
Haykin_ch05_pp3.fm Page 207 Monday, November 26, 202 2:44 PM CHAPTER 5 Information Theory 5. Introduction As mentioned in Chapter and reiterated along the way, the purpose of a communication system is
More informationWideband Fading Channel Capacity with Training and Partial Feedback
Wideband Fading Channel Capacity with Training and Partial Feedback Manish Agarwal, Michael L. Honig ECE Department, Northwestern University 145 Sheridan Road, Evanston, IL 6008 USA {m-agarwal,mh}@northwestern.edu
More informationEE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design
Chapter 4 Receiver Design Chapter 4 Receiver Design Probability of Bit Error Pages 124-149 149 Probability of Bit Error The low pass filtered and sampled PAM signal results in an expression for the probability
More informationEE401: Advanced Communication Theory
EE401: Advanced Communication Theory Professor A. Manikas Chair of Communications and Array Processing Imperial College London Introductory Concepts Prof. A. Manikas (Imperial College) EE.401: Introductory
More informationShannon s noisy-channel theorem
Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for
More informationInteractions of Information Theory and Estimation in Single- and Multi-user Communications
Interactions of Information Theory and Estimation in Single- and Multi-user Communications Dongning Guo Department of Electrical Engineering Princeton University March 8, 2004 p 1 Dongning Guo Communications
More informationNikola Zlatanov, Erik Sippel, Vahid Jamali, and Robert Schober. Abstract
Capacity of the Gaussian Two-Hop Full-Duplex Relay Channel with Residual Self-Interference Nikola Zlatanov, Erik Sippel, Vahid Jamali, and Robert Schober Abstract arxiv:50.07394v6 [cs.it] 6 Nov 06 In this
More information