Digital Communication Systems ECS 452

Size: px
Start display at page:

Download "Digital Communication Systems ECS 452"

Transcription

1 Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong Discrete Channel Office Hours: BKD Monday 4:00-6:00 Wednesday 4:40-6:00

2 Noise & Interference Elements of digital commu. sys. Message Transmitter Information Source Source Encoder Remove redundancy (compression) Channel Encoder Add systematic redundancy to combat errors introduced by the channel Digital Modulator Channel Transmitted Signal Recovered Message Destination Source Decoder Receiver Channel Decoder Digital Demodulator Received Signal 2

3 Noise & Interference System considered previously Message Transmitter Information Source Source Encoder Remove redundancy Channel Encoder Add systematic redundancy Digital Modulator Transmitted Signal Channel Recovered Message Destination Source Decoder Receiver Channel Decoder Digital Demodulator Received Signal 3

4 Noise & Interference System considered in this section Message Transmitter Information Source Source Encoder Remove redundancy Channel Encoder Add systematic redundancy Digital Modulator Equivalent Channel Channel Transmitted Signal X: channel input Recovered Message Destination Source Decoder Receiver Channel Decoder Digital Demodulator Received Signal Y: channel output 4

5 Ex: BSC >> BSC_demo ans = ans = %% Simulation parameters % The number of symbols to be transmitted n = 20; % Channel Input S_X = [0 ]; S_Y = [0 ]; p_x = [ ]; % Channel Characteristics p = 0.; Q = [-p p; p -p]; p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = Elapsed time is seconds. [BSC_demo.m]

6 MATLAB %% Generating the channel input x x = randsrc(,n,[s_x;p_x]); % channel input %% Applying the effect of the channel to create the channel output y y = DMC_Channel_sim(x,S_X,S_Y,Q); % channel output function y = DMC_Channel_sim(x,S_X,S_Y,Q) %% Applying the effect of the channel to create the channel output y y = zeros(size(x)); % preallocation for k = :length(x) % Look at the channel input one by one. Choose the corresponding row % from the Q matrix to generate the channel output. y(k) = randsrc(,,[s_y;q(find(s_x == x(k)),:)]); end [DMC_Channel_sim.m] 6 [DMC_Analysis_demo.m]

7 Rel. freq. from the simulation %% Statistical Analysis % The probability values for the channel inputs p_x % Theoretical probability p_x_sim = hist(x,s_x)/n % Relative frequencies from the simulation % The probability values for the channel outputs q = p_x*q % Theoretical probability q_sim = hist(y,s_y)/n % Relative frequencies from the simulation % The channel transition probabilities from the simulation Q_sim = []; for k = :length(s_x) I = find(x==s_x(k)); LI = length(i); rel_freq_xk = LI/n; yc = y(i); cond_rel_freq = hist(yc,s_y)/li; Q_sim = [Q_sim; cond_rel_freq]; end Q % Theoretical probability Q_sim % Relative frequencies from the simulation 7 [DMC_Analysis_demo.m]

8 MATLAB %% Naive Decoder x_hat = y; %% Error Probability PE_sim = -sum(x==x_hat)/n % Error probability from the simulation % Calculation of the theoretical error probability PC = 0; for k = :length(s_x) t = S_X(k); i = find(s_y == t); if length(i) == PC = PC+ p_x(k)*q(k,i); end end PE_theretical = -PC 8 [DMC_Analysis_demo.m]

9 Ex: BSC >> BSC_demo %% Simulation parameters % The number of symbols to be transmitted n = e4; % Channel Input S_X = [0 ]; S_Y = [0 ]; p_x = [ ]; % Channel Characteristics p = 0.; Q = [-p p; p -p]; p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = Elapsed time is seconds. 9 [BSC_demo.m]

10 Binary Symmetric Channel (BSC) Three equivalent descriptions of BSC: P Y 0 X 0 Q 0 0 p P Y X 0 Q 0 p P Y 0 X Q 0 p P Y X Q p X 0 p p -p -p 0 Y Q y 0 x 0 p p p p 0

11 Tree Diagram and Conditional Probability P A B A P A B =P A B P B B P B P A c B A c P A c B =P A c B P B P B c P A B c A P A B c =P A B c P B c B c P A c B c A c P A c B c =P A c B c P B c

12 Tree Diagram and Total Probability Theorem = = P A = P A B P B + P A B c P B c = = 2

13 Binary Asymmetric Channel (BAC) Three equivalent descriptions of BAC: P Y 0 X 0 Q 0 0 P Y X 0 Q 0 P Y 0 X Q 0 P Y X Q X Y Q y 0 x 0 3

14 4 Ex: BAC %% Simulation parameters % The number of symbols to be transmitted n = 20; % Binary Assymmetric Channel (BAC) % Ex 3.8 in lecture note (.3 in [Z&T, 200]) % Channel Input S_X = [0 ]; S_Y = [0 ]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; x: y: >> BAC_demo ans = ans = p 0 = 0. 5 X p = Y p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = [BAC_demo.m]

15 MATLAB %% Naive Decoder x_hat = y; %% Error Probability PE_sim = -sum(x==x_hat)/n % Error probability from the simulation % Calculation of the theoretical error probability PC = 0; for k = :length(s_x) t = S_X(k); i = find(s_y == t); if length(i) == PC = PC+ p_x(k)*q(k,i); end end PE_theretical = -PC 5 [DMC_Analysis_demo.m]

16 6 Ex: BAC %% Simulation parameters % The number of symbols to be transmitted n = e4; % Binary Assymmetric Channel (BAC) % Ex 3.8 in lecture note (.3 in [Z&T, 200]) % Channel Input S_X = [0 ]; S_Y = [0 ]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; p 0 = 0. 5 X p = Y p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = [BAC_demo.m]

17 7 x: y: Ex: DMC %% Simulation parameters % The number of symbols to be transmitted n = 20; % General DMC % Ex. 3.6 in lecture note % Channel Input S_X = [0 ]; S_Y = [ 2 3]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; >> DMC_demo ans = ans = p 0 = 0. 2 X p = Y p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = [DMC_demo.m]

18 Ex: DMC 8 %% Simulation parameters % The number of symbols to be transmitted n = e4; % General DMC % Ex. 3.6 in lecture note % Channel Input S_X = [0 ]; S_Y = [ 2 3]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; p 0 = 0. 2 X p = Y p_x = p_x_sim = q = q_sim = Q = Q_sim = PE_sim = PE_theretical = [DMC_demo.m]

19 Evaluation of Probability Consider two random variables X and Y. Suppose their joint pmf matrix is P XY, x y Find P X + Y < Step : Find the pairs (x,y) that satisfy the condition x+y < 7 One way to do this is to first construct the matrix of x+y. x y x y

20 Evaluation of Probability Consider two random variables X and Y. Suppose their joint pmf matrix is P XY, x y Find P X + Y < Step 2: Add the corresponding probabilities from the joint pmf (matrix) P X + Y < 7 = = 0.3 x y x y

21 Evaluation of Probability Consider two random variables X and Y. Suppose their joint pmf matrix is P XY, x y Find P X = Y P X = Y = = 0.5 2

22 Sum of two discrete RVs Consider two random variables X and Y. Suppose their joint pmf matrix is P XY, x y Find P X + Y = P X + Y = 7 = 0. x y x y

23 Optimal Decoder for BSC close all; clear all; tic 23 >> BSC_decoder_ALL_demo Decoder_Table_ALL = ans = x 0 0 x.0000 P E Optimal_Detector = 0 Min_PE = Elapsed time is seconds. [BSC_decoder_ALL_demo.m] %% Simulation parameters % Channel Input S_X = [0 ]; S_Y = [0 ]; p0 = 0.8; p = -p0; p_x = [p0 p]; % Channel Characteristics p = 0.; Q = [-p p; p -p]; %% All possible "reasonable" decoders % X_hat = Y; X_hat = -Y; X_hat = ; X_hat = 0 Decoder_Table_ALL = [0 ; 0; ; 0 0] %% Calculate the error probability for each of the decoder PE_ALL = []; for k = :size(decoder_table_all,) Decoder_Table = Decoder_Table_ALL(k,:); PC = 0; for k = :length(s_x) I = (Decoder_Table == S_X(k)); Q_row = Q(k,:); PC = PC + p_x(k)*sum(q_row(i)); end PE_theretical = -PC; PE_ALL = [PE_ALL; PE_theretical]; end %% Display the results [Decoder_Table_ALL PE_ALL] %% Find the optimal detectors [V I] = min(pe_all); Optimal_Detector = Decoder_Table_ALL(I,:) Min_PE = V toc

24 DIY Decoder >> DMC_decoder_DIY_demo ans = ans = ans = PE_sim = PE_theretical = Elapsed time is seconds. %% Simulation parameters % The number of symbols to be transmitted n = 20; % General DMC % Ex. 3.6 in lecture note % Channel Input S_X = [0 ]; S_Y = [ 2 3]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; %% DIY Decoder Decoder_Table = [0 0]; % The decoded values corresponding to the received Y 24 [DMC_decoder_DIY_demo.m]

25 DIY Decoder %% DIY Decoder Decoder_Table = [0 0]; % The decoded values corresponding to the received Y % Decode according to the decoder table x_hat = y; % preallocation for k = :length(s_y) I = (y==s_y(k)); x_hat(i) = Decoder_Table(k); end PE_sim = -sum(x==x_hat)/n % Error probability from the simulation % Calculation of the theoretical error probability PC = 0; for k = :length(s_x) I = (Decoder_Table == S_X(k)); q = Q(k,:); PC = PC+ p_x(k)*sum(q(i)); end PE_theretical = -PC 25 [DMC_decoder_DIY_demo.m]

26 DIY Decoder >> DMC_decoder_DIY_demo PE_sim = PE_theretical = Elapsed time is seconds. %% Simulation parameters % The number of symbols to be transmitted n = e4; % General DMC % Ex. 3.6 in lecture note % Channel Input S_X = [0 ]; S_Y = [ 2 3]; p_x = [ ]; % Channel Characteristics Q = [ ; ]; %% DIY Decoder Decoder_Table = [0 0]; % The decoded values corresponding to the received Y [DMC_decoder_DIY_demo.m]

27 Searching for the Optimal Detector >> DMC_decoder_ALL_demo ans = x x 2 x 3 P E Min_PE = Optimal_Detector = Elapsed time is seconds. 27

28 Guessing Game Consider a random variable X whose pmf is / 4, x,2, p X x / 2, x 3, 0, otherwise. You have to guess the value of X. This game is repeated many times. (Say, 00 times) The goodness of your guess is determined by how often you guess correctly. For example, get 0 each time that you guess correctly. How should you guess? 28

29 Guessing Game 3000 xˆ xˆ xˆ

30 Guessing Game: Version 2 Consider a random variable X whose pmf is / 4, x,2, p X x / 2, x 3, 0, otherwise. You have to guess the value of X. This game is repeated many times. (Say, 00 times) The goodness of your guess is determined by how close you are to the value of X. For example, you pay nothing if your guess is correct but you have to pay X x 2 if your guess is wrong. How should you guess? 30

31 Guessing Game: Version xˆ xˆ xˆ [GuessingGame_2.m]

32 Guessing Game: Version Average Return g [GuessingGame_2_2.m]

33 MAP Decoder %% MAP Decoder P = diag(p_x)*q; % Weight the channel transition probability by the % corresponding prior probability. [V I] = max(p); % For I, the default MATLAB behavior is that when there are % multiple max, the index of the first one is returned. Decoder_Table = S_X(I) % The decoded values corresponding to the received Y %% Decode according to the decoder table x_hat = y; % preallocation for k = :length(s_y) I = (y==s_y(k)); x_hat(i) = Decoder_Table(k); end PE_sim = -sum(x==x_hat)/n % Error probability from the simulation %% Calculation of the theoretical error probability PC = 0; for k = :length(s_x) I = (Decoder_Table == S_X(k)); Q_row = Q(k,:); PC = PC+ p_x(k)*sum(q_row(i)); end PE_theretical = -PC 33 [DMC_decoder_MAP_demo.m]

34 ML Decoder %% ML Decoder [V I] = max(q); % For I, the default MATLAB behavior is that when there are % multiple max, the index of the first one is returned. Decoder_Table = S_X(I) % The decoded values corresponding to the received Y %% Decode according to the decoder table x_hat = y; % preallocation for k = :length(s_y) I = (y==s_y(k)); x_hat(i) = Decoder_Table(k); end PE_sim = -sum(x==x_hat)/n % Error probability from the simulation %% Calculation of the theoretical error probability PC = 0; for k = :length(s_x) I = (Decoder_Table == S_X(k)); Q_row = Q(k,:); PC = PC+ p_x(k)*sum(q_row(i)); end PE_theretical = -PC 34 [DMC_decoder_ML_demo.m]

Digital Communication Systems ECS 452

Digital Communication Systems ECS 452 Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th 3 Discrete Memoryless Channel (DMC) Office Hours: BKD, 6th floor of Sirindhralai building Tuesday 4:20-5:20

More information

ECS 452: Digital Communication Systems 2015/2. HW 1 Due: Feb 5

ECS 452: Digital Communication Systems 2015/2. HW 1 Due: Feb 5 ECS 452: Digital Communication Systems 2015/2 HW 1 Due: Feb 5 Lecturer: Asst. Prof. Dr. Prapun Suksompong Instructions (a) Must solve all non-optional problems. (5 pt) (i) Write your first name and the

More information

4 An Introduction to Channel Coding and Decoding over BSC

4 An Introduction to Channel Coding and Decoding over BSC 4 An Introduction to Channel Coding and Decoding over BSC 4.1. Recall that channel coding introduces, in a controlled manner, some redundancy in the (binary information sequence that can be used at the

More information

Digital Communication Systems ECS 452

Digital Communication Systems ECS 452 Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th 2. Source Coding 1 Office Hours: BKD, 6th floor of Sirindhralai building Tuesday 14:20-15:20 Wednesday 14:20-15:20

More information

Lecture 4 Noisy Channel Coding

Lecture 4 Noisy Channel Coding Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem

More information

Digital Circuits ECS 371

Digital Circuits ECS 371 Digital Circuits ECS 371 Dr. Prapun Suksompong prapun@siit.tu.ac.th Lecture 18 Office Hours: BKD 3601-7 Monday 9:00-10:30, 1:30-3:30 Tuesday 10:30-11:30 1 Announcement Reading Assignment: Chapter 7: 7-1,

More information

ECS 332: Principles of Communications 2012/1. HW 4 Due: Sep 7

ECS 332: Principles of Communications 2012/1. HW 4 Due: Sep 7 ECS 332: Principles of Communications 2012/1 HW 4 Due: Sep 7 Lecturer: Prapun Suksompong, Ph.D. Instructions (a) ONE part of a question will be graded (5 pt). Of course, you do not know which part will

More information

LECTURE 10. Last time: Lecture outline

LECTURE 10. Last time: Lecture outline LECTURE 10 Joint AEP Coding Theorem Last time: Error Exponents Lecture outline Strong Coding Theorem Reading: Gallager, Chapter 5. Review Joint AEP A ( ɛ n) (X) A ( ɛ n) (Y ) vs. A ( ɛ n) (X, Y ) 2 nh(x)

More information

CS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability

More information

ECE Information theory Final

ECE Information theory Final ECE 776 - Information theory Final Q1 (1 point) We would like to compress a Gaussian source with zero mean and variance 1 We consider two strategies In the first, we quantize with a step size so that the

More information

One Lesson of Information Theory

One Lesson of Information Theory Institut für One Lesson of Information Theory Prof. Dr.-Ing. Volker Kühn Institute of Communications Engineering University of Rostock, Germany Email: volker.kuehn@uni-rostock.de http://www.int.uni-rostock.de/

More information

Computer Applications for Engineers ET 601

Computer Applications for Engineers ET 601 Computer Applications for Engineers ET 601 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th Random Variables (Con t) 1 Office Hours: (BKD 3601-7) Wednesday 9:30-11:30 Wednesday 16:00-17:00 Thursday

More information

Cyclic Codes. Saravanan Vijayakumaran August 26, Department of Electrical Engineering Indian Institute of Technology Bombay

Cyclic Codes. Saravanan Vijayakumaran August 26, Department of Electrical Engineering Indian Institute of Technology Bombay 1 / 25 Cyclic Codes Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay August 26, 2014 2 / 25 Cyclic Codes Definition A cyclic shift

More information

Lecture 15: Conditional and Joint Typicaility

Lecture 15: Conditional and Joint Typicaility EE376A Information Theory Lecture 1-02/26/2015 Lecture 15: Conditional and Joint Typicaility Lecturer: Kartik Venkat Scribe: Max Zimet, Brian Wai, Sepehr Nezami 1 Notation We always write a sequence of

More information

HW Solution 12 Due: Dec 2, 9:19 AM

HW Solution 12 Due: Dec 2, 9:19 AM ECS 315: Probability and Random Processes 2015/1 HW Solution 12 Due: Dec 2, 9:19 AM Lecturer: Prapun Suksompong, Ph.D. Problem 1. Let X E(3). (a) For each of the following function g(x). Indicate whether

More information

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122 Lecture 5: Channel Capacity Copyright G. Caire (Sample Lectures) 122 M Definitions and Problem Setup 2 X n Y n Encoder p(y x) Decoder ˆM Message Channel Estimate Definition 11. Discrete Memoryless Channel

More information

Computer Applications for Engineers ET 601

Computer Applications for Engineers ET 601 Computer Applications for Engineers ET 601 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th Random Variables (Con t) 1 Office Hours: (BKD 3601-7) Wednesday 9:30-11:30 Wednesday 16:00-17:00 Thursday

More information

ECS455 Chapter 3 Call Blocking Probability

ECS455 Chapter 3 Call Blocking Probability ECS455 Chapter 3 Call Blocking Probability 1 Dr.Prapun Suksompong prapun.com/ecs455 Office Hours: BKD, 6th floor of Sirindhralai building Tuesday 14:20-15:20 Wednesday 14:20-15:20 Friday 9:15-10:15 Introduction

More information

Lecture 4 Channel Coding

Lecture 4 Channel Coding Capacity and the Weak Converse Lecture 4 Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 15, 2014 1 / 16 I-Hsiang Wang NIT Lecture 4 Capacity

More information

HW Solution 3 Due: July 15

HW Solution 3 Due: July 15 ECS 315: Probability and Random Processes 2010/1 HW Solution 3 Due: July 15 Lecturer: Prapun Suksompong, Ph.D. Instructions (a) A part of ONE question will be graded. Of course, you do not know which problem

More information

CSCI 2570 Introduction to Nanocomputing

CSCI 2570 Introduction to Nanocomputing CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication

More information

Lecture 10: Broadcast Channel and Superposition Coding

Lecture 10: Broadcast Channel and Superposition Coding Lecture 10: Broadcast Channel and Superposition Coding Scribed by: Zhe Yao 1 Broadcast channel M 0M 1M P{y 1 y x} M M 01 1 M M 0 The capacity of the broadcast channel depends only on the marginal conditional

More information

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Tara Javidi These lecture notes were originally developed by late Prof. J. K. Wolf. UC San Diego Spring 2014 1 / 8 I

More information

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results Information Theory Lecture 10 Network Information Theory (CT15); a focus on channel capacity results The (two-user) multiple access channel (15.3) The (two-user) broadcast channel (15.6) The relay channel

More information

Lecture 8: Shannon s Noise Models

Lecture 8: Shannon s Noise Models Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 8: Shannon s Noise Models September 14, 2007 Lecturer: Atri Rudra Scribe: Sandipan Kundu& Atri Rudra Till now we have

More information

PCM Reference Chapter 12.1, Communication Systems, Carlson. PCM.1

PCM Reference Chapter 12.1, Communication Systems, Carlson. PCM.1 PCM Reference Chapter 1.1, Communication Systems, Carlson. PCM.1 Pulse-code modulation (PCM) Pulse modulations use discrete time samples of analog signals the transmission is composed of analog information

More information

Introduction to Information Theory. Part 4

Introduction to Information Theory. Part 4 Introduction to Information Theory Part 4 A General Communication System CHANNEL Information Source Transmitter Channel Receiver Destination 10/2/2012 2 Information Channel Input X Channel Output Y 10/2/2012

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2013, 29 November 2013 Memoryless Sources Arithmetic Coding Sources with Memory Markov Example 2 / 21 Encoding the output

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

Homework Set #3 Rates definitions, Channel Coding, Source-Channel coding

Homework Set #3 Rates definitions, Channel Coding, Source-Channel coding Homework Set # Rates definitions, Channel Coding, Source-Channel coding. Rates (a) Channels coding Rate: Assuming you are sending 4 different messages using usages of a channel. What is the rate (in bits

More information

ECE Information theory Final (Fall 2008)

ECE Information theory Final (Fall 2008) ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1

More information

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 6 I. CHANNEL CODING. X n (m) P Y X 6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2011, 28 November 2011 Memoryless Sources Arithmetic Coding Sources with Memory 2 / 19 Summary of last lecture Prefix-free

More information

Digital communication system. Shannon s separation principle

Digital communication system. Shannon s separation principle Digital communication system Representation of the source signal by a stream of (binary) symbols Adaptation to the properties of the transmission channel information source source coder channel coder modulation

More information

9 THEORY OF CODES. 9.0 Introduction. 9.1 Noise

9 THEORY OF CODES. 9.0 Introduction. 9.1 Noise 9 THEORY OF CODES Chapter 9 Theory of Codes After studying this chapter you should understand what is meant by noise, error detection and correction; be able to find and use the Hamming distance for a

More information

Discrete Random Variable

Discrete Random Variable Discrete Random Variable Outcome of a random experiment need not to be a number. We are generally interested in some measurement or numerical attribute of the outcome, rather than the outcome itself. n

More information

Noisy channel communication

Noisy channel communication Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University

More information

Optimum Soft Decision Decoding of Linear Block Codes

Optimum Soft Decision Decoding of Linear Block Codes Optimum Soft Decision Decoding of Linear Block Codes {m i } Channel encoder C=(C n-1,,c 0 ) BPSK S(t) (n,k,d) linear modulator block code Optimal receiver AWGN Assume that [n,k,d] linear block code C is

More information

16.36 Communication Systems Engineering

16.36 Communication Systems Engineering MIT OpenCourseWare http://ocw.mit.edu 16.36 Communication Systems Engineering Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.36: Communication

More information

Example: sending one bit of information across noisy channel. Effects of the noise: flip the bit with probability p.

Example: sending one bit of information across noisy channel. Effects of the noise: flip the bit with probability p. Lecture 20 Page 1 Lecture 20 Quantum error correction Classical error correction Modern computers: failure rate is below one error in 10 17 operations Data transmission and storage (file transfers, cell

More information

Roll No. :... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/ CODING & INFORMATION THEORY. Time Allotted : 3 Hours Full Marks : 70

Roll No. :... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/ CODING & INFORMATION THEORY. Time Allotted : 3 Hours Full Marks : 70 Name : Roll No. :.... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/2011-12 2011 CODING & INFORMATION THEORY Time Allotted : 3 Hours Full Marks : 70 The figures in the margin indicate full marks

More information

(Classical) Information Theory III: Noisy channel coding

(Classical) Information Theory III: Noisy channel coding (Classical) Information Theory III: Noisy channel coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract What is the best possible way

More information

Information Theory - Entropy. Figure 3

Information Theory - Entropy. Figure 3 Concept of Information Information Theory - Entropy Figure 3 A typical binary coded digital communication system is shown in Figure 3. What is involved in the transmission of information? - The system

More information

Probability review. September 11, Stoch. Systems Analysis Introduction 1

Probability review. September 11, Stoch. Systems Analysis Introduction 1 Probability review Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania aribeiro@seas.upenn.edu http://www.seas.upenn.edu/users/~aribeiro/ September 11, 2015 Stoch.

More information

Covert Communication with Channel-State Information at the Transmitter

Covert Communication with Channel-State Information at the Transmitter Covert Communication with Channel-State Information at the Transmitter Si-Hyeon Lee Joint Work with Ligong Wang, Ashish Khisti, and Gregory W. Wornell July 27, 2017 1 / 21 Covert Communication Transmitter

More information

UNIT I INFORMATION THEORY. I k log 2

UNIT I INFORMATION THEORY. I k log 2 UNIT I INFORMATION THEORY Claude Shannon 1916-2001 Creator of Information Theory, lays the foundation for implementing logic in digital circuits as part of his Masters Thesis! (1939) and published a paper

More information

Information Theory and Coding Techniques

Information Theory and Coding Techniques Information Theory and Coding Techniques Lecture 1.2: Introduction and Course Outlines Information Theory 1 Information Theory and Coding Techniques Prof. Ja-Ling Wu Department of Computer Science and

More information

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding Ch 0 Introduction 0.1 Overview of Information Theory and Coding Overview The information theory was founded by Shannon in 1948. This theory is for transmission (communication system) or recording (storage

More information

Digital Communication Systems ECS 452. Asst. Prof. Dr. Prapun Suksompong 5.2 Binary Convolutional Codes

Digital Communication Systems ECS 452. Asst. Prof. Dr. Prapun Suksompong 5.2 Binary Convolutional Codes Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th 5.2 Binary Convolutional Codes 35 Binary Convolutional Codes Introduced by Elias in 1955 There, it is referred

More information

Welcome to Comp 411! 2) Course Objectives. 1) Course Mechanics. 3) Information. I thought this course was called Computer Organization

Welcome to Comp 411! 2) Course Objectives. 1) Course Mechanics. 3) Information. I thought this course was called Computer Organization Welcome to Comp 4! I thought this course was called Computer Organization David Macaulay ) Course Mechanics 2) Course Objectives 3) Information L - Introduction Meet the Crew Lectures: Leonard McMillan

More information

Physical Layer and Coding

Physical Layer and Coding Physical Layer and Coding Muriel Médard Professor EECS Overview A variety of physical media: copper, free space, optical fiber Unified way of addressing signals at the input and the output of these media:

More information

MATH3302 Coding Theory Problem Set The following ISBN was received with a smudge. What is the missing digit? x9139 9

MATH3302 Coding Theory Problem Set The following ISBN was received with a smudge. What is the missing digit? x9139 9 Problem Set 1 These questions are based on the material in Section 1: Introduction to coding theory. You do not need to submit your answers to any of these questions. 1. The following ISBN was received

More information

Digital Image Processing Lectures 25 & 26

Digital Image Processing Lectures 25 & 26 Lectures 25 & 26, Professor Department of Electrical and Computer Engineering Colorado State University Spring 2015 Area 4: Image Encoding and Compression Goal: To exploit the redundancies in the image

More information

Solutions to Homework Set #3 Channel and Source coding

Solutions to Homework Set #3 Channel and Source coding Solutions to Homework Set #3 Channel and Source coding. Rates (a) Channels coding Rate: Assuming you are sending 4 different messages using usages of a channel. What is the rate (in bits per channel use)

More information

channel of communication noise Each codeword has length 2, and all digits are either 0 or 1. Such codes are called Binary Codes.

channel of communication noise Each codeword has length 2, and all digits are either 0 or 1. Such codes are called Binary Codes. 5 Binary Codes You have already seen how check digits for bar codes (in Unit 3) and ISBN numbers (Unit 4) are used to detect errors. Here you will look at codes relevant for data transmission, for example,

More information

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase

More information

Answers and Solutions to (Even Numbered) Suggested Exercises in Sections of Grimaldi s Discrete and Combinatorial Mathematics

Answers and Solutions to (Even Numbered) Suggested Exercises in Sections of Grimaldi s Discrete and Combinatorial Mathematics Answers and Solutions to (Even Numbered) Suggested Exercises in Sections 6.5-6.9 of Grimaldi s Discrete and Combinatorial Mathematics Section 6.5 6.5.2. a. r = = + = c + e. So the error pattern is e =.

More information

Entropies & Information Theory

Entropies & Information Theory Entropies & Information Theory LECTURE I Nilanjana Datta University of Cambridge,U.K. See lecture notes on: http://www.qi.damtp.cam.ac.uk/node/223 Quantum Information Theory Born out of Classical Information

More information

Channel Coding I. Exercises SS 2017

Channel Coding I. Exercises SS 2017 Channel Coding I Exercises SS 2017 Lecturer: Dirk Wübben Tutor: Shayan Hassanpour NW1, Room N 2420, Tel.: 0421/218-62387 E-mail: {wuebben, hassanpour}@ant.uni-bremen.de Universität Bremen, FB1 Institut

More information

MATH 433 Applied Algebra Lecture 21: Linear codes (continued). Classification of groups.

MATH 433 Applied Algebra Lecture 21: Linear codes (continued). Classification of groups. MATH 433 Applied Algebra Lecture 21: Linear codes (continued). Classification of groups. Binary codes Let us assume that a message to be transmitted is in binary form. That is, it is a word in the alphabet

More information

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011 UCSD ECE53 Handout #40 Prof. Young-Han Kim Thursday, May 9, 04 Homework Set #8 Due: Thursday, June 5, 0. Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process,

More information

ECS455: Chapter 5 OFDM. ECS455: Chapter 5 OFDM. OFDM: Overview. OFDM Applications. Dr.Prapun Suksompong prapun.com/ecs455

ECS455: Chapter 5 OFDM. ECS455: Chapter 5 OFDM. OFDM: Overview. OFDM Applications. Dr.Prapun Suksompong prapun.com/ecs455 ECS455: Chapter 5 OFDM OFDM: Overview Let S = (S 1, S 2,, S ) contains the information symbols. S IFFT FFT Inverse fast Fourier transform Fast Fourier transform 1 Dr.Prapun Suksompong prapun.com/ecs455

More information

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

National University of Singapore Department of Electrical & Computer Engineering. Examination for

National University of Singapore Department of Electrical & Computer Engineering. Examination for National University of Singapore Department of Electrical & Computer Engineering Examination for EE5139R Information Theory for Communication Systems (Semester I, 2014/15) November/December 2014 Time Allowed:

More information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information 4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information Ramji Venkataramanan Signal Processing and Communications Lab Department of Engineering ramji.v@eng.cam.ac.uk

More information

Probability and Random Processes ECS 315

Probability and Random Processes ECS 315 Probability and Random Processes ECS 3 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th Discrete Random Variable Office Hours: Rangsit Library: Tuesday 6:2-7:2 BKD36-7: Thursday 6:-7: Roll a fair

More information

ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS

ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS EC 32 (CR) Total No. of Questions :09] [Total No. of Pages : 02 III/IV B.Tech. DEGREE EXAMINATIONS, APRIL/MAY- 207 Second Semester ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS Time: Three Hours

More information

Space-Time Coding for Multi-Antenna Systems

Space-Time Coding for Multi-Antenna Systems Space-Time Coding for Multi-Antenna Systems ECE 559VV Class Project Sreekanth Annapureddy vannapu2@uiuc.edu Dec 3rd 2007 MIMO: Diversity vs Multiplexing Multiplexing Diversity Pictures taken from lectures

More information

Coding theory: Applications

Coding theory: Applications INF 244 a) Textbook: Lin and Costello b) Lectures (Tu+Th 12.15-14) covering roughly Chapters 1,9-12, and 14-18 c) Weekly exercises: For your convenience d) Mandatory problem: Programming project (counts

More information

Introduction to Wireless & Mobile Systems. Chapter 4. Channel Coding and Error Control Cengage Learning Engineering. All Rights Reserved.

Introduction to Wireless & Mobile Systems. Chapter 4. Channel Coding and Error Control Cengage Learning Engineering. All Rights Reserved. Introduction to Wireless & Mobile Systems Chapter 4 Channel Coding and Error Control 1 Outline Introduction Block Codes Cyclic Codes CRC (Cyclic Redundancy Check) Convolutional Codes Interleaving Information

More information

Solutions to problems from Chapter 3

Solutions to problems from Chapter 3 Solutions to problems from Chapter 3 Manjunatha. P manjup.jnnce@gmail.com Professor Dept. of ECE J.N.N. College of Engineering, Shimoga February 28, 2016 For a systematic (7,4) linear block code, the parity

More information

Information in Biology

Information in Biology Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve

More information

Classical codes for quantum broadcast channels. arxiv: Ivan Savov and Mark M. Wilde School of Computer Science, McGill University

Classical codes for quantum broadcast channels. arxiv: Ivan Savov and Mark M. Wilde School of Computer Science, McGill University Classical codes for quantum broadcast channels arxiv:1111.3645 Ivan Savov and Mark M. Wilde School of Computer Science, McGill University International Symposium on Information Theory, Boston, USA July

More information

On the Use of Division Algebras for Wireless Communication

On the Use of Division Algebras for Wireless Communication On the Use of Division Algebras for Wireless Communication frederique@systems.caltech.edu California Institute of Technology AMS meeting, Davidson, March 3rd 2007 Outline A few wireless coding problems

More information

Lecture 12: November 6, 2017

Lecture 12: November 6, 2017 Information and Coding Theory Autumn 017 Lecturer: Madhur Tulsiani Lecture 1: November 6, 017 Recall: We were looking at codes of the form C : F k p F n p, where p is prime, k is the message length, and

More information

Lecture 1. Introduction

Lecture 1. Introduction Lecture 1. Introduction What is the course about? Logistics Questionnaire Dr. Yao Xie, ECE587, Information Theory, Duke University What is information? Dr. Yao Xie, ECE587, Information Theory, Duke University

More information

Quantitative Biology Lecture 3

Quantitative Biology Lecture 3 23 nd Sep 2015 Quantitative Biology Lecture 3 Gurinder Singh Mickey Atwal Center for Quantitative Biology Summary Covariance, Correlation Confounding variables (Batch Effects) Information Theory Covariance

More information

Shannon's Theory of Communication

Shannon's Theory of Communication Shannon's Theory of Communication An operational introduction 5 September 2014, Introduction to Information Systems Giovanni Sileno g.sileno@uva.nl Leibniz Center for Law University of Amsterdam Fundamental

More information

1 1 0, g Exercise 1. Generator polynomials of a convolutional code, given in binary form, are g

1 1 0, g Exercise 1. Generator polynomials of a convolutional code, given in binary form, are g Exercise Generator polynomials of a convolutional code, given in binary form, are g 0, g 2 0 ja g 3. a) Sketch the encoding circuit. b) Sketch the state diagram. c) Find the transfer function TD. d) What

More information

7 K-D MAP. Tuesday, August 13, :24 PM. ECS452 7 Page 1

7 K-D MAP. Tuesday, August 13, :24 PM. ECS452 7 Page 1 7 K-D MAP Tuesday, August 13, 13 :4 PM ECS45 7 Page 1 5 4 3 1 r -1 - -3-4 -5-6 -4-4 6 8 r 1 ECS45 7 Page ..4.15.3.1..5.1 5 5 5 5 n -5-5 n n -5-5 1 n 1 ECS45 7 Page 3 n -5-5 n n -5-5 1 n 1 s 1 s 3 s ECS45

More information

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing Lecture 7 Agenda for the lecture M-ary hypothesis testing and the MAP rule Union bound for reducing M-ary to binary hypothesis testing Introduction of the channel coding problem 7.1 M-ary hypothesis testing

More information

CMPSCI 240: Reasoning Under Uncertainty

CMPSCI 240: Reasoning Under Uncertainty CMPSCI 240: Reasoning Under Uncertainty Lecture 8 Prof. Hanna Wallach wallach@cs.umass.edu February 16, 2012 Reminders Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/

More information

A Comparison of Cooperative Transmission Schemes for the Multiple Access Relay Channel

A Comparison of Cooperative Transmission Schemes for the Multiple Access Relay Channel A Comparison of Cooperative Transmission Schemes for the ultiple Access elay Channel André Angierski andre.angierski@uni-rostock.de 0/03/7 UNIVESITY OF OSTOCK FACULTY OF COPUTE SCIENCE AND ELECTICAL ENGINEEING

More information

Information Theory for Wireless Communications, Part II:

Information Theory for Wireless Communications, Part II: Information Theory for Wireless Communications, Part II: Lecture 5: Multiuser Gaussian MIMO Multiple-Access Channel Instructor: Dr Saif K Mohammed Scribe: Johannes Lindblom In this lecture, we give the

More information

Channel Coding I. Exercises SS 2017

Channel Coding I. Exercises SS 2017 Channel Coding I Exercises SS 2017 Lecturer: Dirk Wübben Tutor: Shayan Hassanpour NW1, Room N 2420, Tel.: 0421/218-62387 E-mail: {wuebben, hassanpour}@ant.uni-bremen.de Universität Bremen, FB1 Institut

More information

Revision of Lecture 5

Revision of Lecture 5 Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information

More information

LECTURE 13. Last time: Lecture outline

LECTURE 13. Last time: Lecture outline LECTURE 13 Last time: Strong coding theorem Revisiting channel and codes Bound on probability of error Error exponent Lecture outline Fano s Lemma revisited Fano s inequality for codewords Converse to

More information

Lecture 2. Capacity of the Gaussian channel

Lecture 2. Capacity of the Gaussian channel Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN

More information

Turbo Codes. Manjunatha. P. Professor Dept. of ECE. June 29, J.N.N. College of Engineering, Shimoga.

Turbo Codes. Manjunatha. P. Professor Dept. of ECE. June 29, J.N.N. College of Engineering, Shimoga. Turbo Codes Manjunatha. P manjup.jnnce@gmail.com Professor Dept. of ECE J.N.N. College of Engineering, Shimoga June 29, 2013 [1, 2, 3, 4, 5, 6] Note: Slides are prepared to use in class room purpose, may

More information

EE 4TM4: Digital Communications II. Channel Capacity

EE 4TM4: Digital Communications II. Channel Capacity EE 4TM4: Digital Communications II 1 Channel Capacity I. CHANNEL CODING THEOREM Definition 1: A rater is said to be achievable if there exists a sequence of(2 nr,n) codes such thatlim n P (n) e (C) = 0.

More information

Cyclic Redundancy Check Codes

Cyclic Redundancy Check Codes Cyclic Redundancy Check Codes Lectures No. 17 and 18 Dr. Aoife Moloney School of Electronics and Communications Dublin Institute of Technology Overview These lectures will look at the following: Cyclic

More information

Modern Digital Communication Techniques Prof. Suvra Sekhar Das G. S. Sanyal School of Telecommunication Indian Institute of Technology, Kharagpur

Modern Digital Communication Techniques Prof. Suvra Sekhar Das G. S. Sanyal School of Telecommunication Indian Institute of Technology, Kharagpur Modern Digital Communication Techniques Prof. Suvra Sekhar Das G. S. Sanyal School of Telecommunication Indian Institute of Technology, Kharagpur Lecture - 15 Analog to Digital Conversion Welcome to the

More information

Image and Multidimensional Signal Processing

Image and Multidimensional Signal Processing Image and Multidimensional Signal Processing Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ Image Compression 2 Image Compression Goal: Reduce amount

More information

Polar Coding. Part 1 - Background. Erdal Arıkan. Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey

Polar Coding. Part 1 - Background. Erdal Arıkan. Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey Polar Coding Part 1 - Background Erdal Arıkan Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey Algorithmic Coding Theory Workshop June 13-17, 2016 ICERM, Providence, RI

More information

Source Coding. Scalar Quantization

Source Coding. Scalar Quantization Source Coding Source Coding The Communications Toolbox includes some basic functions for source coding. Source coding, also known as quantization or signal formatting, includes the concepts of analog-to-digital

More information

2018/5/3. YU Xiangyu

2018/5/3. YU Xiangyu 2018/5/3 YU Xiangyu yuxy@scut.edu.cn Entropy Huffman Code Entropy of Discrete Source Definition of entropy: If an information source X can generate n different messages x 1, x 2,, x i,, x n, then the

More information

Constellation Shaping for Communication Channels with Quantized Outputs

Constellation Shaping for Communication Channels with Quantized Outputs Constellation Shaping for Communication Channels with Quantized Outputs, Dr. Matthew C. Valenti and Xingyu Xiang Lane Department of Computer Science and Electrical Engineering West Virginia University

More information