ECE 564/645 - Digital Communications, Spring 2018 Midterm Exam #1 March 22nd, 7:00-9:00pm Marston 220
|
|
- Brooke Harris
- 5 years ago
- Views:
Transcription
1 ECE 564/645 - Digital Communications, Spring 08 Midterm Exam # March nd, 7:00-9:00pm Marston 0 Overview The exam consists of four problems for 0 points (ECE 564) or 5 points (ECE 645). The points for each part of each problem are given in brackets - you should spend your two hours accordingly. The exam is closed book, but you are allowed one page-side of notes. Calculators are not allowed. I will provide all necessary blank paper. Testmanship Full credit will be given only to fully justified answers. Giving the steps along the way to the answer will not only earn full credit but also maximize the partial credit should you stumble or get stuck. If you get stuck, attempt to neatly define your approach to the problem and why you are stuck. If part of a problem depends on a previous part that you are unable to solve, explain the method for doing the current part, and, if possible, give the answer in terms of the quantities of the previous part that you are unable to obtain. Start each problem on a new page. Not only will this facilitate grading but also make it easier for you to jump back and forth between problems. If you get to the end of the problem and realize that your answer must be wrong, be sure to write this must be wrong because... so that I will know you recognized such a fact. Academic dishonesty will be dealt with harshly - the minimum penalty will be an F for the course.
2 . Huffman Coding: [5] (a) Let a source have alphabet X = {A, B, C, D, E, F }, with respective probabilities of the letters 0.4, 0.35, 0., 0.05, 0.05, and Find a Huffman code for this source for N = letters/block. Find the rate of your Huffman code. What can you say about the entropy H(X) of the source? [0] (b) I have a source with X = {A, B, C, D, E, F, G}, but I do not know the source probabilities. A friend comes to me and tells me that he/she has a Huffman code for this source with rate 3. bits/symbol. Is this possible?. Consider the -dimensional vector channel r = s + n where r is the received vector, s = s i if message m i is to be sent on (0, T s ), and n = (n, n ) T. Let n and n be independent Gaussian random variables with mean 0 and variance N 0 (i.e. this is an AWGN vector channel). Suppose there are eight possible equally likely messages (M = 8) and s = (, 0) T s = (, 0) T s 3 = (, 0) T s 4 = (, 0) T s 5 = (0, ) T s 6 = (0, ) T s 7 = (0, ) T s 8 = (0, ) T [5] (a) Draw the signal space and decision regions for the MAP receiver. [0] (b) In terms of the units of the vector space (you do not have to convert to E s in this part!) and N 0 : Find the Union Bound to the conditional symbol error probability P (E s ). Remove as many summands as possible from the Union Bound for P (E s ), while still guaranteeing an upper bound is preserved. [5] (c) Convert your answers from the second half of (b) to be in terms of E s and N 0, where E s is the average energy per signal. [5] (d) We want to compare to QPSK: Convert your answer from part (c) to be in terms of E b and N 0. How much better (or worse) it this system s performance at high signal-to-noise ratio (SNR) than QPSK? (I have been intentionally vague about the word performance here; I want you to tell me how you would describe the relative performance of the two systems at high SNR.)
3 [5] (e) We desire to use our 8-ary signal set to transmit 3 bits per symbol (of course). The receiver is still the MAP receiver from (a). Assign the sets of three bits (i.e. 000, 00, 00, 0, 00, 0, 0, ) to the signals so as to minimize the bit error probability at large signal-to-noise ratios. Only minimal justification is required. [5] (f) Without changing the energy of any of the signals, change the signal set to one whose MAP receiver has a smaller symbol error probability P (E) at large Es N 0. An approximate sketch is sufficient, but be sure to justify your answer. 3. Consider the waveform channel: r(t) where is additive white Gaussian noise with power spectral density N 0, r(t) is the received waveform, and = s i (t) when message m i is to be sent during time t (0, T s ). Your boss wants to employ quadrature phase-shift keying (QPSK) over the interval (0, T s ), which would have signal set: s (t) = cos(πf c t) + sin(πf c t) s (t) = cos(πf c t) sin(πf c t), s 3 (t) = cos(πf c t) + sin(πf c t), s 4 (t) = cos(πf c t) sin(πf c t), where f c T s is the carrier frequency. Unfortunately, the in-phase and quadrature branches are not balanced; hence, for some constant a > 0, your actual signal set (over (0, T s )) is: s (t) = cos(πf c t) + a sin(πf c t) s (t) = cos(πf c t) a sin(πf c t), s 3 (t) = cos(πf c t) + a sin(πf c t), s 4 (t) = cos(πf c t) a sin(πf c t), This problem deals with a system employing s (t), s (t), s 3 (t), and s 4 (t). You can assume the four signals are equally likely (i.e. P (s i (t) sent) = /4, for all i) throughout this problem. [0] (a) Find an orthonormal basis φ (t), φ (t),..., φ N (t) for s (t), s (t), s 3 (t), and s 4 (t) with minimum N. (You need not use Gram-Schmidt if you see an easier solution), and (do not forget this part!), give the vector representation of the signals in that basis. [5] (b) Give an optimal receiver that takes as input r(t) and decides the signal most likely to be sent given r(t). You can assume that the receiver knows the value of a. [0] (c) Suppose a = /. Find the exact probability of symbol error (P E ) for your receiver
4 in (b) in terms of E b, the energy per bit, and N 0. [5] (d) At high signal-to-noise ratios (SNRs), how much better or worse (in db) is this system (with a = /) than QPSK? 4. [564 only] In this problem, we consider what happens when the scaling of the correlators at the receiver for an AWGN channel is not as given in class. Consider a binary signal set {s 0 (t), s (t)}, where the information bit selects which of these two signals will be sent over the waveform channel: r(t) where is additive white Gaussian noise with power spectral density N 0, and r(t) is the received waveform. Suppose the orthonormal basis for {s 0 (t), s (t)} requires two basis functions φ (t) and φ (t), and the front-end of the receiver calculates (different than in class because of the out front!): r = r = 0 0 r(t)φ (t)dt r(t)φ (t)dt yielding the vector r = (r, r ) T for us to use to decide which signal was sent. Define the vectors x 0 and x, as follows: the j th component of x 0 and x is given by, respectively: x 0,j = 0 s 0 (t)φ j (t)dt x,j = 0 s (t)φ j (t)dt [5] (a) Suppose x 0 = (, 0) T and x = (0, ) T. Draw the optimal decision regions in r-space. [0] (b) Now, suppose I give you some x 0 and x (not necessarily those in part (a)). Find the probability of error of an optimal receiver (that has as input r, as given above) as a function of x 0 x and N 0. Be sure to justify your answer fully. [5] (c) As in class, let s i be the representation of s i (t) on the orthonormal basis. For any s 0 and s, what is the probability of error of the optimal receiver as a function of s 0 s and N 0?
5 5. [645 only] In this problem, we consider how to deal with a (very) poor front-end to our receiver. Consider the waveform channel: r(t) where is additive white Gaussian noise with power spectral density N 0, r(t) is the received waveform, and = s i (t) when message m i is to be sent during time t (0, T s ). Suppose there are M = 4 possible equally likely messages and the corresponding signals are: s (t) s (t) t t s 3 (t) s 4 (t) t t You, of course, recognize immediately that this is a two-dimensional signal space, and thus you should build a correlation receiver with two basis functions for the front-end of the optimal receiver. But, alas, the person designing the front-end of your receiver does not, and simply multiplies by the function: { t, 0 t x(t) = 0, else and integrates, providing you with the single number r = 0 r(t)x(t)dt. [0] (a) Find the optimal processing of r to decide which signal was sent. (Note: The fractions get a little messy in this part and the next - sorry!) [5] (b) Find the probability of symbol error (choosing the wrong symbol) as a function of the average signal energy E s and N 0. (Remember that the signal energy of s i (t) is 0 s i (t)dt.). [0] (c) For high SNRs, roughly how much performance (in db) is lost by using this (very) poor front-end versus an optimal front-end? (You can use quick d min arguments here.)
ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)
ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling
More informationA Simple Example Binary Hypothesis Testing Optimal Receiver Frontend M-ary Signal Sets Message Sequences. possible signals has been transmitted.
Introduction I We have focused on the problem of deciding which of two possible signals has been transmitted. I Binary Signal Sets I We will generalize the design of optimum (MPE) receivers to signal sets
More informationECE 564/645 - Digital Communication Systems (Spring 2014) Final Exam Friday, May 2nd, 8:00-10:00am, Marston 220
ECE 564/645 - Digital Commuicatio Systems (Sprig 014) Fial Exam Friday, May d, 8:00-10:00am, Marsto 0 Overview The exam cosists of four (or five) problems for 100 (or 10) poits. The poits for each part
More informationA First Course in Digital Communications
A First Course in Digital Communications Ha H. Nguyen and E. Shwedyk February 9 A First Course in Digital Communications 1/46 Introduction There are benefits to be gained when M-ary (M = 4 signaling methods
More informationECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)
ECE 74 - Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE) 1. A Huffman code finds the optimal codeword to assign to a given block of source symbols. (a) Show that cannot be a Huffman
More informationDigital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10
Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,
More informationUTA EE5362 PhD Diagnosis Exam (Spring 2011)
EE5362 Spring 2 PhD Diagnosis Exam ID: UTA EE5362 PhD Diagnosis Exam (Spring 2) Instructions: Verify that your exam contains pages (including the cover shee. Some space is provided for you to show your
More informationSummary: SER formulation. Binary antipodal constellation. Generic binary constellation. Constellation gain. 2D constellations
TUTORIAL ON DIGITAL MODULATIONS Part 8a: Error probability A [2011-01-07] 07] Roberto Garello, Politecnico di Torino Free download (for personal use only) at: www.tlc.polito.it/garello 1 Part 8a: Error
More informationModulation & Coding for the Gaussian Channel
Modulation & Coding for the Gaussian Channel Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology
More informationa) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics.
Digital Modulation and Coding Tutorial-1 1. Consider the signal set shown below in Fig.1 a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics. b) What is the minimum Euclidean
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationUniversity of Illinois ECE 313: Final Exam Fall 2014
University of Illinois ECE 313: Final Exam Fall 2014 Monday, December 15, 2014, 7:00 p.m. 10:00 p.m. Sect. B, names A-O, 1013 ECE, names P-Z, 1015 ECE; Section C, names A-L, 1015 ECE; all others 112 Gregory
More informationDigital Transmission Methods S
Digital ransmission ethods S-7.5 Second Exercise Session Hypothesis esting Decision aking Gram-Schmidt method Detection.K.K. Communication Laboratory 5//6 Konstantinos.koufos@tkk.fi Exercise We assume
More informationEE 121: Introduction to Digital Communication Systems. 1. Consider the following discrete-time communication system. There are two equallly likely
EE 11: Introduction to Digital Communication Systems Midterm Solutions 1. Consider the following discrete-time communication system. There are two equallly likely messages to be transmitted, and they are
More informationChapter 4: Continuous channel and its capacity
meghdadi@ensil.unilim.fr Reference : Elements of Information Theory by Cover and Thomas Continuous random variable Gaussian multivariate random variable AWGN Band limited channel Parallel channels Flat
More informationEE4304 C-term 2007: Lecture 17 Supplemental Slides
EE434 C-term 27: Lecture 17 Supplemental Slides D. Richard Brown III Worcester Polytechnic Institute, Department of Electrical and Computer Engineering February 5, 27 Geometric Representation: Optimal
More informationCOS 341: Discrete Mathematics
COS 341: Discrete Mathematics Midterm Exam Fall 2006 Print your name General directions: This exam is due on Monday, November 13 at 4:30pm. Late exams will not be accepted. Exams must be submitted in hard
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationEE456 Digital Communications
EE456 Digital Communications Professor Ha Nguyen September 5 EE456 Digital Communications Block Diagram of Binary Communication Systems m ( t { b k } b k = s( t b = s ( t k m ˆ ( t { bˆ } k r( t Bits in
More informationChapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code
Chapter 2 Date Compression: Source Coding 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code 2.1 An Introduction to Source Coding Source coding can be seen as an efficient way
More informationEC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY
EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase
More informationELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS
EC 32 (CR) Total No. of Questions :09] [Total No. of Pages : 02 III/IV B.Tech. DEGREE EXAMINATIONS, APRIL/MAY- 207 Second Semester ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS Time: Three Hours
More informationMachine Learning, Fall 2009: Midterm
10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all
More informationLecture 12. Block Diagram
Lecture 12 Goals Be able to encode using a linear block code Be able to decode a linear block code received over a binary symmetric channel or an additive white Gaussian channel XII-1 Block Diagram Data
More informationCommunication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi
Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi Lecture - 41 Pulse Code Modulation (PCM) So, if you remember we have been talking
More informationMITOCW ocw-6-451_ feb k_512kb-mp4
MITOCW ocw-6-451_4-261-09feb2005-220k_512kb-mp4 And what we saw was the performance was quite different in the two regimes. In power-limited regime, typically our SNR is much smaller than one, whereas
More informationApplication of Matched Filter
Application of Matched Filter Lecture No. 12 Dr. Aoife Moloney School of Electronics and Communications Dublin Institute of Technology Overview This lecture will look at the following: Matched filter vs
More informationECE 534 Information Theory - Midterm 2
ECE 534 Information Theory - Midterm Nov.4, 009. 3:30-4:45 in LH03. You will be given the full class time: 75 minutes. Use it wisely! Many of the roblems have short answers; try to find shortcuts. You
More informationCoding theory: Applications
INF 244 a) Textbook: Lin and Costello b) Lectures (Tu+Th 12.15-14) covering roughly Chapters 1,9-12, and 14-18 c) Weekly exercises: For your convenience d) Mandatory problem: Programming project (counts
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationCommunications and Signal Processing Spring 2017 MSE Exam
Communications and Signal Processing Spring 2017 MSE Exam Please obtain your Test ID from the following table. You must write your Test ID and name on each of the pages of this exam. A page with missing
More information3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions
Engineering Tripos Part IIA THIRD YEAR 3F: Signals and Systems INFORMATION THEORY Examples Paper Solutions. Let the joint probability mass function of two binary random variables X and Y be given in the
More informationADVERSARY ANALYSIS OF COCKROACH NETWORK UNDER RAYLEIGH FADING CHANNEL: PROBABILITY OF ERROR AND ADVERSARY DETECTION. A Dissertation by.
ADVERSARY ANALYSIS OF COCKROACH NETWORK UNDER RAYLEIGH FADING CHANNEL: PROBABILITY OF ERROR AND ADVERSARY DETECTION A Dissertation by Tze Chien Wong Master of Science, Wichita State University, 2008 Bachelor
More informationExercise 1. = P(y a 1)P(a 1 )
Chapter 7 Channel Capacity Exercise 1 A source produces independent, equally probable symbols from an alphabet {a 1, a 2 } at a rate of one symbol every 3 seconds. These symbols are transmitted over a
More informationThe Cooper Union Department of Electrical Engineering ECE111 Signal Processing & Systems Analysis Final May 4, 2012
The Cooper Union Department of Electrical Engineering ECE111 Signal Processing & Systems Analysis Final May 4, 2012 Time: 3 hours. Close book, closed notes. No calculators. Part I: ANSWER ALL PARTS. WRITE
More informationLecture 15: Thu Feb 28, 2019
Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:
More informationSection 1.1: Patterns in Division
Section 1.1: Patterns in Division Dividing by 2 All even numbers are divisible by 2. E.g., all numbers ending in 0,2,4,6 or 8. Dividing by 4 1. Are the last two digits in your number divisible by 4? 2.
More informationNAME... Soc. Sec. #... Remote Location... (if on campus write campus) FINAL EXAM EE568 KUMAR. Sp ' 00
NAME... Soc. Sec. #... Remote Location... (if on campus write campus) FINAL EXAM EE568 KUMAR Sp ' 00 May 3 OPEN BOOK exam (students are permitted to bring in textbooks, handwritten notes, lecture notes
More information17.1 Binary Codes Normal numbers we use are in base 10, which are called decimal numbers. Each digit can be 10 possible numbers: 0, 1, 2, 9.
( c ) E p s t e i n, C a r t e r, B o l l i n g e r, A u r i s p a C h a p t e r 17: I n f o r m a t i o n S c i e n c e P a g e 1 CHAPTER 17: Information Science 17.1 Binary Codes Normal numbers we use
More informationANALYSIS OF A PARTIAL DECORRELATOR IN A MULTI-CELL DS/CDMA SYSTEM
ANAYSIS OF A PARTIA DECORREATOR IN A MUTI-CE DS/CDMA SYSTEM Mohammad Saquib ECE Department, SU Baton Rouge, A 70803-590 e-mail: saquib@winlab.rutgers.edu Roy Yates WINAB, Rutgers University Piscataway
More informationChapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code
Chapter 3 Source Coding 3. An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code 3. An Introduction to Source Coding Entropy (in bits per symbol) implies in average
More informationDCSP-3: Minimal Length Coding. Jianfeng Feng
DCSP-3: Minimal Length Coding Jianfeng Feng Department of Computer Science Warwick Univ., UK Jianfeng.feng@warwick.ac.uk http://www.dcs.warwick.ac.uk/~feng/dcsp.html Automatic Image Caption (better than
More informationDigital Modulation 1
Digital Modulation 1 Lecture Notes Ingmar Land and Bernard H. Fleury Navigation and Communications () Department of Electronic Systems Aalborg University, DK Version: February 5, 27 i Contents I Basic
More informationComputing Probability of Symbol Error
Computing Probability of Symbol Error I When decision boundaries intersect at right angles, then it is possible to compute the error probability exactly in closed form. I The result will be in terms of
More informationThis examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of
More informationGEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems
GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems Aids Allowed: 2 8 1/2 X11 crib sheets, calculator DATE: Tuesday
More informationCoding for Digital Communication and Beyond Fall 2013 Anant Sahai MT 1
EECS 121 Coding for Digital Communication and Beyond Fall 2013 Anant Sahai MT 1 PRINT your student ID: PRINT AND SIGN your name:, (last) (first) (signature) PRINT your Unix account login: ee121- Prob.
More informationAdvanced 3 G and 4 G Wireless Communication Prof. Aditya K Jagannathan Department of Electrical Engineering Indian Institute of Technology, Kanpur
Advanced 3 G and 4 G Wireless Communication Prof. Aditya K Jagannathan Department of Electrical Engineering Indian Institute of Technology, Kanpur Lecture - 19 Multi-User CDMA Uplink and Asynchronous CDMA
More informationMidterm II. Introduction to Artificial Intelligence. CS 188 Spring ˆ You have approximately 1 hour and 50 minutes.
CS 188 Spring 2013 Introduction to Artificial Intelligence Midterm II ˆ You have approximately 1 hour and 50 minutes. ˆ The exam is closed book, closed notes except a one-page crib sheet. ˆ Please use
More informationEntropy as a measure of surprise
Entropy as a measure of surprise Lecture 5: Sam Roweis September 26, 25 What does information do? It removes uncertainty. Information Conveyed = Uncertainty Removed = Surprise Yielded. How should we quantify
More informationAP PHYSICS B 2008 SCORING GUIDELINES
AP PHYSICS B 2008 SCORING GUIDELINES General Notes About 2008 AP Physics Scoring Guidelines 1. The solutions contain the most common method of solving the free-response questions and the allocation of
More informationDesigning Information Devices and Systems II Spring 2016 Anant Sahai and Michel Maharbiz Midterm 1. Exam location: 1 Pimentel (Odd SID + 61C)
EECS 16B Designing Information Devices and Systems II Spring 16 Anant Sahai and Michel Maharbiz Midterm 1 Exam location: 1 Pimentel (Odd SID + 61C) PRINT your student ID: PRINT AND SIGN your name:, (last)
More informationAn introduction to basic information theory. Hampus Wessman
An introduction to basic information theory Hampus Wessman Abstract We give a short and simple introduction to basic information theory, by stripping away all the non-essentials. Theoretical bounds on
More information392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions
392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6 Problem Set 2 Solutions Problem 2.1 (Cartesian-product constellations) (a) Show that if A is a K-fold Cartesian
More informationSOLUTIONS TO ECE 6603 ASSIGNMENT NO. 6
SOLUTIONS TO ECE 6603 ASSIGNMENT NO. 6 PROBLEM 6.. Consider a real-valued channel of the form r = a h + a h + n, which is a special case of the MIMO channel considered in class but with only two inputs.
More informationC if U can. Algebra. Name
C if U can Algebra Name.. How will this booklet help you to move from a D to a C grade? The topic of algebra is split into six units substitution, expressions, factorising, equations, trial and improvement
More informationRevision of Lecture 4
Revision of Lecture 4 We have completed studying digital sources from information theory viewpoint We have learnt all fundamental principles for source coding, provided by information theory Practical
More informationMidterm 2. Your Exam Room: Name of Person Sitting on Your Left: Name of Person Sitting on Your Right: Name of Person Sitting in Front of You:
CS70 Discrete Mathematics and Probability Theory, Fall 2018 Midterm 2 8:00-10:00pm, 31 October Your First Name: SIGN Your Name: Your Last Name: Your SID Number: Your Exam Room: Name of Person Sitting on
More informationBASICS OF DETECTION AND ESTIMATION THEORY
BASICS OF DETECTION AND ESTIMATION THEORY 83050E/158 In this chapter we discuss how the transmitted symbols are detected optimally from a noisy received signal (observation). Based on these results, optimal
More information2018/5/3. YU Xiangyu
2018/5/3 YU Xiangyu yuxy@scut.edu.cn Entropy Huffman Code Entropy of Discrete Source Definition of entropy: If an information source X can generate n different messages x 1, x 2,, x i,, x n, then the
More informationChapter 2: Source coding
Chapter 2: meghdadi@ensil.unilim.fr University of Limoges Chapter 2: Entropy of Markov Source Chapter 2: Entropy of Markov Source Markov model for information sources Given the present, the future is independent
More informationThis examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 6 December 2006 This examination consists of
More informationIE 230 Seat # (1 point) Name (clearly) < KEY > Closed book and notes. No calculators. Designed for 60 minutes, but time is essentially unlimited.
Closed book and notes. No calculators. Designed for 60 minutes, but time is essentially unlimited. Cover page, four pages of exam. This test covers through Section 2.7 of Montgomery and Runger, fourth
More information18.310A Final exam practice questions
18.310A Final exam practice questions This is a collection of practice questions, gathered randomly from previous exams and quizzes. They may not be representative of what will be on the final. In particular,
More informationAlgebra Year 10. Language
Algebra Year 10 Introduction In Algebra we do Maths with numbers, but some of those numbers are not known. They are represented with letters, and called unknowns, variables or, most formally, literals.
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science ALGORITHMS FOR INFERENCE Fall 2014
MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.438 ALGORITHMS FOR INFERENCE Fall 2014 Quiz 2 Wednesday, December 10, 2014 7:00pm 10:00pm This is a closed
More informationA Hilbert Space for Random Processes
Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts A Hilbert Space for Random Processes I A vector space for random processes X t that is analogous to L 2 (a, b) is of
More informationECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters
ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009
More informationChapter 2 Signal Processing at Receivers: Detection Theory
Chapter Signal Processing at Receivers: Detection Theory As an application of the statistical hypothesis testing, signal detection plays a key role in signal processing at receivers of wireless communication
More informationECS 332: Principles of Communications 2012/1. HW 4 Due: Sep 7
ECS 332: Principles of Communications 2012/1 HW 4 Due: Sep 7 Lecturer: Prapun Suksompong, Ph.D. Instructions (a) ONE part of a question will be graded (5 pt). Of course, you do not know which part will
More informationTheory of Computation
Theory of Computation Dr. Sarmad Abbasi Dr. Sarmad Abbasi () Theory of Computation / Lecture 3: Overview Decidability of Logical Theories Presburger arithmetic Decidability of Presburger Arithmetic Dr.
More informationCHANNEL CAPACITY CALCULATIONS FOR M ARY N DIMENSIONAL SIGNAL SETS
THE UNIVERSITY OF SOUTH AUSTRALIA SCHOOL OF ELECTRONIC ENGINEERING CHANNEL CAPACITY CALCULATIONS FOR M ARY N DIMENSIONAL SIGNAL SETS Philip Edward McIllree, B.Eng. A thesis submitted in fulfilment of the
More informationHW Solution 3 Due: July 15
ECS 315: Probability and Random Processes 2010/1 HW Solution 3 Due: July 15 Lecturer: Prapun Suksompong, Ph.D. Instructions (a) A part of ONE question will be graded. Of course, you do not know which problem
More informationElectrical Engineering Written PhD Qualifier Exam Spring 2014
Electrical Engineering Written PhD Qualifier Exam Spring 2014 Friday, February 7 th 2014 Please do not write your name on this page or any other page you submit with your work. Instead use the student
More informationLecture 11: Continuous-valued signals and differential entropy
Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components
More informationCommunication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi
Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi Lecture - 3 Brief Review of Signals and Systems My subject for today s discussion
More informationPower Spectral Density of Digital Modulation Schemes
Digital Communication, Continuation Course Power Spectral Density of Digital Modulation Schemes Mikael Olofsson Emil Björnson Department of Electrical Engineering ISY) Linköping University, SE-581 83 Linköping,
More informationECE538 Final Exam Fall 2017 Digital Signal Processing I 14 December Cover Sheet
ECE58 Final Exam Fall 7 Digital Signal Processing I December 7 Cover Sheet Test Duration: hours. Open Book but Closed Notes. Three double-sided 8.5 x crib sheets allowed This test contains five problems.
More informationInformation Sources. Professor A. Manikas. Imperial College London. EE303 - Communication Systems An Overview of Fundamentals
Information Sources Professor A. Manikas Imperial College London EE303 - Communication Systems An Overview of Fundamentals Prof. A. Manikas (Imperial College) EE303: Information Sources 24 Oct. 2011 1
More informationSpring /11/2009
MA 123 Elementary Calculus SECOND MIDTERM Spring 2009 03/11/2009 Name: Sec.: Do not remove this answer page you will return the whole exam. You will be allowed two hours to complete this test. No books
More informationCommunication Theory II
Communication Theory II Lecture 15: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 29 th, 2015 1 Example: Channel Capacity of BSC o Let then: o For
More informationMLC Fall 2014 Written Answer Questions. Model Solutions
MLC Fall 2014 Written Answer Questions Model Solutions 1 MLC Fall 2014 Question 1 Model Solution Learning Objectives: 2(a), (a), 4(f) Textbook References: 5.4, 6.6, 1.5 (a) The EPV of the commission payments
More informationChannel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.
Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationEE 574 Detection and Estimation Theory Lecture Presentation 8
Lecture Presentation 8 Aykut HOCANIN Dept. of Electrical and Electronic Engineering 1/14 Chapter 3: Representation of Random Processes 3.2 Deterministic Functions:Orthogonal Representations For a finite-energy
More informationEfficient Algorithms and Intractable Problems Spring 2016 Alessandro Chiesa and Umesh Vazirani Midterm 2
CS 170 Efficient Algorithms and Intractable Problems Spring 2016 Alessandro Chiesa and Umesh Vazirani Midterm 2 Name: SID: GSI and section time: Write down the names of the students around you as they
More informationHandout 12: Error Probability Analysis of Binary Detection
ENGG 230-B: Principles of Communication Systems 207 8 First Term Handout 2: Error Probability Analysis of Binary Detection Instructor: Wing-Kin Ma November 4, 207 Suggested Reading: Chapter 8 of Simon
More informationCS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability
More informationAnalog Electronics 2 ICS905
Analog Electronics 2 ICS905 G. Rodriguez-Guisantes Dépt. COMELEC http://perso.telecom-paristech.fr/ rodrigez/ens/cycle_master/ November 2016 2/ 67 Schedule Radio channel characteristics ; Analysis and
More informationMAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)
MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)
More informationMath Introduction to Logic Final Exam
Math 2283 - Introduction to Logic Final Exam Assigned: 2018.11.26 Due: 2018.12.10 at 08:00 Instructions: Work on this by yourself, the only person you may contact in any way to discuss or ask questions
More informationDEPARTMENT OF EECS MASSACHUSETTS INSTITUTE OF TECHNOLOGY. 6.02: Digital Communication Systems, Fall Quiz I. October 11, 2012
6.02 Fall 2012, Quiz 2 Page 1 of 12 Name: DEPARTMENT OF EECS MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.02: Digital Communication Systems, Fall 2012 Quiz I October 11, 2012 your section Section Time Recitation
More informationPrinciples of Communications
Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @
More information5 Analog carrier modulation with noise
5 Analog carrier modulation with noise 5. Noisy receiver model Assume that the modulated signal x(t) is passed through an additive White Gaussian noise channel. A noisy receiver model is illustrated in
More informationInformation Theory. Coding and Information Theory. Information Theory Textbooks. Entropy
Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is
More informationTrue or False. Circle T if the statement is always true; otherwise circle F. for all angles θ. T F. 1 sin θ
Math 90 Practice Midterm III Solutions Ch. 8-0 (Ebersole), 3.3-3.8 (Stewart) DISCLAIMER. This collection of practice problems is not guaranteed to be identical, in length or content, to the actual exam.
More informationIntroduction to Algorithms March 10, 2010 Massachusetts Institute of Technology Spring 2010 Professors Piotr Indyk and David Karger Quiz 1
Introduction to Algorithms March 10, 2010 Massachusetts Institute of Technology 6.006 Spring 2010 Professors Piotr Indyk and David Karger Quiz 1 Quiz 1 Do not open this quiz booklet until directed to do
More informationGeneral Notes About 2007 AP Physics Scoring Guidelines
AP PHYSICS C: ELECTRICITY AND MAGNETISM 2007 SCORING GUIDELINES General Notes About 2007 AP Physics Scoring Guidelines 1. The solutions contain the most common method of solving the free-response questions
More informationLecture 7. Union bound for reducing M-ary to binary hypothesis testing
Lecture 7 Agenda for the lecture M-ary hypothesis testing and the MAP rule Union bound for reducing M-ary to binary hypothesis testing Introduction of the channel coding problem 7.1 M-ary hypothesis testing
More informationAnalysis of methods for speech signals quantization
INFOTEH-JAHORINA Vol. 14, March 2015. Analysis of methods for speech signals quantization Stefan Stojkov Mihajlo Pupin Institute, University of Belgrade Belgrade, Serbia e-mail: stefan.stojkov@pupin.rs
More information