Asymptotic Capacity Bounds for Magnetic Recording. Raman Venkataramani Seagate Technology (Joint work with Dieter Arnold)
|
|
- Augustine Pope
- 5 years ago
- Views:
Transcription
1 Asymptotic Capacity Bounds for Magnetic Recording Raman Venkataramani Seagate Technology (Joint work with Dieter Arnold)
2 Outline Problem Statement Signal and Noise Models for Magnetic Recording Capacity Bounds Additive noise dominated channels Jitter noise dominated channels
3 Problem Statement How does the capacity of the system scale with the bit-width? T What is the optimum bit width from a purely information theoretic viewpoint?
4 Signal and Noise Models Signal: the magnetization in the medium is either ±1 Noise: electronic noise and transition noise.
5 Signal and Noise Models x(t) = k a k p(t kt ), p(t) = u(t) u(t T ) r(t) = k a k a k 1 2 h T (t kt t k ) + w(t) a k {+1, 1} t k : transition noise w(t) : electronic noise AWGN with PSD N 0 /2 h T (t) : transition ( 1 +1) response
6 Recording Modes Magnetization profile for longitudinal recording Magnetization profile for perpendicular recording
7 Transition Response Longitudinal recording h T (t) = Perpendicular recording A 1 + (2t/w) 2 h T (t) = A erf ( 2 log 2t ) w The pulse width is sometimes also written as PW50.
8 Capacity The fundamental quantity of interest is the areal capacity of the magnetic medium. In the Shannon sense, this is the number of bits per unit area that can be stored and recovered reliably. Assuming a fixed track width, we can focus on the linear density of bits or linear capacity: C = lim τ 1 τ max({x(t), t [0, τ]}; {r(t), t [0, τ]})
9 Problem Statement How is the optimal bit width capacity per unit length? T for maximum Misconceptions: Very small bit-widths are bad because... There is higher transition noise. There is longer ISI and lower SNR per bit. Truth: We can overcome all these problems with clever coding.
10 Thought Experiment No code T = 1 R = 1 Repetition code T = 0.5 R = 0.5 If we use better code, case 2 will outperform case 1.
11 1 Binary Entropy Function Repetition Code Code Rate Transition Probability Achievable region of code rate and transition probability.
12 A Note about the SNR We define: SNR = E iw 2 where E i = N 0 R Reasons for this definition: dh T (t) 2 dt dt It is finite for both perpendicular and longitudinal modes. It is a function of the head and medium alone. It is independent of or the code we use. T
13 Electronic Noise Dominated Channel x(t) = k a k p(t kt ), p(t) = u(t) u(t T ) r(t) = k (a k a k 1 )h T (t kt ) + w(t) = k a k h D (t kt ) + w(t) h D (t) = h T (t) h T (t T ) : dibit response w(t) : electronic noise AWGN with PSD N 0 /2
14 Sufficient statistics To be exact, we need to pass through a matched filter and sample at baud rate. T w r(t) If, it is almost optimal to use a simple low-pass filter and sample a baud rate:
15 Discrete-time Model We obtain an ISI channel with binary inputs: r n = k a k hn k + w n w n N(0, N 0 /2T ), i.i.d. where, hn h D (nt ) Let the above ISI channel have capacity. Then, the linear capacity is K B (T ) = 1 T C B(T ) C B (T )
16 Simple Upper Bound Suppose we relax the input constraint: {a k = ±1} {a k R, E a 2 k 1} We obtain a Gaussian channel, whose capacity is computed using the water-filling method. C B (T ) C G (T ) C G (T ) = max S x [ ] 1 2 such that 0.5 log (1 + S x[ν] H[ν] 2 S x [ν]dν = 1 N 0 /2T )
17 Lower Bounds Several researchers have computed information rates for binary ISI channels: Monte-Carlo method: compute the quantity I(x N ; r N )/N for a very long sequence of length N with the input generated using an appropriate Markov model. This method becomes unreliable or computationally intense for high recording densities.
18 Shamai-Laroia Conjecture Along any horizontal line, the SNR loss due to ISI is the same for both Gaussian and binary signalling.
19 Shamai-Laroia Conjecture Suppose that the input to the binary ISI channel is independent and uniformly (IUD) distributed C B (T ) C BP SK (2 I G(T ) 1) where I G (T ) is the i.i.d Gaussian information rate. Remark: This bound is also found to be surprisingly tight for non IUD Markov sources up to memory 6 and we have C B (T ) C BP SK (2 C G(T ) 1)
20 Optimal Input Distribution For Gaussian inputs, the optimal input spectrum is not flat. It becomes highly non i.i.d as T 0 For binary inputs, we could mimic the Gaussian water-filling spectrum, but it is neither optimal nor achievable in general. For the binary case, we optimize over firstorder Markov sources to get a lower bound.
21 Markov Source Model The first order model is specified by one probability parameter p T For each we pick the optimal probability to maximize the Shamai-Laroia bound. As T 0 we find that p 0
22 8 7 6 Gauss waterfilling MS1 optimized bits/use/t MS0 i.u.d T SNR 10 0 Linear capacity bounds for longitudinal recording
23 2.5 Gauss waterfilling 2 MS1 optimized bits/use/t MS0 i.u.d T SNR 10 0 Linear capacity bounds for perpendicular recording
24 Jitter Noise Dominated Channel Most general method: pulse-width modulation r(t) = n ( 1) n h T (t σ n w n ) σ n = n t k, w n N(0, σj 2 ) k=1
25 Capacity Since there is no additive noise, we can use a zero forcing equalizer to estimate ˆσ n = σ n + w n Thus, ˆt n = ˆσ n ˆσ n 1 = t n + w n w n 1 These are sufficient statistics and the channel is ISI! Linear capacity: C = max p(t N ) I( ˆT N ; T N ) E n T n
26 Lower Bound on Capacity We can show that the following is a lower bound on the linear capacity: K K 0 where K 0 is the capacity of the ISI-free channel: i.e., ˆt n = t n + w n I( K 0 = max ˆT ; T ) p(t ) E T Remark: This constant is not important. The bound behaves like. 1/σ j
27 Summary From a purely information theoretic viewpoint there are considerable gains at we higher recording densities. This ignores timing recovery and other implementation issues. In today s recording medium transition noise is a dominant noise source and σ j 0.1T We should lower potential gains. T by a factor of 10 to see
Chapter 4: Continuous channel and its capacity
meghdadi@ensil.unilim.fr Reference : Elements of Information Theory by Cover and Thomas Continuous random variable Gaussian multivariate random variable AWGN Band limited channel Parallel channels Flat
More informationComputation of Information Rates from Finite-State Source/Channel Models
Allerton 2002 Computation of Information Rates from Finite-State Source/Channel Models Dieter Arnold arnold@isi.ee.ethz.ch Hans-Andrea Loeliger loeliger@isi.ee.ethz.ch Pascal O. Vontobel vontobel@isi.ee.ethz.ch
More informationECE Information theory Final (Fall 2008)
ECE 776 - Information theory Final (Fall 2008) Q.1. (1 point) Consider the following bursty transmission scheme for a Gaussian channel with noise power N and average power constraint P (i.e., 1/n X n i=1
More informationLecture 15: Thu Feb 28, 2019
Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:
More informationOn the Shamai-Laroia Approximation for the Information Rate of the ISI Channel
On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel Yair Carmon and Shlomo Shamai (Shitz) Department of Electrical Engineering, Technion - Israel Institute of Technology 2014
More informationANALYSIS OF A PARTIAL DECORRELATOR IN A MULTI-CELL DS/CDMA SYSTEM
ANAYSIS OF A PARTIA DECORREATOR IN A MUTI-CE DS/CDMA SYSTEM Mohammad Saquib ECE Department, SU Baton Rouge, A 70803-590 e-mail: saquib@winlab.rutgers.edu Roy Yates WINAB, Rutgers University Piscataway
More informationa) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics.
Digital Modulation and Coding Tutorial-1 1. Consider the signal set shown below in Fig.1 a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics. b) What is the minimum Euclidean
More informationChapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye
Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for
More informationOne Lesson of Information Theory
Institut für One Lesson of Information Theory Prof. Dr.-Ing. Volker Kühn Institute of Communications Engineering University of Rostock, Germany Email: volker.kuehn@uni-rostock.de http://www.int.uni-rostock.de/
More informationDigital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10
Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,
More informationEstimation of the Capacity of Multipath Infrared Channels
Estimation of the Capacity of Multipath Infrared Channels Jeffrey B. Carruthers Department of Electrical and Computer Engineering Boston University jbc@bu.edu Sachin Padma Department of Electrical and
More informationThis examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of
More informationX 1 : X Table 1: Y = X X 2
ECE 534: Elements of Information Theory, Fall 200 Homework 3 Solutions (ALL DUE to Kenneth S. Palacio Baus) December, 200. Problem 5.20. Multiple access (a) Find the capacity region for the multiple-access
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationEE 5345 Biomedical Instrumentation Lecture 12: slides
EE 5345 Biomedical Instrumentation Lecture 1: slides 4-6 Carlos E. Davila, Electrical Engineering Dept. Southern Methodist University slides can be viewed at: http:// www.seas.smu.edu/~cd/ee5345.html EE
More informationLecture 7: Wireless Channels and Diversity Advanced Digital Communications (EQ2410) 1
Wireless : Wireless Advanced Digital Communications (EQ2410) 1 Thursday, Feb. 11, 2016 10:00-12:00, B24 1 Textbook: U. Madhow, Fundamentals of Digital Communications, 2008 1 / 15 Wireless Lecture 1-6 Equalization
More informationAdaptive Modulation with Smoothed Flow Utility
Adaptive Modulation with Smoothed Flow Utility Ekine Akuiyibo Stephen Boyd Electrical Engineering Department, Stanford University ITMANET 05/24/10 The problem choose transmit power(s) and flow rate(s)
More informationEE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design
Chapter 4 Receiver Design Chapter 4 Receiver Design Probability of Bit Error Pages 124-149 149 Probability of Bit Error The low pass filtered and sampled PAM signal results in an expression for the probability
More informationDirect-Sequence Spread-Spectrum
Chapter 3 Direct-Sequence Spread-Spectrum In this chapter we consider direct-sequence spread-spectrum systems. Unlike frequency-hopping, a direct-sequence signal occupies the entire bandwidth continuously.
More informationMulti User Detection I
January 12, 2005 Outline Overview Multiple Access Communication Motivation: What is MU Detection? Overview of DS/CDMA systems Concept and Codes used in CDMA CDMA Channels Models Synchronous and Asynchronous
More informationThe Poisson Channel with Side Information
The Poisson Channel with Side Information Shraga Bross School of Enginerring Bar-Ilan University, Israel brosss@macs.biu.ac.il Amos Lapidoth Ligong Wang Signal and Information Processing Laboratory ETH
More informationInformation Theoretic Imaging
Information Theoretic Imaging WU Faculty: J. A. O Sullivan WU Doctoral Student: Naveen Singla Boeing Engineer: James Meany First Year Focus: Imaging for Data Storage Image Reconstruction Data Retrieval
More informationEE Introduction to Digital Communications Homework 8 Solutions
EE 2 - Introduction to Digital Communications Homework 8 Solutions May 7, 2008. (a) he error probability is P e = Q( SNR). 0 0 0 2 0 4 0 6 P e 0 8 0 0 0 2 0 4 0 6 0 5 0 5 20 25 30 35 40 SNR (db) (b) SNR
More informationLecture 2. Capacity of the Gaussian channel
Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN
More informationConstellation Shaping for Communication Channels with Quantized Outputs
Constellation Shaping for Communication Channels with Quantized Outputs Chandana Nannapaneni, Matthew C. Valenti, and Xingyu Xiang Lane Department of Computer Science and Electrical Engineering West Virginia
More informationELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals
ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Matched Filter Generalized Matched Filter Signal
More informationInformation Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST
Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it
More informationthat efficiently utilizes the total available channel bandwidth W.
Signal Design for Band-Limited Channels Wireless Information Transmission System Lab. Institute of Communications Engineering g National Sun Yat-sen University Introduction We consider the problem of signal
More informationSignal Design for Band-Limited Channels
Wireless Information Transmission System Lab. Signal Design for Band-Limited Channels Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal
More informationLecture 5b: Line Codes
Lecture 5b: Line Codes Dr. Mohammed Hawa Electrical Engineering Department University of Jordan EE421: Communications I Digitization Sampling (discrete analog signal). Quantization (quantized discrete
More informationDoes soft-decision ECC improve endurance?
Flash Memory Summit 2015 Does soft-decision ECC improve endurance? Thomas Parnell IBM Research - Zurich Outline Motivation Why do we need more endurance? Capacity What is channel capacity? How can it be
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationDecision-Point Signal to Noise Ratio (SNR)
Decision-Point Signal to Noise Ratio (SNR) Receiver Decision ^ SNR E E e y z Matched Filter Bound error signal at input to decision device Performance upper-bound on ISI channels Achieved on memoryless
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationLECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs
LECTURE 18 Last time: White Gaussian noise Bandlimited WGN Additive White Gaussian Noise (AWGN) channel Capacity of AWGN channel Application: DS-CDMA systems Spreading Coding theorem Lecture outline Gaussian
More informationADAPTIVE EQUALIZATION AT MULTI-GHZ DATARATES
ADAPTIVE EQUALIZATION AT MULTI-GHZ DATARATES Department of Electrical Engineering Indian Institute of Technology, Madras 1st February 2007 Outline Introduction. Approaches to electronic mitigation - ADC
More information1-bit quantization and oversampling at the receiver: Sequence based communication
Landau et al. EURASIP Journal on Wireless Communications and Networking (208) 208:83 https://doi.org/0.86/s3638-08-07-z RESEARCH -bit quantization and oversampling at the receiver: Sequence based communication
More informationEE401: Advanced Communication Theory
EE401: Advanced Communication Theory Professor A. Manikas Chair of Communications and Array Processing Imperial College London Introductory Concepts Prof. A. Manikas (Imperial College) EE.401: Introductory
More informationExample: Bipolar NRZ (non-return-to-zero) signaling
Baseand Data Transmission Data are sent without using a carrier signal Example: Bipolar NRZ (non-return-to-zero signaling is represented y is represented y T A -A T : it duration is represented y BT. Passand
More informationNew Rank-One Matrix Decomposition Techniques and Applications to Signal Processing
New Rank-One Matrix Decomposition Techniques and Applications to Signal Processing Yongwei Huang Hong Kong Baptist University SPOC 2012 Hefei China July 1, 2012 Outline Trust-region subproblems in nonlinear
More informationUCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011
UCSD ECE53 Handout #40 Prof. Young-Han Kim Thursday, May 9, 04 Homework Set #8 Due: Thursday, June 5, 0. Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process,
More informationSimulation-Based Computation of Information Rates for Channels with Memory
January 12, 2004 Submitted to IEEE Trans. on Information Theory Simulation-Based Computation of Information Rates for Channels with Memory Dieter Arnold, Hans-Andrea Loeliger, Pascal O. Vontobel, Aleksandar
More informationIterative Timing Recovery
Iterative Timing Recovery John R. Barry School of Electrical and Computer Engineering, Georgia Tech Atlanta, Georgia U.S.A. barry@ece.gatech.edu 0 Outline Timing Recovery Tutorial Problem statement TED:
More informationLECTURE 16 AND 17. Digital signaling on frequency selective fading channels. Notes Prepared by: Abhishek Sood
ECE559:WIRELESS COMMUNICATION TECHNOLOGIES LECTURE 16 AND 17 Digital signaling on frequency selective fading channels 1 OUTLINE Notes Prepared by: Abhishek Sood In section 2 we discuss the receiver design
More informationExercises with solutions (Set D)
Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where
More information19. Channel coding: energy-per-bit, continuous-time channels
9. Channel coding: energy-per-bit, continuous-time channels 9. Energy per bit Consider the additive Gaussian noise channel: Y i = X i + Z i, Z i N ( 0, ). (9.) In the last lecture, we analyzed the maximum
More informationSignal Processing for Digital Data Storage (11)
Outline Signal Processing for Digital Data Storage (11) Assist.Prof. Piya Kovintavewat, Ph.D. Data Storage Technology Research Unit Nahon Pathom Rajabhat University Partial-Response Maximum-Lielihood (PRML)
More informationLECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem
LECTURE 15 Last time: Feedback channel: setting up the problem Perfect feedback Feedback capacity Data compression Lecture outline Joint source and channel coding theorem Converse Robustness Brain teaser
More informationmuch more on minimax (order bounds) cf. lecture by Iain Johnstone
much more on minimax (order bounds) cf. lecture by Iain Johnstone http://www-stat.stanford.edu/~imj/wald/wald1web.pdf today s lecture parametric estimation, Fisher information, Cramer-Rao lower bound:
More informationEE4061 Communication Systems
EE4061 Communication Systems Week 11 Intersymbol Interference Nyquist Pulse Shaping 0 c 2015, Georgia Institute of Technology (lect10 1) Intersymbol Interference (ISI) Tx filter channel Rx filter a δ(t-nt)
More informationLecture 5: Asymptotic Equipartition Property
Lecture 5: Asymptotic Equipartition Property Law of large number for product of random variables AEP and consequences Dr. Yao Xie, ECE587, Information Theory, Duke University Stock market Initial investment
More informationThe Fading Number of a Multiple-Access Rician Fading Channel
The Fading Number of a Multiple-Access Rician Fading Channel Intermediate Report of NSC Project Capacity Analysis of Various Multiple-Antenna Multiple-Users Communication Channels with Joint Estimation
More informationCoding for Digital Communication and Beyond Fall 2013 Anant Sahai MT 1
EECS 121 Coding for Digital Communication and Beyond Fall 2013 Anant Sahai MT 1 PRINT your student ID: PRINT AND SIGN your name:, (last) (first) (signature) PRINT your Unix account login: ee121- Prob.
More informationWeiyao Lin. Shanghai Jiao Tong University. Chapter 5: Digital Transmission through Baseband slchannels Textbook: Ch
Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 5: Digital Transmission through Baseband slchannels Textbook: Ch 10.1-10.5 2009/2010 Meixia Tao @ SJTU 1 Topics to be Covered
More informationTiming Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL
T F T I G E O R G A I N S T I T U T E O H E O F E A L P R O G R ESS S A N D 1 8 8 5 S E R V L O G Y I C E E C H N O Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL Aravind R. Nayak
More informationGaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26
Gaussian channel Information theory 2013, lecture 6 Jens Sjölund 8 May 2013 Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Outline 1 Definitions 2 The coding theorem for Gaussian channel 3 Bandlimited
More informationCapacity of the Discrete-Time AWGN Channel Under Output Quantization
Capacity of the Discrete-Time AWGN Channel Under Output Quantization Jaspreet Singh, Onkar Dabeer and Upamanyu Madhow Abstract We investigate the limits of communication over the discrete-time Additive
More informationBinary Transmissions over Additive Gaussian Noise: A Closed-Form Expression for the Channel Capacity 1
5 Conference on Information Sciences and Systems, The Johns Hopkins University, March 6 8, 5 inary Transmissions over Additive Gaussian Noise: A Closed-Form Expression for the Channel Capacity Ahmed O.
More informationLecture 4. Capacity of Fading Channels
1 Lecture 4. Capacity of Fading Channels Capacity of AWGN Channels Capacity of Fading Channels Ergodic Capacity Outage Capacity Shannon and Information Theory Claude Elwood Shannon (April 3, 1916 February
More informationEE 661: Modulation Theory Solutions to Homework 6
EE 66: Modulation Theory Solutions to Homework 6. Solution to problem. a) Binary PAM: Since 2W = 4 KHz and β = 0.5, the minimum T is the solution to (+β)/(2t ) = W = 2 0 3 Hz. Thus, we have the maximum
More informationInteractions of Information Theory and Estimation in Single- and Multi-user Communications
Interactions of Information Theory and Estimation in Single- and Multi-user Communications Dongning Guo Department of Electrical Engineering Princeton University March 8, 2004 p 1 Dongning Guo Communications
More informationThe Continuing Miracle of Information Storage Technology Paul H. Siegel Director, CMRR University of California, San Diego
The Continuing Miracle of Information Storage Technology Paul H. Siegel Director, CMRR University of California, San Diego 10/15/01 1 Outline The Shannon Statue A Miraculous Technology Information Theory
More informationDigital Communications
Digital Communications Chapter 9 Digital Communications Through Band-Limited Channels Po-Ning Chen, Professor Institute of Communications Engineering National Chiao-Tung University, Taiwan Digital Communications:
More informationData Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise.
Data Detection for Controlled ISI *Symbol by symbol suboptimum detection For the duobinary signal pulse h(nt) = 1 for n=0,1 and zero otherwise. The samples at the output of the receiving filter(demodulator)
More informationCapacity of the Discrete Memoryless Energy Harvesting Channel with Side Information
204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener
More informationDispersion of the Gilbert-Elliott Channel
Dispersion of the Gilbert-Elliott Channel Yury Polyanskiy Email: ypolyans@princeton.edu H. Vincent Poor Email: poor@princeton.edu Sergio Verdú Email: verdu@princeton.edu Abstract Channel dispersion plays
More informationECE598: Information-theoretic methods in high-dimensional statistics Spring 2016
ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma
More information18.2 Continuous Alphabet (discrete-time, memoryless) Channel
0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not
More informationMethod for analytically calculating BER (bit error rate) in presence of non-linearity. Gaurav Malhotra Xilinx
Method for analytically calculating BER (bit error rate) in presence of non-linearity Gaurav Malhotra Xilinx Outline Review existing methodology for calculating BER based on linear system analysis. Link
More informationInformation Theory for Dispersion-Free Fiber Channels with Distributed Amplification
Information Theory for Dispersion-Free Fiber Channels with Distributed Amplification Gerhard Kramer Department of Electrical and Computer Engineering Technical University of Munich CLEO-PR, OECC & PGC
More informationOPTIMAL FINITE ALPHABET SOURCES OVER PARTIAL RESPONSE CHANNELS. A Thesis DEEPAK KUMAR
OPTIMAL FINITE ALPHABET SOURCES OVER PARTIAL RESPONSE CHANNELS A Thesis by DEEPAK KUMAR Submitted to the Office of Graduate Studies of Texas A&M University in partial fulfillment of the requirements for
More informationEE456 Digital Communications
EE456 Digital Communications Professor Ha Nguyen September 5 EE456 Digital Communications Block Diagram of Binary Communication Systems m ( t { b k } b k = s( t b = s ( t k m ˆ ( t { bˆ } k r( t Bits in
More information2A1H Time-Frequency Analysis II
2AH Time-Frequency Analysis II Bugs/queries to david.murray@eng.ox.ac.uk HT 209 For any corrections see the course page DW Murray at www.robots.ox.ac.uk/ dwm/courses/2tf. (a) A signal g(t) with period
More informationEE5713 : Advanced Digital Communications
EE5713 : Advanced Digital Communications Week 12, 13: Inter Symbol Interference (ISI) Nyquist Criteria for ISI Pulse Shaping and Raised-Cosine Filter Eye Pattern Equalization (On Board) 20-May-15 Muhammad
More informationAppendix B Information theory from first principles
Appendix B Information theory from first principles This appendix discusses the information theory behind the capacity expressions used in the book. Section 8.3.4 is the only part of the book that supposes
More informationA new analytic approach to evaluation of Packet Error Rate in Wireless Networks
A new analytic approach to evaluation of Packet Error Rate in Wireless Networks Ramin Khalili Université Pierre et Marie Curie LIP6-CNRS, Paris, France ramin.khalili@lip6.fr Kavé Salamatian Université
More informationResidual Versus Suppressed-Carrier Coherent Communications
TDA Progress Report -7 November 5, 996 Residual Versus Suppressed-Carrier Coherent Communications M. K. Simon and S. Million Communications and Systems Research Section This article addresses the issue
More informationOn the Capacity of Markov Sources over Noisy Channels
On the Capacity of Markov Sources over Noisy Channels Aleksandar KavCiC Division of Engineering and Applied Sciences Harvard University, Cambridge, MA 02138 Abstract- We present an expectation-maximization
More informationMemory in Classical Information Theory: A Brief History
Beyond iid in information theory 8- January, 203 Memory in Classical Information Theory: A Brief History sergio verdu princeton university entropy rate Shannon, 948 entropy rate: memoryless process H(X)
More informationInformation Sources. Professor A. Manikas. Imperial College London. EE303 - Communication Systems An Overview of Fundamentals
Information Sources Professor A. Manikas Imperial College London EE303 - Communication Systems An Overview of Fundamentals Prof. A. Manikas (Imperial College) EE303: Information Sources 24 Oct. 2011 1
More informationinformation estimation feedback
relations between information and estimation in the presence of feedback talk at workshop on: Tsachy Weissman Information and Control in Networks LCCC - Lund Center for Control of Complex Engineering Systems
More informationOn the Optimality of Treating Interference as Noise in Competitive Scenarios
On the Optimality of Treating Interference as Noise in Competitive Scenarios A. DYTSO, D. TUNINETTI, N. DEVROYE WORK PARTIALLY FUNDED BY NSF UNDER AWARD 1017436 OUTLINE CHANNEL MODEL AND PAST WORK ADVANTAGES
More informationBASICS OF DETECTION AND ESTIMATION THEORY
BASICS OF DETECTION AND ESTIMATION THEORY 83050E/158 In this chapter we discuss how the transmitted symbols are detected optimally from a noisy received signal (observation). Based on these results, optimal
More informationPerformance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels
Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels Jilei Hou, Paul H. Siegel and Laurence B. Milstein Department of Electrical and Computer Engineering
More informationRevision of Lecture 4
Revision of Lecture 4 We have completed studying digital sources from information theory viewpoint We have learnt all fundamental principles for source coding, provided by information theory Practical
More informationSample Problems for the 9th Quiz
Sample Problems for the 9th Quiz. Draw the line coded signal waveform of the below line code for 0000. (a Unipolar nonreturn-to-zero (NRZ signaling (b Polar nonreturn-to-zero (NRZ signaling (c Unipolar
More informationEstimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition
Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition Seema Sud 1 1 The Aerospace Corporation, 4851 Stonecroft Blvd. Chantilly, VA 20151 Abstract
More informationStochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno
Stochastic Processes M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno 1 Outline Stochastic (random) processes. Autocorrelation. Crosscorrelation. Spectral density function.
More informationPerformance evaluation for ML sequence detection in ISI channels with Gauss Markov Noise
arxiv:0065036v [csit] 25 Jun 200 Performance evaluation for ML sequence detection in ISI channels with Gauss Marov Noise Naveen Kumar, Aditya Ramamoorthy and Murti Salapaa Dept of Electrical and Computer
More informationSquare Root Raised Cosine Filter
Wireless Information Transmission System Lab. Square Root Raised Cosine Filter Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal design
More informationPerformance Analysis of Spread Spectrum CDMA systems
1 Performance Analysis of Spread Spectrum CDMA systems 16:33:546 Wireless Communication Technologies Spring 5 Instructor: Dr. Narayan Mandayam Summary by Liang Xiao lxiao@winlab.rutgers.edu WINLAB, Department
More informationThis examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 6 December 2006 This examination consists of
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your
More informationDigital Modulation 1
Digital Modulation 1 Lecture Notes Ingmar Land and Bernard H. Fleury Navigation and Communications () Department of Electronic Systems Aalborg University, DK Version: February 5, 27 i Contents I Basic
More informationCharacterization of Heat-Assisted Magnetic Recording Channels
DIMACS Series in Discrete Mathematics and Theoretical Computer Science Volume 73, 2007 Characterization of Heat-Assisted Magnetic Recording Channels Rathnakumar Radhakrishnan, Bane Vasić, Fatih Erden,
More informationThe Method of Types and Its Application to Information Hiding
The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,
More informationLecture 1 From Continuous-Time to Discrete-Time
Lecture From Continuous-Time to Discrete-Time Outline. Continuous and Discrete-Time Signals and Systems................. What is a signal?................................2 What is a system?.............................
More informationConstellation Shaping for Communication Channels with Quantized Outputs
Constellation Shaping for Communication Channels with Quantized Outputs, Dr. Matthew C. Valenti and Xingyu Xiang Lane Department of Computer Science and Electrical Engineering West Virginia University
More informationPulse Shaping and ISI (Proakis: chapter 10.1, 10.3) EEE3012 Spring 2018
Pulse Shaping and ISI (Proakis: chapter 10.1, 10.3) EEE3012 Spring 2018 Digital Communication System Introduction Bandlimited channels distort signals the result is smeared pulses intersymol interference
More informationPrinciples of Communications
Principles of Communications Chapter V: Representation and Transmission of Baseband Digital Signal Yongchao Wang Email: ychwang@mail.xidian.edu.cn Xidian University State Key Lab. on ISN November 18, 2012
More information