Quantum rate distortion, reverse Shannon theorems, and source-channel separation
|
|
- Earl Lambert
- 5 years ago
- Views:
Transcription
1 Quantum rate distortion, reverse Shannon theorems, and source-channel separation ilanjana Datta, Min-Hsiu Hsieh, Mark Wilde (1) University of Cambridge,U.K. (2) McGill University, Montreal, Canada
2 Classical Information Theory Shannon s Source CodingTheorem Compression of information emitted by a source Shannon s oisy Channel Coding Theorem Transmission of information through a noisy channel
3 Shannon s Coding Theorems Source Coding Theorem memoryless stochastically independent signals X ~ p X ; x X The fundamental limit of data compression: = Shannon entropy of the source: H ( X ) Channel Coding Theorem X ~ p X input noisy channel memoryless Y output The fundamental limit on the rate of reliable information transmission: = Capacity of the channel C( ) max I( X : Y) p x mutual information
4 Lossless Data Compression s original signal from source E compressed signal s ' s D s ' decompressed signal (exact replica!) Shannon s Source CodingTheorem corresponds to asymptotically lossless data compression multiple ( n) uses of the channel signals: sequences p ( ) e n x1, x2..., xn 0 as n 0 (average probability of error in recovering original signal)
5 Lossless Compression oftentoostringent a condition for cases of multimedia data; e.g. audio, video and still images when storage space is insufficient Why? Typically, a substantial amount of data can be discarded before the information is sufficiently degraded to be noticeable!
6 An Example original image (uncompressed), size KB lossless compression (PG); size 60.1 KB lossy compression; 4.82 KB lossy compression; 1.14 KB
7 a data compression scheme : instead D: Lossy Data Compression recovered data original data D recovered data original data the allowed distortion Rate Distortion Theory the theory of Lossy Data Compression tradeoff rate of compression the allowed distortion fundamental limit on the asymptotic rate of data compression for a given maximum distortion
8 Classical Rate Distortion Theory For a given memoryless information source ; If the maximum allowed distortion is The minimum rate of data compression: R( D) rate distortion function min ( a) D I( X : Y) (Shannon) X ~ p X ; 0 D1, H( X) mutual information ( a) : p YX stochastic maps ; Ε( d( X, Y )) D average distortion Y : d( x, y) : output of a stochastic map distortion measure e.g. pyx : X Y 2 d( x, y) x y
9 Our Aim To obtain rate distortion functions in the quantum realm
10 Storage setting: Scenario for Quantum Rate Distortion for a fixed 0 D 1, H A a memoryless quantum information source
11 Storage setting: Scenario for Quantum Rate Distortion for a fixed 0 D 1 n uses of it n E n compressed state in c H n D recovered signal Rate of data compression achievable if q lim d (, D E ) n R if n R D R Quantum Rate Distortion function dim H n 2 D ( ): inf{ : achievable} c nr (min. rate of compression under given distortion)
12 Quantum Rate Distortion Function Barnum proved: R D min, q ( ) Ic (1) coherent information : CPTP d(, ) D RA R. A. I S R B c. R. B RB, ( ) ; d(, ) 1 F (, ) e entanglement fidelity He conjectured: " " holds in (1) problem!! the coherent information can be negative BUT The rate of a data compression scheme is non-negative!!
13 Storage setting: Equivalent Scenarios for Quantum Rate Distortion n uses of the source n E nrq ( D ) qubits D Communication setting: n uses of the source E nrq ( D ) noiseless qubit channels D
14 Another Scenario for Quantum Rate Distortion Communication setting: (entanglement assisted) shared entanglement n uses of the source E ea nrq ( D) noiseless qubit channels D R ea q ( D) min. number of noiseless qubit channels needed per use of the source in the presence of entanglement for a given distortion D entanglement-assisted quantum rate distortion function
15 Result - I Entanglement-assisted quantum rate distortion function: R ( D) min : CPTP ea q d (, ) D 1 : 2 I R B (2) quantum mutual information RA Proof: R. A. ea q (i) Prove : R ( D) RHS of (2). R. B RB using entropic inequalities (ii) Prove: This bound is achievable, i.e., " " holds
16 The rate Sketch of proof: R 1 min : 2 : CPTP d(, ) D IR B is achievable, Source H A ; Its purification: RA Let the minimizing CPTP map Let U : A BE a Stinespring isometry of RA R. A. (id U ) R E R RBE B entanglement noiseless qubit channel
17 RA R. A. R (id U ) 1 min I R: B 2 To prove: R : CPTP d(, ) D E R RBE B entanglement noiseless qubit channel is achievable ote: B satisfies d(, ) : ( ) D Suffices to prove: 1 : B 2 I R B uses of a noiseless qubit channel per use of the source can be sent to Bob by Alice through
18 How can we transfer B to Bob? initially n U n Asymptotic setting n BE State splitting finally n n B
19 How can we transfer B to Bob? RBE R q qubits from Alice to Bob R RBE E B + LO E B entanglement q 1 2 I R: B : CPTP d(, ) D achievable rate! 1 min : 2 IR B
20 n (i) local preparation of U (ii) State splitting with the help of shared entanglement such that n n BE n B simulating the (output of the) quantum channel n when the input is (using shared entanglement) n a special case of another protocol -- channel simulation Quantum Reverse Shannon Theorem (QRST) achievability of R ea q ( D)
21 Result - II Unassisted Quantum rate distortion function: lim 1 n Rq ( D ) ( n) n min : CPTP n ( n) d(, ) D ( n ) n E ( ) P regularized entanglement of purification For any AB ABR BR : : CPTP E min H (id ) 0 P BR R B R BR where, for any state, H ( ): Von eumann entropy of
22 Unassisted Quantum rate distortion function: R D ( n) n lim min E ( ) P q ( ) n Result - II 1 n : CPTP ( n) d(, ) D Channel simulation (QRST) in the absence of shared entanglement achievability of this rate To simulate output of n when input is n qubits sent at a rate E P ( ) RB B n ( ) B
23 Source Channel Separation Theorems Shannon: connects the two pillars of Classical Info. Theory Is it possible to transmit a classical information source reliably over a classical channel? Yes -- if & only if H ( X) C( ) ( ) ( X ) Implication: data compression code channel code decoding a 2-stage encoding & decoding with the best data compression + error correction codes is optimal!
24 Quantum Source Channel Separation Theorems Theorem : It possible to transmit a quantum information source ( ) Result - III reliably over a quantum channel ( ) if & only if von eumann entropy H ( ) Q ( ) quantum capacity Implication: data compression code channel code decoding a 2-stage encoding & decoding with the best data compression + quantum error correction codes is optimal!
25 H ( ) Q ( ) What if? Suppose Alice & Bob share entanglement, i.e., shared entanglement E D Theorem: it is possible to transmit the source over the channel, if & only if 1 ea q up to some distortion R ( D) I (, ) 2 c D
26 R ea q ( D): Summary entanglement-assisted quantum rate distortion function in terms of the quantum mutual information Rq ( D ): unassisted quantum rate distortion function in terms of regularised entanglement of purification Quantum Source Channel Separation Theorems
27 The following are some extra slides which I have removed/modified to make the talk 20 minute long.
28 Shannon s Coding Theorems Source Coding Theorem memoryless stochastically independent signals X ~ p X ; x X The fundamental limit of data compression: = Shannon entropy of the source: H ( X ) Channel Coding Theorem X ~ p X input noisy channel memoryless Y output The fundamental limit on the rate of reliable information transmission: = Capacity of the channel C( ) max I( X : Y) p x mutual information
29 Rate Distortion Function RD ( ) min I( X: Y) p YX : E( d ( XY, )) D For D 0, Y X R( D) H( X) D R( D) H( X) For 0,
30 Barnum s Conjecture Quantum Rate Distortion function R ( ) q D : CPTP d (, ) D min, Ic Motivation : Rate Distortion Function RD ( ) min I( X: Y) pyx ( y x): E( d ( xy, )) D mutual information similarity in the classical case Capacity of a noisy channel C( ) max I( X : Y) X ~ px p x p Y X Y Quantum capacity of a 1 n n Q ( ) lim max I, noisy quantum channel c n coherent information
31 State Splitting RBE R LO & QC from Alice to Bob R RBE E B E B entanglement BE Resource Inequality : q qq e qq 1 e IB: E ; 2 q 1 2 I R: B : CPTP d(, ) D achievable rate! 1 min : 2 E B IR B
32 Unassisted Quantum rate distortion function: R D ( n) n lim min E ( ) P q ( ) n Result - II 1 n : CPTP ( n) d(, ) D Channel simulation (QRST) in the absence of shared entanglement achievability of this rate To simulate output of n when input is Resource Inequality : E ( ) q q E P ( ) qubits RB P RB B ( ) B n
33 Unassisted Quantum rate distortion function: R D ( n) n lim min E ( ) P q ( ) n 1 n : CPTP ( n) d(, ) D Why the double regularization? Storage setting: n uses of the source n E D whereas in QRST: simulates n D C n " " To prove need to prove the converse done using entropic inequalities
Quantum rate distortion, reverse Shannon theorems, and source-channel separation
Quantum rate distortion, reverse Shannon theorems, and source-channel separation ilanjana Datta, Min-Hsiu Hsieh, and Mark M. Wilde arxiv:08.4940v3 [quant-ph] 9 Aug 202 Abstract We derive quantum counterparts
More informationEntropies & Information Theory
Entropies & Information Theory LECTURE I Nilanjana Datta University of Cambridge,U.K. See lecture notes on: http://www.qi.damtp.cam.ac.uk/node/223 Quantum Information Theory Born out of Classical Information
More informationQuantum Teleportation Pt. 1
Quantum Teleportation Pt. 1 PHYS 500 - Southern Illinois University April 17, 2018 PHYS 500 - Southern Illinois University Quantum Teleportation Pt. 1 April 17, 2018 1 / 13 Types of Communication In the
More informationClassical and Quantum Channel Simulations
Classical and Quantum Channel Simulations Mario Berta (based on joint work with Fernando Brandão, Matthias Christandl, Renato Renner, Joseph Renes, Stephanie Wehner, Mark Wilde) Outline Classical Shannon
More informationStrong converse theorems using Rényi entropies
Strong converse theorems using Rényi entropies Felix Leditzky joint work with Mark M. Wilde and Nilanjana Datta arxiv:1506.02635 5 January 2016 Table of Contents 1 Weak vs. strong converse 2 Rényi entropies
More information9. Distance measures. 9.1 Classical information measures. Head Tail. How similar/close are two probability distributions? Trace distance.
9. Distance measures 9.1 Classical information measures How similar/close are two probability distributions? Trace distance Fidelity Example: Flipping two coins, one fair one biased Head Tail Trace distance
More informationLecture 11: Quantum Information III - Source Coding
CSCI5370 Quantum Computing November 25, 203 Lecture : Quantum Information III - Source Coding Lecturer: Shengyu Zhang Scribe: Hing Yin Tsang. Holevo s bound Suppose Alice has an information source X that
More informationDigital Communications III (ECE 154C) Introduction to Coding and Information Theory
Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Tara Javidi These lecture notes were originally developed by late Prof. J. K. Wolf. UC San Diego Spring 2014 1 / 8 I
More information2015 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media,
2015 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising
More informationLecture 1. Introduction
Lecture 1. Introduction What is the course about? Logistics Questionnaire Dr. Yao Xie, ECE587, Information Theory, Duke University What is information? Dr. Yao Xie, ECE587, Information Theory, Duke University
More informationLecture 8: Shannon s Noise Models
Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 8: Shannon s Noise Models September 14, 2007 Lecturer: Atri Rudra Scribe: Sandipan Kundu& Atri Rudra Till now we have
More informationTrading classical communication, quantum communication, and entanglement in quantum Shannon theory
Trading classical communication, quantum communication, and entanglement in quantum Shannon theory Min-Hsiu Hsieh and Mark M. Wilde arxiv:9.338v3 [quant-ph] 9 Apr 2 Abstract We give trade-offs between
More informationStrong converse theorems using Rényi entropies
Strong converse theorems using Rényi entropies Felix Leditzky a, Mark M. Wilde b, and Nilanjana Datta a a Statistical Laboratory, Centre for Mathematical Sciences, University of Cambridge, Cambridge C3
More informationEntanglement-Assisted Capacity of a Quantum Channel and the Reverse Shannon Theorem
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 10, OCTOBER 2002 2637 Entanglement-Assisted Capacity of a Quantum Channel the Reverse Shannon Theorem Charles H. Bennett, Peter W. Shor, Member, IEEE,
More informationQuantum Hadamard channels (II)
Quantum Hadamard channels (II) Vlad Gheorghiu Department of Physics Carnegie Mellon University Pittsburgh, PA 15213, USA August 5, 2010 Vlad Gheorghiu (CMU) Quantum Hadamard channels (II) August 5, 2010
More informationDegradable Quantum Channels
Degradable Quantum Channels Li Yu Department of Physics, Carnegie-Mellon University, Pittsburgh, PA May 27, 2010 References The capacity of a quantum channel for simultaneous transmission of classical
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationLecture 4 Channel Coding
Capacity and the Weak Converse Lecture 4 Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 15, 2014 1 / 16 I-Hsiang Wang NIT Lecture 4 Capacity
More informationSemidefinite programming strong converse bounds for quantum channel capacities
Semidefinite programming strong converse bounds for quantum channel capacities Xin Wang UTS: Centre for Quantum Software and Information Joint work with Runyao Duan and Wei Xie arxiv:1610.06381 & 1601.06888
More informationQuantum Information Theory Tutorial
Quantum Information Theory Tutorial Mark M. Wilde Hearne Institute for Theoretical Physics, Department of Physics and Astronomy, Center for Computation and Technology, Louisiana State University, Baton
More informationLecture 6: Quantum error correction and quantum capacity
Lecture 6: Quantum error correction and quantum capacity Mark M. Wilde The quantum capacity theorem is one of the most important theorems in quantum hannon theory. It is a fundamentally quantum theorem
More informationQuantum to Classical Randomness Extractors
Quantum to Classical Randomness Extractors Mario Berta, Omar Fawzi, Stephanie Wehner - Full version preprint available at arxiv: 1111.2026v3 08/23/2012 - CRYPTO University of California, Santa Barbara
More informationGilles Brassard. Université de Montréal
Gilles Brassard Université de Montréal Gilles Brassard Université de Montréal VOLUME 76, NUMBER 5 P H Y S I C A L R E V I E W L E T T E R S 29 JANUARY 1996 Purification of Noisy Entanglement and Faithful
More informationInformation Theoretic Limits of Randomness Generation
Information Theoretic Limits of Randomness Generation Abbas El Gamal Stanford University Shannon Centennial, University of Michigan, September 2016 Information theory The fundamental problem of communication
More informationarxiv: v4 [quant-ph] 22 Nov 2012
Quantum trade-off coding for bosonic communication Mark M. Wilde and Patrick Hayden School of Computer Science, McGill University, Montreal, Québec, Canada H3A A7 Saikat Guha Quantum Information Processing
More informationTransmitting and Hiding Quantum Information
2018/12/20 @ 4th KIAS WORKSHOP on Quantum Information and Thermodynamics Transmitting and Hiding Quantum Information Seung-Woo Lee Quantum Universe Center Korea Institute for Advanced Study (KIAS) Contents
More informationQuantum Sphere-Packing Bounds and Moderate Deviation Analysis for Classical-Quantum Channels
Quantum Sphere-Packing Bounds and Moderate Deviation Analysis for Classical-Quantum Channels (, ) Joint work with Min-Hsiu Hsieh and Marco Tomamichel Hao-Chung Cheng University of Technology Sydney National
More informationto mere bit flips) may affect the transmission.
5 VII. QUANTUM INFORMATION THEORY to mere bit flips) may affect the transmission. A. Introduction B. A few bits of classical information theory Information theory has developed over the past five or six
More informationLecture 4 Noisy Channel Coding
Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem
More informationCapacity of a channel Shannon s second theorem. Information Theory 1/33
Capacity of a channel Shannon s second theorem Information Theory 1/33 Outline 1. Memoryless channels, examples ; 2. Capacity ; 3. Symmetric channels ; 4. Channel Coding ; 5. Shannon s second theorem,
More informationInformation Theory. Coding and Information Theory. Information Theory Textbooks. Entropy
Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is
More informationQuantum Information Theory
Quantum Information Theory C. H. Bennett IBM Research Yorktown QIS Workshop Vienna, VA 23 Apr 2009 Nature and importance of Information Distinctive features of Quantum Information Superposition principle,
More informationThe Quantum Reverse Shannon Theorem Based on One-Shot Information Theory
Commun. Math. Phys. 306, 579 65 (0) Digital Object Identifier (DOI) 0.007/s000-0-309-7 Communications in Mathematical Physics The Quantum Reverse Shannon Theorem Based on One-Shot Information Theory Mario
More information1. Basic rules of quantum mechanics
1. Basic rules of quantum mechanics How to describe the states of an ideally controlled system? How to describe changes in an ideally controlled system? How to describe measurements on an ideally controlled
More informationFundamental rate-loss tradeoff for optical quantum key distribution
Fundamental rate-loss tradeoff for optical quantum key distribution Masahiro Takeoka (NICT) Saikat Guha (BBN) Mark M. Wilde (LSU) Quantum Krispy Kreme Seminar @LSU January 30, 2015 Outline Motivation Main
More informationCompression and Coding
Compression and Coding Theory and Applications Part 1: Fundamentals Gloria Menegaz 1 Transmitter (Encoder) What is the problem? Receiver (Decoder) Transformation information unit Channel Ordering (significance)
More informationarxiv:quant-ph/ v2 14 May 2002
arxiv:quant-ph/0106052v2 14 May 2002 Entanglement-Assisted Capacity of a Quantum Channel and the Reverse Shannon Theorem Charles H. Bennett, Peter W. Shor, John A. Smolin, and Ashish V. Thapliyal Abstract
More informationDistributed Lossless Compression. Distributed lossless compression system
Lecture #3 Distributed Lossless Compression (Reading: NIT 10.1 10.5, 4.4) Distributed lossless source coding Lossless source coding via random binning Time Sharing Achievability proof of the Slepian Wolf
More informationQuantum Rate-Distortion Coding of Relevant Information
Quantum Rate-Distortion Coding of Relevant Information Sina Salek 1 2 Daniela Cadamuro 1 3 Philipp Kammerlander 4 and Karoline Wiesner 1 1 School of Mathematics University of Bristol Bristol BS8 1TW United
More informationPerformance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)
Performance-based Security for Encoding of Information Signals FA9550-15-1-0180 (2015-2018) Paul Cuff (Princeton University) Contributors Two students finished PhD Tiance Wang (Goldman Sachs) Eva Song
More informationSemidefinite programming strong converse bounds for quantum channel capacities
Semidefinite programming strong converse bounds for quantum channel capacities Xin Wang UTS: Centre for Quantum Software and Information Joint work with Wei Xie, Runyao Duan (UTS:QSI) QIP 2017, Microsoft
More informationOn Common Information and the Encoding of Sources that are Not Successively Refinable
On Common Information and the Encoding of Sources that are Not Successively Refinable Kumar Viswanatha, Emrah Akyol, Tejaswi Nanjundaswamy and Kenneth Rose ECE Department, University of California - Santa
More informationNoisy channel communication
Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University
More informationLecture 22: Final Review
Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information
More informationSpin chain model for correlated quantum channels
Spin chain model for correlated quantum channels Davide Rossini Scuola Internazionale Superiore di Studi Avanzati SISSA Trieste, Italy in collaboration with: Vittorio Giovannetti (Pisa) Simone Montangero
More informationSending Quantum Information with Zero Capacity Channels
Sending Quantum Information with Zero Capacity Channels Graeme Smith IM Research KITP Quantum Information Science Program November 5, 2009 Joint work with Jon and John Noisy Channel Capacity X N Y p(y
More informationTime Reversal and Exchange Symmetries of Unitary Gate Capacities
Time Reversal and Exchange Symmetries of Unitary Gate Capacities The MIT Faculty has made this article openly available Please share how this access benefits you Your story matters Citation As Published
More informationUNIT I INFORMATION THEORY. I k log 2
UNIT I INFORMATION THEORY Claude Shannon 1916-2001 Creator of Information Theory, lays the foundation for implementing logic in digital circuits as part of his Masters Thesis! (1939) and published a paper
More informationReliable Computation over Multiple-Access Channels
Reliable Computation over Multiple-Access Channels Bobak Nazer and Michael Gastpar Dept. of Electrical Engineering and Computer Sciences University of California, Berkeley Berkeley, CA, 94720-1770 {bobak,
More informationInformation-theoretic treatment of tripartite systems and quantum channels. Patrick Coles Carnegie Mellon University Pittsburgh, PA USA
Information-theoretic treatment of tripartite systems and quantum channels Patrick Coles Carnegie Mellon University Pittsburgh, PA USA ArXiv: 1006.4859 a c b Co-authors Li Yu Vlad Gheorghiu Robert Griffiths
More informationMultimedia. Multimedia Data Compression (Lossless Compression Algorithms)
Course Code 005636 (Fall 2017) Multimedia Multimedia Data Compression (Lossless Compression Algorithms) Prof. S. M. Riazul Islam, Dept. of Computer Engineering, Sejong University, Korea E-mail: riaz@sejong.ac.kr
More informationLecture: Quantum Information
Lecture: Quantum Information Transcribed by: Crystal Noel and Da An (Chi Chi) November 10, 016 1 Final Proect Information Find an issue related to class you are interested in and either: read some papers
More informationarxiv:quant-ph/ v1 9 May 2005
Quantum information can be negative Michal Horodecki 1, Jonathan Oppenheim & Andreas Winter 3 1 Institute of Theoretical Physics and Astrophysics, University of Gdańsk, 80 95 Gdańsk, Poland Department
More informationDigital communication system. Shannon s separation principle
Digital communication system Representation of the source signal by a stream of (binary) symbols Adaptation to the properties of the transmission channel information source source coder channel coder modulation
More informationLecture 16 Oct 21, 2014
CS 395T: Sublinear Algorithms Fall 24 Prof. Eric Price Lecture 6 Oct 2, 24 Scribe: Chi-Kit Lam Overview In this lecture we will talk about information and compression, which the Huffman coding can achieve
More informationCSCI 2570 Introduction to Nanocomputing
CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication
More informationQuantum Information Theory and Cryptography
Quantum Information Theory and Cryptography John Smolin, IBM Research IPAM Information Theory A Mathematical Theory of Communication, C.E. Shannon, 1948 Lies at the intersection of Electrical Engineering,
More informationInformation Complexity and Applications. Mark Braverman Princeton University and IAS FoCM 17 July 17, 2017
Information Complexity and Applications Mark Braverman Princeton University and IAS FoCM 17 July 17, 2017 Coding vs complexity: a tale of two theories Coding Goal: data transmission Different channels
More informationOther Topics in Quantum Information
p. 1/23 Other Topics in Quantum Information In a course like this there is only a limited time, and only a limited number of topics can be covered. Some additional topics will be covered in the class projects.
More informationProblem Set: TT Quantum Information
Problem Set: TT Quantum Information Basics of Information Theory 1. Alice can send four messages A, B, C, and D over a classical channel. She chooses A with probability 1/, B with probability 1/4 and C
More informationMAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)
MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)
More informationBlock 2: Introduction to Information Theory
Block 2: Introduction to Information Theory Francisco J. Escribano April 26, 2015 Francisco J. Escribano Block 2: Introduction to Information Theory April 26, 2015 1 / 51 Table of contents 1 Motivation
More informationarxiv:quant-ph/ v1 6 Nov 2003
Tema con variazioni: quantum channel capacity arxiv:quant-ph/0311037v1 6 Nov 2003 Submitted to: New J. Phys. Dennis Kretschmann and Reinhard F Werner Institut für Mathematische Physik, Technische Universität
More information(Classical) Information Theory III: Noisy channel coding
(Classical) Information Theory III: Noisy channel coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract What is the best possible way
More informationQuantum Information Theory
Chapter 5 Quantum Information Theory Quantum information theory is a rich subject that could easily have occupied us all term. But because we are short of time (I m anxious to move on to quantum computation),
More informationarxiv: v1 [quant-ph] 3 Jan 2008
A paradigm for entanglement theory based on quantum communication Jonathan Oppenheim 1 1 Department of Applied Mathematics and Theoretical Physics, University of Cambridge U.K. arxiv:0801.0458v1 [quant-ph]
More informationPing Pong Protocol & Auto-compensation
Ping Pong Protocol & Auto-compensation Adam de la Zerda For QIP seminar Spring 2004 02.06.04 Outline Introduction to QKD protocols + motivation Ping-Pong protocol Security Analysis for Ping-Pong Protocol
More informationCoding for Discrete Source
EGR 544 Communication Theory 3. Coding for Discrete Sources Z. Aliyazicioglu Electrical and Computer Engineering Department Cal Poly Pomona Coding for Discrete Source Coding Represent source data effectively
More informationMAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A
MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK DEPARTMENT: ECE SEMESTER: IV SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A 1. What is binary symmetric channel (AUC DEC
More informationSquashed entanglement
Squashed Entanglement based on Squashed Entanglement - An Additive Entanglement Measure (M. Christandl, A. Winter, quant-ph/0308088), and A paradigm for entanglement theory based on quantum communication
More informationInformation Theory - Entropy. Figure 3
Concept of Information Information Theory - Entropy Figure 3 A typical binary coded digital communication system is shown in Figure 3. What is involved in the transmission of information? - The system
More informationChapter 14: Quantum Information Theory and Photonic Communications
Chapter 14: Quantum Information Theory and Photonic Communications Farhan Rana (MIT) March, 2002 Abstract The fundamental physical limits of optical communications are discussed in the light of the recent
More informationQuantum Achievability Proof via Collision Relative Entropy
Quantum Achievability Proof via Collision Relative Entropy Salman Beigi Institute for Research in Fundamental Sciences (IPM) Tehran, Iran Setemper 8, 2014 Based on a joint work with Amin Gohari arxiv:1312.3822
More information18.2 Continuous Alphabet (discrete-time, memoryless) Channel
0-704: Information Processing and Learning Spring 0 Lecture 8: Gaussian channel, Parallel channels and Rate-distortion theory Lecturer: Aarti Singh Scribe: Danai Koutra Disclaimer: These notes have not
More informationELEMENT OF INFORMATION THEORY
History Table of Content ELEMENT OF INFORMATION THEORY O. Le Meur olemeur@irisa.fr Univ. of Rennes 1 http://www.irisa.fr/temics/staff/lemeur/ October 2010 1 History Table of Content VERSION: 2009-2010:
More informationSchur-Weyl duality, quantum data compression, tomography
PHYSICS 491: Symmetry and Quantum Information April 25, 2017 Schur-Weyl duality, quantum data compression, tomography Lecture 7 Michael Walter, Stanford University These lecture notes are not proof-read
More informationShannon s noisy-channel theorem
Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for
More information5 Mutual Information and Channel Capacity
5 Mutual Information and Channel Capacity In Section 2, we have seen the use of a quantity called entropy to measure the amount of randomness in a random variable. In this section, we introduce several
More informationUniformly Additive Entropic Formulas
Uniformly Additive Entropic Formulas Andrew Cross, Ke Li, Graeme Smith JILA and Department of Physics, University of Colorado Boulder QMATH2016 Georgia Tech October 9, 2016 Information theory: optimal
More informationClassical data compression with quantum side information
Classical data compression with quantum side information I Devetak IBM TJ Watson Research Center, Yorktown Heights, NY 10598, USA arxiv:quant-ph/0209029v3 18 Apr 2003 A Winter Department of Computer Science,
More informationarxiv:quant-ph/ v1 2 Dec 2005
A Resource Framework for Quantum Shannon Theory arxiv:quant-ph/0512015v1 2 Dec 2005 I. Devetak Department of Electrical Engineering Systems, University of Southern California, Los Angeles, CA 90089, USA
More informationTurbo Compression. Andrej Rikovsky, Advisor: Pavol Hanus
Turbo Compression Andrej Rikovsky, Advisor: Pavol Hanus Abstract Turbo codes which performs very close to channel capacity in channel coding can be also used to obtain very efficient source coding schemes.
More informationChapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University
Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission
More informationStrong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach
Strong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach Silas Fong (Joint work with Vincent Tan) Department of Electrical & Computer Engineering National University
More information4. Quantization and Data Compression. ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak
4. Quantization and Data Compression ECE 32 Spring 22 Purdue University, School of ECE Prof. What is data compression? Reducing the file size without compromising the quality of the data stored in the
More informationCommon Information. Abbas El Gamal. Stanford University. Viterbi Lecture, USC, April 2014
Common Information Abbas El Gamal Stanford University Viterbi Lecture, USC, April 2014 Andrew Viterbi s Fabulous Formula, IEEE Spectrum, 2010 El Gamal (Stanford University) Disclaimer Viterbi Lecture 2
More informationSource Coding with Lists and Rényi Entropy or The Honey-Do Problem
Source Coding with Lists and Rényi Entropy or The Honey-Do Problem Amos Lapidoth ETH Zurich October 8, 2013 Joint work with Christoph Bunte. A Task from your Spouse Using a fixed number of bits, your spouse
More informationEfficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel
Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel Ahmad Abu Al Haija ECE Department, McGill University, Montreal, QC, Canada Email: ahmad.abualhaija@mail.mcgill.ca
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationBasic Principles of Video Coding
Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion
More informationInformation measures in simple coding problems
Part I Information measures in simple coding problems in this web service in this web service Source coding and hypothesis testing; information measures A(discrete)source is a sequence {X i } i= of random
More informationFRAMES IN QUANTUM AND CLASSICAL INFORMATION THEORY
FRAMES IN QUANTUM AND CLASSICAL INFORMATION THEORY Emina Soljanin Mathematical Sciences Research Center, Bell Labs April 16, 23 A FRAME 1 A sequence {x i } of vectors in a Hilbert space with the property
More informationGaussian Bosonic Channels: conjectures, proofs, and bounds
IICQI14 Gaussian Bosonic Channels: conjectures, proofs, and bounds Vittorio Giovannetti NEST, Scuola Normale Superiore and Istituto Nanoscienze-CNR A. S. Holevo (Steklov, Moskov) R. Garcia-Patron (University
More informationAnalyzing Large Communication Networks
Analyzing Large Communication Networks Shirin Jalali joint work with Michelle Effros and Tracey Ho Dec. 2015 1 The gap Fundamental questions: i. What is the best achievable performance? ii. How to communicate
More informationLecture 6 I. CHANNEL CODING. X n (m) P Y X
6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder
More informationEntropy in Classical and Quantum Information Theory
Entropy in Classical and Quantum Information Theory William Fedus Physics Department, University of California, San Diego. Entropy is a central concept in both classical and quantum information theory,
More informationIntroduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.
L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission
More informationEntanglement Manipulation
Entanglement Manipulation Steven T. Flammia 1 1 Perimeter Institute for Theoretical Physics, Waterloo, Ontario, N2L 2Y5 Canada (Dated: 22 March 2010) These are notes for my RIT tutorial lecture at the
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More information