Quantum Information Theory. Renato Renner ETH Zurich

Similar documents
Quantum Cryptography. Chris1an Schaffner. Research Center for Quantum So8ware

Secrecy and the Quantum

Lecture 1: Perfect Secrecy and Statistical Authentication. 2 Introduction - Historical vs Modern Cryptography

Parallel Repetition of entangled games on the uniform distribution

PERFECTLY secure key agreement has been studied recently

Security Implications of Quantum Technologies

5th March Unconditional Security of Quantum Key Distribution With Practical Devices. Hermen Jan Hupkes

Quantum Cryptography

Enigma Marian Rejewski, Jerzy Róz ycki, Henryk Zygalski

+ = OTP + QKD = QC. ψ = a. OTP One-Time Pad QKD Quantum Key Distribution QC Quantum Cryptography. θ = 135 o state 1

Introduction to Quantum Key Distribution

Physics is becoming too difficult for physicists. David Hilbert (mathematician)

Cryptography CS 555. Topic 25: Quantum Crpytography. CS555 Topic 25 1

Transmitting and Hiding Quantum Information

Lecture 11: Quantum Information III - Source Coding

Cryptography and Security Final Exam

Entanglement and information

9. Distance measures. 9.1 Classical information measures. Head Tail. How similar/close are two probability distributions? Trace distance.

Quantum Cryptography. Areas for Discussion. Quantum Cryptography. Photons. Photons. Photons. MSc Distributed Systems and Security

Ping Pong Protocol & Auto-compensation

Cryptography. Lecture 2: Perfect Secrecy and its Limitations. Gil Segev

3/10/11. Which interpreta/on sounds most reasonable to you? PH300 Modern Physics SP11

Perfectly secure cipher system.

APPLICATIONS. Quantum Communications

Cellular automata, entropy and box- coun4ng dimension

Tutorial on Quantum Computing. Vwani P. Roychowdhury. Lecture 1: Introduction

Problem Set: TT Quantum Information

10. Physics from Quantum Information. I. The Clifton-Bub-Halvorson (CBH) Theorem.

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Quantum Error Correcting Codes and Quantum Cryptography. Peter Shor M.I.T. Cambridge, MA 02139

Secret-Key Agreement over Unauthenticated Public Channels Part I: Definitions and a Completeness Result

arxiv:quant-ph/ v1 13 Jan 2003

Single and Entangled photons. Edward Pei

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

The BB84 cryptologic protocol

An Introduction to Quantum Information. By Aditya Jain. Under the Guidance of Dr. Guruprasad Kar PAMU, ISI Kolkata

Entropies & Information Theory

An Introduction to Quantum Information and Applications

Quantum Technologies for Cryptography

Notes for Lecture 17

Lecture 2: Perfect Secrecy and its Limitations

Quantum Communication. Serge Massar Université Libre de Bruxelles

Topics. Probability Theory. Perfect Secrecy. Information Theory

W3203 Discrete Mathema1cs. Number Theory. Spring 2015 Instructor: Ilia Vovsha. hcp://

Ph 219/CS 219. Exercises Due: Friday 3 November 2006

(Classical) Information Theory III: Noisy channel coding

CS120, Quantum Cryptography, Fall 2016

Lecture 1: Introduction to Public key cryptography

Cryptography and Security Final Exam

Elementary cellular automata

Research, Development and Simulation of Quantum Cryptographic Protocols

LECTURE NOTES ON Quantum Cryptography

Technical Report Communicating Secret Information Without Secret Messages

10 - February, 2010 Jordan Myronuk

Chapter 13: Photons for quantum information. Quantum only tasks. Teleportation. Superdense coding. Quantum key distribution

arxiv:quant-ph/ v2 3 Oct 2000

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013

Bit-Commitment and Coin Flipping in a Device-Independent Setting

Entanglement and Quantum Teleportation

Other Topics in Quantum Information

Quantum sampling of mixed states

Chapter 2 : Perfectly-Secret Encryption

Quantum Cryptography

A probabilistic quantum key transfer protocol

An Introduction. Dr Nick Papanikolaou. Seminar on The Future of Cryptography The British Computer Society 17 September 2009

Realization of B92 QKD protocol using id3100 Clavis 2 system

CS6750: Cryptography and Communica7on Security

My brief introduction to cryptography

CS 282A/MATH 209A: Foundations of Cryptography Prof. Rafail Ostrosky. Lecture 4

Stop Conditions Of BB84 Protocol Via A Depolarizing Channel (Quantum Cryptography)

Quantum Teleportation Pt. 1

Research Proposal for Secure Double slit experiment. Sandeep Cheema Security Analyst, Vichara Technologies. Abstract

Outline. Computer Science 418. Number of Keys in the Sum. More on Perfect Secrecy, One-Time Pad, Entropy. Mike Jacobson. Week 3

Theory of Computation Chapter 12: Cryptography

Zero-Knowledge Against Quantum Attacks

Information-theoretic Secrecy A Cryptographic Perspective

Example: sending one bit of information across noisy channel. Effects of the noise: flip the bit with probability p.

Seminar 1. Introduction to Quantum Computing

Network Security Based on Quantum Cryptography Multi-qubit Hadamard Matrices

Quantum Pseudo-Telepathy

EPR paradox, Bell inequality, etc.

Quantum Information & Quantum Computation

Leftovers from Lecture 3

Lecture 9 - Symmetric Encryption

Lecture 11 September 30, 2015

A Matlab Realization of Shor s Quantum Factoring Algorithm

Cryptography. P. Danziger. Transmit...Bob...

A New Wireless Quantum Key Distribution Protocol based on Authentication And Bases Center (AABC)

We can't solve problems by using the same kind of thinking we used when we created them.!

Quantum technology popular science description

Quantum Information Theory and Cryptography

Quantum key distribution for the lazy and careless

Quantum Wireless Sensor Networks

Lecture 19: Public-key Cryptography (Diffie-Hellman Key Exchange & ElGamal Encryption) Public-key Cryptography

Lecture Notes on Secret Sharing

Quantum Information Transfer and Processing Miloslav Dušek

A Piggybank Protocol for Quantum Cryptography

CHALMERS GÖTEBORGS UNIVERSITET. TDA352 (Chalmers) - DIT250 (GU) 11 April 2017, 8:30-12:30

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

Lecture Notes 20: Zero-Knowledge Proofs

Transcription:

Quantum Information Theory Renato Renner ETH Zurich

Do quantum computers exist? Geordie Rose and his D- Wave quantum computer

Commercial devices Quantum cryptography device by id Quan(que, in opera@on at the FIFA World Cup compe@@on in Durban

Commercial devices Quan@s Random Number Generator (4 Mbits/s)

Is the world deterministic? Jedenfalls bin ich überzeugt, daß der nicht würfelt. Albert Einstein, in a letter to Max Born It took several decades until the inherent non-deterministic nature of quantum theory was accepted as a physical fact. Mein Ziel war zu beweisen, dass niemand, nicht mal Gott, den Verlauf der Welt voraussagen kann. Ernst Specker (mathematician, ETH Zurich)

Research in quantum information Nobel prize 2012 Serge Haroche David Wineland

Research in quantum information Research groups in quantum informa@on science (from hkp://www.quan@ki.org)

Research in quantum information Swiss network consis@ng of more than 300 scien@sts QSIT

Quantum Information Theory

What is quantum? quantum physics classical physics general relativity 10-15 m 10-9 m 10-3 m 10 3 m 10 9 m 10 15 m 10 21 m 1 fm 1 pm 1 nm 1 µm 1 mm 1 m 1 km 1 Mm 1 Gm 1 Tm 1 Pm 1 Em 1 Zm

Quantum physics deviates from our day-today experience about the world around us Example 1: Entanglement polariza@on measurement along direc@on ψ ψ ψ

Quantum physics deviates from our day-today experience about the world around us Example 2: No-cloning principle ψ ψ reproduc@on ψ

Quantum physics deviates from our day-today experience about the world around us Example 2: No-cloning principle ψ ψ reproduc@on ψ No- cloning theorem [WooKers und Zurek, 1982] No physical device can copy the state of a quantum system (for arbitrary states ψ).

The no- cloning principle does not seem to be valid for macroscopic objects reproduc@on

Quantum Information Theory

What is informa@on?

Two different approaches Kolmogorov s no@on: based on the theory of computa@on Shannon s no@on: based on probability theory

Common feature of both approaches The mathema@cal theory should be independent of how informa@on is represented.

Common feature of both approaches The mathema@cal theory should be independent of how informa@on is represented. As we shall see, this idea is doomed to fail

Kolmogorov s no@on of informa@on

What is informa@on? For each pixel (x0,y0) on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) } The algorithm reproduces this picture.

What is informa@on? For each pixel on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) } Informa@on can be represented in various equivalent ways.

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m.

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m. Examples

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m. Examples (1) m = 0000000000000000000000000000000000000

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m. Examples (1) m = 0000000000000000000000000000000000000 (2) m = 0000000000000100000000000010000000001

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m. Examples (1) m = 0000000000000000000000000000000000000 (2) m = 0000000000000100000000000010000000001 (3) m = 1592653589793238462643383279502884197

Kolmogorov s no@on of informa@on Defini1on [Kolmogorov]: The informa@on content of a message m is the length (in number of bits) of the shortest program that outputs m. Examples (1) m = 0000000000000000000000000000000000000 (2) m = 0000000000000100000000000010000000001 (3) m = 1592653589793238462643383279502884197 (4) m = 3845879501648135484764749358418500147

Kolmogorov s no@on of informa@on

Kolmogorov s no@on of informa@on Kolmogorov s defini@on of informa@on content has some remarkable proper@es:

Kolmogorov s no@on of informa@on Kolmogorov s defini@on of informa@on content has some remarkable proper@es: Model- independent: it is independent of the underlying programming language (up to an addi@ve constant).

Kolmogorov s no@on of informa@on Kolmogorov s defini@on of informa@on content has some remarkable proper@es: Model- independent: it is independent of the underlying programming language (up to an addi@ve constant). Incomputable: There is no algorithm that takes as input a message m and outputs its informa@on content.

Shannon s no@on of informa@on

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m].

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m]. Examples

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m]. Examples (1) m: the lokery numbers

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m]. Examples (1) m: the lokery numbers (2) m: message whether you have won the lokery

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m]. Examples (1) m: the lokery numbers (2) m: message whether you have won the lokery (3) m = π

Shannon s no@on of informa@on Defini1on [Shannon]: The informa@on content S(m) of a message m is equal to the nega@ve logarithm of its probability Pr[m], i.e., S(m) = log 2 Pr[m]. Examples (1) m: the lokery numbers (2) m: message whether you have won the lokery (3) m = π (4) m: random bitstring of length n

Shannon s no@on of informa@on

Shannon s no@on of informa@on Some remarks on Shannon s defini@on of informa@on content :

Shannon s no@on of informa@on Some remarks on Shannon s defini@on of informa@on content : Probabilis@c defini@on: Requires an underlying probability distribu@on P M on the set of messages M.

Shannon s no@on of informa@on Some remarks on Shannon s defini@on of informa@on content : Probabilis@c defini@on: Requires an underlying probability distribu@on P M on the set of messages M. Easily computable.

Shannon s no@on of informa@on Some remarks on Shannon s defini@on of informa@on content : Probabilis@c defini@on: Requires an underlying probability distribu@on P M on the set of messages M. Easily computable. Widely used in modern informa@on theory (in theory and prac@ce).

The idea of informa@on compression Suppose that we want to transmit the picture over a communica@on channel with limited capacity. Sender Receiver 1000 bit channel

The idea of informa@on compression Sender Receiver Compression Decoding For each pixel on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) } 1000 bit channel For each pixel on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) }

M The idea of informa@on compression For each pixel on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) } 1000 bits For each pixel on the screen do: { x = 0 y = 0 itera@on = 0 max_itera@on = 1000 while ( x*x + y*y <= (2*2) AND itera@on < max_itera@on ) { xtemp = x*x - y*y + x0 y = 2*x*y + y0 x = xtemp itera@on = itera@on + 1 } if ( itera@on == max_itera@on ) then color = black else color = itera@on plot(x0,y0,color) } M For abcdef Hijklmnop Qrstuvw Xzy xyz For abcdef Hijklmnop Qrstuvw Xzy xyz 200 bits For abcdef Hijklmnop Qrstuvw Xzy xyz For abcdef Hijklmnop Qrstuvw Xzy xyz Quan@fy informa@on content of a message M by the size (in # bits) of the minimal compression.

Shannon entropy and compression M For abcdef Hijklmnop Qrstuvw Xzy xyz For abcdef Hijklmnop Qrstuvw Xzy xyz H(M) bits For abcdef Hijklmnop Qrstuvw Xzy xyz For abcdef Hijklmnop Qrstuvw Xzy xyz Theorem [Shannon 1948] The Shannon entropy H(M) corresponds to the minimum (average) compression length of M.

Compression according to Shannon

Opera@onal relevance Given a physical object, how much informa@on is required to describe it? Given a physical device, what is the maximum amount of informa@on that can be stored reliably?

Why are such ques@ons interes@ng? Technological applica@ons (informa@on processing and transmission) Simulatability of physical systems

Why are such ques@ons interes@ng? (cont d) Development of physical theories Used in other areas of science (biology, finances, linguis@cs, )

Linking Quantum Physics and Information Theory

Informa@on is physical Rolf Landauer: informa@on is always represented by the state of a physical system. If informa@on is represented by a quantum system then it is by defini@on quantum informa@on.

Independence of informa@on carriers According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M Representa1on of a message M Each value M=m is represented by a different physical state of the system.

Independence of informa@on carriers According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M Representa1on of a bit Each value of a bit ( 0 or 1 ) is represented by two different (perfectly dis@nguishable) states of the informa@on carrier.

Independence of informa@on carriers? According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M

Independence of informa@on carriers? According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M But does this paradigm also apply to informa@on stored in quantum devices? [Photos: QSIT]

Classical informa@on Classically, information may always be represented as a sequence of binary numbers (the bits). 0-1

Quantum informa@on Quantum information is represented as the state of a quantum system, such as the polarization degree of freedom of a photon. ψ

Qubit Although the smallest possible unit of quantum informa@on is a that represented on a two- level system (a qubit), there is a con@nuum of states. ψ The state of a qubit is generally represented as a vector in C 2.

Qubits The state of a single system is specified by a 2- dimensional vector ψ C 2 ψ The state of n qubits is specified by a 2 n - dimensional vector ψ C 2n ρ ρ ρ

Comparison: bits vs qubits 36 qubits 2 36 coordinates > 100 GByte Ψ

Independence of informa@on carriers? According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M But does this paradigm also apply to informa@on stored in quantum devices? [Photos: QSIT]

Toy example

Toy example 1. N collabora@ng players sixng in a room

Toy example 1. N collabora@ng players sixng in a room 2. 2 of them selected at random and put in separated rooms

Toy example C 1. N collabora@ng players sixng in a room 2. 2 of them selected at random and put in separated rooms 3. N- 2 remaining players announce a bit C of their choice

Toy example B1 C B2 1. N collabora@ng players sixng in a room 2. 2 of them selected at random and put in separated rooms 3. N- 2 remaining players announce a bit C of their choice 4. separated players output bits B 1 and B 2

Toy example B1 C B2 1. N collabora@ng players sixng in a room 2. 2 of them selected at random and put in separated rooms 3. N- 2 remaining players announce a bit C of their choice 4. separated players output bits B 1 and B 2 Game is won if B 1 B 2.

Maximum winning probability Strategies B=0 B=1 B=C B=1-C

Maximum winning probability Each player may choose one of the following four strategies (in case he is selected). Strategies B=0 B=1 B=C B=1-C (The strategy defines how the output B is derived from the input C.)

Maximum winning probability Each player may choose one of the following four strategies (in case he is selected). Strategies B=0 B=1 B=C B=1-C (The strategy defines how the output B is derived from the input C.) The game cannot be won if the two selected players follow iden@cal strategies.

Maximum winning probability Each player may choose one of the following four strategies (in case he is selected). Strategies B=0 B=1 B=C B=1-C (The strategy defines how the output B is derived from the input C.) The game cannot be won if the two selected players follow iden@cal strategies. This happens with probability 1/4 (for N large).

Maximum winning probability Each player may choose one of the following four strategies (in case he is selected). Strategies B=0 B=1 B=C B=1-C (The strategy defines how the output B is derived from the input C.) The game cannot be won if the two selected players follow iden@cal strategies. This happens with probability 1/4 (for N large). Hence, the game is lost with probability (at least) 1/4.

What did we prove? Claim For any possible strategy, the game is lost with probability at least 1/4.

What did we prove? Claim For any possible strategy, the game is lost with probability at least 1/4. Addi1onal implicit assump1on All informa@on is encoded and processed classically.

Quantum strategies are stronger The game can be won with probability 1 if the players can use an internal quantum device. Note: all communica@on during the game is s@ll purely classical.

Quantum strategies are stronger B1 C B2 The game can be won with probability 1 if the players can use an internal quantum device. Note: all communica@on during the game is s@ll purely classical.

Quantum strategy

Quantum strategy 1. N players start with correlated state Ψ = 0 N + 1 N

Quantum strategy 1. N players start with correlated state Ψ = 0 N + 1 N 2. keep state stored

Quantum strategy C 1. N players start with correlated state Ψ = 0 N + 1 N 2. keep state stored 3. all remaining players measure in diagonal basis and choose C as the xor of their measurement results

Quantum strategy B1 C B2 1. N players start with correlated state Ψ = 0 N + 1 N 2. keep state stored 3. all remaining players measure in diagonal basis and choose C as the xor of their measurement results 4. separated players determine B1 and B2 by measuring in either the diagonal or the circular basis, depending on C.

What can we learn from this example?

What can we learn from this example? Quantum mechanics allows us to win games that cannot be won in a classical world (examples known as pseudo telepathy games ). (Telepathy is obviously dangerous from a cryptographic point of view. )

What can we learn from this example? Quantum mechanics allows us to win games that cannot be won in a classical world (examples known as pseudo telepathy games ). (Telepathy is obviously dangerous from a cryptographic point of view. ) There is no physical principle that allows us to rule out quantum strategies.

What can we learn from this example? Quantum mechanics allows us to win games that cannot be won in a classical world (examples known as pseudo telepathy games ). (Telepathy is obviously dangerous from a cryptographic point of view. ) There is no physical principle that allows us to rule out quantum strategies. It is, in general, unavoidable to take into account quantum effects.

Independence of informa@on carriers? According to Shannon s theory, informa@on is independent of the physical informa@on carriers. M M M But does this paradigm also apply to informa@on stored in quantum devices? No! [Photos: QSIT]

Shannon s impossibility result C C = enc(m,s) M = dec(c,s) Theorem For informa@on- theore@cally secure encryp@on, the key S needs to be at least as long as the message M. In par@cular, One- Time- Pad encryp(on is op@mal.

Proof of Shannon s theorem Let M be a uniformly distributed n- bit message, S a secret key, and C the ciphertext. Requirements H(M SC) = 0, since M is determined by S, C. H(M C) = H(M) = n, since M is indep. of C. Hence H(S) I(M : S C) = H(M C) - H(M SC) = n.

Shannon s impossibility result Theorem [Shannon, 1949] Two par@es connected via an insecure channel cannot exchange any messages secretly (even if they have methods for authen@ca@on). A M B E

BenneK and Brassard s possibility result If informa@on cannot be cloned, then it can also not be stolen (without leaving traces). C.H. BenneK [Photo: ETH Zurich] G. Brassard [Photo: ETH Zurich]

BenneK and Brassard s possibility result If informa@on cannot be cloned, then it can also not be stolen (without leaving traces). C.H. BenneK [Photo: ETH Zurich] G. Brassard [Photo: ETH Zurich] This was the inven@on of quantum cryptography.

Quantum cryptography M ρ ρ M A B E Idea: Use no- cloning principle to verify secrecy.

One- @me- pad encryp@on Let M {0,1} be a message bit and S {0,1} a key bit. S C = M + S (mod 2) S M A C B M E +S (mod 2) +S (mod 2) M C M

One- @me- pad encryp@on Let M {0,1} be a message bit and S {0,1} a key bit. S C = M + S (mod 2) S M A C E B M Theorem If S is uniformly distributed then C is uncorrelated to M.

No- cloning principle provides security B A X or Z C H(X B) + H(Z C) 1 Idea Check sta@s@cally that H(X B) is small. The generalized uncertainty principle then implies that H(Z C) is large. [Tomamichel et al., Nature Communica@ons, 2012]

Quantum cryptography A M B E Protocol 1. Use quantum communica@on to generate a key (the no- cloning principle guarantees that it is secure) 2. Use one- @me- pad encryp@on to send message M.

An apparent contradic@on Theorem [BenneK and Brassard, 1984] Two par@es connected via an insecure channel can exchange messages secretly (provided they have a method for authen@ca@on). A M B E

An apparent contradic@on Theorem [BenneK and Brassard, 1984] Two par@es connected via an insecure channel can exchange messages secretly (provided they have a method for authen@ca@on). M Theorem [Shannon, 1949] A Two par@es connected via an insecure channel cannot exchange any messages secretly (even if they have methods for authen@ca@on). E B

Proof of Shannon s theorem Let M be a uniformly distributed n- bit message, S a secret key, and C the ciphertext.. Requirements H(M SC) = 0, since M determined by S, C. H(M C) = H(M) = n, since M indep. of C. Hence H(S) I(M : S C) = H(M C) - H(M SC) = n.

Proof of Shannon s theorem Let M be a uniformly distributed n- bit message, S a secret key, and C the ciphertext.. Requirements H(M SC Bob ) H(M SC) = 0, since M determined by S, C. H(M C) = H(M) = n, since M indep. of C. Hence H(S) I(M : S C) = H(M C) - H(M SC) = n.

Proof of Shannon s theorem Let M be a uniformly distributed n- bit message, S a secret key, and C the ciphertext.. Requirements H(M SC Bob ) H(M SC) = 0, since M determined by S, C. H(M C) = H(M) = n, since M indep. of C. Hence H(M C Eve ) H(S) I(M : S C) = H(M C) - H(M SC) = n.

Proof of Shannon s theorem No cloning: Let M be a uniformly distributed n- bit C Bob message, C Eve in general S a secret key, and C the ciphertext.. Requirements H(M SC Bob ) H(M SC) = 0, since M determined by S, C. H(M C) = H(M) = n, since M indep. of C. Hence H(M C Eve ) H(S) I(M : S C) = H(M C) - H(M SC) = n.

Proper@es of entangled qubits α β ρ ρ X Y Pr[X = Y] = Cos 2 (α- β) Pr[X Y] = Sin 2 (α- β) (α- β) 2 (for small angle differences)

Proper@es of entangled qubits α β ρ ρ X Y Pr[X = Y] = Cos 2 (α- β) Pr[X Y] = Sin 2 (α- β) (α- β) 2 (for small angle differences) Note: If the le par@cle is measured with angle α and gives output 0 (or 1) then the right par@cle behaves as if it was prepared along α (or α+π/2).

Conclusions

Conclusions Quantum vs. classical physical objects: The joint state space of object A and object B is not simply the cartesian product of the two individual state spaces.

Conclusions Quantum vs. classical physical objects: The joint state space of object A and object B is not simply the cartesian product of the two individual state spaces. Information is physical: Since information is physical, the physical properties of the underlying information carriers have to be taken into account when describing the laws of information.

Conclusions Quantum vs. classical physical objects: The joint state space of object A and object B is not simply the cartesian product of the two individual state spaces. Information is physical: Since information is physical, the physical properties of the underlying information carriers have to be taken into account when describing the laws of information. Implications: The resulting laws of information are fundamentally different from the corresponding classical laws. Examples include the no-cloning principle, which has applications, e.g., in cryptography.

Many thanks for your attention