Introduction to Information Theory. B. Škorić, Physical Aspects of Digital Security, Chapter 2

Similar documents
Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Communication Theory and Engineering

Entropies & Information Theory

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Lecture 2: August 31

Information Theory. David Rosenberg. June 15, New York University. David Rosenberg (New York University) DS-GA 1003 June 15, / 18

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Information in Biology

Topics. Probability Theory. Perfect Secrecy. Information Theory

Principles of Communications

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

Information in Biology

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

5 Mutual Information and Channel Capacity

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013

Lecture 1: Introduction, Entropy and ML estimation

Complex Systems Methods 2. Conditional mutual information, entropy rate and algorithmic complexity

CS 630 Basic Probability and Information Theory. Tim Campbell

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

Lecture 5 - Information theory

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

Lecture 22: Final Review

Solutions to Set #2 Data Compression, Huffman code and AEP

Classification & Information Theory Lecture #8

Noisy channel communication

Information-theoretic Secrecy A Cryptographic Perspective

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Computing and Communications 2. Information Theory -Entropy

Example: Letter Frequencies

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases

Exercise 1. = P(y a 1)P(a 1 )

Chaos, Complexity, and Inference (36-462)

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information


Solutions to Homework Set #1 Sanov s Theorem, Rate distortion

William Stallings Copyright 2010

Homework Set #2 Data Compression, Huffman code and AEP

RSA RSA public key cryptosystem

ECE 4400:693 - Information Theory

Lecture 1: Perfect Secrecy and Statistical Authentication. 2 Introduction - Historical vs Modern Cryptography

Solutions to Homework Set #3 Channel and Source coding

3F1 Information Theory, Lecture 1

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

Quantitative Biology Lecture 3

Quantum Information Theory and Cryptography

The Method of Types and Its Application to Information Hiding

Quantitative Approaches to Information Protection

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Information Theory - Entropy. Figure 3

U.C. Berkeley CS276: Cryptography Luca Trevisan February 5, Notes for Lecture 6

Shannon s noisy-channel theorem

Lecture 3: Channel Capacity

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

Stream ciphers I. Thomas Johansson. May 16, Dept. of EIT, Lund University, P.O. Box 118, Lund, Sweden

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

SIGNATURE SCHEMES & CRYPTOGRAPHIC HASH FUNCTIONS. CIS 400/628 Spring 2005 Introduction to Cryptography

Example: Letter Frequencies

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

Gambling and Information Theory

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

3F1 Information Theory, Lecture 3

Example: Letter Frequencies

9. Distance measures. 9.1 Classical information measures. Head Tail. How similar/close are two probability distributions? Trace distance.

Lecture 5: Asymptotic Equipartition Property

Lecture 11: Quantum Information III - Source Coding

Quantum Entanglement and Cryptography. Deepthi Gopal, Caltech

ECE Information theory Final

CS 591, Lecture 2 Data Analytics: Theory and Applications Boston University

Revision of Lecture 5

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information

Information Theory Primer:

COMPSCI 650 Applied Information Theory Jan 21, Lecture 2

Information measures in simple coding problems

1 Recommended Reading 1. 2 Public Key/Private Key Cryptography Overview RSA Algorithm... 2

5199/IOC5063 Theory of Cryptology, 2014 Fall

Lecture 2: Perfect Secrecy and its Limitations

PERFECTLY secure key agreement has been studied recently

Lecture 18: Quantum Information Theory and Holevo s Bound

(Classical) Information Theory III: Noisy channel coding

Lecture 4 Noisy Channel Coding

Part I. Entropy. Information Theory and Networks. Section 1. Entropy: definitions. Lecture 5: Entropy

ELEC546 Review of Information Theory

Lecture 15: Conditional and Joint Typicaility

Problem Set: TT Quantum Information

Lecture 2. Capacity of the Gaussian channel

Note that the new channel is noisier than the original two : and H(A I +A2-2A1A2) > H(A2) (why?). min(c,, C2 ) = min(1 - H(a t ), 1 - H(A 2 )).

Information Theory: Entropy, Markov Chains, and Huffman Coding

Computer Science A Cryptography and Data Security. Claude Crépeau

LECTURE 3. Last time:

Concentration of Measure Effects in Quantum Information. Patrick Hayden (McGill University)

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Block 2: Introduction to Information Theory

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Basics in Cryptology. Outline. II Distributed Cryptography. Key Management. Outline. David Pointcheval. ENS Paris 2018

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

Error Reconciliation in QKD. Distribution

Transcription:

Introduction to Information Theory B. Škorić, Physical Aspects of Digital Security, Chapter 2 1

Information theory What is it? - formal way of counting information bits Why do we need it? - often used in crypto - best way to talk about physical sources of randomness 2

Measuring ignorance Experiment X with unpredictable outcome known possible outcomes xi known probabilities pi. We want to measure the unpredictability of the experiment should increase with #outcomes should be maximal for uniform pi 3

Measuring ignorance plaintext Encrypt ciphertext Decrypt plaintext Eve How much does Eve know about the plaintext given that she has seen the ciphertext? 4

Measuring common knowledge noise data data How much of the info that Alice sends does actually reach Bob? ( channel capacity ) 5

Data compression Yawn Bla bla blather bla dinner bla bla bla bla blep bla bla bla bla bla bla bla tonight bla blabla bla blather bla bla blather bla bla blab bla dinner tonight How far can a body of data be compressed? 6

plaatje Claude Shannon (1916-2001) Groundbreaking work in 1948 7

Shannon entropy Notation: H(X) or H( {p1,p2,...} ) Formal requirements for counting information bits Sub-additivity: H(X,Y) H(X) + H(Y) - equality only when X and Y independent Expansibility: extra outcome xj with pj=0 has no effect Normalization: - entropy of (1/2, 1/2) is 1 bit; of (1,0,0,...) is zero. Unique expression satisfying all requirements: H(X) = Σx px log2(px) 8

Examples of Shannon entropy H(X) = Σx px log(px) Uniform distribution X (1/n,..., 1/n). H(X) = log n. Binary entropy function. X (p, 1-p). h(p) = h(1-p) = p log p (1-p) log(1-p) 9

Mini quiz H(X) = Σx px log(px) X (1/2, 1/4, 1/4). Compute H(X). Which yes/no questions would you ask to quickly determine the outcome of the experiment? How many questions do you need on average? 10

Interpretation of Shannon entropy 1) Average number of binary questions needed to decide on outcome of experiment (lower bound) 2) Lower bound on compression size 3) Theory of types : Given pmf S = (s1,...,sq); N independent experiments; Type T sequences with exactly ni = N si occurrences of i. t T. Pr[X = t] = q i=1 s n i i = q i=1 s Ns i i =2 P i Ns i log s i =2 NH(S). 11

Relative entropy Measure of distance between distributions - asymmetric D(P Q) = - non-negative x - zero only when distr. are identical Interpretation - when you think distr. is Q, but actually it is P, #questions H(P) + D(P Q). X P(x) log P(x) Q(x) 12

Time for an exercise Prove that relative entropy is non-negative, D(P Q) 0. Prove sub-additive property, H(X,Y) H(X) + H(Y) 13

Conditional entropy Joint distribution (X,Y) P. Conditional probability: py x = pxy / px. Entropy of Y for given X=x: H(Y X=x) = Σy py x log(py x) Conditional entropy of Y given X: H(Y X) = Ex[H(Y X=x)] = -Σxy pxy log(py x) 14

Conditional entropy Watch out: Y X is not a RV, but a set of RVs. Y X=x is a RV Chain rule: pxy = px py x = py px y. H(X,Y) = H(X) + H(Y X) = H(Y) + H(X Y). 15

Some properties Conditioning reduces entropy: H(X Y) H(X) Proof: chain rule H(X Y) = H(X,Y) H(Y) sub-additive property H(X,Y) H(X) + H(Y) 16

Mutual information H(X) H(Y) H(X Y) I(X;Y) H(Y X) Mutual information I(X;Y) - overlap between X and Y - non-negative - binary questions in common Total area is H(X,Y) I(X;X) = H(X) 17

Mutual information H(X) H(Y) H(X Y) I(X;Y) H(Y X) I(X;Y) = H(X) H(X Y) = H(Y) H(Y X) = H(X,Y) H(X Y) H(Y X) = H(X) + H(Y) H(X,Y) = D( XY X Y) I(X;Y) = Σxy pxy log(pxy / px py) 18

Uses of mutual information noise N data X Channel capacity I(X;Y) Y=X+N Noisy broadcast X Z Y Secret key generation capacity I(X;Y Z) 19

Uses of mutual information / cond. entropies Communication theory (error correction,...) Crypto Biometrics Statistics (classification, hypothesis testing,...) Image processing Statistical physics Econometrics... 20

Other information measures Min-entropy Hmin(X) = log(pmax) also called guessing entropy crypto: pmax is Prob[correct guess in one go] Rényi entropy α>1 collision entropy H (X) = 1 1 log x X p x. we will encounter α=2 lim α H α (X) = Hmin(X) lim α 1 H α (X) = H(X) 21

... 22