Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151)

Size: px
Start display at page:

Download "Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151)"

Transcription

1 Kotebe Metropolitan University Department of Computer Science and Technology Multimedia (CoSc 4151) Chapter Three Multimedia Data Compression Part I: Entropy in ordinary words Claude Shannon developed the core elements of information theory in the 1940s. There are many aspects to this theory and much has been done beyond Shannon (see Chaitin and Kolmogorov). However, a key question I want to pose is this: Suppose you are given a sequence of symbols (say bits (on/off or heads/tails), or letters of the alphabet). How much information is in the sequence? Shannon showed that information is, in a useful sense, a measure of entropy. More entropy, more information. This may seem very counterintuitive, but think about it like this: Say you've seen a long chain of symbols in the sequence so far. Based on all you've seen so far, you could now compute the probability of each possible next symbol. Suppose one symbol was much more probable than the others. Then you would be less surprised to see it next, than the others. Information is amount of data that must be communicated in order to bridge between what you would expect next and what is actually coming next. If the next symbol will unsurprising to you based on seeing the previous symbols, then it can be communicated with less data than some symbol that would more surprising. You can measure the average surprise over all the different symbols and use it to determine how much information per symbol actually exists. This information is measured in *bits* (ones/zeros) Think of a flipping a coin again and again and sending the outcomes to your friend over the telephone. If it's a fair coin, the information content of both "head" and "tail" is one bit. You need to tell your friend, on every flip, 1 bit of information: head or tail. Suppose it's a coin that comes up 80% of the time as heads. Then, when you flip it an comes up heads, you can send *less than* 1 bit of information (keep in mind, this is averaged over many flips) for a head and more than 1 bit of information for a tail. On average, though, over both heads and tails, you could send less than one 1 bit of information per flip. Suppose it's a double-headed coin. Then, at the beginning, you'd just tell your friend this, and your average bits per flip would be zero. Suppose the coin followed a pattern, say HHTTHHTTHHTT...: then you could tell your friend a fixed amount of information: "HHTT [repeat]", which, amortized over many flips would approach zero bits per flip. Very little information. Shannon's classic experiment on computing the information content (entropy) of letters of the alphabet within an English sentence is a fun thing to play with: With 27 characters (A-Z + space), English text can at most have about 4.8 bits per character of information (see if you'd like to compute for yourself). But the various measures (including the above game), show that actual English text has closer to 1 bit per character of information (and probably closer to 0.5 bits/ character). What about Amharic and other Ethiopian languages? So, what does this all have to do with computer science? Well, one easy connection is compression. When you compress a file, you are re- encoding its sequence of symbols such that the file size gets as close as possible to its information content/entropy. That re-encoded file will have much more "surprise" than the original. If you measure the information content of its symbols, it will be much closer to the Shannon limit. If you look at a compressed file on your computer (jpg, mpg, zip, etc), you'll 1

2 find that its bits are very close to the random sequence you might expect from flipping a fair coin. You should now understand, btw, why you should never believe someone who claims that you can shrink a file further by compressing it a second time. Part II: Information and Entropy Information Equation I(p) = log b p Where: p = probability of the event happening b = base (base 2 is mostly used in information theory) *unit of information is determined by base base 2 = bits; base 3 = trits; base 10 = Hartleys; base e = nats; Example of Calculating Information Coin Toss There are two probabilities in fair coin, which are head(.5) and tail(.5). So if you get either head or tail you will get 1 bit of information through following formula. I(head) = - log (.5) = 1 bit Another Example Balls in the bin The information you will get by choosing a ball from the bin are calculated as following. I (red ball) = - log(4/9) = bits I (yellow ball) = - log(2/9) = bits I (green ball) = - log(3/9) = bits Then, what is Entropy? - Entropy is simply the average (expected) amount of the information from the event. Entropy Equation n Entropy = i p i log b p i Where n = number of different outcomes How the entropy equation was derived n I = i (N p i ) log b p i Where: I = total information from N occurrences N = number of occurrences (N*Pi) = Approximated number that the certain result will come out in N occurrence 2

3 n Entropy = i p i log b p i So when you look at the difference between the total Information from N occurrences and the Entropy equation, only thing that changed in the place of N. The N is moved to the right, which means that I/N is Entropy. Therefore, Entropy is the average (expected) amount of information in a certain event. Let s look at the previous example again Calculating the entropy In this example there are three outcomes possible when you choose the ball, it can be either red, yellow, or green. (n = 3) So the equation will be following. n Entropy = i p i log b p i Entropy = - (4/9) log(4/9) + -(2/9) log(2/9) + - (3/9) log(3/9) = Therefore, you are expected to get information each time you choose a ball from the bin Does Entropy have range from 0 to 1? No. However, the range is set based on the number of outcomes. Equation for calculating the range of Entropy: 0 Entropy log(n), where n is number of outcomes Entropy 0(minimum entropy) occurs when one of the probabilities is 1 and rest are 0 s Entropy log(n) (maximum entropy) occurs when all the probabilities have equal values of 1/n. What is the difference between Information and knowledge? Information different from knowledge Concerned with abstract possibilities, not their meaning Information = reduction in uncertainty Imagine: #1 you re about to observe the outcome of a coin flip #2 you re about to observe the outcome of a die roll There is more uncertainty in #2 Next: 1. You observed outcome of #1 uncertainty reduced to zero. 2. You observed outcome of #2 uncertainty reduced to zero. = more information was provided by the outcome in #2 3

4 Information is Additive 1 I (k fair coin tosses) = 1 2 k = k bits So: Random word from a1 00,000 word vocabulary: o I(word) = log100,000 = 16.61bits o A 1000 word document from same source I(document) = 16,610bits o A 480x640 pixel, 16-grayscale video picture: I(picture) = 307,200 * log16 = 1,228,800 bits A (VGA) picture is worth (a lot more than) a 1000 words! (In reality, both are gross over estimates.) Part III: Alternative Explanations of Entropy (THIS IS VERY IMPORTANT) n Entropy = i p i log b p i 1. Average amount of information provided per symbol 2. Average amount surprise when observing a symbol 3. Uncertainty an observer has before seeing the symbol 4. Average number of bits needed to communicate each symbol Shannon: there are codes that will communicate these symbols with efficiency arbitrarily close to Entropy bits/symbol; there are no codes that will do it with efficiency less than Entropy bits/symbol NOTE: RELATE THE ABOVE INTERPRETATIONS WITH SOLUTIONS OF PREVIOS PROBLEMS Special Case: k = 2 Flipping a coin with P ( head ) = p, P ( tail ) = 1 - p Entropy = p log 2 1 p + (p 1) log 2 1 (p 1) Notice: Zero uncertainty/information/surprise at edges. Maximum info at 0.5 (1bit). Drops off quickly. So a sequence of (independent) 0 s-and-1 s can provide up to 1 bit of information per digit, provided the 0 s and 1 s are equally likely at any point. If they are not equally likely the sequence provides less information and can be compressed. Entropy doesn t tell you the amount of information you get on the happening of a particular event (one occurrence of head in coin toss), rather it tells you the average amount of information you get (per event head in coin toss) out of repeated occurrences of that event such as happening of head out of repeated occurrences of coin toss experiment. At this point, be sure you understand the concept and examples of entropy, it is very useful in information science. The concept of Entropy is very wide and crosses several disciplines, it is not limited to Information Science. For those from other discipline (or interested in being interdisciplinary), you can entertain by connecting the concept of entropy to other discipline/s you want. If you want to read further, please visit this link (especially read pages 15 to 34) 4

5 Modeling and Compression We will look at models whose purpose is primarily compression of multimedia data. We are interested in modeling multimedia data. To model means to replace something complex with a simpler (= shorter) analog. Some models help understand the original phenomenon/data better: Example: Laws of physics Huge arrays of astronomical observations (e.g. Tycho Brahe s logbooks) summarised in a few characters (e.g. Kepler, Newton): F = G M 1M 2 r 2. This model helps us understand gravity better. Is an example of tremendous compression of data.

6 Recap: The Need for Compression Raw video, image, and audio files can be very large. Example: One minute of uncompressed audio. Audio Type 44.1 KHz KHz KHz 16 Bit Stereo: 10.1 MB 5.05 MB 2.52 MB 16 Bit Mono: 5.05 MB 2.52 MB 1.26 MB 8 Bit Mono: 2.52 MB 1.26 MB 630 KB Example: Uncompressed images. Image Type File Size 512 x 512 Monochrome 0.25 MB 512 x bit colour image 0.25 MB 512 x bit colour image 0.75 MB

7 Recap: The Need for Compression Example: Videos (involves a stream of audio plus video imagery). Raw Video uncompressed image frames 512x512 True Colour at 25 FPS = 1125 MB/min. HDTV ( ) Gigabytes per minute uncompressed, True Colour at 25 FPS = 8.7 GB/min. Relying on higher bandwidths is not a good option M25 Syndrome: traffic will always increase to fill the current bandwidth limit whatever this is. Compression HAS TO BE part of the representation of audio, image, and video formats.

8 Basics of Information Theory Suppose we have an information source (random variable) S which emits symbols {s 1, s 2,..., s n } with probabilities p 1, p 2,..., p n. According to Shannon, the entropy of S is defined as: H(S) = 1 p i log 2, p i i where p i is the probability that symbol s i will occur. When a symbol with probability p i is transmitted, it reduces the amount of uncertainty in the receiver by a factor of 1 pi. log 2 1 pi = log 2 p i indicates the amount of information conveyed by s i, i.e., the number of bits needed to code s i (Shannon s coding theorem).

9 Entropy Example Example: Entropy of a fair coin. The coin emits symbols s 1 = heads and s 2 = tails with p 1 = p 2 = 1/2. Therefore, the entropy if this source is: H(coin) = (1/2 log 2 1/2 + 1/2 log 2 1/2) = (1/ /2 1) = ( 1/2 1/2) = 1 bit. Example: Grayscale image In an image with uniform distribution of gray-level intensity (and all pixels independent), i.e. p i = 1/256, then The # of bits needed to code each gray level is 8 bits. The entropy of this image is 8.

10 Entropy Example Example: Breakfast order #1. Alice: What do you want for breakfast: pancakes or eggs? I am unsure, because you like them equally (p 1 = p 2 = 1/2)... Bob: I want pancakes. Question: How much information has Bob communicated to Alice?

11 Entropy Example Example: Breakfast order #1. Alice: What do you want for breakfast: pancakes or eggs? I am unsure, because you like them equally (p 1 = p 2 = 1/2)... Bob: I want pancakes. Question: How much information has Bob communicated to Alice? Answer: He has reduced the uncertainty by a factor of 2, therefore 1 bit.

12 Entropy Example Example: Breakfast order #2. Alice: What do you want for breakfast: pancakes, eggs, or salad? I am unsure, because you like them equally (p 1 = p 2 = p 3 = 1/3)... Bob: Eggs. Question: What is Bob s entropy assuming he behaves like a random variable = how much information has Bob communicated to Alice?

13 Entropy Example Example: Breakfast order #2. Alice: What do you want for breakfast: pancakes, eggs, or salad? I am unsure, because you like them equally (p 1 = p 2 = p 3 = 1/3)... Bob: Eggs. Question: What is Bob s entropy assuming he behaves like a random variable = how much information has Bob communicated to Alice? Answer: H(Bob) = 3 i=1 1 3 log 2 3 = log bits.

14 Entropy Example Example: Breakfast order #3. Alice: What do you want for breakfast: pancakes, eggs, or salad? I am unsure, because you like them equally (p 1 = p 2 = p 3 = 1/3)... Bob: Dunno. I definitely do not want salad. Question: How much information has Bob communicated to Alice?

15 Entropy Example Example: Breakfast order #3. Alice: What do you want for breakfast: pancakes, eggs, or salad? I am unsure, because you like them equally (p 1 = p 2 = p 3 = 1/3)... Bob: Dunno. I definitely do not want salad. Question: How much information has Bob communicated to Alice? Answer: He has reduced her uncertainty by a factor of 3/2 (leaving 2 out of 3 equal options), therefore transmitted log 2 3/ bits.

16 Shannon s Experiment (1951) Estimated entropy for English text: H English bits/letter. (If all letters and space were equally probable, then it would be H 0 = log bits/letter.) External link: Shannon s original 1951 paper. External link: Java applet recreating Shannon s experiment.

17 Shannon s coding theorem Shannon 1948 Basically: The optimal code length for an event with probability p is L(p) = log 2 p ones and zeros (or generally, log b p if instead we use b possible values for codes). External link: Shannon s original 1948 paper.

18 Shannon vs Kolmogorov What if we have a finite string? Shannon s entropy is a statistical measure of information. We can cheat and regard a string as infinitely long sequence of i.i.d. random variables. Shannon s theorem then approximately applies. Kolmogorov Complexity: Basically, the length of the shortest program that ouputs a given string. Algorithmical measure of information. K(S) is not computable! Practical algorithmic compression is hard.

19 Compression in Multimedia Data Compression basically employs redundancy in the data: Temporal in 1D data, 1D signals, audio, between video frames etc. Spatial correlation between neighbouring pixels or data items. Spectral e.g. correlation between colour or luminescence components. This uses the frequency domain to exploit relationships between frequency of change in data. Psycho-visual exploit perceptual properties of the human visual system.

20 Lossless vs Lossy Compression Compression can be categorised in two broad ways: Lossless Compression: after decompression gives an exact copy of the original data. Example: Entropy encoding schemes (Shannon-Fano, Huffman coding), arithmetic coding, LZW algorithm (used in GIF image file format). Lossy Compression: after decompression gives ideally a close approximation of the original data, ideally perceptually lossless. Example: Transform coding FFT/DCT based quantisation used in JPEG/MPEG differential encoding, vector quantisation.

21 Why Lossy Compression? Lossy methods are typically applied to high resoultion audio, image compression. Have to be employed in video compression (apart from special cases). Basic reason: Compression ratio of lossless methods (e.g. Huffman coding, arithmetic coding, LZW) is not high enough for audio/video. By cleverly making a small sacrifice in terms of fidelity of data, we can often achieve very high compression ratios. Cleverly = sacrifice information that is psycho-physically unimportant.

22 Lossless Compression Algorithms Repetitive Sequence Suppression. Run-Length Encoding (RLE). Pattern Substitution. Entropy Encoding: Shannon-Fano Algorithm. Huffman Coding. Arithmetic Coding. Lempel-Ziv-Welch (LZW) Algorithm.

Lecture 10 : Basic Compression Algorithms

Lecture 10 : Basic Compression Algorithms Lecture 10 : Basic Compression Algorithms Modeling and Compression We are interested in modeling multimedia data. To model means to replace something complex with a simpler (= shorter) analog. Some models

More information

Information. = more information was provided by the outcome in #2

Information. = more information was provided by the outcome in #2 Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information channels and coding will not discuss those here.. Information 2. Entropy 3. Mutual

More information

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University A Gentle Tutorial on Information Theory and Learning Roni Rosenfeld Mellon University Mellon Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information

More information

Multimedia. Multimedia Data Compression (Lossless Compression Algorithms)

Multimedia. Multimedia Data Compression (Lossless Compression Algorithms) Course Code 005636 (Fall 2017) Multimedia Multimedia Data Compression (Lossless Compression Algorithms) Prof. S. M. Riazul Islam, Dept. of Computer Engineering, Sejong University, Korea E-mail: riaz@sejong.ac.kr

More information

Image and Multidimensional Signal Processing

Image and Multidimensional Signal Processing Image and Multidimensional Signal Processing Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ Image Compression 2 Image Compression Goal: Reduce amount

More information

Lecture 1: Shannon s Theorem

Lecture 1: Shannon s Theorem Lecture 1: Shannon s Theorem Lecturer: Travis Gagie January 13th, 2015 Welcome to Data Compression! I m Travis and I ll be your instructor this week. If you haven t registered yet, don t worry, we ll work

More information

CMPT 365 Multimedia Systems. Lossless Compression

CMPT 365 Multimedia Systems. Lossless Compression CMPT 365 Multimedia Systems Lossless Compression Spring 2017 Edited from slides by Dr. Jiangchuan Liu CMPT365 Multimedia Systems 1 Outline Why compression? Entropy Variable Length Coding Shannon-Fano Coding

More information

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Multimedia Communications. Mathematical Preliminaries for Lossless Compression Multimedia Communications Mathematical Preliminaries for Lossless Compression What we will see in this chapter Definition of information and entropy Modeling a data source Definition of coding and when

More information

Information and Entropy. Professor Kevin Gold

Information and Entropy. Professor Kevin Gold Information and Entropy Professor Kevin Gold What s Information? Informally, when I communicate a message to you, that s information. Your grade is 100/100 Information can be encoded as a signal. Words

More information

Chapter 2: Source coding

Chapter 2: Source coding Chapter 2: meghdadi@ensil.unilim.fr University of Limoges Chapter 2: Entropy of Markov Source Chapter 2: Entropy of Markov Source Markov model for information sources Given the present, the future is independent

More information

Source Coding Techniques

Source Coding Techniques Source Coding Techniques. Huffman Code. 2. Two-pass Huffman Code. 3. Lemple-Ziv Code. 4. Fano code. 5. Shannon Code. 6. Arithmetic Code. Source Coding Techniques. Huffman Code. 2. Two-path Huffman Code.

More information

ECE472/572 - Lecture 11. Roadmap. Roadmap. Image Compression Fundamentals and Lossless Compression Techniques 11/03/11.

ECE472/572 - Lecture 11. Roadmap. Roadmap. Image Compression Fundamentals and Lossless Compression Techniques 11/03/11. ECE47/57 - Lecture Image Compression Fundamentals and Lossless Compression Techniques /03/ Roadmap Preprocessing low level Image Enhancement Image Restoration Image Segmentation Image Acquisition Image

More information

repetition, part ii Ole-Johan Skrede INF Digital Image Processing

repetition, part ii Ole-Johan Skrede INF Digital Image Processing repetition, part ii Ole-Johan Skrede 24.05.2017 INF2310 - Digital Image Processing Department of Informatics The Faculty of Mathematics and Natural Sciences University of Oslo today s lecture Coding and

More information

An introduction to basic information theory. Hampus Wessman

An introduction to basic information theory. Hampus Wessman An introduction to basic information theory Hampus Wessman Abstract We give a short and simple introduction to basic information theory, by stripping away all the non-essentials. Theoretical bounds on

More information

6.02 Fall 2012 Lecture #1

6.02 Fall 2012 Lecture #1 6.02 Fall 2012 Lecture #1 Digital vs. analog communication The birth of modern digital communication Information and entropy Codes, Huffman coding 6.02 Fall 2012 Lecture 1, Slide #1 6.02 Fall 2012 Lecture

More information

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H. Problem sheet Ex. Verify that the function H(p,..., p n ) = k p k log p k satisfies all 8 axioms on H. Ex. (Not to be handed in). looking at the notes). List as many of the 8 axioms as you can, (without

More information

4. Quantization and Data Compression. ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak

4. Quantization and Data Compression. ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak 4. Quantization and Data Compression ECE 32 Spring 22 Purdue University, School of ECE Prof. What is data compression? Reducing the file size without compromising the quality of the data stored in the

More information

1. Basics of Information

1. Basics of Information 1. Basics of Information 6.004x Computation Structures Part 1 Digital Circuits Copyright 2015 MIT EECS 6.004 Computation Structures L1: Basics of Information, Slide #1 What is Information? Information,

More information

UNIT I INFORMATION THEORY. I k log 2

UNIT I INFORMATION THEORY. I k log 2 UNIT I INFORMATION THEORY Claude Shannon 1916-2001 Creator of Information Theory, lays the foundation for implementing logic in digital circuits as part of his Masters Thesis! (1939) and published a paper

More information

Digital communication system. Shannon s separation principle

Digital communication system. Shannon s separation principle Digital communication system Representation of the source signal by a stream of (binary) symbols Adaptation to the properties of the transmission channel information source source coder channel coder modulation

More information

Lecture 7: DecisionTrees

Lecture 7: DecisionTrees Lecture 7: DecisionTrees What are decision trees? Brief interlude on information theory Decision tree construction Overfitting avoidance Regression trees COMP-652, Lecture 7 - September 28, 2009 1 Recall:

More information

Text Compression. Jayadev Misra The University of Texas at Austin December 5, A Very Incomplete Introduction to Information Theory 2

Text Compression. Jayadev Misra The University of Texas at Austin December 5, A Very Incomplete Introduction to Information Theory 2 Text Compression Jayadev Misra The University of Texas at Austin December 5, 2003 Contents 1 Introduction 1 2 A Very Incomplete Introduction to Information Theory 2 3 Huffman Coding 5 3.1 Uniquely Decodable

More information

! Where are we on course map? ! What we did in lab last week. " How it relates to this week. ! Compression. " What is it, examples, classifications

! Where are we on course map? ! What we did in lab last week.  How it relates to this week. ! Compression.  What is it, examples, classifications Lecture #3 Compression! Where are we on course map?! What we did in lab last week " How it relates to this week! Compression " What is it, examples, classifications " Probability based compression # Huffman

More information

Entropy as a measure of surprise

Entropy as a measure of surprise Entropy as a measure of surprise Lecture 5: Sam Roweis September 26, 25 What does information do? It removes uncertainty. Information Conveyed = Uncertainty Removed = Surprise Yielded. How should we quantify

More information

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding Ch 0 Introduction 0.1 Overview of Information Theory and Coding Overview The information theory was founded by Shannon in 1948. This theory is for transmission (communication system) or recording (storage

More information

Compressing Kinetic Data From Sensor Networks. Sorelle A. Friedler (Swat 04) Joint work with David Mount University of Maryland, College Park

Compressing Kinetic Data From Sensor Networks. Sorelle A. Friedler (Swat 04) Joint work with David Mount University of Maryland, College Park Compressing Kinetic Data From Sensor Networks Sorelle A. Friedler (Swat 04) Joint work with David Mount University of Maryland, College Park Motivation Motivation Computer Science Graphics: Image and video

More information

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 203 Vazirani Note 2 Random Variables: Distribution and Expectation We will now return once again to the question of how many heads in a typical sequence

More information

God doesn t play dice. - Albert Einstein

God doesn t play dice. - Albert Einstein ECE 450 Lecture 1 God doesn t play dice. - Albert Einstein As far as the laws of mathematics refer to reality, they are not certain; as far as they are certain, they do not refer to reality. Lecture Overview

More information

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 202 Vazirani Note 4 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected in, randomly

More information

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols. Universal Lossless coding Lempel-Ziv Coding Basic principles of lossless compression Historical review Variable-length-to-block coding Lempel-Ziv coding 1 Basic Principles of Lossless Coding 1. Exploit

More information

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet)

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet) Compression Motivation Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet) Storage: Store large & complex 3D models (e.g. 3D scanner

More information

Entropies & Information Theory

Entropies & Information Theory Entropies & Information Theory LECTURE I Nilanjana Datta University of Cambridge,U.K. See lecture notes on: http://www.qi.damtp.cam.ac.uk/node/223 Quantum Information Theory Born out of Classical Information

More information

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS Lesson 5 Other Coding Techniques Instructional Objectives At the end of this lesson, the students should be able to:. Convert a gray-scale image into bit-plane

More information

Compression. What. Why. Reduce the amount of information (bits) needed to represent image Video: 720 x 480 res, 30 fps, color

Compression. What. Why. Reduce the amount of information (bits) needed to represent image Video: 720 x 480 res, 30 fps, color Compression What Reduce the amount of information (bits) needed to represent image Video: 720 x 480 res, 30 fps, color Why 720x480x20x3 = 31,104,000 bytes/sec 30x60x120 = 216 Gigabytes for a 2 hour movie

More information

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information. L65 Dept. of Linguistics, Indiana University Fall 205 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission rate

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 28 Information theory answers two fundamental questions in communication theory: What is the ultimate data compression? What is the transmission

More information

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of

More information

Multimedia Information Systems

Multimedia Information Systems Multimedia Information Systems Samson Cheung EE 639, Fall 2004 Lecture 3 & 4: Color, Video, and Fundamentals of Data Compression 1 Color Science Light is an electromagnetic wave. Its color is characterized

More information

Compression. Encryption. Decryption. Decompression. Presentation of Information to client site

Compression. Encryption. Decryption. Decompression. Presentation of Information to client site DOCUMENT Anup Basu Audio Image Video Data Graphics Objectives Compression Encryption Network Communications Decryption Decompression Client site Presentation of Information to client site Multimedia -

More information

Reduce the amount of data required to represent a given quantity of information Data vs information R = 1 1 C

Reduce the amount of data required to represent a given quantity of information Data vs information R = 1 1 C Image Compression Background Reduce the amount of data to represent a digital image Storage and transmission Consider the live streaming of a movie at standard definition video A color frame is 720 480

More information

Information & Correlation

Information & Correlation Information & Correlation Jilles Vreeken 11 June 2014 (TADA) Questions of the day What is information? How can we measure correlation? and what do talking drums have to do with this? Bits and Pieces What

More information

CIS 2033 Lecture 5, Fall

CIS 2033 Lecture 5, Fall CIS 2033 Lecture 5, Fall 2016 1 Instructor: David Dobor September 13, 2016 1 Supplemental reading from Dekking s textbook: Chapter2, 3. We mentioned at the beginning of this class that calculus was a prerequisite

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Spring 206 Rao and Walrand Note 6 Random Variables: Distribution and Expectation Example: Coin Flips Recall our setup of a probabilistic experiment as

More information

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria Source Coding Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Asymptotic Equipartition Property Optimal Codes (Huffman Coding) Universal

More information

Multimedia Networking ECE 599

Multimedia Networking ECE 599 Multimedia Networking ECE 599 Prof. Thinh Nguyen School of Electrical Engineering and Computer Science Based on lectures from B. Lee, B. Girod, and A. Mukherjee 1 Outline Digital Signal Representation

More information

CSE 421 Greedy: Huffman Codes

CSE 421 Greedy: Huffman Codes CSE 421 Greedy: Huffman Codes Yin Tat Lee 1 Compression Example 100k file, 6 letter alphabet: File Size: ASCII, 8 bits/char: 800kbits 2 3 > 6; 3 bits/char: 300kbits better: 2.52 bits/char 74%*2 +26%*4:

More information

BASIC COMPRESSION TECHNIQUES

BASIC COMPRESSION TECHNIQUES BASIC COMPRESSION TECHNIQUES N. C. State University CSC557 Multimedia Computing and Networking Fall 2001 Lectures # 05 Questions / Problems / Announcements? 2 Matlab demo of DFT Low-pass windowed-sinc

More information

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019 Lecture 11: Information theory DANIEL WELLER THURSDAY, FEBRUARY 21, 2019 Agenda Information and probability Entropy and coding Mutual information and capacity Both images contain the same fraction of black

More information

Information Theory (Information Theory by J. V. Stone, 2015)

Information Theory (Information Theory by J. V. Stone, 2015) Information Theory (Information Theory by J. V. Stone, 2015) Claude Shannon (1916 2001) Shannon, C. (1948). A mathematical theory of communication. Bell System Technical Journal, 27:379 423. A mathematical

More information

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006) MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK SATELLITE COMMUNICATION DEPT./SEM.:ECE/VIII UNIT V PART-A 1. What is binary symmetric channel (AUC DEC 2006) 2. Define information rate? (AUC DEC 2007)

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK DEPARTMENT: ECE SEMESTER: IV SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A 1. What is binary symmetric channel (AUC DEC

More information

Stream Codes. 6.1 The guessing game

Stream Codes. 6.1 The guessing game About Chapter 6 Before reading Chapter 6, you should have read the previous chapter and worked on most of the exercises in it. We ll also make use of some Bayesian modelling ideas that arrived in the vicinity

More information

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE General e Image Coder Structure Motion Video x(s 1,s 2,t) or x(s 1,s 2 ) Natural Image Sampling A form of data compression; usually lossless, but can be lossy Redundancy Removal Lossless compression: predictive

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2013, 29 November 2013 Memoryless Sources Arithmetic Coding Sources with Memory Markov Example 2 / 21 Encoding the output

More information

Applications of Information Theory in Science and in Engineering

Applications of Information Theory in Science and in Engineering Applications of Information Theory in Science and in Engineering Some Uses of Shannon Entropy (Mostly) Outside of Engineering and Computer Science Mário A. T. Figueiredo 1 Communications: The Big Picture

More information

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar Introduction to Information Theory By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar Introduction [B.P. Lathi] Almost in all the means of communication, none produces error-free communication.

More information

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Tara Javidi These lecture notes were originally developed by late Prof. J. K. Wolf. UC San Diego Spring 2014 1 / 8 I

More information

A Mathematical Theory of Communication

A Mathematical Theory of Communication A Mathematical Theory of Communication Ben Eggers Abstract This paper defines information-theoretic entropy and proves some elementary results about it. Notably, we prove that given a few basic assumptions

More information

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 Please submit the solutions on Gradescope. EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018 1. Optimal codeword lengths. Although the codeword lengths of an optimal variable length code

More information

3F1 Information Theory, Lecture 3

3F1 Information Theory, Lecture 3 3F1 Information Theory, Lecture 3 Jossy Sayir Department of Engineering Michaelmas 2011, 28 November 2011 Memoryless Sources Arithmetic Coding Sources with Memory 2 / 19 Summary of last lecture Prefix-free

More information

to mere bit flips) may affect the transmission.

to mere bit flips) may affect the transmission. 5 VII. QUANTUM INFORMATION THEORY to mere bit flips) may affect the transmission. A. Introduction B. A few bits of classical information theory Information theory has developed over the past five or six

More information

Introduction to Information Theory. Part 3

Introduction to Information Theory. Part 3 Introduction to Information Theory Part 3 Assignment#1 Results List text(s) used, total # letters, computed entropy of text. Compare results. What is the computed average word length of 3 letter codes

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Hidden Markov Models, I. Examples. Steven R. Dunbar. Toy Models. Standard Mathematical Models. Realistic Hidden Markov Models.

Hidden Markov Models, I. Examples. Steven R. Dunbar. Toy Models. Standard Mathematical Models. Realistic Hidden Markov Models. , I. Toy Markov, I. February 17, 2017 1 / 39 Outline, I. Toy Markov 1 Toy 2 3 Markov 2 / 39 , I. Toy Markov A good stack of examples, as large as possible, is indispensable for a thorough understanding

More information

CS 124 Math Review Section January 29, 2018

CS 124 Math Review Section January 29, 2018 CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to

More information

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase

More information

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc. Chapter 14 From Randomness to Probability Copyright 2012, 2008, 2005 Pearson Education, Inc. Dealing with Random Phenomena A random phenomenon is a situation in which we know what outcomes could happen,

More information

CMPT 365 Multimedia Systems. Final Review - 1

CMPT 365 Multimedia Systems. Final Review - 1 CMPT 365 Multimedia Systems Final Review - 1 Spring 2017 CMPT365 Multimedia Systems 1 Outline Entropy Lossless Compression Shannon-Fano Coding Huffman Coding LZW Coding Arithmetic Coding Lossy Compression

More information

Intro to Information Theory

Intro to Information Theory Intro to Information Theory Math Circle February 11, 2018 1. Random variables Let us review discrete random variables and some notation. A random variable X takes value a A with probability P (a) 0. Here

More information

log 2 N I m m log 2 N + 1 m.

log 2 N I m m log 2 N + 1 m. SOPHOMORE COLLEGE MATHEMATICS OF THE INFORMATION AGE SHANNON S THEOREMS Let s recall the fundamental notions of information and entropy. To repeat, Shannon s emphasis is on selecting a given message from

More information

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips 6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips Hey, everyone. Welcome back. Today, we're going to do another fun problem that

More information

CSEP 521 Applied Algorithms Spring Statistical Lossless Data Compression

CSEP 521 Applied Algorithms Spring Statistical Lossless Data Compression CSEP 52 Applied Algorithms Spring 25 Statistical Lossless Data Compression Outline for Tonight Basic Concepts in Data Compression Entropy Prefix codes Huffman Coding Arithmetic Coding Run Length Coding

More information

When working with probabilities we often perform more than one event in a sequence - this is called a compound probability.

When working with probabilities we often perform more than one event in a sequence - this is called a compound probability. + Independence + Compound Events When working with probabilities we often perform more than one event in a sequence - this is called a compound probability. Compound probabilities are more complex than

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

Classification & Information Theory Lecture #8

Classification & Information Theory Lecture #8 Classification & Information Theory Lecture #8 Introduction to Natural Language Processing CMPSCI 585, Fall 2007 University of Massachusetts Amherst Andrew McCallum Today s Main Points Automatically categorizing

More information

Real-Time Audio and Video

Real-Time Audio and Video MM- Multimedia Payloads MM-2 Raw Audio (uncompressed audio) Real-Time Audio and Video Telephony: Speech signal: 2 Hz 3.4 khz! 4 khz PCM (Pulse Coded Modulation)! samples/sec x bits = 64 kbps Teleconferencing:

More information

Chapter 8: An Introduction to Probability and Statistics

Chapter 8: An Introduction to Probability and Statistics Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including

More information

MITOCW watch?v=vjzv6wjttnc

MITOCW watch?v=vjzv6wjttnc MITOCW watch?v=vjzv6wjttnc PROFESSOR: We just saw some random variables come up in the bigger number game. And we're going to be talking now about random variables, just formally what they are and their

More information

Image Compression - JPEG

Image Compression - JPEG Overview of JPEG CpSc 86: Multimedia Systems and Applications Image Compression - JPEG What is JPEG? "Joint Photographic Expert Group". Voted as international standard in 99. Works with colour and greyscale

More information

Compression and Coding

Compression and Coding Compression and Coding Theory and Applications Part 1: Fundamentals Gloria Menegaz 1 Transmitter (Encoder) What is the problem? Receiver (Decoder) Transformation information unit Channel Ordering (significance)

More information

Information and Entropy

Information and Entropy Information and Entropy Shannon s Separation Principle Source Coding Principles Entropy Variable Length Codes Huffman Codes Joint Sources Arithmetic Codes Adaptive Codes Thomas Wiegand: Digital Image Communication

More information

BASICS OF COMPRESSION THEORY

BASICS OF COMPRESSION THEORY BASICS OF COMPRESSION THEORY Why Compression? Task: storage and transport of multimedia information. E.g.: non-interlaced HDTV: 0x0x0x = Mb/s!! Solutions: Develop technologies for higher bandwidth Find

More information

Lecture 16. Error-free variable length schemes (contd.): Shannon-Fano-Elias code, Huffman code

Lecture 16. Error-free variable length schemes (contd.): Shannon-Fano-Elias code, Huffman code Lecture 16 Agenda for the lecture Error-free variable length schemes (contd.): Shannon-Fano-Elias code, Huffman code Variable-length source codes with error 16.1 Error-free coding schemes 16.1.1 The Shannon-Fano-Elias

More information

Shannon's Theory of Communication

Shannon's Theory of Communication Shannon's Theory of Communication An operational introduction 5 September 2014, Introduction to Information Systems Giovanni Sileno g.sileno@uva.nl Leibniz Center for Law University of Amsterdam Fundamental

More information

ECE 587 / STA 563: Lecture 5 Lossless Compression

ECE 587 / STA 563: Lecture 5 Lossless Compression ECE 587 / STA 563: Lecture 5 Lossless Compression Information Theory Duke University, Fall 2017 Author: Galen Reeves Last Modified: October 18, 2017 Outline of lecture: 5.1 Introduction to Lossless Source

More information

Information Theory. Week 4 Compressing streams. Iain Murray,

Information Theory. Week 4 Compressing streams. Iain Murray, Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 4 Compressing streams Iain Murray, 2014 School of Informatics, University of Edinburgh Jensen s inequality For convex functions: E[f(x)]

More information

Implementation of Lossless Huffman Coding: Image compression using K-Means algorithm and comparison vs. Random numbers and Message source

Implementation of Lossless Huffman Coding: Image compression using K-Means algorithm and comparison vs. Random numbers and Message source Implementation of Lossless Huffman Coding: Image compression using K-Means algorithm and comparison vs. Random numbers and Message source Ali Tariq Bhatti 1, Dr. Jung Kim 2 1,2 Department of Electrical

More information

Image Compression. Fundamentals: Coding redundancy. The gray level histogram of an image can reveal a great deal of information about the image

Image Compression. Fundamentals: Coding redundancy. The gray level histogram of an image can reveal a great deal of information about the image Fundamentals: Coding redundancy The gray level histogram of an image can reveal a great deal of information about the image That probability (frequency) of occurrence of gray level r k is p(r k ), p n

More information

MATH STUDENT BOOK. 12th Grade Unit 9

MATH STUDENT BOOK. 12th Grade Unit 9 MATH STUDENT BOOK 12th Grade Unit 9 Unit 9 COUNTING PRINCIPLES MATH 1209 COUNTING PRINCIPLES INTRODUCTION 1. PROBABILITY DEFINITIONS, SAMPLE SPACES, AND PROBABILITY ADDITION OF PROBABILITIES 11 MULTIPLICATION

More information

University of California, Berkeley, Statistics 134: Concepts of Probability. Michael Lugo, Spring Exam 1

University of California, Berkeley, Statistics 134: Concepts of Probability. Michael Lugo, Spring Exam 1 University of California, Berkeley, Statistics 134: Concepts of Probability Michael Lugo, Spring 2011 Exam 1 February 16, 2011, 11:10 am - 12:00 noon Name: Solutions Student ID: This exam consists of seven

More information

Chapter 2 Source Models and Entropy. Any information-generating process can be viewed as. computer program in executed form: binary 0

Chapter 2 Source Models and Entropy. Any information-generating process can be viewed as. computer program in executed form: binary 0 Part II Information Theory Concepts Chapter 2 Source Models and Entropy Any information-generating process can be viewed as a source: { emitting a sequence of symbols { symbols from a nite alphabet text:

More information

Image Compression. Qiaoyong Zhong. November 19, CAS-MPG Partner Institute for Computational Biology (PICB)

Image Compression. Qiaoyong Zhong. November 19, CAS-MPG Partner Institute for Computational Biology (PICB) Image Compression Qiaoyong Zhong CAS-MPG Partner Institute for Computational Biology (PICB) November 19, 2012 1 / 53 Image Compression The art and science of reducing the amount of data required to represent

More information

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY BURTON ROSENBERG UNIVERSITY OF MIAMI Contents 1. Perfect Secrecy 1 1.1. A Perfectly Secret Cipher 2 1.2. Odds Ratio and Bias 3 1.3. Conditions for Perfect

More information

Cryptography and Network Security Prof. D. Mukhopadhyay Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Cryptography and Network Security Prof. D. Mukhopadhyay Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Cryptography and Network Security Prof. D. Mukhopadhyay Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Module No. # 01 Lecture No. # 08 Shannon s Theory (Contd.)

More information

CS4800: Algorithms & Data Jonathan Ullman

CS4800: Algorithms & Data Jonathan Ullman CS4800: Algorithms & Data Jonathan Ullman Lecture 22: Greedy Algorithms: Huffman Codes Data Compression and Entropy Apr 5, 2018 Data Compression How do we store strings of text compactly? A (binary) code

More information

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics Probability Rules MATH 130, Elements of Statistics I J. Robert Buchanan Department of Mathematics Fall 2018 Introduction Probability is a measure of the likelihood of the occurrence of a certain behavior

More information

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39 Entropy Probability and Computing Presentation 22 Probability and Computing Presentation 22 Entropy 1/39 Introduction Why randomness and information are related? An event that is almost certain to occur

More information

2018/5/3. YU Xiangyu

2018/5/3. YU Xiangyu 2018/5/3 YU Xiangyu yuxy@scut.edu.cn Entropy Huffman Code Entropy of Discrete Source Definition of entropy: If an information source X can generate n different messages x 1, x 2,, x i,, x n, then the

More information

P [(E and F )] P [F ]

P [(E and F )] P [F ] CONDITIONAL PROBABILITY AND INDEPENDENCE WORKSHEET MTH 1210 This worksheet supplements our textbook material on the concepts of conditional probability and independence. The exercises at the end of each

More information