An Introduction to Lossless Audio Compression

Similar documents
LOSSLESS audio coding is designed for applications

8/13/16. Data analysis and modeling: the tools of the trade. Ø Set of numbers. Ø Binary representation of numbers. Ø Floating points.

SCALABLE AUDIO CODING USING WATERMARKING

encoding without prediction) (Server) Quantization: Initial Data 0, 1, 2, Quantized Data 0, 1, 2, 3, 4, 8, 16, 32, 64, 128, 256

Number Representation and Waveform Quantization

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments

DSP Design Lecture 2. Fredrik Edman.

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments. Tutorial 1. Acknowledgement and References for lectures 1 to 5

Introduction p. 1 Compression Techniques p. 3 Lossless Compression p. 4 Lossy Compression p. 5 Measures of Performance p. 5 Modeling and Coding p.

Compression and Coding

Order Adaptive Golomb Rice Coding for High Variability Sources

Elements of Floating-point Arithmetic

Multimedia Networking ECE 599

THE newest video coding standard is known as H.264/AVC

SPEECH ANALYSIS AND SYNTHESIS

BASIC COMPRESSION TECHNIQUES

Huffman Coding. C.M. Liu Perceptual Lab, College of Computer Science National Chiao-Tung University

Analysis of methods for speech signals quantization

ECE 372 Microcontroller Design

Image Compression. Qiaoyong Zhong. November 19, CAS-MPG Partner Institute for Computational Biology (PICB)

Adaptive Karhunen-Loeve Transform for Enhanced Multichannel Audio Coding

Selective Use Of Multiple Entropy Models In Audio Coding

Fast Progressive Wavelet Coding

What does such a voltage signal look like? Signals are commonly observed and graphed as functions of time:

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

ON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

Fault Tolerance Technique in Huffman Coding applies to Baseline JPEG

Elements of Floating-point Arithmetic

Multimedia Systems Giorgio Leonardi A.A Lecture 4 -> 6 : Quantization

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

Linear Prediction Coding. Nimrod Peleg Update: Aug. 2007

Reduce the amount of data required to represent a given quantity of information Data vs information R = 1 1 C

Can the sample being transmitted be used to refine its own PDF estimate?

Chapter 1 Error Analysis

Lecture 12. Block Diagram

INTEGER WAVELET TRANSFORM BASED LOSSLESS AUDIO COMPRESSION

Time-domain representations

Multimedia Communications. Differential Coding

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

Compression methods: the 1 st generation

Basic Principles of Video Coding

Lecture 7 Predictive Coding & Quantization

Data Compression Techniques

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

Chapter 4 Number Representations

Image Data Compression

9. Datapath Design. Jacob Abraham. Department of Electrical and Computer Engineering The University of Texas at Austin VLSI Design Fall 2017

Predictive Coding. Prediction Prediction in Images

Predictive Coding. Prediction

Progressive Wavelet Coding of Images

CMPT 889: Lecture 3 Fundamentals of Digital Audio, Discrete-Time Signals

L. Yaroslavsky. Fundamentals of Digital Image Processing. Course

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

The information loss in quantization

How do computers represent numbers?

1 Floating point arithmetic

Algorithms (II) Yu Yu. Shanghai Jiaotong University

WaveNet: A Generative Model for Raw Audio

CSE 408 Multimedia Information System Yezhou Yang

Numbering Systems. Contents: Binary & Decimal. Converting From: B D, D B. Arithmetic operation on Binary.

Lecture 9 Video Coding Transforms 2

Lecture 7: Linear Prediction

Converting DCT Coefficients to H.264/AVC

Communications II Lecture 9: Error Correction Coding. Professor Kin K. Leung EEE and Computing Departments Imperial College London Copyright reserved

Data Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise.

Compression. Encryption. Decryption. Decompression. Presentation of Information to client site

ECE472/572 - Lecture 11. Roadmap. Roadmap. Image Compression Fundamentals and Lossless Compression Techniques 11/03/11.

Review of Quantization. Quantization. Bring in Probability Distribution. L-level Quantization. Uniform partition

Applied Mathematics 205. Unit 0: Overview of Scientific Computing. Lecturer: Dr. David Knezevic

Statistical Analysis and Distortion Modeling of MPEG-4 FGS

SIGNAL COMPRESSION Lecture 7. Variable to Fix Encoding

SIMON FRASER UNIVERSITY School of Engineering Science


BASICS OF COMPRESSION THEORY

Topic 3. Design of Sequences with Low Correlation

Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER? Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER?

FPGA Resource Utilization Estimates for NI crio LabVIEW FPGA Version: 8.6 NI-RIO Version: 3.0 Date: 8/5/2008

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Lecture 1 : Data Compression and Entropy

Digital Signal Processing

Speech Coding. Speech Processing. Tom Bäckström. October Aalto University

Rounding Transform. and Its Application for Lossless Pyramid Structured Coding ABSTRACT

A Lossless Image Coder With Context Classification, Adaptive Prediction and Adaptive Entropy Coding

On Common Information and the Encoding of Sources that are Not Successively Refinable

Integer Least Squares: Sphere Decoding and the LLL Algorithm

Data Compression Techniques (Spring 2012) Model Solutions for Exercise 2

Source Coding. Scalar Quantization

2018/5/3. YU Xiangyu

Error Detection & Correction

Implementation of CCSDS Recommended Standard for Image DC Compression

Chapter 12 Variable Phase Interpolation

DESIGN OF QUANTIZED FIR FILTER USING COMPENSATING ZEROS

FPGA Resource Utilization Estimates for NI PXI-7854R. LabVIEW FPGA Version: 8.6 NI-RIO Version: 3.0 Date: 8/5/2008

Intraframe Prediction with Intraframe Update Step for Motion-Compensated Lifted Wavelet Video Coding

Pulse-Code Modulation (PCM) :

4. Quantization and Data Compression. ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak

Digital Systems Roberto Muscedere Images 2013 Pearson Education Inc. 1

ENGINEERING TRIPOS PART IIB: Technical Milestone Report

USAGE OF OLD MAPS AND THEIR DISTRIBUTION WITHIN THE FRAME OF WEB MAP SERVICES

Transcription:

An Introduction to Lossless Audio Compression Florin Ghido Department of Signal Processing Tampere University of Technology SGN-2306 Signal Compression 1 / 31

Introduction and Definitions Digital Audio and Lossless Compression Definitions Types of Lossless Compressors Unused Bits Removal Affine Transforms Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction and Integer Mappings Golomb-Rice Coding Adaptive Arithmetic Coding Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References 2 / 31

Digital Audio and Lossless Compression Definitions Types of Lossless Compressors About digital audio Digital audio is recorded by conversion of the physical sound signal into an electrical signal, which is measured F = 1/T times a second using a discrete number of levels L = 2 B. F is called the sampling rate or sampling frequency and B is the number of data bits used to represent a value. Also, there is a number of synchronized audio channels C to recreate the spatial impression. The entire process is reversed when playing sound from a recording. Advantages of digital audio over analog audio: no degradation in time or generation degradation (for copies); advanced digital signal processing possible in digital domain; higher data density compared to analog recording methods; possibility of hierarchical coding, and online distribution. 3 / 31

Digital Audio and Lossless Compression Definitions Types of Lossless Compressors Digital audio data types Typical values for F, B, and C are: for F, 8000 Hz (telephony), 16000 Hz (wide band telephony), 22050 Hz (half of CD), 32000 Hz (DAT), 44100 Hz (CD, most used), 48000 Hz (DAT and DVD), 96000 Hz and 192000 Hz (DVD-Audio, HD-DVD, and Blu-ray); for B, 8 bit integer unsigned, 8 bit integer signed, 16 bit integer signed, 24 bit integer signed, 32 bit integer signed (rarely used), and 32 bit (floating-point). All signed values are in two s complement format; for C, 1 channel (mono), 2 channels (stereo), and multichannel which has a larger number of discrete channels (2.1, 3.0, 3.1, 4.0, 5.0, 5.1, 6.1, 7.1) - the suffix indicates if there is a discrete low frequency enhancement (LFE) channel. 4 / 31

Digital Audio and Lossless Compression Definitions Types of Lossless Compressors Lossless audio compression Main characteristics: Lossless signal reconstruction (bit identical reconstruction); Expected average compression is 2:1 to 3:1 (for Audio-CDs); Wide area of applications: archiving, transcoding, distribution, online music stores, portable music players, DVD-Audio. Desired qualities: Low decoder complexity (w.r.t. hardware architecture, ISA); Configurable encoder complexity (from medium to very high); Support for most bit depths (including floating-point data), sampling rates, and number of channels; Scalable, streamable, embeddable in container formats. 5 / 31

Digital Audio and Lossless Compression Definitions Types of Lossless Compressors Definition of audio signals Audio signals are discrete time and discrete levels (integer or less commonly floating-point valued) unidimensional data. Typically, for compression the samples are grouped in independent blocks with length N from a fraction of a second up to several seconds; this is done for practical reasons in order to allow seeking to an arbitrary location and to limit damage in case of data corruption. We denote the audio signal, mono or stereo, by: x 0, x 1, x 2,, x N 1 (1) y 0, y 1, y 2,, y N 1 (2) where, for a stereo signal, x i is the left channel and y i is the right channel. When processing data in order, x i is processed before y i. 6 / 31

Digital Audio and Lossless Compression Definitions Types of Lossless Compressors Properties of a data compression algorithm Depending on the type of processing involved in a compression algorithm (either for prediction or entropy coding), we can roughly categorize them into several classes: fixed, where a hard-coded predictor (for example polynomial predictor) or ad-hoc entropy coding parameters are used; semi-fixed, where the predictor or entropy coding parameters are hard-coded, but determined from a large training corpus; forward adaptive (asymmetrical), where the predictor or entropy coding parameters are computed for each small data block, and transmitted as side information; backward adaptive (symmetrical), where the predictor or entropy coding parameters are recomputed from time to time, but only using past data; lookahead may be used to select improved meta-parameters. 7 / 31

Unused Bits Removal Affine Transforms Why data preprocessing? Data preprocessing is a bijective transform (typically from the forward adaptive class) applied to a data block before prediction and entropy coding in order to improve compressibility by taking advantage of special properties of the data: several unused least-significant bits in the values; smaller range of values than possible in the representing data type; some values are not possible, due to processing in the sensor hardware or previous data manipulation; for floating-point, convert to integers and also save additional information needed for reconstruction. 8 / 31

Unused Bits Removal Affine Transforms Unused least significant bits removal This is the simplest and straightforward preprocessing method intended to remove the zero least significant bits common to all the sample values in the block. This situation would happen when the hardware sensor is, for example, 12 bits and data is packed into 16 bit values, or the hardware sensor is 18 or 20 bits and data is packed into 24 bit values. 1. Set mask = 0. 2. For i {0,, N 1} mask = mask bitor x i 3. Set unused = number of zero least significant bits in mask 9 / 31

Unused Bits Removal Affine Transforms Affine integer transforms A more advanced preprocessing method is intended to remove any integer scaling operation, combined with addition of an offset. We try to find the nonnegative integers α, β with α 2 and β < α, such that x i = αz i + β, where z i is an integer sequence. This includes, as a particular case, the unused bits removal method. 1. Set α = x 1 x 0. 2. For i {1,, N 1} α = cmmdc(α, x i x 0 ) 3. Set β = (x 1 x 0 ) modulo α 10 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Prediction and the simplest case Prediction is used in lossless audio compression in order to decrease the average amplitude of the audio signal, therefore reducing significantly its entropy. The most widely used class of predictors is that of linear predictors, which try to estimate the value of the current sample x k by means of a linear combination of the M previous samples, x k 1, x k 2,, x k M. M is called the prediction order, and the prediction is denoted by ˆx k. The simplest predictors are fixed polynomial predictors, which assume the signal can be modeled by a small order polynomial: order 1, ˆx k = x k 1 order 2, ˆx k = 2x k 1 x k 2 order 3, ˆx k = 3x k 1 3x k 2 + x k 3 11 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Forward adaptive prediction If we assume the signal is stationary on an interval of 10-50 milliseconds (which corresponds to a number of samples from a few hundreds to a few thousands), we can compute the best predictor for that group of samples (called a frame); it is necessary, however, to save the prediction coefficients as side information (for the decoder), and moreover they must be saved with some limited precision (quantized), in order to minimize the size overhead. There is a tradeoff between prediction performance (sum of squared errors) and precision of the prediction coefficients; our target is to obtain the smallest possible overall compressed size. 12 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Forward adaptive prediction (cont.) If the frame extends from sample p to sample q, using a predictor w of order M, the criterion to be minimized is: q M 1 J(w) = (x i w j x i j 1 ) 2 (3) i=p The optimal predictor w can be found using the Levinson-Durbin algorithm. The algorithm takes the autocorrelation vector r of length M + 1, defined as q r j = x i x i j, for j {0,, M} (4) i=p where the values x i outside the defined range are considered zero, and produces a vector of intermediary (reflection) coefficients k and the prediction coefficients w. j=0 13 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Quantization of prediction coefficients The prediction coefficients w must be quantized to a given number of fractional bits cb, before being saved as side information and used for prediction. We can write for the quantized scaled integer coefficients w, the relation w j = floor(2 cb w j + 0.5) (5) that is, scaling with the corresponding power of 2 and rounding to the nearest integer. A problem with this classical approach is that there is no simple range the values w j can take, and no clear bound on the penalty in terms of the criterion J(w) due to quantization. 14 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Quantization of prediction coefficients (cont.) The optimal number of fractional bits cb to be used varies typically between 6 and 12, depending on the frame content. If we want bounded ranges for the prediction coefficients and a better representation (less prediction penalty for the same amount of side information), the alternative is to quantize and save the reflection coefficients as k j = floor(2 cb k j + 0.5) (6) There is a bijective mapping between reflection coefficients and prediction coefficients (in fact, the prediction coefficients are normally computed from the reflection coefficients), and we have that k i 1. The optimal number of fractional bits cb to be used for reflection coefficients varies between 5 and 8. 15 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Backward adaptive prediction We can also compute from time to time, or for each sample the best prediction coefficients for recent data and use those coefficients for prediction of the next sample. This approach has the advantage of not requiring to see in advance (no lookahead needed) - can be applied in an online setting, and also no coefficients must be saved as side information. As a downside, the decoder has also to mimic the same computations done by the encoder, which can make the entire process slower. The most used methods of computing the coefficients are simple adaptive gradient methods (like Least Mean Squares or Normalized Least Mean Squares), or the (Recursive) Least Squares method. 16 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Backward adaptive prediction (cont.) In the case of adaptive gradient methods, we illustrate below the usage of the Least Mean Squares algorithm for prediction, where µ is a small constant such that µ < 1/(2Mr 0 ), as 1. For j {0,, M 1} // initialize prediction coefficients Set w j = 0 Set e j = x i 2. For i {M,, N 1} 2.1 Set p = 0 2.2 For j {0,, M 1} // compute prediction for x i Set p = p + w j x i j 1 2.3 Set e i = x i p // e i is the prediction error 2.4 For j {0,, M 1} // update prediction coefficients Set w j = w j + µ e i x i j 1 17 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Backward adaptive prediction (cont.) In the case of (Recursive) Least Squares method, the criterion J LS (w) = i n=0 M 1 α i n (x n w j x n j 1 ) 2 (7) is used at each sample x i or from time to time for recomputing the prediction coefficients w. α is a positive constant with α < 1, the exponential forgetting factor, which puts more emphasis in the criterion on the prediction performance in the recent past. Efficient techniques exist (the Recursive formulation of the Least Squares is one of them) such that the new prediction coefficients are computed taking advantage of the previous coefficients, at much lower complexity than computing them from the start. j=0 18 / 31

Fixed Predictors Forward Adaptive Prediction Quantization of Prediction Coefficients Backward Adaptive Prediction Stereo Prediction Stereo prediction In the case of a stereo signal, we can obtain better predictions if we predict a sample using the previous values from the same channel and also the previous values (and the current value, if available) from the other channel. The prediction equations for stereo predictor composed of an intrachannel predictor of order M and an interchannel predictor of order L are, for the left channel ˆx i = M 1 j=0 L 1 w j x i j 1 + w j+m y i j 1 (8) and for the right channel (note that x i can be used for prediction) ŷ i = M 1 j=0 w j=0 L 1 j y i j 1 + w j+m x i j (9) j=0 19 / 31

and Integer Mappings Golomb-Rice Coding Adaptive Arithmetic Coding and Bijective Integer Mappings Entropy coding is used to take advantage of the reduced amplitude of the prediction residuals from the prediction step and from the fact that the prediction residuals are generally distributed according to a symmetrical two sided Laplacian (or geometric) distribution. It is easy to convert the signed prediction residuals to unsigned values, by using the following bijective integer mapping { u(x) = 2x for x 0 2x 1 for x < 0 (10) The signed values, 2, 1, 0, 1, 2, get rearranged to 0, 1, 1, 2, 2,, and therefore we obtain a one-sided geometric distribution. If x is B bits signed, u(x) is also B bits, but unsigned. 20 / 31

Golomb-Rice coding Introduction and Definitions and Integer Mappings Golomb-Rice Coding Adaptive Arithmetic Coding Golomb and Golomb-Rice codes are two class of codes, described earlier in the Signal Compression course, which are well suited for coding one-sided geometric distributions. Golomb codes use a parameter s to divide each value to be coded into bins of equal width s, splitting a value u into m = floor(u/s) and l = u mod s. The value of m is coded in unary (i.e. m one bits followed by a zero bit), and l is coded using a special code which uses either floor(log 2 (s)) or ceil(log 2 (s)) bits. Golomb-Rice codes further limit the value of parameter s to be a power of two s = 2 p, eliminating the need for an integer division, and coding l raw using exactly p bits. 21 / 31

Golomb-Rice coding (cont.) and Integer Mappings Golomb-Rice Coding Adaptive Arithmetic Coding There are two ways of computing s and p, forward-adaptive and backward-adaptive. The first method is to compute the optimum s or p for each short block of a few hundred samples and save it as side information. The suggested values are s E(u(x)) and p log 2 (E(u(x))), but we can also check for values around the estimated integer value, and select the one which provides the best compression. The second method is to estimate s i or p i recursively, by keeping an exponentially weighted running average U i, as U i+1 = βu i + (1 β)u(x i ) (11) where β is a positive constant, the exponential forgetting factor, with β < 1, and computing s i = U i or p i = floor(log 2 (U i )). 22 / 31

Adaptive arithmetic coding and Integer Mappings Golomb-Rice Coding Adaptive Arithmetic Coding This improved method is based on the general principles of Golomb coding. The parameter s i is estimated recursively, as in Golomb coding. The difference is that l (the least significant part) is coded with equal probability 1/s, and m (the most significant part) is not coded in unary, instead it is using an adaptive technique. A counter table with T entries is maintained (typically, T = 32). If a value for m < T 1 is encountered, the symbol m from the table is encoded and its counter incremented, followed by raw encoding of l. However, if a value for m T 1 is encountered, the symbol T 1 from the table (escape indicator) is encoded and its counter incremented, followed by raw encoding of u (the entire value). 23 / 31

Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Adaptive signal segmentation for compression In the case of forward adaptive prediction, where prediction coefficients are saved as side information, it is important that each frame to be nearly stationary, otherwise the prediction coefficients will be suboptimal. We want to adaptively split the signal into nonequal frames so that the overall compressed size will be minimized (the minimum description length principle). One method starts from frames of size 2 R G (G is the grid size), and checks if compressing the two halves of size 2 R 1 G leads to better compression. If yes, the same process is applied recursively for each half, as long as the resulting frame size is at least G. Another method considers frames of arbitrary size kg, with k k max, the maximum allowed frame size, and uses a dynamic programming algorithm (evaluating a function which gives the size of a continuous frame) to compute the optimal segmentation. 24 / 31

Compression of floating-point audio Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Floating-point audio is used in the professional audio workflow to keep intermediate processing results, and also for some audio restoration projects, where no added distortion is wanted. The most used format is the 32 bit IEEE floating-point format, which consists of 1 sign bit s, 8 bits of exponent e, and 23 mantissa bits s. The real value of a finite normalized (not denormalized, infinity, or NaN) floating-point number is ( 1) s 1.s 2 e 128. The integer based lossless audio compression methods cannot be applied directly to floating-point numbers. One efficient and simple method to be able to use the integer based compression methods is to approximate the floating-point values with integers and also save some additional information needed to correctly restore the values. 25 / 31

2.8 x decoding speed (20.91 vs. 7.38 real-time) 26 / 31 Introduction and Definitions Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Standardization status and state of the art Existing MPEG-4 standards Audio Lossless Coding (ALS) - lossless only Scalable to Lossless (SLS) - AAC core (optional) to lossless Lossless Coding of 1-bit Oversampled Audio (DST) - for DSD Compression and speed performance of the top two compressors, OptimFROG 4.600ex and MPEG-4 ALS V19, on a private 51.6 GB CD-Audio corpus (consisting of 82 Audio-CDs): mode maximumcompression vs. RLSLMS best mode, BGMC 1.04% better compression (52.64% vs. 53.68%) same encoding speed (0.33 vs. 0.31 real-time) 6.7 x decoding speed (2.03 vs. 0.30 real-time) mode normal vs. opt. comp., LTP, MCC, max. order 1023 0.14% better compression (54.37% vs. 54.50%) 30.5 x encoding speed (7.34 vs. 0.24 real-time)

Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Overall compressed size vs. average encoding speed for asymmetrical ALS-V19 and OFR-AS OLD/NEW v0.300 Overall compressed size (%, lower is better) 59.6 59.5 59.4 59.3 59.2 59.1 59 58.9 58.8 58.7 ALS V19 OFA NEW OFA OLD 58.6 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 Average encoding speed (X real time, higher is better) 27 / 31

Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Overall compressed size vs. average decoding speed for asymmetrical ALS-V19 and OFR-AS OLD/NEW v0.300 Overall compressed size (%, lower is better) 59.6 59.5 59.4 59.3 59.2 59.1 59 58.9 58.8 58.7 ALS V19 OFA NEW OFA OLD 58.6 5 10 15 20 25 30 35 40 45 50 55 Average decoding speed (X real time, higher is better) 28 / 31

Further reading and references I Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References T. Robinson. SHORTEN: Simple lossless and near-lossless waveform compression. In Technical Report CUED/F-INFENG/TR.156, Cambridge University Engineering Department, Cambridge, UK, December 1994. M. Hans and R.W. Schafer, Lossless Compression of Digital Audio, IEEE Signal Processing Magazine, vol. 18, issue 4, pp. 21-32, July 2001. Simon Haykin, Adaptive Filter Theory, Prentice Hall, 4th edition, September 2001. 29 / 31

Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References Further reading and references II F. Ghido, An Asymptotically Optimal Predictor for Stereo Lossless Audio Compression, in Proceedings of the Data Compression Conference, pp. 429, Snowbird, Utah, March 2003. T. Liebchen and Y.A. Reznik, MPEG-4 ALS: an Emerging Standard for Lossless Audio Coding, in Proceedings of the Data Compression Conference, pp. 439-448, Snowbird, Utah, March 2004. F. Ghido, OptimFROG Lossless Audio Compressor (symmetrical), available on Internet at http://www.losslessaudio.org/, July 2006, version 4.600ex. 30 / 31

Further reading and references III Signal Segmentation for Compression Compression of Floating-Point Audio Compression Results and Comparison Further Reading and References F. Ghido and I. Tabus, Adaptive Design of the Preprocessing Stage for Stereo Lossless Audio Compression, in Proceedings of 122th Audio Engineering Society Convention (AES 122), Convention Paper 7085, Vienna, Austria, May 2007. ISO/IEC, ISO/IEC 14496-3:2005/Amd 2: 2006, Audio Lossless Coding (ALS), new audio profiles and BSAC extensions, available on Internet at http://www.nue.tu-berlin.de/mp4als, November 2007, reference software version RM19. 31 / 31

Some Overall Results Compression versus decoding speed: total compression in percent (lower is better) and average decoding speed in multiples of real-time (higher is better).