A MODEL OF JERKINESS FOR TEMPORAL IMPAIRMENTS IN VIDEO TRANSMISSION. S. Borer. SwissQual AG 4528 Zuchwil, Switzerland

Size: px
Start display at page:

Download "A MODEL OF JERKINESS FOR TEMPORAL IMPAIRMENTS IN VIDEO TRANSMISSION. S. Borer. SwissQual AG 4528 Zuchwil, Switzerland"

Transcription

1 A MODEL OF JERKINESS FOR TEMPORAL IMPAIRMENTS IN VIDEO TRANSMISSION S. Borer SwissQual AG 4528 Zuchwil, Switzerland ABSTRACT Transmission of digital videos over band limited and possibly error prone channels is an important source of temporal impairments, such as reduced frame rates, frame freezings, and frame droppings. This often results in a non fluent and non smooth presentation of video during playback. Perceptually, this is called jerkiness. In this paper a model of perceived jerkiness is proposed. It can be applied to video sequences showing any pattern of jerkiness, with constant or variable frame rate. It includes a parametrisation of the viewing condition. Thus, it can be used to predict jerkiness from small to large resolutions. The model is compared to data from subjective experiments containing a large variety of temporal degradations and spanning the whole range of currently used video resolutions from QCIF up to HD. On this data it shows excellent performance. The model does not require any comparison to a reference signal. Therefore, it can be applied to no-reference monitoring approaches. Index Terms Jerkiness, Video Quality, Digital Video Transmission 1. INTRODUCTION Digital video transmission systems consist of a chain of a video encoder, a network, a decoder, and a video player. Due to limited bandwidth of the transmission the encoder can reduce the frame rate of the video to save on bandwidth. On the other hand, due to a possible error-prone network video transmission may be delayed or information can get lost. As a result, during playback, the video player may repeat frames, drop frames, or display frames exceedingly long. All this introduces temporal impairments of the video playback. Technically, temporal variations in the video playback are called jitter, and the exceedingly long display of a frame is called freezing. For an end-user, these impairments may lead to perceptual degradations called jerkiness. In this paper we present a jerkiness measure. It uses information from a video sequence captured at the output of a video player. It is a so-called no-reference measure, i.e. no information about the video before encoding and transmission is used. Our proposed jerkiness measure shows a high correspondence with subjective ratings. A typical application for this no-reference jerkiness model is quality monitoring. Several other studies about the relation between temporal impairments and perceptual quality exist: The effect of frame rate reduction for low-bitrate transmission on perceived video quality was studied in [3] for videos of QCIF resolution. Furthermore, the impact of frame rate reduction, jitter, and freezing was studied in [2] for the larger VGA resolution. Both studies used the same test methodology as the data in this paper, and therefore, the model predictions can be compared to their data. The relation between short and long freezing on perceived quality was studied in [6], using a different subjective test methodology. Using this data, a model was proposed in [5]. A different model was proposed in [4] based on a compressed video sequence. As the mathematical forms as well as the evaluation methodologies are different, it is difficult to compare the two. Our model is different to the above two models in three aspects: First, it includes a dependence on frame resolution and viewing condition. Therefore, it can be applied for any frame size from QCIF up to HD. Second, the formulation is more general and can be used for any fixed or variable frame rate video sequence. Third, the mathematical form of the jerkiness measure is different. The paper is organised as follows: First, section 2 introduces notation and concepts. The central part is section 3, where the jerkiness model is introduced. Section 4 describes the model dependence on the viewing condition and the frame resolution. In section 5 the performance of the jerkiness model is evaluated. Finally, section 6 concludes with a summary of the main results and future research topics. 2. NOTATION AND CONCEPTS The input of the jerkiness model is a video sequence. The video sequence is captured at the output of the video player. Each new frame is shown by the player as an image and is captured together with a time stamp, the display start of the /10/$ IEEE 218 QoMEX 2010

2 frame. Depending on the situation, the captured video sequences may have a variable frame rate. More formally, a video sequence is a sequence of the form v =(f i,t i ) i=1,..n, (1) where f i denotes frame i of the video, which will be displayed starting at time t i up to time t i+1, and n is the total number of frames. If the total number of frames in the sequence is clear from the context, or is of minor importance, the shorthand notation (f i,t i ) will be used. The times t i will be referred to as time stamps, and motion intensity m i +1 (v) t i =t i +1 t i Δt i = t i+1 t i (2) is called the display time of frame i. For each video sequence as input the output of the jerkiness model is a positive value, the estimated or predicted jerkiness. Therefore, the term jerkiness measure will also be used. The jerkiness measure will have two main variables, the display time of frames and the motion intensity. The next paragraph describes how motion is measured. Ideally, motion would be measured by calculating some statistic of the velocity distribution of the moving objects in the video sequence. As this might be very complicated, a very simple motion measure will be used here. For a video sequence v =(f i,t i ) the motion intensity is the root mean squared interframe difference over the whole frame m i+1 (v) = (f i+1 (x) f i (x)) 2, (3) x where f i (x) denotes the pixel value of the Y-component of a frame i at location x. Note that the frames are used in full resolution for small resolutions only, here QCIF. For large resolutions, the Y- component of the frame is first subsampled to QCIF resolution before motion calculation. 3. MEASURING JERKINESS This section is the central part of the paper, introducing the jerkiness model. Let (f i,t i ) denote a video sequence. Repeated frames will be removed from the video sequence beforehand. For ease of presentation, suppose that frames that are perceived as a repetition are exact repetitions, i.e. they are equal. If this is not the case, an approach for the detection of perceived repeated frames can be found in [1]. Recall that t i denotes the display time start of frame f i, and t i+1 is the display time end of frame f i, equal to the display time start of the following frame. For motivation, consider a sample video sequence containing an isolated freezing event as shown in figure 1. The plot shows the motion intensity. The interval of zero motion corresponds to the freezing of the video sequence. The time Fig. 1. The plot shows the motion intensity for a short video sequence with freezing. The dashed lines show the freezing duration and the motion intensity at the end of the freezing. main variables of the jerkiness measure are the display time Δt i = t i+1 t i (horizontal dashed line) of the freezing frame and the motion intensity m i+1 (vertical dashed line) at the end of the freezing interval. This corresponds to the perception of the freezing of the frame and the jump, i.e. sudden large motion, at the end of the freezing interval. The considerations for the choice of the jerkiness measure (5) are the following: First, if the motion intensity is larger, then the jerkiness is larger. Thus, jerkiness should be a monotone function, say μ of motion intensity. Second, if there is no motion, then jerkiness is zero. Thus, μ(0) = 0 and jerkiness is a product of μ and a second part. Third, dependence of jerkiness on motion intensity might saturate for large motion intensities. On the other hand, for very small motion intensity values the perceptual impact of jitter might be small. Thus, μ is chosen to be a parametrised S-shaped or sigmoid function, see appendix 8 for details. Fourth, jerkiness depends on the frame display time Δt i. The larger the display time the larger the jerkiness. Thus, jerkiness depends monotonously on display time. Fifth, for very small display times there is no jerkiness. Thus, jerkiness is a product of a display time dependent part and the motion dependent part μ, where the display time dependent part yields small values for small values of the display time. Last, for very large frame display times it is difficult to make useful subjective experiments. Thus, a simple approach is a close to linear relationship, e.g. if a 10sec video sequence freezes almost 10 seconds or if it freezes twice almost 5 seconds should give a similar jerkiness value. This can be modelled by Δt i τ(δt i ), (4) for an S-shaped function τ. This is close to linear in the saturated region of τ, is monotone, and in the region where τ is almost zero the term (4) is almost zero. 219

3 zoom Table 1. Viewing distance (in multiples of display height) and corresponding viewing angle (under which the display height is seen) as recommended in [8]. resolution viewing distance viewing angle (in display height) (of display height) QCIF CIF SD HD display time in s Fig. 2. The plot shows the S-shaped transform τ used for the dependency of the jerkiness measure on the frame display time, for a CIF video under standard viewing conditions. The above motivates the following definition of the jerkiness measure: Jerkiness is the sum (over all frame times) of the product of relative display time times a monotone function of display time times a monotone function of motion intensity at display time end, J(v) = 1 Δt i τ α (Δt i ) μ(m i+1 (v)). (5) T i Here, the video sequence is of the form (1), the display time is defined in (2), the motion intensity is defined in (3), and T is the total video duration. The functions τ α and μ are S- shaped functions, as described in appendix 8. They have 3 parameters, the position and slope of the inflection point. For μ these are kept fix. For τ the 3 parameters are reparametrised by a single parameter α depending on the viewing angle. Details and motivation for this reparametrisation are described in section 4. Figure 2 shows the transformation τ α under typical viewing conditions for CIF resolution. For frame display times below 50ms the value is almost zero, thus jerkiness values will be very small. The inflection point is shown as a black dot, close to 100ms, and is mapped to 0.05, which will produce moderate jerkiness values. This is consistent with findings of [6], where the authors found that at display times around 100ms there is a high probability for subjects to detect a small degradation. The functions τ α and μ take values in [0, 1), therefore the jerkiness measure will take values in [0, 1), where zero means no jerkiness. Under normal circumstances, dropping more frames will result in a higher motion intensity at the end of the frame drop, and therefore a higher jerkiness value. Thus, the lower the frame rate, the higher the jerkiness value. In addition, all function are continuous. Therefore, intuitively speaking, two similar video sequences will have similar jerkiness values. Note that one key idea is already present in [5], the dependence of jerkiness on the motion intensity at the end of the jerkiness interval. Contrary to our formulation as a product of display time dependent and motion dependent part, the authors used a additive combination. This has a different interpretation: Jerkiness is measured if either a long display time or high motion is present. In our formulation, jerkiness is measured if both a long display time and motion is present. 4. RESOLUTION DEPENDENCE In this section the jerkiness model is completed by including into the model the dependence on the video resolution, from QCIF up to HD. In this context, resolution means not only the video resolution itself but includes the corresponding viewing condition. Recall that the jerkiness model contains a product of two S-shaped functions. One, μ, accounts for the motion dependent part, the other, τ α, accounts for part of the display time dependent part. Recall that motion intensity is measured on subsampled frames for larger resolutions. Therefore, the motion dependent part already contains implicitly a frame resolution dependent component. In this section, the resolution dependence of the display time dependent part τ α will be explained. There may be different reasons for a resolution dependent perception of jerkiness: different viewing angles, as typically the viewing angle for HD is much larger than for QCIF; expectations might be different, viewers in a subjective experiment may react much stronger on temporal impairments in a HD experiment than by watching and rating QCIF video sequences, and there might be many more reasons. In this paper, only the dependence on the viewing angle will be taken into account. Table 1 shows the typical viewing angle and viewing distance used in the subjective experiments performed for the projects of the video quality expert group (VQEG), see [8]. It can be seen, that the viewing angle is 220

4 larger for larger resolutions. E.g. the viewing angle for HD is about 2-3 times as large as the one for QCIF. The dependence of jerkiness on the resolution will be included in the model using the following hypothesis: Suppose two video sequences of different resolution having the same content are viewed under typical viewing distances. Approximately, the same jerkiness value should result, if the temporal sampling rates are adjusted, such that the moving objects have the same angular displacement from one frame to the next frame. In the jerkiness model, equation (5), the resolution dependent part is the S-shaped function τ α, which is parametrised by the position (p x,p y ) and slope q of the inflection point, see equation (6) in the appendix. If the viewing angle is increased by a factor c, then the angular displacement is increased by a factor c, thus, under the hypothesis above, the temporal resolution has to be increased by decreasing the display time by 1/c. Thus τ α is compressed by a factor c along the time axis, i.e. the new parameters of the inflection point are the position (p x /c, p y ) and slope q c. On the other hand, for smaller viewing angles, the S-shaped function is stretched along the time axis. Here, the following factors are used: Given the parameters (p x,p y ) and slope q for QCIF, the factor c for CIF is 1.18,for SD it is 1.54 and for HD it is These factors are derived from the viewing angles given in table 1. This model of resolution dependence implies the following for a fixed temporal impairment: The larger the resolution, the larger the viewing angle under typical viewing conditions, and therefore, the larger the angular displacement of a moving object before and after the impairment, hence, the higher the jerkiness value. 5. MODEL EVALUATION To understand the correspondence between modelled jerkiness and perceived quality subjective tests were performed. This section starts with describing the subjective datasets. Then, one dataset of CIF resolution was used for parameter estimation. Finally, the other datasets were used for model evaluation Subjective Data Four different subjective datasets are used. Two datasets, one QCIF and one CIF video resolution, containing reduced frame rate and freezed frame conditions as a result of transmission errors. Two datasets for the large SD and HD resolutions containing frame freezes only, as reduced frame rates are not common for these resolutions. In more detail, the datasets are composed as follows: 1. QCIF: jerkiness cif qcif sd hd mos Fig. 3. The scatter plot shows the MOS values of the subjective data against the models jerkiness predictions. Data is shown for formats QCIF, CIF, SD, HD. 2. CIF: (a) Contents: Four contents: speaker, fishes, actors, sport. (b) Conditions: reduced frame rates: 25fps, 12.5fps, 8fps, 5fps. Freezing, 1,2,or 3 freezing intervals at random positions of fixed duration. All uncompressed. (a) Contents: The same contents as in QCIF dataset, plus an average motion movie sequence. (b) Conditions: frame rates of 30fps or 25fps, 12.5fps, 5fps. Freezing, 1,2,or 3 freezing intervals at random positions of fixed duration. All uncompressed. 3. SD: 4. HD: (a) Contents: Five contents: interview, music video, comedy, action movie, sport. (b) Conditions: Freezing as a result of 0.5%, 1%, 2% of random (bursty) packet loss. Freezing as a result of 0.06% of uniform packet loss. All 25fps and encoded at 4MBit/s. (a) Contents: The five contents of SD: interview, music video, comedy, action movie, sport. (b) Conditions: Freezing as a result of 0.5%, 1%, 2% of random (bursty) packet loss. Freezing as a result of 0.06% of uniform packet loss. All 25fps and encoded at 16MBit/s. 221

5 Note that the datasets were created using a full matrix design, i.e. for each content and each condition one sample was generated. Each dataset contained a low motion content, speaker or interview, and at least one high motion content, sport. Additional samples containing other types of distortions were added to generate equilibrated datasets, which can be used for subjective testing. Note that SD and HD data was compressed at a high bitrate. This was done because the data was generated with additional intentions in mind, outside of the scope of this paper. Nevertheless, it was checked that there were no spatial artifacts influencing the present analysis. Subjective test were performed following the recommendation [7] and best practise of [8]. Of course, viewing distance was adjusted to match the values in table 1. For each subjective test 24 subjects were used. The subjects rated the videos using the scores 1(bad), 2(poor), 3(fair), 4(good), 5(excellent). The subjective ratings were averaged, to generate the mean opinion score (MOS) Parameter Estimation Model parameters were estimated using the CIF dataset. Recall the jerkiness measure, equation (5). There are six parameters, three for each S-shaped function τ α and μ. The following parameters were determined: For μ the parameters (p x,p y,q) = (5, 0.5, 0.25) and for τ α the parameters (p x,p y,q)=(0.12, 0.05, 1.5) for QCIF resolution. The parameters for τ α for the other resolutions are given by (p x,p y,q)=(0.12/c, 0.05, 1.5 c), as explained in section 4, where the values of c are given Model Evaluation Keeping these parameters fixed, jerkiness was calculated and compared to the subjective ratings. Figure 3 shows the result. The scatter plot shows all data samples for the four resolutions. There is a non-linear relationship between the mean opinion score and the predicted jerkiness. To quantify the correspondence of the predicted jerkiness to subjective score, the predicted values were fitted to the MOS using a 3rd order polynomial. The residual RMSE is shown in table 2. By comparison, [5] reports similar error values for CIF, the errors for QCIF in [4] are considerably larger. Note that this comparison serves as an indication only, as video data is different as well as subjective testing methodologies. Furthermore, model predictions can be compared to the values reported in other studies: In [2], for VGA resolution at 12.5fps an average content has a score of 4.0, and for a 5fps sequence a score of Our model predicts for a high motion scene at 12fps a score of 3.73 and for a 5fps sequence a score of For a high motion QCIF sequence jerkiness prediction at 8.5fps is 3.68, and a value of 3.8 is reported in [3]. Table 2. Jerkiness values are fitted to MOS using a 3rd order polynomial. The residual root mean squared error for all data samples except CIF from figure 3 and for each resolution separately are shown in the table. Recall that the CIF data was used to fit model parameters. Resolution RMSE All but CIF 0.26 QCIF 0.27 SD 0.27 HD 0.22 (CIF 0.31) 6. CONCLUSION In this paper a model of perceived jerkiness was proposed. It is general enough to be applicable for small to large resolutions, variable frame rate video sequences containing any type of temporal impairment. The model showed excellent correspondence to the subjective data. As it has relatively few parameters, model over-fitting can almost be excluded. In addition, the model is consistent with findings of subjective experiments performed by others: Frame dropping experiments [6], and frame rate reduction experiments [3],[2]. Unfortunately, a detailed comparison to other models is difficult. First of all, the models in [5] and [4] can be applied for one resolution only, CIF and QCIF, respectively. Second, in [4] the model input are compressed video sequences only. Furthermore, the performance can not be compared to [5], as their data is not available and the values of the model parameters are not given in the paper. Still open is a systematic analysis of the relation of jerkiness to motion intensity. The data of the subjective experiments used in this paper were not designed to analyse this relation in more detail. Note that in the present approach the jerkiness measure depends on the possibly large jump of the motion intensity at the end of a freezing interval. It does not depend on the motion intensity at the beginning of a freezing interval., i.e. the sudden stopping of motion of the video playback. If an additional dependence on the sudden stopping of motion improves performance is still an open question for research. 7. ACKNOWLEDGEMENT Many thanks go to Alexander Raake and Marie Neige Garcia from Deutsche Telekom Laboratories involved in the design, execution, and evaluation of the subjective tests for the SD and HD data. 222

6 S(x) (0.12,0.1,1.5) (0.5,0.2,10) x Fig. 4. The plot shows two sample S-shaped functions. The dot shows the inflection point. The parameters (p x,p y,q) are given in parenthesis. 8. APPENDIX This appendix describes the details of the S-shaped functions used for the jerkiness model. The function, denoted here by S, is parametrised by the location (p x,p y ) and slope q of the inflection point, in detail [4] Kai-Chieh Yang, Khaled El-Maleh, Clark C. Guest, and Pankaj K. Das, Perceptual temporal quality metric for compressed video, in Proc. of the Second International Workshop on Video Processing and Quality Metrics for Consumer Electronics (VPQM), [5] Ricardo R. Pastrana-Vidal and Jean-Charles Gicquel, Automatic quality assessment of video fluidity impairments using a no-reference metric, in Proc. of the Second International Workshop on Video Processing and Quality Metrics for Consumer Electronics (VPQM), [6] R.R. Pastrana-Vidal, J.C. Gicquel, C. Colomes, and H. Cherifi, Sporadic frame dropping impact on quality perception, in Proceedings of SPIE, 2004, vol. 5292, p [7] ITU-T, Subjective video quality assessment methods for multimedia applications, ITU-T Recommendation P.910, [8] VQEG, HDTV and Multimedia project, { a x b if x<= p x S(x) = d 1+exp( c (x p +1 d else, (6) x)) where a = p y /p (qpx/py) x, b = qp x /p y, c =4q/d,and d = 2(1 p y ). The S-shaped function starts at the origin, grows polynomially up to the inflection point and saturates exponentially towards 1. See figure 4 for two sample plots using different parameters. 9. REFERENCES [1] Stephen Wolf, A no reference (nr) and reduced reference (rr) metric for detecting dropped video frames, in Proc. of the Second International Workshop on Video Processing and Quality Metrics for Consumer Electronics (VPQM), [2] Q. Huynh-Thu and M. Ghanbari, Impact of jitter and jerkiness on perceived video quality, in Proc. of the Second International Workshop on Video Processing and Quality Metrics for Consumer Electronics (VPQM), [3] Q. Huynh-Thu and M. Ghanbari, Perceived quality of the variation of the video temporal resolution for low bit rate coding, in Picture Coding Symposium,

no reference video quality assessment metrics for multimedia: state of the art of signal-based

no reference video quality assessment metrics for multimedia: state of the art of signal-based Effect of transmission performance on Multimedia Quality of Service 17-19 June 2008 - Prague, Czec Republic no reference video quality assessment metrics for multimedia: state of the art of signal-based

More information

Analysis of Rate-distortion Functions and Congestion Control in Scalable Internet Video Streaming

Analysis of Rate-distortion Functions and Congestion Control in Scalable Internet Video Streaming Analysis of Rate-distortion Functions and Congestion Control in Scalable Internet Video Streaming Min Dai Electrical Engineering, Texas A&M University Dmitri Loguinov Computer Science, Texas A&M University

More information

ANALYSIS AND PREDICTION OF JND-BASED VIDEO QUALITY MODEL

ANALYSIS AND PREDICTION OF JND-BASED VIDEO QUALITY MODEL ANALYSIS AND PREDICTION OF JND-BASED VIDEO QUALITY MODEL Haiqiang Wang, Xinfeng Zhang, Chao Yang and C.-C. Jay Kuo University of Southern California, Los Angeles, California, USA {haiqianw, xinfengz, yangchao}@usc.edu,

More information

SYDE 575: Introduction to Image Processing. Image Compression Part 2: Variable-rate compression

SYDE 575: Introduction to Image Processing. Image Compression Part 2: Variable-rate compression SYDE 575: Introduction to Image Processing Image Compression Part 2: Variable-rate compression Variable-rate Compression: Transform-based compression As mentioned earlier, we wish to transform image data

More information

Multimedia Databases. Previous Lecture Video Abstraction Video Abstraction Example 6/20/2013

Multimedia Databases. Previous Lecture Video Abstraction Video Abstraction Example 6/20/2013 Previous Lecture Multimedia Databases Hidden Markov Models (continued from last lecture) Introduction into Video Retrieval Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität

More information

Multimedia Databases. Wolf-Tilo Balke Younès Ghammad Institut für Informationssysteme Technische Universität Braunschweig

Multimedia Databases. Wolf-Tilo Balke Younès Ghammad Institut für Informationssysteme Technische Universität Braunschweig Multimedia Databases Wolf-Tilo Balke Younès Ghammad Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de Previous Lecture Hidden Markov Models (continued from

More information

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 10 Video Retrieval Shot Detection 10 Video Retrieval

More information

encoding without prediction) (Server) Quantization: Initial Data 0, 1, 2, Quantized Data 0, 1, 2, 3, 4, 8, 16, 32, 64, 128, 256

encoding without prediction) (Server) Quantization: Initial Data 0, 1, 2, Quantized Data 0, 1, 2, 3, 4, 8, 16, 32, 64, 128, 256 General Models for Compression / Decompression -they apply to symbols data, text, and to image but not video 1. Simplest model (Lossless ( encoding without prediction) (server) Signal Encode Transmit (client)

More information

Multimedia Databases Video Abstraction Video Abstraction Example Example 1/8/ Video Retrieval Shot Detection

Multimedia Databases Video Abstraction Video Abstraction Example Example 1/8/ Video Retrieval Shot Detection 10 Video Retrieval Shot Detection Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 10 Video Retrieval

More information

Single Frame Rate-Quantization Model for MPEG-4 AVC/H.264 Video Encoders

Single Frame Rate-Quantization Model for MPEG-4 AVC/H.264 Video Encoders Single Frame Rate-Quantization Model for MPEG-4 AVC/H.264 Video Encoders Tomasz Grajek and Marek Domański Poznan University of Technology Chair of Multimedia Telecommunications and Microelectronics ul.

More information

Intraframe Prediction with Intraframe Update Step for Motion-Compensated Lifted Wavelet Video Coding

Intraframe Prediction with Intraframe Update Step for Motion-Compensated Lifted Wavelet Video Coding Intraframe Prediction with Intraframe Update Step for Motion-Compensated Lifted Wavelet Video Coding Aditya Mavlankar, Chuo-Ling Chang, and Bernd Girod Information Systems Laboratory, Department of Electrical

More information

h 8x8 chroma a b c d Boundary filtering: 16x16 luma H.264 / MPEG-4 Part 10 : Intra Prediction H.264 / MPEG-4 Part 10 White Paper Reconstruction Filter

h 8x8 chroma a b c d Boundary filtering: 16x16 luma H.264 / MPEG-4 Part 10 : Intra Prediction H.264 / MPEG-4 Part 10 White Paper Reconstruction Filter H.264 / MPEG-4 Part 10 White Paper Reconstruction Filter 1. Introduction The Joint Video Team (JVT) of ISO/IEC MPEG and ITU-T VCEG are finalising a new standard for the coding (compression) of natural

More information

Statistical Analysis and Distortion Modeling of MPEG-4 FGS

Statistical Analysis and Distortion Modeling of MPEG-4 FGS Statistical Analysis and Distortion Modeling of MPEG-4 FGS Min Dai Electrical Engineering Texas A&M University, TX 77843 Dmitri Loguinov Computer Science Texas A&M University, TX 77843 Hayder Radha Hayder

More information

Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms

Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms Flierl and Girod: Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms, IEEE DCC, Mar. 007. Half-Pel Accurate Motion-Compensated Orthogonal Video Transforms Markus Flierl and Bernd Girod Max

More information

6. H.261 Video Coding Standard

6. H.261 Video Coding Standard 6. H.261 Video Coding Standard ITU-T (formerly CCITT) H-Series of Recommendations 1. H.221 - Frame structure for a 64 to 1920 kbits/s channel in audiovisual teleservices 2. H.230 - Frame synchronous control

More information

Redundancy Allocation Based on Weighted Mismatch-Rate Slope for Multiple Description Video Coding

Redundancy Allocation Based on Weighted Mismatch-Rate Slope for Multiple Description Video Coding 1 Redundancy Allocation Based on Weighted Mismatch-Rate Slope for Multiple Description Video Coding Mohammad Kazemi, Razib Iqbal, Shervin Shirmohammadi Abstract Multiple Description Coding (MDC) is a robust

More information

CSE 126 Multimedia Systems Midterm Exam (Form A)

CSE 126 Multimedia Systems Midterm Exam (Form A) University of California, San Diego Inst: Prof P. V. Rangan CSE 126 Multimedia Systems Midterm Exam (Form A) Spring 2003 Solution Assume the following input (before encoding) frame sequence (note that

More information

A Video Codec Incorporating Block-Based Multi-Hypothesis Motion-Compensated Prediction

A Video Codec Incorporating Block-Based Multi-Hypothesis Motion-Compensated Prediction SPIE Conference on Visual Communications and Image Processing, Perth, Australia, June 2000 1 A Video Codec Incorporating Block-Based Multi-Hypothesis Motion-Compensated Prediction Markus Flierl, Thomas

More information

Human Visual System Based Adaptive Inter Quantization

Human Visual System Based Adaptive Inter Quantization Human Visual System Based Adaptive Inter Quantization Jin Li 1, Jari Koivusaari 1,Jarma akala 1,Moncef Gabbouj 1 and Hexin Chen 2 Department of Information echnology, ampere University of echnology ampere,

More information

Proc. of NCC 2010, Chennai, India

Proc. of NCC 2010, Chennai, India Proc. of NCC 2010, Chennai, India Trajectory and surface modeling of LSF for low rate speech coding M. Deepak and Preeti Rao Department of Electrical Engineering Indian Institute of Technology, Bombay

More information

Modelling of produced bit rate through the percentage of null quantized transform coefficients ( zeros )

Modelling of produced bit rate through the percentage of null quantized transform coefficients ( zeros ) Rate control strategies in H264 Simone Milani (simone.milani@dei.unipd.it) with the collaboration of Università degli Studi di adova ST Microelectronics Summary General scheme of the H.264 encoder Rate

More information

Phase-Correlation Motion Estimation Yi Liang

Phase-Correlation Motion Estimation Yi Liang EE 392J Final Project Abstract Phase-Correlation Motion Estimation Yi Liang yiliang@stanford.edu Phase-correlation motion estimation is studied and implemented in this work, with its performance, efficiency

More information

System Model. Real-Time systems. Giuseppe Lipari. Scuola Superiore Sant Anna Pisa -Italy

System Model. Real-Time systems. Giuseppe Lipari. Scuola Superiore Sant Anna Pisa -Italy Real-Time systems System Model Giuseppe Lipari Scuola Superiore Sant Anna Pisa -Italy Corso di Sistemi in tempo reale Laurea Specialistica in Ingegneria dell Informazione Università di Pisa p. 1/?? Task

More information

IMPACTS OF SPATIAL OBSERVATION ERROR CORRELATION IN ATMOSPHERIC MOTION VECTORS ON DATA ASSIMILATION

IMPACTS OF SPATIAL OBSERVATION ERROR CORRELATION IN ATMOSPHERIC MOTION VECTORS ON DATA ASSIMILATION Proceedings for the 13 th International Winds Workshop 27 June - 1 July 2016, Monterey, California, USA IMPACTS OF SPATIAL OBSERVATION ERROR CORRELATION IN ATMOSPHERIC MOTION VECTORS ON DATA ASSIMILATION

More information

Rate-Constrained Multihypothesis Prediction for Motion-Compensated Video Compression

Rate-Constrained Multihypothesis Prediction for Motion-Compensated Video Compression IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL 12, NO 11, NOVEMBER 2002 957 Rate-Constrained Multihypothesis Prediction for Motion-Compensated Video Compression Markus Flierl, Student

More information

Design of IP networks with Quality of Service

Design of IP networks with Quality of Service Course of Multimedia Internet (Sub-course Reti Internet Multimediali ), AA 2010-2011 Prof. Pag. 1 Design of IP networks with Quality of Service 1 Course of Multimedia Internet (Sub-course Reti Internet

More information

Sound Recognition in Mixtures

Sound Recognition in Mixtures Sound Recognition in Mixtures Juhan Nam, Gautham J. Mysore 2, and Paris Smaragdis 2,3 Center for Computer Research in Music and Acoustics, Stanford University, 2 Advanced Technology Labs, Adobe Systems

More information

MODERN video coding standards, such as H.263, H.264,

MODERN video coding standards, such as H.263, H.264, 146 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 16, NO. 1, JANUARY 2006 Analysis of Multihypothesis Motion Compensated Prediction (MHMCP) for Robust Visual Communication Wei-Ying

More information

Perceptual Feedback in Multigrid Motion Estimation Using an Improved DCT Quantization

Perceptual Feedback in Multigrid Motion Estimation Using an Improved DCT Quantization IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 10, NO. 10, OCTOBER 2001 1411 Perceptual Feedback in Multigrid Motion Estimation Using an Improved DCT Quantization Jesús Malo, Juan Gutiérrez, I. Epifanio,

More information

LORD: LOw-complexity, Rate-controlled, Distributed video coding system

LORD: LOw-complexity, Rate-controlled, Distributed video coding system LORD: LOw-complexity, Rate-controlled, Distributed video coding system Rami Cohen and David Malah Signal and Image Processing Lab Department of Electrical Engineering Technion - Israel Institute of Technology

More information

SSIM-Inspired Perceptual Video Coding for HEVC

SSIM-Inspired Perceptual Video Coding for HEVC 2012 IEEE International Conference on Multimedia and Expo SSIM-Inspired Perceptual Video Coding for HEVC Abdul Rehman and Zhou Wang Dept. of Electrical and Computer Engineering, University of Waterloo,

More information

Fast Frame-Based Scene Change Detection in the Compressed Domain for MPEG-4 Video

Fast Frame-Based Scene Change Detection in the Compressed Domain for MPEG-4 Video Fast Frame-Based Scene Change Detection in the Compressed Domain for MPEG-4 Video Jens Brandt, Jens Trotzky, Lars Wolf IBR Technische Universität Braunschweig Germany Future Multimedia Networking - FMN

More information

Optimal Media Streaming in a Rate-Distortion Sense For Guaranteed Service Networks

Optimal Media Streaming in a Rate-Distortion Sense For Guaranteed Service Networks Optimal Media Streaming in a Rate-Distortion Sense For Guaranteed Service Networks Olivier Verscheure and Pascal Frossard Jean-Yves Le Boudec IBM Watson Research Center Swiss Federal Institute of Tech.

More information

New image-quality measure based on wavelets

New image-quality measure based on wavelets Journal of Electronic Imaging 19(1), 118 (Jan Mar 2) New image-quality measure based on wavelets Emil Dumic Sonja Grgic Mislav Grgic University of Zagreb Faculty of Electrical Engineering and Computing

More information

A Systematic Description of Source Significance Information

A Systematic Description of Source Significance Information A Systematic Description of Source Significance Information Norbert Goertz Institute for Digital Communications School of Engineering and Electronics The University of Edinburgh Mayfield Rd., Edinburgh

More information

Title: Quantifying Response Time in Carbon Monoxide Monitors to Produce High-

Title: Quantifying Response Time in Carbon Monoxide Monitors to Produce High- Title: Quantifying Response Time in Carbon Monoxide Monitors to Produce High- Resolution Measurements Authors: D.B. Young, R.M. Jones Division of Environmental and Occupational Health Sciences School of

More information

Enhanced SATD-based cost function for mode selection of H.264/AVC intra coding

Enhanced SATD-based cost function for mode selection of H.264/AVC intra coding SIViP (013) 7:777 786 DOI 10.1007/s11760-011-067-z ORIGINAL PAPER Enhanced SATD-based cost function for mode selection of H.6/AVC intra coding Mohammed Golam Sarwer Q. M. Jonathan Wu Xiao-Ping Zhang Received:

More information

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO

INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND AUDIO INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC9/WG11 CODING OF MOVING PICTURES AND AUDIO ISO/IEC JTC1/SC9/WG11 MPEG 98/M3833 July 1998 Source:

More information

Video Coding With Linear Compensation (VCLC)

Video Coding With Linear Compensation (VCLC) Coding With Linear Compensation () Arif Mahmood Zartash Afzal Uzmi Sohaib Khan School of Science and Engineering Lahore University of Management Sciences, Lahore, Pakistan {arifm, zartash, sohaib}@lums.edu.pk

More information

WaveNet: A Generative Model for Raw Audio

WaveNet: A Generative Model for Raw Audio WaveNet: A Generative Model for Raw Audio Ido Guy & Daniel Brodeski Deep Learning Seminar 2017 TAU Outline Introduction WaveNet Experiments Introduction WaveNet is a deep generative model of raw audio

More information

Part 2: Video Coding Techniques

Part 2: Video Coding Techniques art 2: Video Coding Techniques Outline Vincent Roca and Christoph Neumann {firstname.name}@inrialpes.fr lanète project; INRIA Rhône-Alpes MIS 03, Napoli, November 2003 Copyright 2003, INRIA; all rights

More information

Distributed Arithmetic Coding

Distributed Arithmetic Coding Distributed Arithmetic Coding Marco Grangetto, Member, IEEE, Enrico Magli, Member, IEEE, Gabriella Olmo, Senior Member, IEEE Abstract We propose a distributed binary arithmetic coder for Slepian-Wolf coding

More information

Convolutional Codes Klaus von der Heide

Convolutional Codes Klaus von der Heide Convolutional Codes Klaus von der Heide Convolutional codes encode a stream of symbols into n streams of symbols. 1/n = R is called the code rate. A second important parameter is the constraint length

More information

Vector Quantization Encoder Decoder Original Form image Minimize distortion Table Channel Image Vectors Look-up (X, X i ) X may be a block of l

Vector Quantization Encoder Decoder Original Form image Minimize distortion Table Channel Image Vectors Look-up (X, X i ) X may be a block of l Vector Quantization Encoder Decoder Original Image Form image Vectors X Minimize distortion k k Table X^ k Channel d(x, X^ Look-up i ) X may be a block of l m image or X=( r, g, b ), or a block of DCT

More information

THE newest video coding standard is known as H.264/AVC

THE newest video coding standard is known as H.264/AVC IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 17, NO. 6, JUNE 2007 765 Transform-Domain Fast Sum of the Squared Difference Computation for H.264/AVC Rate-Distortion Optimization

More information

Estimation-Theoretic Delayed Decoding of Predictively Encoded Video Sequences

Estimation-Theoretic Delayed Decoding of Predictively Encoded Video Sequences Estimation-Theoretic Delayed Decoding of Predictively Encoded Video Sequences Jingning Han, Vinay Melkote, and Kenneth Rose Department of Electrical and Computer Engineering University of California, Santa

More information

Convergence of Time Decay for Event Weights

Convergence of Time Decay for Event Weights Convergence of Time Decay for Event Weights Sharon Simmons and Dennis Edwards Department of Computer Science, University of West Florida 11000 University Parkway, Pensacola, FL, USA Abstract Events of

More information

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS Michael A. Lexa and Don H. Johnson Rice University Department of Electrical and Computer Engineering Houston, TX 775-892 amlexa@rice.edu,

More information

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments Dr. Jian Zhang Conjoint Associate Professor NICTA & CSE UNSW COMP9519 Multimedia Systems S2 2006 jzhang@cse.unsw.edu.au

More information

STATISTICS FOR EFFICIENT LINEAR AND NON-LINEAR PICTURE ENCODING

STATISTICS FOR EFFICIENT LINEAR AND NON-LINEAR PICTURE ENCODING STATISTICS FOR EFFICIENT LINEAR AND NON-LINEAR PICTURE ENCODING Item Type text; Proceedings Authors Kummerow, Thomas Publisher International Foundation for Telemetering Journal International Telemetering

More information

Math 345: Applied Mathematics VIII: The Wobble Problem

Math 345: Applied Mathematics VIII: The Wobble Problem Math 345: Applied Mathematics VIII: The Wobble Problem Marcus Pendergrass November 12, 2008 Old audio recordings often suffer from so-called wow and flutter distortion. On a tape player, for instance,

More information

Motion Vector Prediction With Reference Frame Consideration

Motion Vector Prediction With Reference Frame Consideration Motion Vector Prediction With Reference Frame Consideration Alexis M. Tourapis *a, Feng Wu b, Shipeng Li b a Thomson Corporate Research, 2 Independence Way, Princeton, NJ, USA 855 b Microsoft Research

More information

Introduction to statistics

Introduction to statistics Introduction to statistics Literature Raj Jain: The Art of Computer Systems Performance Analysis, John Wiley Schickinger, Steger: Diskrete Strukturen Band 2, Springer David Lilja: Measuring Computer Performance:

More information

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments. Tutorial 1. Acknowledgement and References for lectures 1 to 5

Lecture 2: Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments. Tutorial 1. Acknowledgement and References for lectures 1 to 5 Lecture : Introduction to Audio, Video & Image Coding Techniques (I) -- Fundaments Dr. Jian Zhang Conjoint Associate Professor NICTA & CSE UNSW COMP959 Multimedia Systems S 006 jzhang@cse.unsw.edu.au Acknowledgement

More information

Evaluation of H.265 and H.264 for Panoramas Video under Different Map Projections

Evaluation of H.265 and H.264 for Panoramas Video under Different Map Projections Evaluation of H.265 and H.264 for Panoramas Video under Different Map Projections Bijia Li 1,2, Li Song 1,2, Rong Xie 1,2, Nam Ling 3 1Institute of Image Communication and Network Engineering, Shanghai

More information

A Complete Video Coding Chain Based on Multi-Dimensional Discrete Cosine Transform

A Complete Video Coding Chain Based on Multi-Dimensional Discrete Cosine Transform RADIOENGINEERING, VOL. 19, NO. 3, SEPTEMBER 2010 421 A Complete Video Coding Chain Based on Multi-Dimensional Discrete Cosine Transform Tomas FRYZA Department of Radio Electronics, Brno University of Technology,

More information

Capacity Region of the Permutation Channel

Capacity Region of the Permutation Channel Capacity Region of the Permutation Channel John MacLaren Walsh and Steven Weber Abstract We discuss the capacity region of a degraded broadcast channel (DBC) formed from a channel that randomly permutes

More information

Image Filtering. Slides, adapted from. Steve Seitz and Rick Szeliski, U.Washington

Image Filtering. Slides, adapted from. Steve Seitz and Rick Szeliski, U.Washington Image Filtering Slides, adapted from Steve Seitz and Rick Szeliski, U.Washington The power of blur All is Vanity by Charles Allen Gillbert (1873-1929) Harmon LD & JuleszB (1973) The recognition of faces.

More information

Lecture 7 Predictive Coding & Quantization

Lecture 7 Predictive Coding & Quantization Shujun LI (李树钧): INF-10845-20091 Multimedia Coding Lecture 7 Predictive Coding & Quantization June 3, 2009 Outline Predictive Coding Motion Estimation and Compensation Context-Based Coding Quantization

More information

Running head: ANATOMY OF A JERK 1

Running head: ANATOMY OF A JERK 1 Running head: ANATOMY OF A JERK 1 Anatomy of a Jerk: A Dissertation on the Effects of the Third Derivative on Graphs in the Two Dimensional Cartesian Plane Michael Xing Amy Goodrum and Anna Hunt Summer

More information

CS Sampling and Aliasing. Analog vs Digital

CS Sampling and Aliasing. Analog vs Digital CS 112 - Sampling and Aliasing Aditi Majumder, CS 112, Winter 2007 Slide 1 Analog vs Digital God has created the world analog Man has created digital world Aditi Majumder, CS 112, Winter 2007 Slide 2 1

More information

Compression and Coding

Compression and Coding Compression and Coding Theory and Applications Part 1: Fundamentals Gloria Menegaz 1 Transmitter (Encoder) What is the problem? Receiver (Decoder) Transformation information unit Channel Ordering (significance)

More information

SPEECH ANALYSIS AND SYNTHESIS

SPEECH ANALYSIS AND SYNTHESIS 16 Chapter 2 SPEECH ANALYSIS AND SYNTHESIS 2.1 INTRODUCTION: Speech signal analysis is used to characterize the spectral information of an input speech signal. Speech signal analysis [52-53] techniques

More information

Variable Length Codes for Degraded Broadcast Channels

Variable Length Codes for Degraded Broadcast Channels Variable Length Codes for Degraded Broadcast Channels Stéphane Musy School of Computer and Communication Sciences, EPFL CH-1015 Lausanne, Switzerland Email: stephane.musy@ep.ch Abstract This paper investigates

More information

CSE 408 Multimedia Information System Yezhou Yang

CSE 408 Multimedia Information System Yezhou Yang Image and Video Compression CSE 408 Multimedia Information System Yezhou Yang Lots of slides from Hassan Mansour Class plan Today: Project 2 roundup Today: Image and Video compression Nov 10: final project

More information

Multiuser Successive Refinement and Multiple Description Coding

Multiuser Successive Refinement and Multiple Description Coding Multiuser Successive Refinement and Multiple Description Coding Chao Tian Laboratory for Information and Communication Systems (LICOS) School of Computer and Communication Sciences EPFL Lausanne Switzerland

More information

+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing

+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing Linear recurrent networks Simpler, much more amenable to analytic treatment E.g. by choosing + ( + ) = Firing rates can be negative Approximates dynamics around fixed point Approximation often reasonable

More information

Intra Frame Coding for Advanced Video Coding Standard to reduce Bitrate and obtain consistent PSNR Using Gaussian Pulse

Intra Frame Coding for Advanced Video Coding Standard to reduce Bitrate and obtain consistent PSNR Using Gaussian Pulse Intra Frame Coding for Advanced Video Coding Standard to reduce Bitrate and obtain consistent PSNR Using Gaussian Pulse Manjanaik.N1, Dr.Manjunath.R2 Electronics and Communication Engineering, Jain University,

More information

On Scalable Coding in the Presence of Decoder Side Information

On Scalable Coding in the Presence of Decoder Side Information On Scalable Coding in the Presence of Decoder Side Information Emrah Akyol, Urbashi Mitra Dep. of Electrical Eng. USC, CA, US Email: {eakyol, ubli}@usc.edu Ertem Tuncel Dep. of Electrical Eng. UC Riverside,

More information

Animation Curves and Splines 1

Animation Curves and Splines 1 Animation Curves and Splines 1 Animation Homework Set up a simple avatar E.g. cube/sphere (or square/circle if 2D) Specify some key frames (positions/orientations) Associate a time with each key frame

More information

University of Siena. Multimedia Security. Watermark extraction. Mauro Barni University of Siena. M. Barni, University of Siena

University of Siena. Multimedia Security. Watermark extraction. Mauro Barni University of Siena. M. Barni, University of Siena Multimedia Security Mauro Barni University of Siena : summary Optimum decoding/detection Additive SS watermarks Decoding/detection of QIM watermarks The dilemma of de-synchronization attacks Geometric

More information

I. Pre-Lab Introduction

I. Pre-Lab Introduction I. Pre-Lab Introduction Please complete the following pages before the lab by filling in the requested items. A. Atomic notation: Atoms are composed of a nucleus containing neutrons and protons surrounded

More information

MDS codec evaluation based on perceptual sound attributes

MDS codec evaluation based on perceptual sound attributes MDS codec evaluation based on perceptual sound attributes Marcelo Herrera Martínez * Edwar Jacinto Gómez ** Edilberto Carlos Vivas G. *** submitted date: March 03 received date: April 03 accepted date:

More information

3drs e3drs fs e3drs fs Rate (kbps) Mother and Daughter (b) Miss America (a) 140.

3drs e3drs fs e3drs fs Rate (kbps) Mother and Daughter (b) Miss America (a) 140. oise{robust Recursive otion Estimation for H.263{based videoconferencing systems Stefano Olivieriy, Gerard de Haan z, and Luigi Albaniy yphilips S.p.A, Philips Research onza Via Philips, 12, 252 onza (I,

More information

Dominant Feature Vectors Based Audio Similarity Measure

Dominant Feature Vectors Based Audio Similarity Measure Dominant Feature Vectors Based Audio Similarity Measure Jing Gu 1, Lie Lu 2, Rui Cai 3, Hong-Jiang Zhang 2, and Jian Yang 1 1 Dept. of Electronic Engineering, Tsinghua Univ., Beijing, 100084, China 2 Microsoft

More information

Resource Allocation for Video Streaming in Wireless Environment

Resource Allocation for Video Streaming in Wireless Environment Resource Allocation for Video Streaming in Wireless Environment Shahrokh Valaee and Jean-Charles Gregoire Abstract This paper focuses on the development of a new resource allocation scheme for video streaming

More information

Fast Intra Coding Method of H.264 for Video Surveillance System

Fast Intra Coding Method of H.264 for Video Surveillance System 76 IJCSNS International Journal of Computer Science and Network Security, OL.7 No.10, October 2007 Fast Intra Coding Method of.264 for ideo Surveillance System Jang-Seon Ryu and Eung-Tea Kim, Korea Polytechnic

More information

Low Bit-Rate Speech Codec Based on a Long-Term Harmonic Plus Noise Model

Low Bit-Rate Speech Codec Based on a Long-Term Harmonic Plus Noise Model PAPERS Journal of the Audio Engineering Society Vol. 64, No. 11, November 216 ( C 216) DOI: https://doi.org/1.17743/jaes.216.28 Low Bit-Rate Speech Codec Based on a Long-Term Harmonic Plus Noise Model

More information

REED-SOLOMON CODE SYMBOL AVOIDANCE

REED-SOLOMON CODE SYMBOL AVOIDANCE Vol105(1) March 2014 SOUTH AFRICAN INSTITUTE OF ELECTRICAL ENGINEERS 13 REED-SOLOMON CODE SYMBOL AVOIDANCE T Shongwe and A J Han Vinck Department of Electrical and Electronic Engineering Science, University

More information

18/10/2017. Image Enhancement in the Spatial Domain: Gray-level transforms. Image Enhancement in the Spatial Domain: Gray-level transforms

18/10/2017. Image Enhancement in the Spatial Domain: Gray-level transforms. Image Enhancement in the Spatial Domain: Gray-level transforms Gray-level transforms Gray-level transforms Generic, possibly nonlinear, pointwise operator (intensity mapping, gray-level transformation): Basic gray-level transformations: Negative: s L 1 r Generic log:

More information

The Choice of MPEG-4 AAC encoding parameters as a direct function of the perceptual entropy of the audio signal

The Choice of MPEG-4 AAC encoding parameters as a direct function of the perceptual entropy of the audio signal The Choice of MPEG-4 AAC encoding parameters as a direct function of the perceptual entropy of the audio signal Claus Bauer, Mark Vinton Abstract This paper proposes a new procedure of lowcomplexity to

More information

Lifting Parameterisation of the 9/7 Wavelet Filter Bank and its Application in Lossless Image Compression

Lifting Parameterisation of the 9/7 Wavelet Filter Bank and its Application in Lossless Image Compression Lifting Parameterisation of the 9/7 Wavelet Filter Bank and its Application in Lossless Image Compression TILO STRUTZ Deutsche Telekom AG, Hochschule für Telekommunikation Institute of Communications Engineering

More information

Digital communication system. Shannon s separation principle

Digital communication system. Shannon s separation principle Digital communication system Representation of the source signal by a stream of (binary) symbols Adaptation to the properties of the transmission channel information source source coder channel coder modulation

More information

Scheduling of Frame-based Embedded Systems with Rechargeable Batteries

Scheduling of Frame-based Embedded Systems with Rechargeable Batteries Scheduling of Frame-based Embedded Systems with Rechargeable Batteries André Allavena Computer Science Department Cornell University Ithaca, NY 14853 andre@cs.cornell.edu Daniel Mossé Department of Computer

More information

Modeling the Time-varying Subjective Quality of HTTP Video Streams with Rate Adaptations

Modeling the Time-varying Subjective Quality of HTTP Video Streams with Rate Adaptations Modeling the Time-varying Subjective Quality of HTTP Video Streams with Rate Adaptations Chao Chen, Student Member, IEEE, Lark Kwon Choi, Student Member, IEEE, Gustavo de Veciana, Fellow, IEEE, Constantine

More information

Principles of Communications

Principles of Communications Principles of Communications Chapter V: Representation and Transmission of Baseband Digital Signal Yongchao Wang Email: ychwang@mail.xidian.edu.cn Xidian University State Key Lab. on ISN November 18, 2012

More information

Rate Adaptation for Buffer Underflow Avoidance in Multimedia Signal Streaming

Rate Adaptation for Buffer Underflow Avoidance in Multimedia Signal Streaming Rate Adaptation for Buffer Underflow Avoidance in Multimedia Signal Streaming Matteo Petracca, Fabio De Vito, Juan Carlos De Martin Dipartimento di Automatica e Informatica Politecnico di Torino C.so Duca

More information

Introduction to Uncertainty and Treatment of Data

Introduction to Uncertainty and Treatment of Data Introduction to Uncertainty and Treatment of Data Introduction The purpose of this experiment is to familiarize the student with some of the instruments used in making measurements in the physics laboratory,

More information

Behaviour Recognition using the Event Calculus

Behaviour Recognition using the Event Calculus Behaviour Recognition using the Event Calculus Alexander Artikis and George Paliouras Abstract We present a system for recognising human behaviour given a symbolic representation of surveillance videos.

More information

Measuring Deterministic Jitter with a K28.5 Pattern and an Oscilloscope

Measuring Deterministic Jitter with a K28.5 Pattern and an Oscilloscope Application Note: HFAN-4.5.0 Rev1; 04/08 Measuring Deterministic Jitter with a K28.5 Pattern and an Oscilloscope [A version of this application note has been published in the July, 2002 issue of Communications

More information

Predictive Coding. Prediction Prediction in Images

Predictive Coding. Prediction Prediction in Images Prediction Prediction in Images Predictive Coding Principle of Differential Pulse Code Modulation (DPCM) DPCM and entropy-constrained scalar quantization DPCM and transmission errors Adaptive intra-interframe

More information

CS578- Speech Signal Processing

CS578- Speech Signal Processing CS578- Speech Signal Processing Lecture 7: Speech Coding Yannis Stylianou University of Crete, Computer Science Dept., Multimedia Informatics Lab yannis@csd.uoc.gr Univ. of Crete Outline 1 Introduction

More information

Validation Report for Precipitation products from Cloud Physical Properties (PPh-PGE14: PCPh v1.0 & CRPh v1.0)

Validation Report for Precipitation products from Cloud Physical Properties (PPh-PGE14: PCPh v1.0 & CRPh v1.0) Page: 1/26 Validation Report for Precipitation SAF/NWC/CDOP2/INM/SCI/VR/15, Issue 1, Rev. 0 15 July 2013 Applicable to SAFNWC/MSG version 2013 Prepared by AEMET Page: 2/26 REPORT SIGNATURE TABLE Function

More information

Old painting digital color restoration

Old painting digital color restoration Old painting digital color restoration Michail Pappas Ioannis Pitas Dept. of Informatics, Aristotle University of Thessaloniki GR-54643 Thessaloniki, Greece Abstract Many old paintings suffer from the

More information

Predictive Coding. Prediction

Predictive Coding. Prediction Predictive Coding Prediction Prediction in Images Principle of Differential Pulse Code Modulation (DPCM) DPCM and entropy-constrained scalar quantization DPCM and transmission errors Adaptive intra-interframe

More information

Expected Performance From WIYN Tip-Tilt Imaging

Expected Performance From WIYN Tip-Tilt Imaging Expected Performance From WIYN Tip-Tilt Imaging C. F. Claver 3 September 1997 Overview Image motion studies done at WIYN show that a significant improvement to delivered image quality can be obtained from

More information

AN ADAPTIVE PERCEPTUAL QUANTIZATION METHOD FOR HDR VIDEO CODING

AN ADAPTIVE PERCEPTUAL QUANTIZATION METHOD FOR HDR VIDEO CODING AN ADAPTIVE PERCEPTUAL QUANTIZATION METHOD FOR HDR VIDEO CODING Y. Liu, N. Sidaty, W. Hamidouche, O. Déforges IETR Lab, CNRS 6164 INSA de Rennes, France G. Valenzise 1 and E. Zerman 2 1 L2S UMR 8506 CNRS,

More information

Multi-Hypothesis based Distributed Video Coding using LDPC Codes

Multi-Hypothesis based Distributed Video Coding using LDPC Codes Multi-Hypothesis based Distributed Video Coding using LDPC Codes Kiran Misra, Shirish Karande, Hayder Radha Department of Electrical and Computer Engineering 2120, Engineering Building Michigan State University

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER 2006 5177 The Distributed Karhunen Loève Transform Michael Gastpar, Member, IEEE, Pier Luigi Dragotti, Member, IEEE, and Martin Vetterli,

More information

Reduce the amount of data required to represent a given quantity of information Data vs information R = 1 1 C

Reduce the amount of data required to represent a given quantity of information Data vs information R = 1 1 C Image Compression Background Reduce the amount of data to represent a digital image Storage and transmission Consider the live streaming of a movie at standard definition video A color frame is 720 480

More information