Identification and separation of noises with spectro-temporal patterns

Similar documents
NMF WITH SPECTRAL AND TEMPORAL CONTINUITY CRITERIA FOR MONAURAL SOUND SOURCE SEPARATION. Julian M. Becker, Christian Sohn and Christian Rohlfing

Non-negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs

Time-Series Analysis for Ear-Related and Psychoacoustic Metrics

Non-Negative Matrix Factorization And Its Application to Audio. Tuomas Virtanen Tampere University of Technology

Convention Paper Presented at the 128th Convention 2010 May London, UK

SINGLE CHANNEL SPEECH MUSIC SEPARATION USING NONNEGATIVE MATRIX FACTORIZATION AND SPECTRAL MASKS. Emad M. Grais and Hakan Erdogan

Nonnegative Matrix Factor 2-D Deconvolution for Blind Single Channel Source Separation

On Spectral Basis Selection for Single Channel Polyphonic Music Separation

Analysis of polyphonic audio using source-filter model and non-negative matrix factorization

Single Channel Music Sound Separation Based on Spectrogram Decomposition and Note Classification

ESTIMATING TRAFFIC NOISE LEVELS USING ACOUSTIC MONITORING: A PRELIMINARY STUDY

Convolutive Non-Negative Matrix Factorization for CQT Transform using Itakura-Saito Divergence

Nonnegative Matrix Factorization with Markov-Chained Bases for Modeling Time-Varying Patterns in Music Spectrograms

MULTI-RESOLUTION SIGNAL DECOMPOSITION WITH TIME-DOMAIN SPECTROGRAM FACTORIZATION. Hirokazu Kameoka

A Nonlinear Psychoacoustic Model Applied to the ISO MPEG Layer 3 Coder

CONVOLUTIVE NON-NEGATIVE MATRIX FACTORISATION WITH SPARSENESS CONSTRAINT

Gaussian Processes for Audio Feature Extraction

Generalized Constraints for NMF with Application to Informed Source Separation

Non-negative Matrix Factorization: Algorithms, Extensions and Applications

SOUND SOURCE SEPARATION BASED ON NON-NEGATIVE TENSOR FACTORIZATION INCORPORATING SPATIAL CUE AS PRIOR KNOWLEDGE

ORTHOGONALITY-REGULARIZED MASKED NMF FOR LEARNING ON WEAKLY LABELED AUDIO DATA. Iwona Sobieraj, Lucas Rencker, Mark D. Plumbley

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement

COMP 546. Lecture 21. Cochlea to brain, Source Localization. Tues. April 3, 2018

Response-Field Dynamics in the Auditory Pathway

LONG-TERM REVERBERATION MODELING FOR UNDER-DETERMINED AUDIO SOURCE SEPARATION WITH APPLICATION TO VOCAL MELODY EXTRACTION.

Automatic Rank Determination in Projective Nonnegative Matrix Factorization

REVIEW OF SINGLE CHANNEL SOURCE SEPARATION TECHNIQUES

Non-Negative Tensor Factorisation for Sound Source Separation

744 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 4, MAY 2011

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION

L6: Short-time Fourier analysis and synthesis

Discovering Convolutive Speech Phones using Sparseness and Non-Negativity Constraints

Matrix Factorization for Speech Enhancement

ACCOUNTING FOR PHASE CANCELLATIONS IN NON-NEGATIVE MATRIX FACTORIZATION USING WEIGHTED DISTANCES. Sebastian Ewert Mark D. Plumbley Mark Sandler

Sparseness Constraints on Nonnegative Tensor Decomposition

Nonnegative Tensor Factorization with Smoothness Constraints

Feature extraction 1

POLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS

Lecture 7: Pitch and Chord (2) HMM, pitch detection functions. Li Su 2016/03/31

Dept. Electronics and Electrical Engineering, Keio University, Japan. NTT Communication Science Laboratories, NTT Corporation, Japan.

NONNEGATIVE MATRIX FACTORIZATION WITH TRANSFORM LEARNING. Dylan Fagot, Herwig Wendt and Cédric Févotte

Combination Tones As a Phenomenon of the Central Level in Auditory System

Non-Negative Matrix Factorization with Quasi-Newton Optimization

David Weenink. First semester 2007

LECTURE NOTES IN AUDIO ANALYSIS: PITCH ESTIMATION FOR DUMMIES

COMP 546. Lecture 20. Head and Ear. Thurs. March 29, 2018

TIME-DEPENDENT PARAMETRIC AND HARMONIC TEMPLATES IN NON-NEGATIVE MATRIX FACTORIZATION

FACTORS IN FACTORIZATION: DOES BETTER AUDIO SOURCE SEPARATION IMPLY BETTER POLYPHONIC MUSIC TRANSCRIPTION?

MDS codec evaluation based on perceptual sound attributes

Non-negative matrix factorization and term structure of interest rates

IN real-world audio signals, several sound sources are usually

Sound Recognition in Mixtures

Dominant Feature Vectors Based Audio Similarity Measure

OPTIMAL WEIGHT LEARNING FOR COUPLED TENSOR FACTORIZATION WITH MIXED DIVERGENCES

Design Criteria for the Quadratically Interpolated FFT Method (I): Bias due to Interpolation

Last time: small acoustics

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

SPEECH ANALYSIS AND SYNTHESIS

Independent Component Analysis and Unsupervised Learning

c Springer, Reprinted with permission.

Environmental Sound Classification in Realistic Situations

A COMPUTATIONAL SOFTWARE FOR NOISE MEASUREMENT AND TOWARD ITS IDENTIFICATION

An Improved Cumulant Based Method for Independent Component Analysis

Speech Signal Representations

Deep NMF for Speech Separation

t Tao Group Limited, Reading, RG6 IAZ, U.K. wenwu.wang(ieee.org t Samsung Electronics Research Institute, Staines, TW1 8 4QE, U.K.

ACOUSTIC SCENE CLASSIFICATION WITH MATRIX FACTORIZATION FOR UNSUPERVISED FEATURE LEARNING. Victor Bisot, Romain Serizel, Slim Essid, Gaël Richard

THE task of identifying the environment in which a sound

Linear and Non-Linear Responses to Dynamic Broad-Band Spectra in Primary Auditory Cortex

Covariance smoothing and consistent Wiener filtering for artifact reduction in audio source separation

A METHOD OF ICA IN TIME-FREQUENCY DOMAIN

ACOUSTIC CLARITY AND AUDITORY ROOM SIZE PERCEPTION. Densil Cabrera 1. Sydney, NSW 2006, Australia

CEPSTRAL ANALYSIS SYNTHESIS ON THE MEL FREQUENCY SCALE, AND AN ADAPTATIVE ALGORITHM FOR IT.

Characterisation of the directionality of reflections in small room acoustics

DIMENSIONALITY REDUCTION METHODS IN INDEPENDENT SUBSPACE ANALYSIS FOR SIGNAL DETECTION. Mijail Guillemard, Armin Iske, Sara Krause-Solberg

Cochlear modeling and its role in human speech recognition

Measurement of Temporal and Spatial Factors of a Flushing Toilet Noise in a Downstairs Bedroom

Speech enhancement based on nonnegative matrix factorization with mixed group sparsity constraint

TIME-DEPENDENT PARAMETRIC AND HARMONIC TEMPLATES IN NON-NEGATIVE MATRIX FACTORIZATION

EUSIPCO

COLLABORATIVE SPEECH DEREVERBERATION: REGULARIZED TENSOR FACTORIZATION FOR CROWDSOURCED MULTI-CHANNEL RECORDINGS. Sanna Wager, Minje Kim

The effect of boundary shape to acoustic parameters

CS229 Project: Musical Alignment Discovery

Topic 6. Timbre Representations

Independent Component Analysis and Unsupervised Learning. Jen-Tzung Chien

Sound Engineering Test and Analysis

Acoustic Quantities. LMS Test.Lab. Rev 12A

Sparse and Shift-Invariant Feature Extraction From Non-Negative Data

CRYSTALLIZATION SONIFICATION OF HIGH-DIMENSIONAL DATASETS

On monotonicity of multiplicative update rules for weighted nonnegative tensor factorization

Author manuscript, published in "7th International Symposium on Computer Music Modeling and Retrieval (CMMR 2010), Málaga : Spain (2010)"

Short Term Memory and Pattern Matching with Simple Echo State Networks

Single Channel Signal Separation Using MAP-based Subspace Decomposition

Acoustic MIMO Signal Processing

Linear Prediction 1 / 41

Multiresolution Analysis

Bayesian Estimation of Time-Frequency Coefficients for Audio Signal Enhancement

1 Wind Turbine Acoustics. Wind turbines generate sound by both mechanical and aerodynamic

University of Colorado at Boulder ECEN 4/5532. Lab 2 Lab report due on February 16, 2015

B-Spline Smoothing of Feature Vectors in Nonnegative Matrix Factorization

Transcription:

PROCEEDINGS of the 22 nd International Congress on Acoustics Soundscape, Psychoacoustics and Urban Environment: Paper ICA2016-532 Identification and separation of noises with spectro-temporal patterns Roland Sottek HEAD acoustics GmbH, Ebertstr. 30a, 52134 Herzogenrath, Germany, roland.sottek@head-acoustics.de Abstract Acoustic signals often contain perceptually detectable noise patterns with spectro-temporal structures, causing sensations like roughness (due to modulated signal components) and tonality. Technical sounds or environmental noises are often composed of several such components. It is assumed that perceptual evaluations of such complex scenarios show larger deviations because test participants concentrate on different components depending on their preference. Therefore, it is desirable to identify and possibly separate these components allowing for an investigation of each individual noise pattern. The goal is to recognize the composition of all components corresponding to their pitch and modulation rate. Such information could be used for further development and improvement of calculation methods for psychoacoustic parameters. This paper presents different approaches based on time-frequency analyses as well as on the hearing model of Sottek evaluating a three-dimensional autocorrelation analysis as a function of time, lag, and frequency band. The extension to the third dimension allows for a better consideration of modulated signals. Keywords: Blind source separation, time-frequency analyses, hearing model

Identification and separation of noises with spectrotemporal patterns 1 Introduction Acoustic signals often contain noise components with spectro-temporal structures like modulated sounds. These perceptually detectable noise patterns may cause sensations, for instance roughness and tonality. Technical sounds or environmental noises are often composed of several such components. Under these circumstances, preliminary listening test results show more variations compared to tests with single modulated components, indicating that test subjects might be focusing on different components [1]. Thus, it would be desirable to simulate this behavior by an automatic procedure to identify and evaluate each component separately, and, if possible, even to separate the sound of interest into the different components as time signals for further processing, e.g., for listening. It is often assumed that the overall sound can be represented as a linear combination of unknown components with unknown weights. There are several methods available for such a blind source separation using different constraints. For example, the Independent Subspace Analysis (ISA) [2] estimates the different components using the spectrogram (magnitude of Short-Time Fourier Transform: STFT) of the mixture assuming statistically independent sources. An extended approach takes the complex spectrogram (magnitude and phase information, CISA) into account [3]. Another common method for blind source separation is the Nonnegative Matrix Factorization (NMF) using a timefrequency analysis like the STFT as a two-dimensional array to be factorized in a product of two nonnegative matrices containing the sources and the weights under the assumption of different constraints like sparseness of the matrices [4]. However, in all the above approaches, details about modulations often remain hidden. Therefore, a new method based on the three-dimensional autocorrelation analysis (dimensions: time, lag, and frequency band) of the hearing model of Sottek [5] is taken for the separation using a Nonnegative Tensor Factorization (NTF). The extension to the third dimension allows for a better consideration of modulated signals. 2 Hearing model of Sottek Figure 1 displays the basic hearing model structure for the separation of modulated components using a NTF. 2.1 Outer and middle ear filtering The pre-processing consists of filtering with outer and middle ear transfer functions: the input signal p(t) is processed by a filter corresponding to the ear s transmission characteristic a 0 of Zwicker s loudness model [6] in series with a high-pass filter of 1 st -order (attenuation of 6.5 db/decade for frequencies below 500 Hz). Thus, the outer and middle ear filter 2

approximates the inverted equal loudness contour at 100 phon. It considers, for instance, the loudness elevation due to the cavum conchae resonance of the ear near 4 khz. time signal p(t) outer and middle ear filtering auditory filter bank with i=1..n critical bands 1 i n formation of envelope autocorrelation function (ACF) consideration of specific loudness noise reduction spectral transform and weighting 1 i n modulation tensor NTF separated components Figure 1: Basic hearing model structure including the determination of the autocorrelation function as function of time, lag, and frequency band (see also Figure 2) for the separation of modulated components using a NTF 2.2 Auditory filtering bank An auditory filter bank consisting of overlapping asymmetric filters 1 models the frequencydependent critical bandwidths and the tuning curves of the frequency-to-place transform of the inner ear, which mediates the firing of the auditory hair cells as the traveling wave from an incoming sound event progresses along the basilar membrane. The inconstant ratio of bandwidth versus frequency of the auditory filter bank conveys a high frequency resolution at low frequencies and a high time resolution at high frequencies, with a very small product of time and frequency resolution at all frequencies, which empowers, for example, human hearing s recognition of short-duration low-frequency events. The impulse responses of the auditory filters are chosen as modulated low-pass filters (F is the center frequency of the corresponding critical band) 1 h F (t) = 2 ε(t) (n 1)! 1 τ(f) ( t n 1 τ(f) ) exp ( t ) exp (j2πft) (1) τ(f) where n is the filter order ( n = 5 is recommended) and τ(f) is a time constant (inversely proportional to the bandwidth of the critical band filter). 1 The degree of overlapping depends on the application: e. g., for roughness calculation an overlap of half a critical bandwidth was used, for loudness calculation even no overlap is required, and for a spectral deconvolution algorithm several hundreds of filters were used [5]. 3

2.3 Formation of envelope Subsequent rectification accounts for the fact that nerves fire only when the basilar membrane vibrates in a specific direction. The firing rates of the nerve cells are limited to a maximum frequency. This is modeled applying 3 rd -order low-pass filters with a cutoff frequency of 120 Hz, and leads to smooth envelope signals. 2 The sampling rate of the envelope signals can be decimated by a factor of 16, yielding a reduced rate of 3 khz for signals sampled at 48 khz. 2.4 Autocorrelation function (ACF) In early publications, Licklider assumed that human pitch perception is based on both spectral and temporal cues [7]. According to Licklider, the neuronal processing in human hearing applies a running autocorrelation analysis of the critical band signals. Under this assumption, psychoacoustic tonality phenomena like difference-tone perception or the missing fundamental phenomenon can be explained. This work inspired the idea to use the sliding autocorrelation function as a processing block in the hearing model for the calculation of roughness and fluctuation strength [5] and later for other psychoacoustic quantities like tonality [8]. freq. band i lag autocorrelation function at a particular time time t hearing model spectrum vs. time operation vs. freq. operation vs. lag Figure 2: Sliding autocorrelation function applied to the band-pass signals. The hearing model spectrum vs. time is accessible as ACF(τ = 0) as a function of time and frequency band. Different operations can be applied vs. lag and vs. frequency bands for further processing. Figure 2 displays an example of the sliding autocorrelation function as described in equation (2) φ ss (t, τ, i) = s i (t ) s i (t τ) w i(t τ)dt (2) where τ is the lag, and s i is the band-pass signal of the i th critical band. The weighting function w i(t) determines how the past signal values contribute to the actual value of the autocorrelation function. Defining w i (t) = w i( t), equation (2) can be expressed advantageously as a convolution product; w i (t) is taken as an impulse response of a 3 rd -order low-pass filter: φ ss (t, τ, i) = [s i (t) s i (t τ)] w i (t) (3) 2 Similarly, the envelope can be calculated as the magnitude of the analytic input signal using the Hilbert transform. 4

The calculation of the ACF according to equation (3) is highly efficient compared to the integral formulation in equation (2), but still time consuming. Therefore, the sliding ACF is calculated block-wise using the DFT to shorten computing time. In this case, the weighting of the past signal values is a little different. First, there is a low-pass effect due to averaging over the block length. Second, additional 3 rd -order low-pass filters are applied to the ACF time curve for each frequency band and lag of interest. The equivalent rectangular duration of the 3 rd -order low-pass filters amounts to 5 ms. For the evaluation of envelope fluctuations a DFT length of 512 is chosen (corresponding to a duration of approximately 170.7 ms at a sampling rate of 3 khz). The calculation is performed for temporally-overlapping blocks: An overlap of about 75% leads to approximately 24 ACF calculations per second. 2.5 Consideration of specific loudness The nonlinearity between specific loudness and sound pressure was reconsidered in the hearing model according to results of several listening tests [9]. Further improvements for higher levels above approximately 80 db were achieved by introducing a nonlinearity function according to equation (4) y(p ) = ( p v 0 ) (1 + ( p ) p 0 p ti N i=1 v i v i 1 α i α i ) (4) with ν 0 = 1 and N further exponents ν i. The thresholds p ti were set to 15 db, 25 db, 35 db, 45 db, 55 db, 65 db, 75 db, and 85 db; all α i were set to 1.5. The N = 8 exponents (ν 1 = 0.6602, ν 2 = 0.0864, ν 3 = 0.6384, ν 4 = 0.0328, ν 5 = 0.4068, ν 6 = 0.2082, ν 7 = 0.3994, and ν 8 = 0.6434) were achieved by applying a nonlinear-optimization procedure in order to minimize the mean squared error between the results of the loudness matching experiment and the results of the model calculation. The specific loudness is calculated by applying the nonlinearity to the square root of ACF(τ = 0), because equation (4) is defined for pressure input signals, whereas ACF(τ = 0) corresponds to the energy of the input signal. The specific loudness in each band is set to zero if a frequency-dependent specific loudness threshold is not exceeded. These threshold values are chosen to be the specific loudness values corresponding to the lower threshold of hearing for pure tones minus a fixed value (for the first attempt 6 db). This makes it possible to take into account loudness summation for complex sounds at threshold in combination with a lower threshold of total loudness. This simple but effective approach must be studied and perhaps optimized later together with the different weighting of the loudness of tonal and non-tonal components. 3 Afterwards, the entire ACF is normalized for each lag with the same factor, i.e., the ratio of the specific loudness and ACF(τ = 0). 3 Recent investigations showed that the existing loudness procedures underestimate the loudness of tonal signals [8]. 5

Frequency Frequency 22 nd International Congress on Acoustics, ICA 2016 2.6 Noise reduction In general, ACF(τ = 0) corresponds to the overall signal energy, i.e., noise and tonal energy. The tonal energy can be estimated, e.g., as the 99%-percentile value of the ACF in a certain lag range, thus the noise energy can be taken as the difference between ACF(τ = 0) and the tonal energy [8]. In order to reduce the noise, which has a strong impact to the ACF mainly at very small lags, the ACF is attenuated by applying a symmetric window function around τ = 0 (for τ 4 ms). The window function shows a minimum (tonal energy divided by ACF(τ = 0)) at τ = 0 and is increasing trigonometrically to a value of 1 at τ = 4 ms. 2.7 Spectral transform and weighting The ACF is an even function of the lag with a real-valued nonnegative spectrum. The spectrum of the ACF as a function of modulation rate is calculated by means of a DFT. Only half of the spectral data must be considered because of its symmetry: i.e., 256 values. The noise reduction described in section 2.6 may cause small negative values of the spectrum, which are set to zero. In order to simulate the perception of roughness, the modulation spectrum can be weighted with respect to modulation rate in conjunction with a loudness-dependent weighting factor for each frequency band. 3 Nonnegative matrix and tensor factorization 3.1 Nonnegative matrix factorization Many scientists have investigated NMF, but it became popular by the work of Lee and Seung [4] for the detection of patterns in two-dimensional data. The main purpose of NMF is to factorize a matrix V (dimension FxN) with nonnegative values into a product of two nonnegative matrices W (dimension FxK) and H (dimension KxN) of low rank (Figure 3) with K components, (each matrix contains one vector per component). Figure 3 shows as an example a matrix V (result of a time-frequency analysis) factorized into a product of two nonnegative matrices W and H, containing the frequency pattern and the time pattern, respectively. V W H = * Time Time Component 1 Component 2 Figure 3: NMF: factorization of a two-dimensional nonnegative matrix V into a product of two nonnegative matrices; W: frequency pattern (activation in frequency bands), H: time pattern (activation in time) of low rank; representation of a component by two vectors In general, there is no unique solution and the factorization process leads to an approximation V of the initial matrix V: 6

V V = W H (5) The factorization can be achieved by an iterative procedure minimizing a cost function D, e.g., a metric for the distance between V and V. 4 The overall distance D is the sum of all scalar (element-wise) distances d: F N D(V V ) = d(v(f, n) V (f, n)) f=1 n=1 (6) A typical measure for d is the β-divergence [10]-[12]: d β (x y) = { x xβ 1 y β 1 β 1 xβ y β β β R {0,1} x log ( x ) x + y y β = 1 x y log (x y ) 1 β = 0 (7) The distance metric for d β is known as Euclidian distance (scaled by 0.5) for β = 2, Kullback- Leibner distance for β = 1, and Itakura-Saito distance for β = 0. In case of the Euclidian distance small values of x and y lead to small values of d β=2. The Itakura-Saito distance metric is scale invariant (d β=0 (λ x λ y) = d β=0 (x y)), thus small and large values of x and y have the same relevance for d β=0. The Kullback-Leibner distance represents a compromise. The iterative factorization starts with the initialization of W and H, e.g., with nonnegative random numbers. Usually, different random initializations are tested, and the initial values showing the lowest value of D are selected as a starting point (i = 0) for the iterative procedure [13]. The update rules to minimize D are given by [11]: H i = H i 1 WT (((W H i 1 ). ^(β 2)). V) W T ((W H i 1 ). ^(β 1)) W i = W i 1 (((W i 1 H). ^(β 2)). V) H T ((W i 1 H). ^(β 1)) H T (8) 4 The cost function can be extended using additional constraints like sparseness. 7

Note that. ^,., and the fraction line denote element-wise exponentiation, multiplication, and division, respectively. The iteration stops if a certain criterion is fulfilled, e.g., if the difference of D between two iteration steps is below a given threshold. The factorization provides for each component a vector of W and a corresponding vector of H allowing for the generation of a submatrix (a spectrogram) for the component of interest. A disadvantage of NMF is that a-priori the number of relevant components must be known, but there are strategies available for an automatic detection of relevant components [14]. The separated components are often created using the Wiener reconstruction, an approach similar to Wiener filtering [11]. This method uses for all components the same phase information. Time signals can be resynthesized based on inverse transforms, e.g., in the case of a STFT by an ISTFT (Inverse Short-Time Fourier Transform). A Hann window is used for the synthesis in order to reduce artifacts due to the transition between successive blocks [1]. 3.2 Nonnegative tensor factorization The NTF is an extension of NMF for the detection of patterns in three-dimensional data. A tensor V (dimension FxNxI ) with nonnegative values is factorized into a product of three nonnegative matrices W (dimension FxK), H (dimension NxK), and Q (dimension IxK) of low rank (Figure 4) with K components (each matrix contains one vector per component). N K F N K I K F I V Figure 4: NTF: factorization of a three-dimensional nonnegative matrix V (tensor) into a product of three sparse nonnegative matrices; W: frequency pattern (activation in frequency bands), H: modulation pattern (activation in modulation rates), Q: time pattern (activation in time) Similar to the NMF method, the NTF factorizes the tensor by minimizing a cost function using also the β-divergence as defined in equation (7). Additionally, the multiplicative update rules can be extended to the three-dimensional case. For details on the algorithm, please refer to [15]. Figure 4 shows a scheme factorizing a three-dimensional nonnegative matrix V (tensor) in three sparse nonnegative matrices. In case of the presented autocorrelation analysis of the hearing model of Sottek (Figure 2), the three dimensions are frequency bands ( F = 24 for nonoverlapping critical bands or F = 47 if an overlap of half a critical bandwidth is used), modulation rates (N = 256 for a DFT length of 512), and time (I 24 T, T is the signal duration in s). In principle, time signals can also be resynthesized based on a Wiener reconstruction in case of the presented NTF approach. The first results are promising [1]. W H Q 8

4 Application example The number of the separated components for the following example was determined automatically using the algorithm described in [14]. The components were determined by means of minimizing the distance metric D based on the Kullback-Leibner distance d β=1. Figure 5 shows results obtained by NTF using the presented hearing model for a sound recorded from a hard disk drive. The several modulated components could not be resolved in the same way by NMF [1]. Comp. 1 Comp. 2 Comp. 3 Comp. 4 Comp. 5 Comp. 1 Comp. 2 Comp. 1 Comp. 2 Modulation rate / Hz Comp. 3 Comp. 4 Comp. 5 Comp. 3 Comp. 4 Comp. 5 Figure 5: NTF: factorization of hard disk noise; top left picture shows the activation with respect to modulation rate for the five found components. The other pictures show the spectrogram (upper part) and the modulation spectrum vs. critical bands (lower part) for each component. Modulated signal shares in different frequency bands that show the same pattern regarding modulation rates are not resolved and are considered as one component. This was tested for a mixture of five modulated components (different frequency bands and various conditions of modulation rates (same/different). It is promising to see that the same phenomenon is reported by some participants of preliminary listening tests. They consider such uniform modulated signals covering more than one frequency band similarly as one sound object, but more elaborate tests are needed to assure this observation [1]. 5 Conclusions NMF and NTF are promising methods for blind source separation of audio signals, whereas NTF has the advantage to consider additional dimensions to frequency and time, e.g., in this paper the modulation rate of a signal component that is important from the perceptional point of view. Further investigations are ongoing to work on the separation of modulated as well as tonal components, especially for the challenge of tones with varying frequency. Additionally, investigations for improving the auralization of the modulation tensor components using the hearing model will be performed. 9

Acknowledgments The author wishes to thank his former student Stephan Drewes for his investigations on algorithms for the detection of acoustic patterns [1]. References [1] Drewes, S. Untersuchung von Verfahren zur Detektion akustischer Muster. Diploma thesis, RWTH Aachen, 2013. [2] Casey, M.A.; Westner, A. Separation of Mixed Audio Sources by Independent Subspace Analysis. Proc. International Computer Music Conference (ICMC) 2000, Berlin, Germany, August 27 - September 1, 2000, pp. 154-161. [3] Philippen, B.; Sottek, R.; Spiertz, M. Blinde Quellentrennung von Fahrzeuggeräuschen. Proc. Daga 2010, Berlin, Germany, March 15-18, 2010, pp. 111-112. [4] Lee, D.D.; Seung, H.S. Learning the parts of objects with non-negative matrix factorization. Nature, Vol. 401, 1999, pp. 788 791. [5] Sottek, R. Modelle zur Signalverarbeitung im menschlichen Gehör. Dissertation, RWTH Aachen, 1993. [6] Fastl, H.; Zwicker, E. Psychoacoustics: Facts and Models. Springer, Berlin, Heidelberg, New York, 3rd edition, 2006. [7] Licklider, J.C.R. Duplex Theory of Pitch Perception. Cellular and Molecular Life Sciences, Vol. 7(4), 1951, pp. 128-134. [8] Sottek, R. Progress in calculating tonality of technical sounds. Proc. Internoise, Melbourne, Australia, November 16-19, 2014. In CD-ROM. [9] Bierbaums, T.; Sottek, R. Modellierung der zeitvarianten Lautheit mit einem Gehörmodell. Proc. DAGA 2012, Darmstadt, Germany, March 19-22, 2012, pp. 591-592. [10] Cichocki, A.; Zdunek, R.; Amari, S. Csiszár s Divergences for Non-negative Matrix Factorization: Family of New Algorithms. Proc. 6th International Conference on Independent Component Analysis and Blind Source Separation, Vol. 3889, Charleston, SC, USA, March 5-8, 2006, pp. 32 39. [11] Févotte, C.; Bertin, N.; Durrieu, J.-L. Nonnegative Matrix Factorization with the Itakura-Saito Divergence: With Application to Music Analysis. Neural Computation, Vol. 21(3), 2009, pp. 793 830. [12] Févotte, C.; Idier, J. Algorithms for Nonnegative Matrix Factorization with the β -Divergence. Neural Computation, Vol. 23(9), 2011, pp. 2421 2456. [13] Cichocki, A.; Zdunek, R.; Phan, A.-H.; Amari, S. Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-way Data Analysis and Blind Source Separation. John Wiley & Sons Ltd, Chichester, 2009. [14] Tan, V.; Févotte, C. Automatic Relevance Determination in Nonnegative Matrix Factorization with the β -Divergence. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 35(7), 2013, pp. 1592 1605. [15] Févotte, C.; Ozerov, A. Notes on Nonnegative Tensor Factorization of the Spectrogram for Audio Source Separation: Statistical Insights and Towards Self-Clustering of the Spatial Cues. Proc. 7th International Symposium on Computer Music Modeling and Retrieval (CMMR), Vol. 6684, Málaga, Spain, June 21-24, 2010, pp. 102 115. 10