THE hidden Markov model (HMM)-based parametric

Size: px
Start display at page:

Download "THE hidden Markov model (HMM)-based parametric"

Transcription

1 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY Modeling Spectral Envelopes Using Restricted Boltzmann Macines and Deep Belief Networks for Statistical Parametric Speec Syntesis Zen-Hua Ling, Member, IEEE, Li Deng, Fellow, IEEE, and Dong Yu, Senior Member, IEEE Abstract Tis paper presents a new spectral modeling metod for statistical parametric speec syntesis. In te conventional metods, ig-level spectral parameters, suc as mel-cepstra or line spectral pairs, are adopted as te features for idden Markov model (HMM) based parametric speec syntesis. Our proposed metod described in tis paper improves te conventional metod in two ways. First, distributions of low-level, un-transformed spectral envelopes (extracted by te STRAIGHT vocoder) are used as te parameters for syntesis. Second, instead of using single Gaussian distribution, we adopt te grapical models wit multiple idden variables, including restricted Boltzmann macines (RBM) and deep belief networks (DBN), to represent te distribution of te low-level spectral envelopes at eac HMM state. At te syntesis time, te spectral envelopes are predicted from te RBM-HMMs or te DBN-HMMs of te input sentence following te maximum output probability parameter generation criterion wit te constraints of te dynamic features. A Gaussian approximation is applied to te marginal distribution of te visible stocastic variables in te RBM or DBN at eac HMM state in order to acieve a closed-form solution to te parameter generation problem. Our experimental results sow tat bot RBM-HMM and DBN-HMM are able to generate spectral envelope parameter sequences better tan te conventional Gaussian- HMM wit superior generalization capabilities and tat DBN- HMM and RBM-HMM perform similarly due possibly to te use of Gaussian approximation. As a result, our proposed metod can significantly alleviate te over-smooting effect and improve te naturalness of te conventional HMM-based speec syntesis system using mel-cepstra. Index Terms Speec syntesis, idden Markov model, restricted Boltzmann macine, deep belief network, spectral envelope I. INTRODUCTION THE idden Markov model (HMM)-based parametric speec syntesis metod as become a mainstream speec syntesis metod in recent years [2], [3]. In tis metod, te spectrum, F0 and segment durations are modeled simultaneously witin a unified HMM framework [2]. At syntesis time, tese parameters are predicted so as to Tis work was partially funded by te National Nature Science Foundation of Cina (Grant No ) and te Cina Scolarsip Council Young Teacer Study Abroad Project. Te paper is te expanded version of te conference paper publised in ICASSP-2013 [1]. Z.-H. Ling is wit te National Engineering Laboratory of Speec and Language Information Processing, University of Science and Tecnology of Cina, Hefei, , Cina ( zling@ustc.edu.cn). Tis work was carried out wen e was a visiting scolar at te Department of Electrical Engineering, University of Wasington, WA, USA. L. Deng and D. Yu are wit Microsoft Researc, One Microsoft Way, Redmond WA, 98052, USA ( deng@microsoft.com; dongyu@microsoft.com). maximize teir output probabilities from te HMM of te input sentence. Te constraints of te dynamic features are considered during parameter generation in order to guarantee te smootness of te generated spectral and F0 trajectories [4]. Finally, te predicted parameters are sent to a speec syntesizer to reconstruct te speec waveforms. Tis metod is able to syntesize igly intelligible and smoot speec sounds [5], [6]. However, te quality of te syntetic speec is degraded due to tree main factors: limitations of te parametric syntesizer itself, inadequacy of acoustic modeling used in te syntesizer, and te over-smooting effect of parameter generation [7]. Many improved approaces ave been proposed to overcome te disadvantages of tese tree factors. In terms of te speec syntesizer, STRAIGHT [8], as a ig-performance speec vocoder, as been widely used in current HMM-based speec syntesis systems. It follows te source-filter model of speec production. In order to represent te excitation and vocal tract caracteristics separately, F0 and a smoot spectral envelope witout periodicity interference are extracted at eac frame. Ten, mel-cepstra [5] or line spectral pairs [6] can be derived from te spectral envelopes of training data for te following HMM modeling. During syntesis, te generated spectral parameters are used eiter to reconstruct speec waveforms directly or to recover te spectral envelopes for furter speec reconstruction by STRAIGHT. Acoustic modeling is anoter key component of te HMMbased parametric speec syntesis. In te common spectral modeling metods, te probability density functions (PDF) of eac HMM state is represented by a single Gaussian distribution wit diagonal covariance matrix and te distribution parameters are estimated under te maximum likeliood (ML) criterion [2]. Because te single Gaussian distributions are used as te state PDFs, te outputs of maximum output probability parameter generation tend to distribute near te modes (also te means) of te Gaussians, wic are estimated by averaging observations wit similar context descriptions in te ML training. Altoug tis averaging process improves te robustness of parameter generation, te detailed caracteristics of te spectral parameters are lost. Terefore, te reconstructed spectral envelopes are over-smooted, wic leads to a muffled voice quality in te syntetic speec. Te existing refinements on acoustic modeling include increasing te number of Gaussians for eac HMM state [4], reformulating HMM as a trajectory model [9], improving te model training criterion by minimizing te generation error [10].

2 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY In order to alleviate te over-smooting effect, many improved parameter generation metods ave also been proposed, suc as modifying te parameter generation criterion by integrating a global variance model [11] or minimizing model divergences [12], post-filtering after parameter generation [6], [13], using real speec parameters or segments to generate te speec waveform [14], [15], or sampling trajectories from te predictive distribution [16], [17], and so on. In tis paper, we propose a new spectral modeling metod wic copes wit te first two factors mentioned above. First, te raw spectral envelopes extracted by te STRAIGHT vocoder are utilized directly witout furter deriving spectral parameters from tem during feature extraction. Comparing wit te ig-level 1 spectral parameters, suc as mel-cepstra or line spectral pairs, te low-level spectral envelopes are more pysically meaningful and more directly related wit te subjective perception on te speec quality. Tus, te influence of spectral parameter extraction on te spectral modeling can be avoided. Similar approac can be found in [18], were te spectral envelopes derived from te armonic amplitudes are adopted to replace te mel-cepstra for HMM-based Arabic speec syntesis and te naturalness improvement can be acieved. Second, te grapical models wit multiple idden layers, suc as restricted Boltzmann macines (RBM) [19] and deep belief networks (DBN) [20], are introduced to represent te distribution of te spectral envelopes at eac HMM state instead of single Gaussian distribution. An RBM is a bipartite undirected grapical model wit a two-layer arcitecture and a DBN contains more idden layers, wic can be estimated using a stack of RBMs. Bot of tese two models are better in describing te distribution of ig-dimensional observations wit cross-dimension correlations, i.e., te spectral envelopes, tan te single Gaussian distribution and Gaussian mixture model (GMM). Te acoustic modeling metod wic describes te production, perception and distribution of speec signals is always an important researc topic in speec signal processing [21]. In recent years, RBMs and DBNs ave been successfully applied to modeling speec signals, suc as spectrogram coding [22], speec recognition [23], [24], and acoustic-articulatory inversion mapping [25], were tey mainly act as te pre-training metods for a deep autoencoder or a deep neural network (DNN). Te arcitectures used in deep learning as applied to speec processing ave been motivated by te multi-layered structures in bot speec production and perception involving ponological features, motor control, articulatory dynamics, and acoustic and auditory parameters [26], [27]. Te approaces of applying RBMs, DBNs, and oter deep learning metods to te statistical parametric speec syntesis ave also been studied very recently [1], [28] [30]. In [28], a DNN-based statistical parametric speec syntesis metod is presented, wic maps te input context information towards te acoustic features using a neural network wit deep structures. In [29], a DNN wic is pre-trained by te DBN learning is adopted as a feature 1 Here, te level refers to te steps of signal processing procedures involved in te spectral feature extraction. Te ig-level spectral parameters are commonly derived from te low-level ones by functional representation and parameterization. extractor for te Gaussian process based F0 contour prediction. Furtermore, RBMs and DBNs can be used as density models instead of te DNN initialization metods for te speec syntesis application. In [30], a single DBN model is trained to represent te joint distribution between te tonal syllable ID and te acoustic features. In [1], a set of RBMs are estimated to describe te distributions of te spectral envelopes in te context-dependent HMM states. In tis paper, we extend our previous work in [1] by incorporating te dynamic features of spectral envelopes into te RBM modeling and developing RBMs to DBNs wic as more layers of idden units. Tis paper is organized as follows. In Section II, we will briefly review te basic tecniques of RBMs and DBNs. In Section III, we will describe te details of our proposed metod. Section IV reports our experimental results. Section V gives te conclusion and te discussion on our future work. II. RESTRICTED BOLTZMANN MACHINES AND DEEP BELIEF NETWORKS A. Restricted Boltzmann macines An RBM is a kind of bipartite undirected grapical model (i.e. Markov random field) wic is used to describe te dependency among a set of random variables using a twolayer arcitecture [19]. In tis model, te visible stocastic units v = [v 1,..., v V ] are connected to te idden stocastic units = [ 1,..., H ] as sown in Fig. 1.a), were V and H are te numbers of units of te visible and idden layers respectively, and ( ) means te matrix transpose. Assuming v {0, 1} V and {0, 1} H are bot binary stocastic variables, te energy function of te state {v, } is defined as V H V H E(v, ) = a i v i b j j w ij v i j, (1) i=1 j=1 i=1 j=1 were w ij represents te symmetric interaction between v i and j, a i and b j are bias terms. Te model parameters are composed of a = [a 1,..., a V ], b = [b 1,..., b H ], and W = {w ij } V H. Te joint distribution over te visible and idden units is defined as were P (v, ) = 1 exp ( E(v, )), (2) Z Z = v exp ( E(v, )) (3) is te partition function wic can be estimated using te annealed importance sampling (AIS) metod [31]. Terefore, te probability density function over te visible vector v can be calculated as P (v) = 1 exp ( E(v, )). (4) Z Given a training set, te RBM model parameters {a, b, W} can be estimated by maximum likeliood learning using te contrastive divergence (CD) algoritm [32]. RBM can also be applied to model te distribution of real-valued data (e.g. te speec parameters) by adopting its

3 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY a) b) v Fig. 1. Te grapical model representations for a) an RBM and b) a treeidden-layer DBN. Gaussian-Bernoulli form, wic means v R V are real-valued and {0, 1} H are binary. Tus, te energy function of te state {v, } is defined as V (v i a i ) 2 H V H v i E(v, ) = b j j w ij j, σ i i=1 2σ 2 i 2 v j=1 3 1 i=1 j=1 were te variance parameters σi 2 are commonly fixed to a predetermined value instead of learning from te training data [33]. B. Deep belief networks A deep belief network (DBN) is a probabilistic generative model wic is composed of many layers of idden units [20]. Te grapical model representation for a tree-idden-layer DBN is sown in Fig. 1.b). In tis model, eac layer captures te correlations among te activities of idden features in te layer below. Te top two layers of te DBN form an undirected bipartite grap. Te lower layers form a directed grap wit a top-down direction to generate te visible units. Matematically, te joint distribution over te visible and all idden units can be written as P (v, 1,..., L ) = P (v 1 )P ( 1 2 ) P ( L 2 L 1 )P ( L 1, L ), (6) were l = [ l 1,..., l H l ] is te idden stocastic vector of te l-t idden layer, H l is te dimensionality of l, and L is te number of idden layers. Te joint distribution P ( L 1, L ) is represented by an RBM as (2) wit te weigt matrix W L and te bias vectors a L and b L. P (v 1 ) and P ( l 1 l ), l {2, 3,..., L 1} are represented by sigmoid belief networks [34]. Eac sigmoid belief network is described by a weigt matrix W l and a bias vector a l. Assuming v are real-valued and l, l {1, 2,..., L} are binary, te dependency between v and 1 in te sigmoid belief network is described by (5) P (v 1 ) = N (v; W a 1, Σ) (7) were N ( ) denotes a Gaussian distribution; Σ = diag{σi 2} and turns to an identity matrix wen σi 2 are fixed to 1 during model training. For l {2, 3,..., L 1}, te dependency between two adjacent idden layers is represented by P ( l 1 i = 1 l ) = g(a l i + j w l ij l j) (8) were g(x) = 1/(1+exp( x)) is te sigmoid function. For an L-idden-layer DBN, its model parameters are composed of {a 1, W 1,..., a L 1, W L 1, a L, b L, W L }. Furter, te marginal distribution of te visible variables for a DBN can be written as P (v) = P (v, 1,..., L ). (9) 1 L Given te training samples of te visible units, it is difficult to estimate te model parameters of a DBN directly under te maximum likeliood criterion due to te complex model structure wit multiple idden layers. Terefore, a greedy learning algoritm as been proposed and popularly applied to train te DBN in a layer-by-layer manner [20]. A stack of RBMs are used in tis algoritm. Firstly, it estimates te parameters {a 1, b 1, W 1 } of te first layer RBM to model te visible training data. Ten, it freezes te parameters {a 1, W 1 } of te first layer and draws samples from P ( 1 v) to train te next layer RBM {a 2, b 2, W 2 }, were P ( 1 j = 1 v) = g(b 1 j + wijv 1 i ). (10) i Tis training procedure is conducted recursively until it reaces te top layer and gets {a L, b L, W L }. It as been proved tat tis greedy learning algoritm can improve te lower bound on te log-likeliood of te training samples by adding eac new idden layer [20], [31]. Once te model parameters are estimated, to calculate te log-probability tat a DBN assigns to training or test data by (9) directly is also computationally intractable. A lower bound on te log-probability can be estimated by combining te AIS-based partition function estimation wit approximate inference [31]. III. SPECTRAL ENVELOPE MODELING USING RBMS AND DBNS A. HMM-based parametric speec syntesis At first, te conventional HMM-based parametric speec syntesis metod is briefly reviewed. It consists of a training stage and a syntesis stage. During training, te F0 and spectral parameters are extracted from te waveforms contained in te training set. Ten a set of context-dependent HMMs are estimated to maximize te likeliood function for te training acoustic features. Here o = [o 1, o 2,..., o T ] is te observation feature sequence and T is te lengt of te sequence. Te observation feature vector o t R 3D for te t-t frame typically consists of static acoustic parameters c t R D and teir delta and acceleration components as o t = [ c t, c t, 2 c ] t, (11) were D is te dimension of te static component; te dynamic components are commonly calculated as c t = 0.5c t+1 0.5c t 1 t [2, T 1], (12) c 1 = c 2, c T = c T 1 (13)

4 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY and 2 c t = c t+1 2c t + c t 1 t [2, T 1], (14) 2 c 1 = 2 c 2, 2 c T = 2 c T 1. (15) Terefore, te complete feature sequence o can be considered to be a linear transform of te static feature sequence c = [c 1, c 2,..., c T ] as o = Mc, (16) were M R 3T D T D is determined by te delta and acceleration calculation functions in (12)-(15) [4]. A multi-space probability distribution (MSD) [35] is applied to incorporate a distribution for F0 into te probabilistic framework of te HMM considering tat F0 is only defined for voiced speec frames. In order to deal wit te data-sparsity problem of te context-dependent model training wit extensive context features, a decision-tree-based model clustering tecnique tat uses a minimum description lengt (MDL) criterion [36] to guide te tree construction is adopted after initial training of te context-dependent HMMs. Next, a state alignment is conducted using te trained HMMs to train context-dependent state duration probabilities [2] for state duration prediction. A single-mixture Gaussian distribution is used to model te duration probability for eac state. A decision-tree-based model clustering tecnique is similarly applied to tese duration distributions. At te syntesis stage, te maximum output probability parameter generation algoritm is used to generate acoustic parameters [4]. Te result of front-end linguistic analysis on te input text is used to determine te sentence HMM λ. Te state sequence q = {q 1, q 2,..., q T } is predicted using te trained state duration probabilities [2]. Ten, te sequence of speec features are predicted by maximizing P (o λ, q). Considering te constraints between static and dynamic features as in (16), te parameter generation criterion can be rewritten as c = arg max P (Mc λ, q), (17) c were c is te generated static feature sequence. If te emission distribution of eac HMM state is represented by a single Gaussian distribution, te closed-form solution to (17) can be derived. By setting we obtain P (Mc q, λ) c = 0, (18) c = ( M U 1 q M ) 1 M U 1 q m q, (19) were m q = [µ q 1,..., µ q T ] and U q = diag(σ q1,..., Σ qt ) are te mean vector and covariance matrix of te sentence as decided by te state sequence q [4]. B. Spectral envelope modeling and generation using RBM and DBN In tis paper, we improve te conventional spectral modeling metod in te HMM-based parametric speec syntesis from two aspects. First, te raw spectral envelopes extracted by te STRAIGHT vocoder are modeled directly witout F0 and Spectral Parameter Extraction Clustered CD-HMM Training Gaussian HMMs (Spectral Parameters) Training Syntesis Speec Corpus Fully CD-HMM Training Decision-Tree-based Model Clustering Input Text Spectral Envelope Extraction State Alignment Parameter Generation & Syntesizer Syntetic Speec Context-Dependent RBM/DBN Training RBM/DBN-HMMs (Spectral Envelopes) Gaussian Approximation Fig. 2. Flowcart of our proposed metod. Te modules in solid lines represent te procedures of te conventional HMM-based speec syntesis using ig-level spectral parameters, were CD-HMM stands for Context- Dependent HMM. Te modules in das lines describe te add-on procedures of our proposed metod for modeling te spectral envelopes using RBMs or DBNs. furter deriving ig-level spectral parameters. Second, te RBM and DBN models are adopted to replace te single Gaussian distribution at eac HMM state. In order to simplify te model training wit ig-dimensional spectral features, te decision trees for model clustering and te state alignment results are assumed to be given wen te spectral envelopes are modeled. Tus, we can focus on comparing te performance of different models on te clustered model estimation. In current implementation, te conventional context-dependent model training using ig-level spectral parameters and single Gaussian state PDFs is conducted at first to acieve te model clustering and state alignment results. Te flowcart of our proposed metod is sown in Fig. 2. During te acoustic feature extraction using STRAIGHT vocoder, te original linear frequency spectral envelopes 2 are stored besides te spectral parameters. Te context-dependent HMMs for conventional spectral parameters and F0 features are firstly estimated according to te metod introduced in Section III.A. A single Gaussian distribution is used to model te spectral parameters at eac HMM state. Next, a state alignment to te acoustic features is performed. Te state boundaries are used to gater te spectral envelopes for eac clustered context-dependent state. Similar to te ig-level spectral parameters, te feature vector of te spectral envelope at eac frame consists of static, velocity, and acceleration components as (11)-(15). Ten, an RBM or a DBN is esti- 2 Te mel-frequency spectral envelopes can also be used ere to represent te speec perception properties. In tis paper, we adopt te linear frequency spectral envelope because it is te most original description of te vocal tract caracters witout any prior knowledge and assumption on te spectral parameterization and speec perception.

5 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY mated under te maximum likeliood criterion for eac state according to te metods introduced in Section II. Te model estimation of te RBMs or te DBNs is conducted only once using te fixed state boundaries. Finally, te context-dependent RBM-HMMs or DBN-HMMs can be constructed for modeling te spectral envelopes. At syntesis time, te same criterion in (17) is followed to generate te spectral envelopes. Te optimal sequence of spectral envelopes are estimated by maximizing te output probability from te RBM-HMM or te DBN-HMM of te input sentence. Wen single Gaussian distributions are adopted as te state PDFs, tere is a closed-form solution as sown in (19) to tis maximum output probability parameter generation wit te constraints of dynamic features once te state sequence as been determined [4]. However, te marginal distribution defined in (4) for an RBM or in (9) for a DBN is muc more complex tan a single Gaussian, wic makes te closed-form solution impractical. Terefore, a Gaussian approximation is applied before parameter generation to simply te problem. For eac HMM state, a Gaussian distribution N (v; µ, Σ) is constructed, were v is te spectral envelope feature vector containing static, velocity, and acceleration components; µ = arg max log P (v) (20) v is te mode estimated for eac RBM or DBN and P (v) is defined as (4) or (9); Σ is a diagonal covariance matrix estimated by calculating te sample covariances given te training samples of te state. Tese Gaussian distributions are used to replace te RBMs or te DBNs as te state PDFs at syntesis time. Terefore, te conventional parameter generation algoritm wit te constraints of dynamic features can be followed to predict te spectral envelopes by solving a group of linear equations (19). By incorporating te dynamic features of te spectral envelopes during model training and parameter generation, temporally smoot spectral trajectories can be generated at syntesis time. Te detailed algoritms of te mode estimation in (20) for an RBM and a DBN model will be introduced in te following subsections. C. Estimating RBM mode Here, we consider te RBM of te Gaussian-Bernoulli form because te spectral envelope features are real-valued. Given te estimated model parameters {a, b, W} of an RBM, te probability density function (4) over te visible vector v can be furter calculated as 3 P (v) = 1 exp ( E(v, )) Z ( ) = 1 V (v i a i ) 2 exp + b + v W Z 2 i=1 ) = 1 V ( Z exp (v i a i ) 2 2 = 1 Z exp ( i=1 H j=1 j {0,1} exp(b j j + v w j j ) ) V (v i a i ) 2 H ( 1 + exp(bj + v w j ) ), 2 i=1 j=1 (21) were w j denotes te j-t column of matrix W. Because tere is no closed-form solution to solve (20) for an RBM, te gradient descent algoritm is adopted ere, i.e., v (i+1) = v (i) + α log P (v) v v=v (i), (22) were i denotes te number of iteration; α is te step size; log P (v) H exp(b j + v w j ) = (v a) + v 1 + exp(b j + v w j ) w j. (23) j=1 Tus, te estimated mode of te RBM model is determined by a non-linear transform of te model parameters {a, b, W}. In contrast to te single Gaussian distribution, tis mode is no longer te Gaussian mean wic is estimated by averaging te corresponding training vectors under te maximum likeliood criterion. Because te likeliood of an RBM is multimodal, te gradient descent optimization in (22) only leads to a local maximum and te result is sensitive to te initialization of v (0). In order to find a representative v (0), we firstly calculate te means of te conditional distributions P ( v) for all training vectors v. Tese means are averaged and made binary using a fixed tresold of 0.5 to get (0). Ten, te initial v (0) for te iteratively updating in (22) is set as te mean of P (v (0) ). D. Estimating DBN mode Estimating te mode of a DBN model is more complex tan dealing wit an RBM. Te marginal distribution of te visible variables in (9) can be rewritten as P (v) = P (v, 1,..., L ) (24) 1 L = 1 P (v 1 )P ( 1 ), (25) were P (v 1 ) is described in (7) and P ( 1 ) can be calculated by applying P ( l 1 ) = l P ( l 1 l )P ( l ). (26) 3 Te variance parameters σi 2 in (5) are fixed to 1 to simplify te notation.

6 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY recursively for eac l from L 1 to 2. Te conditional distribution P ( l 1 l ) is represented by (8) and P ( L 1 ) is te marginal distribution (4) of te RBM representing te top two idden layers. Similar to te RBM mode estimation, te gradient descent algoritm can be applied ere to optimize (25) once te values of P ( 1 ) are determined for all possible 1. However, tis will lead to an exponential complexity wit respect to te number of idden units at eac idden layer due to te summation in (25) and (26). Tus, suc optimization becomes impractical unless te number of idden units is reasonably small. In order to get a practical solution to te DBN mode estimation, an approximation is made to (24) in tis paper. Te summation over all possible values of te idden units is simplified by considering only te optimal idden vectors at eac layer, i.e. were P (v) P (v, 1,..., L ), (27) { L 1, L } = argmax log P ( L 1, L ) (28) { L 1, L } and for eac l {L 1,..., 3, 2} l 1 = arg max l 1 log P ( l 1 l ). (29) Te joint distribution P ( L 1, L ) in (28) is modeled by a Bernoulli-Bernoulli RBM according to te definition of DBN in Section II.B. Because L 1 and L are bot binary stocastic vectors, te iterated conditional modes (ICM) [37] algoritm is adopted to solve (28). Tis algoritm determines te configuration tat maximizes te joint probability of a Markov random field by iteratively maximizing te probability of eac variable conditioned on te rest. Applying te ICM algoritm ere, we just update L by maximizing P ( L L 1 ) and update L 1 by maximizing P ( L 1 L ) iteratively. Bot of te two conditional distributions are multivariate Bernoulli distribution witout cross-dimension correlation [20]. Te optimal configuration at eac step can be determined simply by applying a tresold of 0.5 for eac binary unit. Te initial L 1 of te iteratively updating is set to be te L 1 wic is obtained by solving (28) for te DBN wit L 1 idden layers. For eac l from L 1 to 2, (29) can be solved recursively according to te conditional distribution in (8). After { 1,..., L } are determined, te mode of te DBN can be estimated by substituting (27) into (20). Considering P (v 1 ) is a Gaussian distribution as (7), we ave µ arg max log P (v, 1,..., L ) v = arg max log P (v 1 ) v = W a 1. (30) A. Experimental conditions IV. EXPERIMENTS A 1-our Cinese speec database produced by a professional female speaker was used in our experiments. It consisted Cumulative Probability A B C Number of Frames Fig. 3. Te cumulative probability curve for te number of frames belonging to eac context-dependent state. Te arrows indicate te numbers of frames of te tree example states used for te analysis in Section IV.B and Fig. 4. of 1,000 sentences togeter wit te segmental and prosodic labels. 800 sentences were selected randomly for training and te remaining 200 sentences were used as a test set. Te waveforms were recorded in 16kHz/16bit format. Wen constructing te baseline system, 41-order melcepstra (including 0-t coefficient for frame power) were derived from te spectral envelope by STRAIGHT analysis at 5ms frame sift. Te F0 and spectral features consisted of static, velocity, and acceleration components. A 5-state left-torigt HMM structure wit no skips was adopted to train te context-dependent pone models. Te covariance matrix of te single Gaussian distribution at eac HMM state was set to be diagonal. After te decision-tree-based model clustering, we got 1,612 context-dependent states in total for te mel-cepstral stream. Te model parameters of tese states were estimated by maximum likeliood training. In te spectral envelope modeling, te FFT lengt of te STRAIGHT analysis was set to 1024 wic led to = 1539 visible units in te RBMs and DBNs corresponding to te spectral amplitudes witin te frequency range of [0, π] togeter wit teir dynamic components. After te HMMs for te mel-cepstra and F0 features were trained, a state alignment was conducted on te training set and te test set to assign te frames to eac state for te spectral envelope modeling and testing. Te cumulative probability curve for te number of frames belonging to eac context-dependent state is illustrated in Fig. 3. From tis figure, we can see tat te numbers of training samples vary a lot among different states. For eac context-dependent state, te logaritmized spectral amplitudes at eac frequency point were normalized to zero mean and unit variance. CD learning wit 1-step Gibbs sampling (CD1) was adopted for te RBM training and te learning rate was Te batc size was set to 10 and 200 epocs were executed for estimating eac RBM. Te DBNs were estimated following te greedy layer-by-layer training algoritm introduced in Section II.B.

7 ave. log-prob. ave. log-prob. ave. log-prob. ave. log-prob. ave. log-prob. ave. log-prob. JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY (a) -80 number of GMM mixtures number of GMM mixtures number of GMM mixtures GMM(train) GMM(test) RBM(train) RBM(test) number of RBM idden units number of RBM idden units number of RBM idden units (b) -700 number of GMM mixtures number of GMM mixtures number of GMM mixtures GMM(train) GMM(test) RBM(train) RBM(test) number of RBM idden units number of RBM idden units number of RBM idden units Fig. 4. Te average log-probabilities on te training and test sets wen modeling (a) te mel-cepstra and (b) te spectral envelopes of state A (left column), state B (middle column), and state C (rigt column) using different models. Te number of training samples belonging to tese tree selected states are indicated in Fig. 3. B. Comparison between GMM and RBM as state PDFs At first we compared te performance of te GMM and te RBM in modeling te distribution of mel-cepstra and spectral envelopes for an HMM state. Tree representative states were selected for tis experiment, wic ave 270, 650, 1530 training frames and 60, 130, 410 test frames respectively. As sown in Fig. 3, te numbers of training frames of tese tree states correspond to te 0.1, 0.5, and 0.9 cumulative probabilities wic are calculated over te numbers of te training frames of all te 1,612 context-dependent states. GMMs and RBMs were trained under te maximum likeliood criterion to model tese tree states. Te covariance matrices in te GMMs were set to be diagonal and te number of Gaussian mixtures varied from 1 to 64. Te number of idden units in te RBMs varied from 1 to 1, 000. Te average log-probabilities on te training and test sets for different models and states are sown in Fig. 4 for te mel-cepstra and te spectral envelopes respectively. Examining te difference between te training and test log-probabilities for bot te melcepstra and te spectral envelopes, we see tat te GMMs ave a clear tendency of over-fitting wit te increasing of model complexity. Tis over-fitting effect becomes less significant wen a larger training set is available. On te oter and, te RBM sows consistently good generalization ability wit te increasing of te number of idden units. Tis can be attribute to utilizing te binary idden units wic create a information bottleneck and act as an effective regularizer during model training. Te differences between te test log-probabilities of te best GMM or RBM models and te single Gaussian distributions for te tree states are listed in Table I. From Fig. 4 and Table I, we can see tat te model accuracy improvements obtained by using te density models tat are more complex tan a single Gaussian distribution are relatively small wen te mel-cepstra are used for spectral modeling. Once te spectral envelopes are used, suc improvements become muc more significant for bot te GMM and RBM models. Besides, te RBM also gives muc iger log-probability to te test data tan te GMM wen modeling te spectral envelopes. Tese results can be attributed to tat te mel-cepstral analysis is a kind of decorrelation processing to te spectrums. A GMM wit multiple components is able to describe te interdimensional correlations of a multivariate distribution to some extend even if te diagonal covariance matrices are used. An RBM wit H idden units can be considered as a GMM wit 2 H structured mixture components according to (21). Terefore, it is good at analyzing te latent patterns embedded in te ig-dimensional raw data wit inter-dimensional correlations. Fig. 5 sows te estimated weigt matrices W in te RBMs wen modeling te spectral envelopes for te tree states. We can see tat te weigt matrices are somewat sparse, indicating eac idden unit tries to capture te caracteristics of te spectral envelope in some specific frequency bands. Tis is similar to a frequency analysis for spectral envelopes wic makes use of te amplitudes of te critical frequency bands context-dependently. For te spectral envelope modeling, we can furter improve te model accuracy by training RBMs layer-by-layer and constructing a DBN. Te average log-probabilities on te training and test sets wen modeling te spectral envelopes using an RBM and a two-idden-layer DBN are compared in Table II. Here, te lower bound estimation [31] to te log-

8 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY TABLE I THE DIFFERENCES BETWEEN THE TEST LOG-PROBABILITIES OF THE BEST GMM OR RBM MODELS AND THE SINGLE GAUSSIAN DISTRIBUTIONS FOR THE THREE SELECTED STATES. THE NUMBERS IN THE BRACKETS INDICATE THE NUMBERS OF GAUSSIAN MIXTURES FOR THE GMMS AND THE NUMBERS OF HIDDEN UNITS THE RBMS WHICH LEAD TO THE HIGHEST LOG-PROBABILITIES ON THE TEST SET. State mel-cepstra spectral envelope GMM RBM GMM RBM A 2.40(4) 9.32(200) (8) (1000) B 4.80(8) 7.73(200) (8) (1000) C 1.80(8) 3.14(1000) (32) (1000) TABLE III SUMMARY OF DIFFERENT SYSTEMS CONSTRUCTED IN THE EXPERIMENTS. System Spectral Features State PDF Baseline mel-cepstra single Gaussian GMM(1) spectral envelope single Gaussian GMM(8) spectral envelope GMM, 8 mixtures RBM(10) spectral envelope RBM, 10 idden units RBM(50) spectral envelope RBM, 50 idden units DBN(50-50) spectral envelope 2-idden-layer DBN, 50 idden units eac layer DBN( ) spectral envelope 3-idden-layer DBN, 50 idden units eac layer Frequency (Hz) State A State B State C TABLE IV AVERAGE LOG-PROBABILITIES OF THE SAMPLE MEANS AND THE ESTIMATED MODES FOR THE FOUR RBM OR DBN BASED SYSTEMS. System Sample Means PDF Modes RBM(10) RBM(50) DBN(50-50) DBN( ) Hidden Unit No Hidden Unit No Hidden Unit No. Fig. 5. Visualization of te estimated weigt matrices W in te RBMs wen modeling te spectral envelopes for te tree states. Te number of idden units is 50 and only te first 513 rows of te weigt matrices are drawn. Eac column in te gray-scale figures corresponds to te weigts connecting one idden unit wit te 513 visible units wic compose te static component of te spectral envelope feature vector. probability of a DBN is adopted. From tis table, we can observe a monotonic increase of test log-probabilities by using more idden layers. C. System construction Seven systems were constructed wose performance we compared in our experiments. Te definitions of tese systems are explained in Table III. As sown in Table I, te model accuracy improvement acieved by adopting te distributions more complicated tan te single Gaussian is not significant wen te mel-cepstra are used as spectral features. Terefore, we focus on te performance of spectral envelope modeling TABLE II THE AVERAGE LOG-PROBABILITIES ON THE TRAINING AND TEST SETS WHEN MODELING THE SPECTRAL ENVELOPES USING AN RBM OF 50 HIDDEN UNITS AND A TWO-HIDDEN-LAYER DBN OF 50 HIDDEN UNITS AT EACH LAYER. State RBM(50) DBN(50-50) train test train test A B C using different forms of state PDFs in our experiments. Considering te computational complexity of training state PDFs for all context-dependent states, te maximum number of idden units in te RBM and DBN models were set to 50. All tese systems sared te same decision trees for model clustering and te same state boundaries wic were derived from te Baseline system. Te F0 and duration models of te seven systems were identical. D. Mode estimation for te RBMs and DBNs Wen constructing te RBM(10), RBM(50), DBN(50-50), and DBN( ) systems, te mode of eac RBM or DBN trained for a context-dependent state was estimated for Gaussian approximation following te metods introduced in Section III.C and III.D. For eac system, te average logprobabilities of te estimated modes and te sample means were calculated. Te results are listed in Table IV. From tis table, we see tat te estimated modes ave muc iger logprobabilities tan te sample means known to ave te igest probability for a single Gaussian distribution. Tis means tat wen te RBMs or te DBNs are adopted to represent te state PDFs, te feature vector wit te igest output probability is not te sample means anymore. Tis implies te superiority of RBM and DBN over single Gaussian distribution in alleviating te over-smooting problem during parameter generation under te maximum output probability criterion. Te spectral envelopes recovered from te modes of different systems for one HMM state 4 are illustrated in Fig. 6. Here, only te static components of te spectral envelope feature vectors are drawn. Te mode of te GMM(1) system is just te Gaussian mean vector. Te mode of te GMM(8) system is approximated as te Gaussian mean of te mixture wit te igest mixture weigt. Comparing GMM(8) wit GMM(1), we can see tat using more Gaussian mixtures 4 Tis state is not one of te tree states used in Section IV.B.

9 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY Norm. Spec. Amp. Norm. Spec. Amp Frequency (Hz) GMM(1) GMM(8) RBM(10) RBM(50) GMM(1) RBM(50) DBN(50 50) DBN( ) Frequency(Hz) Fig. 6. Te spectral envelopes recovered from te modes of different systems for one HMM state. can elp alleviate te over-smooting effect on te spectral envelope. Besides, te estimated state mode of te RBM and DBN based systems ave sarper formant structures tan te GMM-based ones. Comparing RBM(50) wit RBM(10), we can see te advantages of using more idden units in an RBM. Wile te differences among te estimated modes of te RBM(50), DBN(50-50), and DBN( ) systems are less significant. We will investigate te performance of tese systems furter by te following subjective evaluation. E. Subjective evaluation Because te mel-cepstrum extraction can be considered as a kind of linear transform to te logaritmized spectral envelope, te spectral envelope recovered from te mean of te melcepstra in a state is very close to te one recovered from te mean of te corresponding logaritmized spectral envelopes. Terefore, te Baseline and te GMM(1) systems ad very similar syntetic results and te Baseline system was adopted as a representative for tese two systems in te subjective evaluation to simplify te test design. For te GMM(8) system, te EM-based parameter generation algoritm [4] could be applied to predict te spectral envelope trajectories by iteratively updating. In order to get a closed-form solution, we made a single Gaussian approximation to te GMMs at syntesis time by only using te Gaussian mixture wit te igest mixture weigt at eac HMM state. Te first subjective evaluation was to compare among te Baseline, GMM(8), RBM(10), and RBM(50) systems. Fifteen sentences out of te training database were selected and syntesized using tese four systems respectively. 5 Five groups of preference tests were conducted and eac one was to make comparison between two of te four systems as sown in eac row of Table V. Eac of te pairs of syntetic sentences were evaluated in random order by five Cinese-native listeners. 5 Some examples of te syntetic speec using te seven systems listed in Table III can be found at ttp://staff.ustc.edu.cn/~zling/dbnsyn/demo.tml. TABLE V SUBJECTIVE PREFERENCE SCORES (%) AMONG SPEECH SYNTHESIZED USING THE Baseline, GMM(8), RBM(10), AND RBM(50) SYSTEMS, WHERE N/P DENOTES NO PREFERENCE AND p MEANS THE p-value OF t-test BETWEEN THESE TWO SYSTEMS. Baseline GMM(8) RBM(10) RBM(50) N/P p TABLE VI SUBJECTIVE PREFERENCE SCORES (%) AMONG THE RBM(50), DBN(50-50), AND DBN( ) SYSTEMS. RBM(50) DBN(50-50) DBN( ) N/P p Te listeners were asked to identify wic sentence in eac pair sounded more natural. Table V summarizes te preference scores among tese four systems and te p-values given by t- test. From tis table, we can see tat introducing te density models tat are more complex tan single Gaussian, suc as GMM and RBM, to model te spectral envelopes at eac H- MM state can acieve significantly better naturalness tan te single Gaussian distribution based metods. Compared wit te GMM(8) system, te RBM(50) system as muc better preference in naturalness. Tis demonstrates te superiority of RBM over GMM in modeling te spectral envelope features. A comparison between te spectral envelopes generated by te Baseline system and te RBM(50) system is sown in Fig. 7. From tis figure, we can observe te enanced formant structures after modeling te spectral envelopes using RBMs. Besides, we can also find in Table V tat te performance of te RBM-based systems is influenced by te number of idden units used in te model definition wen comparing RBM(10) wit RBM(50). Tese results are consistent wit te formant sarpness of te estimated modes for different systems sown in Fig. 6. In order to investigate te effect of extending RBM to DBN wit more idden layers, anoter subjective evaluation was conducted among te RBM(50), DBN(50-50), and DBN( ) systems. Anoter fifteen sentences out of te training database were used and two groups of preference tests were conducted by five Cinese-native listeners. Te results are sown in Table VI. We can see tat tere is no significant differences among tese tree systems at 0.05 significance level. Altoug we can improve te model accuracy by introducing more idden layers as sown in Table II, te naturalness of syntetic speec can not be improved correspondingly. One possible reason is te approximation we make in (27) wen estimating te DBN mode. F. Objective evaluation Besides te subjective evaluation, we also calculated te spectral distortions on te test set between te spectral en-

10 JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY a) Baseline 8000 Frequency (Hz) Frequency (Hz) b) RBM(50) Frame Frame Fig. 7. Te spectrograms of a segment of syntetic speec using a) te Baseline system and b) te RBM(50) system. Tese spectrograms are not calculated by STFT analysis on te syntetic waveform. For te Baseline system, te spectrogram is drawn based on te spectral envelopes recovered from te generated mel-cepstra. For te RBM(50) system, te spectrogram is drawn based on te generated spectral envelopes directly. velopes generated by te systems listed in Table III and te ones extracted from te natural recordings. Te syntetic spectral envelopes used te state boundaries of te natural recordings to simplify te frame alignment. Ten, te natural and syntetic spectral envelopes at eac frame were normalized to te same power and te calculation of te spectral distortion between tem followed te metod introduced in [38]. For te Baseline system, te generated mel-cepstra were converted to spectral envelopes before te calculation. Te average spectral distortions of all te systems are listed in Table VII. We can see tat te objective evaluation results are inconsistent wit te subjective preference scores sown in Table V. For example, te RBM(50) system as significant better naturalness tan te Baseline system in te subjective evaluation, wile its average spectral distortion is te igest. Te reason is tat te spectral distortion in [38] is a Euclidean distance between two logaritmized spectral envelopes, wic treats eac dimension of te spectral envelopes independently and equally. However, te superiority of our proposed metod is to provide better representation of te cross-dimension correlations for te spectral envelope modeling, wic can not be reflected by tis spectral distortion measurement. Similar inconsistency between subjective evaluation results and objective acoustic distortions for speec syntesis as been observed in [12], [39]. V. CONCLUSION AND FUTURE WORK We ave proposed an RBM and DBN based spectral envelope modeling metod for statistical parametric speec syntesis in tis paper. Te spectral envelopes extracted by STRAIGHT vocoder are modeled by an RBM or a DBN for eac HMM state. At te syntesis time, te mode vectors of te trained RBMs and DBNs are estimated and used in place of te Gaussian means for parameter generation. Our experimental results sow te superiority of RBM and DBN over Gaussian mixture model in describing te distribution TABLE VII AVERAGE SPECTRAL DISTORTIONS (SD) ON TEST SET BETWEEN THE SPECTRAL ENVELOPES GENERATED BY THE SYSTEMS LISTED IN TABLE III AND THE ONES EXTRACTED FROM THE NATURAL RECORDINGS. system ave. SD (db) Baseline 3.85 GMM(1) 3.77 GMM(8) 3.86 RBM(10) 3.89 RBM(50) 4.11 DBN(50-50) 4.10 DBN( ) 4.10 of spectral envelopes as density models and in mitigating te over-smooting effect of te syntetic speec. As we discussed in Section I, tere are also some oter approaces tat can significantly reduce te over-smooting and improve te quality of te syntetic speec, suc as te GV-based parameter generation [11] and te post-filtering tecniques [6], [13]. In tis paper, we focus on te acoustic modeling to tackle te over-smooting problem. It worts to investigate alternative parameter generation and post-filtering algoritms tat are appropriate for our proposed spectral envelope modeling metod in te future. Tis paper only makes some preliminary exploration on applying te ideas of deep learning into statistical parametric speec syntesis. Tere are still several issues in te current implementation tat require furter investigation. First, it is wort examining te system performance wen te number of idden units in te RBMs keeps increasing. As sown in Fig. 3, te training samples are distributed among many contextdependent HMM states in a igly unbalanced manner. Tus, it may be difficult to optimize te model complexity for all states simultaneously. An alternative solution is to train te joint distribution between te observations and te context labels using a single network [20] wic is estimated using all training samples. Similar approac for te statistical parametric speec syntesis as been studied in [30], were te joint distribution between te tonal syllable ID and te spectral and excitation features is modeled using a multi-distribution DBN. Second, increasing te number of idden layers in te DBNs didn t acieve improvement in our subjective evaluation. A better algoritm to estimate te mode of a DBN wit less approximation is necessary. We plan as our future work to implement te Gaussian approximation according to (25) wen te number of idden units is reasonably small and compare its performance wit our current implementation. Anoter strategy is to adopt te sampling outputs rater tan te model modes during parameter generation. As better density models, te RBM and DBN are more appropriate tan te GMM for generating acoustic features by sampling, wic may elp make te syntetic speec less monotonic and boring. Tird, in te work presented in tis paper, te decision tress for model clustering are still constructed using melcepstra and single Gaussian state PDF. To extend te RBM and DBN modeling from PDF estimation for te clustered states to model clustering for te fully context-dependent states will also be a task of our future work. Besides te spectral

MULTI-DISTRIBUTION DEEP BELIEF NETWORK FOR SPEECH SYNTHESIS. Shiyin Kang, Xiaojun Qian and Helen Meng

MULTI-DISTRIBUTION DEEP BELIEF NETWORK FOR SPEECH SYNTHESIS. Shiyin Kang, Xiaojun Qian and Helen Meng MULTI-DISTRIBUTION DEEP BELIEF NETORK FOR SPEECH SYNTHESIS Siyin Kang, Xiaojun Qian and Helen Meng Human Computer Communications Laboratory, Department of Systems Engineering and Engineering Management,

More information

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks Reading Group on Deep Learning Session 4 Unsupervised Neural Networks Jakob Verbeek & Daan Wynen 206-09-22 Jakob Verbeek & Daan Wynen Unsupervised Neural Networks Outline Autoencoders Restricted) Boltzmann

More information

Probabilistic Graphical Models Homework 1: Due January 29, 2014 at 4 pm

Probabilistic Graphical Models Homework 1: Due January 29, 2014 at 4 pm Probabilistic Grapical Models 10-708 Homework 1: Due January 29, 2014 at 4 pm Directions. Tis omework assignment covers te material presented in Lectures 1-3. You must complete all four problems to obtain

More information

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Moammad Ali Keyvanrad a, Moammad Medi Homayounpour a a Laboratory for Intelligent Multimedia Processing (LIMP), Computer

More information

Natural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs

Natural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs Natural Language Understanding Lecture 12: Recurrent Neural Networks and LSTMs Recap: probability, language models, and feedforward networks Simple Recurrent Networks Adam Lopez Credits: Mirella Lapata

More information

Minimizing D(Q,P) def = Q(h)

Minimizing D(Q,P) def = Q(h) Inference Lecture 20: Variational Metods Kevin Murpy 29 November 2004 Inference means computing P( i v), were are te idden variables v are te visible variables. For discrete (eg binary) idden nodes, exact

More information

Learning based super-resolution land cover mapping

Learning based super-resolution land cover mapping earning based super-resolution land cover mapping Feng ing, Yiang Zang, Giles M. Foody IEEE Fellow, Xiaodong Xiuua Zang, Siming Fang, Wenbo Yun Du is work was supported in part by te National Basic Researc

More information

LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION

LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION Xiaowen Dong, Dorina Tanou, Pascal Frossard and Pierre Vandergeynst Media Lab, MIT, USA xdong@mit.edu Signal Processing Laboratories, EPFL,

More information

Overdispersed Variational Autoencoders

Overdispersed Variational Autoencoders Overdispersed Variational Autoencoders Harsil Sa, David Barber and Aleksandar Botev Department of Computer Science, University College London Alan Turing Institute arsil.sa.15@ucl.ac.uk, david.barber@ucl.ac.uk,

More information

232 Calculus and Structures

232 Calculus and Structures 3 Calculus and Structures CHAPTER 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS FOR EVALUATING BEAMS Calculus and Structures 33 Copyrigt Capter 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS 17.1 THE

More information

Notes on Neural Networks

Notes on Neural Networks Artificial neurons otes on eural etwors Paulo Eduardo Rauber 205 Consider te data set D {(x i y i ) i { n} x i R m y i R d } Te tas of supervised learning consists on finding a function f : R m R d tat

More information

Lecture 16 Deep Neural Generative Models

Lecture 16 Deep Neural Generative Models Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed

More information

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS Po-Ceng Cang National Standard Time & Frequency Lab., TL, Taiwan 1, Lane 551, Min-Tsu Road, Sec. 5, Yang-Mei, Taoyuan, Taiwan 36 Tel: 886 3

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Speaker Representation and Verification Part II. by Vasileios Vasilakakis Speaker Representation and Verification Part II by Vasileios Vasilakakis Outline -Approaches of Neural Networks in Speaker/Speech Recognition -Feed-Forward Neural Networks -Training with Back-propagation

More information

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk

More information

RECOGNITION of online handwriting aims at finding the

RECOGNITION of online handwriting aims at finding the SUBMITTED ON SEPTEMBER 2017 1 A General Framework for te Recognition of Online Handwritten Grapics Frank Julca-Aguilar, Harold Moucère, Cristian Viard-Gaudin, and Nina S. T. Hirata arxiv:1709.06389v1 [cs.cv]

More information

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning Long Term Time Series Prediction wit Multi-Input Multi-Output Local Learning Gianluca Bontempi Macine Learning Group, Département d Informatique Faculté des Sciences, ULB, Université Libre de Bruxelles

More information

INTRODUCTION AND MATHEMATICAL CONCEPTS

INTRODUCTION AND MATHEMATICAL CONCEPTS INTODUCTION ND MTHEMTICL CONCEPTS PEVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips of sine,

More information

2.11 That s So Derivative

2.11 That s So Derivative 2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point

More information

Adaptive Neural Filters with Fixed Weights

Adaptive Neural Filters with Fixed Weights Adaptive Neural Filters wit Fixed Weigts James T. Lo and Justin Nave Department of Matematics and Statistics University of Maryland Baltimore County Baltimore, MD 150, U.S.A. e-mail: jameslo@umbc.edu Abstract

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

Efficient algorithms for for clone items detection

Efficient algorithms for for clone items detection Efficient algoritms for for clone items detection Raoul Medina, Caroline Noyer, and Olivier Raynaud Raoul Medina, Caroline Noyer and Olivier Raynaud LIMOS - Université Blaise Pascal, Campus universitaire

More information

Exemplar-based voice conversion using non-negative spectrogram deconvolution

Exemplar-based voice conversion using non-negative spectrogram deconvolution Exemplar-based voice conversion using non-negative spectrogram deconvolution Zhizheng Wu 1, Tuomas Virtanen 2, Tomi Kinnunen 3, Eng Siong Chng 1, Haizhou Li 1,4 1 Nanyang Technological University, Singapore

More information

Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution

Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution International Journal of Clinical Medicine Researc 2016; 3(5): 76-80 ttp://www.aascit.org/journal/ijcmr ISSN: 2375-3838 Estimating Peak Bone Mineral Density in Osteoporosis Diagnosis by Maximum Distribution

More information

How to Find the Derivative of a Function: Calculus 1

How to Find the Derivative of a Function: Calculus 1 Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te

More information

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

Technology-Independent Design of Neurocomputers: The Universal Field Computer 1

Technology-Independent Design of Neurocomputers: The Universal Field Computer 1 Tecnology-Independent Design of Neurocomputers: Te Universal Field Computer 1 Abstract Bruce J. MacLennan Computer Science Department Naval Postgraduate Scool Monterey, CA 9393 We argue tat AI is moving

More information

Symmetry Labeling of Molecular Energies

Symmetry Labeling of Molecular Energies Capter 7. Symmetry Labeling of Molecular Energies Notes: Most of te material presented in tis capter is taken from Bunker and Jensen 1998, Cap. 6, and Bunker and Jensen 2005, Cap. 7. 7.1 Hamiltonian Symmetry

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note

More information

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features Heiga ZEN (Byung Ha CHUN) Nagoya Inst. of Tech., Japan Overview. Research backgrounds 2.

More information

Dedicated to the 70th birthday of Professor Lin Qun

Dedicated to the 70th birthday of Professor Lin Qun Journal of Computational Matematics, Vol.4, No.3, 6, 4 44. ACCELERATION METHODS OF NONLINEAR ITERATION FOR NONLINEAR PARABOLIC EQUATIONS Guang-wei Yuan Xu-deng Hang Laboratory of Computational Pysics,

More information

Introduction to Machine Learning. Recitation 8. w 2, b 2. w 1, b 1. z 0 z 1. The function we want to minimize is the loss over all examples: f =

Introduction to Machine Learning. Recitation 8. w 2, b 2. w 1, b 1. z 0 z 1. The function we want to minimize is the loss over all examples: f = Introduction to Macine Learning Lecturer: Regev Scweiger Recitation 8 Fall Semester Scribe: Regev Scweiger 8.1 Backpropagation We will develop and review te backpropagation algoritm for neural networks.

More information

INTRODUCTION AND MATHEMATICAL CONCEPTS

INTRODUCTION AND MATHEMATICAL CONCEPTS Capter 1 INTRODUCTION ND MTHEMTICL CONCEPTS PREVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips

More information

3.1 Extreme Values of a Function

3.1 Extreme Values of a Function .1 Etreme Values of a Function Section.1 Notes Page 1 One application of te derivative is finding minimum and maimum values off a grap. In precalculus we were only able to do tis wit quadratics by find

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Deep unsupervised learning

Deep unsupervised learning Deep unsupervised learning Advanced data-mining Yongdai Kim Department of Statistics, Seoul National University, South Korea Unsupervised learning In machine learning, there are 3 kinds of learning paradigm.

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

The Verlet Algorithm for Molecular Dynamics Simulations

The Verlet Algorithm for Molecular Dynamics Simulations Cemistry 380.37 Fall 2015 Dr. Jean M. Standard November 9, 2015 Te Verlet Algoritm for Molecular Dynamics Simulations Equations of motion For a many-body system consisting of N particles, Newton's classical

More information

Pre-Calculus Review Preemptive Strike

Pre-Calculus Review Preemptive Strike Pre-Calculus Review Preemptive Strike Attaced are some notes and one assignment wit tree parts. Tese are due on te day tat we start te pre-calculus review. I strongly suggest reading troug te notes torougly

More information

Bayesian ML Sequence Detection for ISI Channels

Bayesian ML Sequence Detection for ISI Channels Bayesian ML Sequence Detection for ISI Cannels Jill K. Nelson Department of Electrical and Computer Engineering George Mason University Fairfax, VA 030 Email: jnelson@gmu.edu Andrew C. Singer Department

More information

Problem Solving. Problem Solving Process

Problem Solving. Problem Solving Process Problem Solving One of te primary tasks for engineers is often solving problems. It is wat tey are, or sould be, good at. Solving engineering problems requires more tan just learning new terms, ideas and

More information

New Distribution Theory for the Estimation of Structural Break Point in Mean

New Distribution Theory for the Estimation of Structural Break Point in Mean New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University

More information

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY (Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative

More information

5.1 We will begin this section with the definition of a rational expression. We

5.1 We will begin this section with the definition of a rational expression. We Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go

More information

Flavius Guiaş. X(t + h) = X(t) + F (X(s)) ds.

Flavius Guiaş. X(t + h) = X(t) + F (X(s)) ds. Numerical solvers for large systems of ordinary differential equations based on te stocastic direct simulation metod improved by te and Runge Kutta principles Flavius Guiaş Abstract We present a numerical

More information

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these. Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra

More information

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2

More information

MAT244 - Ordinary Di erential Equations - Summer 2016 Assignment 2 Due: July 20, 2016

MAT244 - Ordinary Di erential Equations - Summer 2016 Assignment 2 Due: July 20, 2016 MAT244 - Ordinary Di erential Equations - Summer 206 Assignment 2 Due: July 20, 206 Full Name: Student #: Last First Indicate wic Tutorial Section you attend by filling in te appropriate circle: Tut 0

More information

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES Cristoper J. Roy Sandia National Laboratories* P. O. Box 5800, MS 085 Albuquerque, NM 8785-085 AIAA Paper 00-606 Abstract New developments

More information

Department of Statistics & Operations Research, Aligarh Muslim University, Aligarh, India

Department of Statistics & Operations Research, Aligarh Muslim University, Aligarh, India Open Journal of Optimization, 04, 3, 68-78 Publised Online December 04 in SciRes. ttp://www.scirp.org/ournal/oop ttp://dx.doi.org/0.436/oop.04.34007 Compromise Allocation for Combined Ratio Estimates of

More information

Predicting the Intelligibility of Noisy and Non-Linearly Processed Binaural Speech

Predicting the Intelligibility of Noisy and Non-Linearly Processed Binaural Speech 1 Predicting te Intelligibility of Noisy and Non-Linearly Processed Binaural Speec Asger Heidemann Andersen, Jan Mark de Haan, Zeng-Hua Tan, and Jesper Jensen Abstract Objective speec intelligibility measures

More information

Financial Econometrics Prof. Massimo Guidolin

Financial Econometrics Prof. Massimo Guidolin CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis

More information

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties

More information

Higher Derivatives. Differentiable Functions

Higher Derivatives. Differentiable Functions Calculus 1 Lia Vas Higer Derivatives. Differentiable Functions Te second derivative. Te derivative itself can be considered as a function. Te instantaneous rate of cange of tis function is te second derivative.

More information

Quantum Numbers and Rules

Quantum Numbers and Rules OpenStax-CNX module: m42614 1 Quantum Numbers and Rules OpenStax College Tis work is produced by OpenStax-CNX and licensed under te Creative Commons Attribution License 3.0 Abstract Dene quantum number.

More information

Bounds on the Moments for an Ensemble of Random Decision Trees

Bounds on the Moments for an Ensemble of Random Decision Trees Noname manuscript No. (will be inserted by te editor) Bounds on te Moments for an Ensemble of Random Decision Trees Amit Durandar Received: Sep. 17, 2013 / Revised: Mar. 04, 2014 / Accepted: Jun. 30, 2014

More information

Discontinuous Galerkin Methods for Relativistic Vlasov-Maxwell System

Discontinuous Galerkin Methods for Relativistic Vlasov-Maxwell System Discontinuous Galerkin Metods for Relativistic Vlasov-Maxwell System He Yang and Fengyan Li December 1, 16 Abstract e relativistic Vlasov-Maxwell (RVM) system is a kinetic model tat describes te dynamics

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

IEOR 165 Lecture 10 Distribution Estimation

IEOR 165 Lecture 10 Distribution Estimation IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat

More information

ESTIMATION OF TIME-DOMAIN FREQUENCY STABILITY BASED ON PHASE NOISE MEASUREMENT

ESTIMATION OF TIME-DOMAIN FREQUENCY STABILITY BASED ON PHASE NOISE MEASUREMENT ESTIMATION OF TIME-DOMAIN FREQUENCY STABILITY BASED ON PHASE NOISE MEASUREMENT P. C. Cang, H. M. Peng, and S. Y. Lin National Standard Time & Frequenc Laborator, TL, Taiwan, Lane 55, Min-Tsu Road, Sec.

More information

On the Identifiability of the Post-Nonlinear Causal Model

On the Identifiability of the Post-Nonlinear Causal Model UAI 9 ZHANG & HYVARINEN 647 On te Identifiability of te Post-Nonlinear Causal Model Kun Zang Dept. of Computer Science and HIIT University of Helsinki Finland Aapo Hyvärinen Dept. of Computer Science,

More information

[db]

[db] Blind Source Separation based on Second-Order Statistics wit Asymptotically Optimal Weigting Arie Yeredor Department of EE-Systems, el-aviv University P.O.Box 3900, el-aviv 69978, Israel Abstract Blind

More information

Mixtures of Gaussians with Sparse Structure

Mixtures of Gaussians with Sparse Structure Mixtures of Gaussians with Sparse Structure Costas Boulis 1 Abstract When fitting a mixture of Gaussians to training data there are usually two choices for the type of Gaussians used. Either diagonal or

More information

Cubic Functions: Local Analysis

Cubic Functions: Local Analysis Cubic function cubing coefficient Capter 13 Cubic Functions: Local Analysis Input-Output Pairs, 378 Normalized Input-Output Rule, 380 Local I-O Rule Near, 382 Local Grap Near, 384 Types of Local Graps

More information

Spike train entropy-rate estimation using hierarchical Dirichlet process priors

Spike train entropy-rate estimation using hierarchical Dirichlet process priors publised in: Advances in Neural Information Processing Systems 26 (23), 276 284. Spike train entropy-rate estimation using ierarcical Diriclet process priors Karin Knudson Department of Matematics kknudson@mat.utexas.edu

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

INTEGRATING IMPERFECTION OF INFORMATION INTO THE PROMETHEE MULTICRITERIA DECISION AID METHODS: A GENERAL FRAMEWORK

INTEGRATING IMPERFECTION OF INFORMATION INTO THE PROMETHEE MULTICRITERIA DECISION AID METHODS: A GENERAL FRAMEWORK F O U N D A T I O N S O F C O M P U T I N G A N D D E C I S I O N S C I E N C E S Vol. 7 (0) No. DOI: 0.478/v009-0-000-0 INTEGRATING IMPERFECTION OF INFORMATION INTO THE PROMETHEE MULTICRITERIA DECISION

More information

REVIEW LAB ANSWER KEY

REVIEW LAB ANSWER KEY REVIEW LAB ANSWER KEY. Witout using SN, find te derivative of eac of te following (you do not need to simplify your answers): a. f x 3x 3 5x x 6 f x 3 3x 5 x 0 b. g x 4 x x x notice te trick ere! x x g

More information

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5 THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION NEW MODULAR SCHEME introduced from te examinations in 009 MODULE 5 SOLUTIONS FOR SPECIMEN PAPER B THE QUESTIONS ARE CONTAINED IN A SEPARATE FILE

More information

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL Georgeta Budura Politenica University of Timisoara, Faculty of Electronics and Telecommunications, Comm. Dep., georgeta.budura@etc.utt.ro Abstract:

More information

Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks

Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks Jeremy C. Weiss University of Wisconsin-Madison Madison, WI, US jcweiss@cs.wisc.edu Sriraam Natarajan Indiana University

More information

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous

More information

158 Calculus and Structures

158 Calculus and Structures 58 Calculus and Structures CHAPTER PROPERTIES OF DERIVATIVES AND DIFFERENTIATION BY THE EASY WAY. Calculus and Structures 59 Copyrigt Capter PROPERTIES OF DERIVATIVES. INTRODUCTION In te last capter you

More information

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to

More information

Time (hours) Morphine sulfate (mg)

Time (hours) Morphine sulfate (mg) Mat Xa Fall 2002 Review Notes Limits and Definition of Derivative Important Information: 1 According to te most recent information from te Registrar, te Xa final exam will be eld from 9:15 am to 12:15

More information

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed.

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed. Preface Here are my online notes for my course tat I teac ere at Lamar University. Despite te fact tat tese are my class notes, tey sould be accessible to anyone wanting to learn or needing a refreser

More information

Chapter 2 Ising Model for Ferromagnetism

Chapter 2 Ising Model for Ferromagnetism Capter Ising Model for Ferromagnetism Abstract Tis capter presents te Ising model for ferromagnetism, wic is a standard simple model of a pase transition. Using te approximation of mean-field teory, te

More information

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means

More information

Artificial Neural Network Model Based Estimation of Finite Population Total

Artificial Neural Network Model Based Estimation of Finite Population Total International Journal of Science and Researc (IJSR), India Online ISSN: 2319-7064 Artificial Neural Network Model Based Estimation of Finite Population Total Robert Kasisi 1, Romanus O. Odiambo 2, Antony

More information

Investigating Euler s Method and Differential Equations to Approximate π. Lindsay Crowl August 2, 2001

Investigating Euler s Method and Differential Equations to Approximate π. Lindsay Crowl August 2, 2001 Investigating Euler s Metod and Differential Equations to Approximate π Lindsa Crowl August 2, 2001 Tis researc paper focuses on finding a more efficient and accurate wa to approximate π. Suppose tat x

More information

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx. Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions

More information

Effect of the Dependent Paths in Linear Hull

Effect of the Dependent Paths in Linear Hull 1 Effect of te Dependent Pats in Linear Hull Zenli Dai, Meiqin Wang, Yue Sun Scool of Matematics, Sandong University, Jinan, 250100, Cina Key Laboratory of Cryptologic Tecnology and Information Security,

More information

Average Rate of Change

Average Rate of Change Te Derivative Tis can be tougt of as an attempt to draw a parallel (pysically and metaporically) between a line and a curve, applying te concept of slope to someting tat isn't actually straigt. Te slope

More information

University of Cambridge. MPhil in Computer Speech Text & Internet Technology. Module: Speech Processing II. Lecture 2: Hidden Markov Models I

University of Cambridge. MPhil in Computer Speech Text & Internet Technology. Module: Speech Processing II. Lecture 2: Hidden Markov Models I University of Cambridge MPhil in Computer Speech Text & Internet Technology Module: Speech Processing II Lecture 2: Hidden Markov Models I o o o o o 1 2 3 4 T 1 b 2 () a 12 2 a 3 a 4 5 34 a 23 b () b ()

More information

Journal of Computational and Applied Mathematics

Journal of Computational and Applied Mathematics Journal of Computational and Applied Matematics 94 (6) 75 96 Contents lists available at ScienceDirect Journal of Computational and Applied Matematics journal omepage: www.elsevier.com/locate/cam Smootness-Increasing

More information

Exercises for numerical differentiation. Øyvind Ryan

Exercises for numerical differentiation. Øyvind Ryan Exercises for numerical differentiation Øyvind Ryan February 25, 2013 1. Mark eac of te following statements as true or false. a. Wen we use te approximation f (a) (f (a +) f (a))/ on a computer, we can

More information

Simulation and verification of a plate heat exchanger with a built-in tap water accumulator

Simulation and verification of a plate heat exchanger with a built-in tap water accumulator Simulation and verification of a plate eat excanger wit a built-in tap water accumulator Anders Eriksson Abstract In order to test and verify a compact brazed eat excanger (CBE wit a built-in accumulation

More information

Mathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative

Mathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative Matematics 5 Workseet 11 Geometry, Tangency, and te Derivative Problem 1. Find te equation of a line wit slope m tat intersects te point (3, 9). Solution. Te equation for a line passing troug a point (x

More information

EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS

EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Statistica Sinica 24 2014, 395-414 doi:ttp://dx.doi.org/10.5705/ss.2012.064 EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Jun Sao 1,2 and Seng Wang 3 1 East Cina Normal University,

More information

Hidden Markov Models and Gaussian Mixture Models

Hidden Markov Models and Gaussian Mixture Models Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian

More information

MANY scientific and engineering problems can be

MANY scientific and engineering problems can be A Domain Decomposition Metod using Elliptical Arc Artificial Boundary for Exterior Problems Yajun Cen, and Qikui Du Abstract In tis paper, a Diriclet-Neumann alternating metod using elliptical arc artificial

More information

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT LIMITS AND DERIVATIVES Te limit of a function is defined as te value of y tat te curve approaces, as x approaces a particular value. Te limit of f (x) as x approaces a is written as f (x) approaces, as

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Click here to see an animation of the derivative

Click here to see an animation of the derivative Differentiation Massoud Malek Derivative Te concept of derivative is at te core of Calculus; It is a very powerful tool for understanding te beavior of matematical functions. It allows us to optimize functions,

More information

An Empirical Bayesian interpretation and generalization of NL-means

An Empirical Bayesian interpretation and generalization of NL-means Computer Science Tecnical Report TR2010-934, October 2010 Courant Institute of Matematical Sciences, New York University ttp://cs.nyu.edu/web/researc/tecreports/reports.tml An Empirical Bayesian interpretation

More information