MULTI-DISTRIBUTION DEEP BELIEF NETWORK FOR SPEECH SYNTHESIS. Shiyin Kang, Xiaojun Qian and Helen Meng

Size: px
Start display at page:

Download "MULTI-DISTRIBUTION DEEP BELIEF NETWORK FOR SPEECH SYNTHESIS. Shiyin Kang, Xiaojun Qian and Helen Meng"

Transcription

1 MULTI-DISTRIBUTION DEEP BELIEF NETORK FOR SPEECH SYNTHESIS Siyin Kang, Xiaojun Qian and Helen Meng Human Computer Communications Laboratory, Department of Systems Engineering and Engineering Management, Te Cinese University of Hong Kong, Hong Kong SAR, Cina ABSTRACT Deep belief network (DBN) as been sown to be a good generative model in tasks suc as and-written digit image generation. Previous work on DBN in te speec community mainly focuses on using te generatively pre-trained DBN to initialize a discriminative model for better acoustic modeling in speec recognition (SR). To fully utilize its generative nature, we propose to model te speec parameters including spectrum and F0 simultaneously and generate tese parameters from DBN for speec syntesis. Compared wit te predominant HMM-based approac, objective evaluation sows tat te spectrum generated from DBN as less distortion. Subjective results also confirm te advantage of te spectrum from DBN, and te overall quality is comparable to tat of context-independent HMM. Index Terms Speec syntesis, Deep belief network 1. INTRODUCTION Te past decade as witnessed te success of HMM-based text-to-speec (TTS) syntesis [1]. Te core underpinning tecniques include: (1) te adoption of multi-space distribution HMMs in estimating te statistical beavior of speec parameters [2]; (2) te parameter generation algoritm wic uses dynamic features to smoot te originally piece-wise constant speec parameters drawn from te HMM states [3]. Tis work is our first attempt in using deep belief network (DBN) to syntesize speec. DBN is a probabilistic generative model wic is composed of multiple layers of stocastic idden variables [4]. Previously, DBN as been sown to learn very good generative and discriminative models on ig-dimensional data suc as andwritten digits [4],facial pictures [5], uman motion [6] and large vocabulary speec recognition [7]. Tis motivates us to use DBN to model te wide-band spectrogram and F0 contour for speec syntesis. In tis paper, te basic linguistic unit is te Mandarin tonal syllable. e keep te speec vocoding framework as in conventional statistical parametric speec syntesis [8]. DBN is used to model te joint probability of syllables and speec parameters. Te major difference between our approac and te prevalent HMM-based approac is: Instead of using a variable sequence of states to represent te time dynamics of eac syllable, we evenly sample a fixed number of frames witin te delimited syllable boundary wit ig resolution. Besides, in contrast to previous application of DBN in acoustic modeling for SR, we model te continuous spectrum, discrete voice/unvoiced decision and te multi-space F0 pattern simultaneously rater tan only te continuous spectrum [9]. Te rest of te paper is organized as follows: e will introduce DBN in te context of speec syntesis in Section 2. Te procedures to syntesize speec using a DBN will be described in Section 3. Te experiments and results will be sown in Section 4. Finally we present te conclusions and te future directions. 2. MULTI-DISTRIBUTION DEEP BELIEF NETORK (MD-DBN) Speec production teory describes ow te linguistic message undergoes a series of neuromuscular and articulatory processes before acoustic realization. To mimic tis sequential process, we attempt to model speec production wit a directed belief network. However, it is difficult to learn a multi-layered belief network layer by layer, wic involves inferring te posterior of idden units immediately above. Te insigt of [4] states tat tis inference can be greatly simplified, i.e., just deriving te posterior from an up-pass, if we assume te factorial prior coming from te upper layers is defined by a restricted Boltzmann macine (RBM) wic sares te same connection weigts wit te current layer. Tis insigt also enables a layer-wise greedy construction of a DBN from bottom-up using RBMs as te building blocks. To build an MD-DBN for speec syntesis, we use tree types of RBMs: (1) mixed Gaussian-Bernoulli RBMs, for spectrum, log-f0 and voiced-unvoiced representation wit assumed Gaussian or Bernoulli distributions; (2) mixed Categorical-Bernoulli RBMs, for capturing te correspondence between syllable identities and te binary data derived from te speec representation; and (3) Bernoulli RBMs, wic are used to encode binary data.

2 2.1. Bernoulli RBM (B-RBM) A B-RBM is an undirected grapical model wit one layer of stocastic visible binary units v and one layer of stocastic idden binary units. Tere is no interaction between units in te same layer and is tus restricted. It defines te energy of a visible-idden configuration (v, ) as follows: E(v, ; Θ) = T v b T v a T, (1) were Θ = {, a, b} is te set of parameters of a RBM and Θ will be omitted for clarity ereafter. w ij is te weigt of te symmetric connection between te idden unit i and te visible unit j, wile a i and b j are teir bias terms. Te distribution of te (v, ) configuration is: Pr(v, ) = exp( E(v, ))/Z (Z is te normalization term), i.e., te iger te energy, te lower te probability. Te nice property of tis setting is tat te two conditionals Pr( i = 1 v) and Pr(v j = 1 ) can be obtained easily using te fact tat i and v j can only be eiter 0 or 1: Pr( i = 1 v) = σ( j Pr(v j = 1 ) = σ( i w ij v j + a i ), (2) w ij i + b j ), (3) were σ(x) = (1 + e x ) 1. To optimize te log-likeliood of v in a first-order approac, we need te gradient of log Pr(v) wit respect to any θ in Θ: E(v, ) θ E(v, ) θ Pr( v) Pr(, v). v (4) Given te instantiated observation v, te expectation of derivatives in te first term in Eqn. (4) can be easily computed. Unfortunately, te second term in Eqn. (4) involves a summation over all possible v and is intractable. A widely applied metod tat approximates tis summation is te Gibbs sampler wic (optionally starts from v) proceeds in a Markov cain as follows: v (0) v, (0) Pr( v (0) ); (5a) v (1) Pr(v (0) ), (1) Pr( v (1) ); (5b) Contrastive divergence (CD) training [4] makes two furter approximations: (1) tat te cain starts from te clamped v and is run for only k steps (CD-k); (2) te summation is replaced by a single sample. In particular, CD-1 measures te discrepancy between v and its reconstruction to present a direction for optimization. Starting from a training frame v (0), we only sample (0) in Eqn. (5a) and use te expectations for Eqn. (2) and Eqn. (3) to replace te random samples v (1) and (1) in Eqn. (5b) for stability Mixed Gaussian-Bernoulli RBM (GB-RBM) Speec parameters for syntesis include te spectrum and te log-f0 wit assumed Gaussian distribution, and te voicedunvoiced switces wic are essentially binary. To model tese parameters simultaneously, we design te following energy function for GB-RBM: E(v g, v b, ) = T g v g (vg µ) T (v g µ) T b v b b T v b a T, (6) were v g and v b are te Gaussian units and te Bernoulli units in te visible layer, g and b are te respective weigt matrices, and µ is te mean of v g. Te conditional Pr( v g, v b ) can be similarly derived as: Pr( i = 1 v g, v b ) = σ( j w g ij vg j + j w b ijv b j + a i ), (7) and Pr(v b j = 1 ) follows Eqn. (3). Te conditional Pr(v g j ) involves an integral over te continuous vg. e can sow tat: Pr(v g j ) = N (vg j ; i w g ij i + µ, 1). (8) Here we ave assumed tat te data is normalized to ave unit variance. Given tese defined conditional probabilities, CD-1 training is te same as described in Section Mixed Categorical-Bernoulli RBM (CB-RBM) As mentioned previously tat te posterior of idden units can be inferred directly from an up-pass, to associate te inferred posteriors v b wit teir corresponding indexed syllable label l c, we need to define te energy of te CB-RBM as follows: E(l c, v b, ) = T l l c b ct l c T b v b b bt v b a T. (9) To make l c {0, 1} K follow a categorical distribution, i.e. representing te syllable identity using te 1-out-of-K code (K is te number of tonal syllables in te TTS system), we restrict l c to ave only K 1-out-of-K codes. It can be sown tat Pr(lj c = 1 ) is defined by te soft-max: Pr(lj c = 1 ) = exp( i wc ij i + b c j ) k exp( i wc ik i + b c (10) k ). Te conditionals Pr(vj b = 1 ) and Pr( i = 1 l c, v b ) take te same form as Eqn. (3) & (7) respectively in te CD-k training procedure.

3 3. DBN-BASED SPEECH SYNTHESIS 3.1. Training Stage Given a syllable s start and end times, we extract 50 uniformly-spaced frames of 24-order Mel-Generalized Cepstrum coefficients (MGCs)[10] plus log-energy, 200 uniformly-spaced frames of voiced/unvoiced (V/UV) decisions and te corresponding log-f0 values witin te syllable s boundary. Te voiced and unvoiced frames are assigned 1s and 0s in teir V/UV units, respectively. Te log-f0 values for te unvoiced frames are set to be all 0 a dummy value for log-f0. Bot MGC and log-f0 ave been normalized to ave zero mean and unit variance. Te MGCs, log-f0 and V/UV units are concatenated to form a 1650-dimensional super-vector for eac syllable, wic is used as te visible layer v for GB-RBM. Te posteriors of te idden layer Pr( (1) i v g, v b ) for all i yielded from a simple up-pass can be used as te visible data for training te immediate upper-layer B-RBM. Likewise, we stack up as many layers of B-RBMs as we want in a similar fasion. Te joint distribution of te top idden layer s posteriors (recursively propagated from below) and te associated syllable label is modeled by a CB-RBM Syntesis Stage For an arbitrary text prompt, we look up te caracters in a dictionary to find out teir tonal syllable pronunciations. Starting from a clamped 1-out-of-K coded syllable label l c and an initial all-zero (N 1) in te top-layer CB-RBM, we calculate te conditionals Pr( (N) i = 1 l c, (N 1) ) for all = 1 (N) ) for all j in layer i in layer (N), Pr( (N 1) j (N 1), in alternative fasion. Tis procedure continues until convergence or a maximum number of iterations is reaced., l c ) is ten recursively passed down in te DBN to Pr( (N 1) j give out Pr(v g j (1) ) and Pr(vj b (1) ) for te Gaussian units and te Bernoulli units respectively in te GB-RBM at te bottom. Te diagram of parameter generation from a DBN is sown in Fig. (1). For eac V/UV frame, we make a voicing decision depending on weter Pr(vj b (1) ) > 0.5. Te log-f0 (if te corresponding V/UV decision is voiced) and MGC parameters are recovered via scaling by te standard deviation and offsetting by te mean. Te duration of eac syllable is te average estimated from te training data. e apply a 25-point median filter on log-f0 trajectory to reduce noise. Ten bot MGCs and log-f0 parameter sequences are interpolated using cubic spline on te utterance level and decimated to yield parameter sequences wit a constant 5-ms frame sift. Te resulting speec parameter sequences are ten fed to te Mel Log Spectral Approximation (MLSA) filter [11] for te final output signal. l c Tonal Syllable ID g v l MGCs & log-f0 ( N ) (1) V/UV ( N ) ( N -1) (1) Fig. 1. Arcitecture of te MD-DBN for speec syntesis Experiment Setup 4. EXPERIMENTS A manually transcribed Mandarin corpus recorded from a female speaker is used for te experiments. Te training set contains 1,000 utterances wit a total lengt of 80.9 minutes, including 23,727 syllable samples. All tese samples are partitioned into 1,364 classes of tonal syllables. Anoter test set wit 100 utterances is used for model arcitecture determination and te objective evaluation. Te RBMs are trained using stocastic gradient descent wit a mini-batc size of 200 training samples. For GB-RBM, 400 epocs are executed wit a learning rate of 0.01 wile for B-RBMs and CB-RBM 200 epocs are executed wit a learning rate of 0.1. During te weigt updates, we apply a 0.9 momentum and a weigt decay. Te training procedure is accelerated by an NVIDIA Tesla M2090 GPU system. For an MD-DBN wit 4 idden layers and 2000 units per layer, te training takes about 1.1 ours. Eac epoc of training GB-RBM, B-RBM and CB-RBM takes 3.5s, 3.7s and 5.7s respectively. A single GPU system runs at about 8 times faster tan an 8-core 2.4 GHz Intel Xeon E CPU system HMM-based Syntesis Baseline e build a Mandarin HMM-based speec syntesis system on te same training set using a standard recipe [12]. MGCs and log-f0 togeter wit teir and 2 are modeled by multi-stream HMMs. Eac syllable HMM as a left-to-rigt topology wit 10 states. Initially, 416 mono-syllable HMMs are estimated as te seed for 1,364 tonal syllable HMMs. In te syntesis stage, speec parameters including MGCs and log-f0 are obtained by te maximum likeliood parameter generation algoritm [3], and are later used to produce te speec waveform troug te MLSA filter. Te syllable durations are te same as tose in te DBN approac. For a b v

4 fair comparison, no contextual information or post-processing voice enancement tecniques are incorporated Optimizing MD-DBN Arcitecture Finding an adequate arcitecture for MD-DBN is important in DBN-based speec syntesis. However, tweaking based on subjective evaluation is not practical due to te vast number of combinations of MD-DBN dept and layer widt. One commonly-used objective metod in voice conversion as well as speec syntesis is to employ spectral distortion [13] between generated speec and target speec. Here we use Mel-Generalized Cepstral Distortion (MGCD) to determine te MD-DBN arcitecture. MGCD is te Euclidean distance between te MGCs of syntesized speec and tat of original speec recording. All te test set prompts are syntesized to compute te average MGCDs for HMM baseline and MD-DBN wit different number of idden layers and different number of units in eac idden layer. As te syllable duration of te speec from te test set and tat of HMM and MD-DBN can be different, we align te MGCs according to te syllable boundary. MGC Distortion idden layers 4 idden layers 5 idden layers 6 idden layers idt of te idden layers Fig. 2. MGCD as a function of te widt of te idden layers. 3H-7H: No. of idden layers. Te MGCD of te HMM baseline is 0.223, wile te DBN approac arcives better result wit a minimal MGCD of As sown in Fig. (2), te MGC distortions are ig wen te layer widt is 1,000, and too many units (4,000) in te idden layer causes large variance of MGCD. Te MD- DBN wit 4 idden layers and idden layer widt of 2,000 units consistently gives te best MGCD. Hence we will be using tese parameters for te remaining of te evaluation Subjective Evaluation A Mean Opinion Scoring (MOS) test is conducted to compare te subjective perception between te DBN approac and te HMM baseline, togeter wit a ybrid approac (MIX) using MGCs from DBN and log-f0 from HMM. In te MOS test, eac of te 10 experienced listeners is asked to rate 10 utterances syntesized by te DBN and te HMM using a 5- point scale (5:excellent, 4:good, 3:fair, 2:poor, 1:bad). Te MOS result is sown in Table (1). System MOS HMM 2.86 DBN 2.88 MIX: DBN MGCs + HMM Log-F Table 1. MOS test result. Altoug te overall MOS score sows a draw between te DBN and te HMM, te two sets of syntesized speec sounds different 1. itout post-processing, HMM s voice sounds quite muffled, but te prosody remains smoot and stable. Te DBN voice sounds muc clearer tan te HMM baseline, and te prosody is lively. However, it seems tat tonal and U/VU decision errors ave a relatively iger cance to occur in te DBN approac, wic can be te main reason tat lowers its MOS score. Te MIX approac gets igest score, wic probably suggests tat te spectrum can be captured by DBN appropriately but HMM does a better job in F0 modeling. Te cross-comparison reveals more details. it te same log-f0 curve (HMM vs. MIX), MGCs generated by DBN leads to iger MOS scores. Tis result agrees wit te objective MGCD measure (MGCD HMM >MGCD DBN ). it te same MGCs (DBN vs. MIX), log-f0 curve generated by HMM results in iger MOS scores. It can be seen tat listeners prefer smooter F0 patterns wen tere is no igerlevel prosody control of te syntesized speec. 5. CONCLUSIONS AND FUTURE ORK e ave described a DBN model wit a multi-distribution visible layer for statistical parametric speec syntesis, in wic te spectrum and F0 are modeled simultaneously in a unified framework. It is sown tat DBN models spectrum better tan HMM and acieves an overall performance tat is comparable to context-independent HMM syntesis. Future directions for improvement include: (1) introducing context information to improve te prosody; and (2) better modeling of te F0 contour wit iger resolution of frame sampling. 6. ACKNOLEDGMENT Te autors would like to tank Dr. Li Deng from MSR and Prof. Fei Sa from USC for fruitful excanges. 1 Te syntesized speec samples can be downloaded from ttp:// sykang/icassp2013/

5 7. REFERENCES [1] T. Yosimura, K. Tokuda, T. Masuko, T. Kobayasi, and T. Kitamura, Simultaneous modeling of spectrum, pitc and duration in HMM-based speec syntesis, in Eurospeec, 1999, pp [13] S. Desai, E. V. Ragavendra, B. Yegnanarayana, A.. Black, and K. Praallad, Voice conversion using artificial neural networks, in ICASSP, 2009, pp [2] K. Tokuda, T. Mausko, N. Miyazaki, and T. Kobayasi, Multi-space probability distribution HMM, IEICE Trans. Inf. & Syst., vol. E85-D, pp , [3] K. Tokuda, T. Yosimura, T. Masuko, T. Kobayasi, and T. Kitamura, Speec parameter generation algoritms for HMM-based speec syntesis, in ICASSP, 2000, pp [4] G. E. Hinton, S. Osindero, and Y.. Te, A fast learning algoritm for deep belief nets, Neural Computation, vol. 18, no. 7, pp , [5] Josua M Susskind, Geoffrey E Hinton, Javier R Movellan, and Adam K Anderson, Generating facial expressions wit deep belief nets, Affective Computing, Emotion Modelling, Syntesis and Recognition, pp , [6] G.. Taylor, G. E. Hinton, and S. T. Roweis, Two distributed-state models for generating igdimensional time series, Journal of Macine Learning Researc, vol. 12, pp , [7] George E Dal, Dong Yu, Li Deng, and Alex Acero, Context-dependent pre-trained deep neural networks for large-vocabulary speec recognition, Audio, Speec, and Language Processing, IEEE Transactions on, vol. 20, no. 1, pp , [8] A.. Black, H. Zen, and K. Tokuda, Statistical parametric speec syntesis, in ICASSP, 2007, pp [9] A. Moamed, G. E. Dal, and G. E. Hinton, Acoustic modeling using deep belief networks, IEEE Transactions on Audio, Speec & Language Processing, vol. 20, no. 1, pp , [10] K. Tokuda, T. Kobayasi, T. Masuko, and S. Imai, Melgeneralized cepstral analysis - a unified approac to speec spectral estimation, in ICSLP, [11] T. Fukada, K. Tokuda, T. Kobayasi, and S. Imai, An adaptive algoritm for mel-cepstral analysis of speec, in ICASSP, 1992, pp [12] Z. Suang, S. Kang, Q. Si, Y. Qin, and L. Cai, Syllable HMM based mandarin TTS and comparison wit concatenative TTS, in INTERSPEECH, 2009, pp

THE hidden Markov model (HMM)-based parametric

THE hidden Markov model (HMM)-based parametric JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY 2007 1 Modeling Spectral Envelopes Using Restricted Boltzmann Macines and Deep Belief Networks for Statistical Parametric Speec Syntesis Zen-Hua Ling,

More information

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks Reading Group on Deep Learning Session 4 Unsupervised Neural Networks Jakob Verbeek & Daan Wynen 206-09-22 Jakob Verbeek & Daan Wynen Unsupervised Neural Networks Outline Autoencoders Restricted) Boltzmann

More information

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy

Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Deep Belief Network Training Improvement Using Elite Samples Minimizing Free Energy Moammad Ali Keyvanrad a, Moammad Medi Homayounpour a a Laboratory for Intelligent Multimedia Processing (LIMP), Computer

More information

Natural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs

Natural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs Natural Language Understanding Lecture 12: Recurrent Neural Networks and LSTMs Recap: probability, language models, and feedforward networks Simple Recurrent Networks Adam Lopez Credits: Mirella Lapata

More information

Minimizing D(Q,P) def = Q(h)

Minimizing D(Q,P) def = Q(h) Inference Lecture 20: Variational Metods Kevin Murpy 29 November 2004 Inference means computing P( i v), were are te idden variables v are te visible variables. For discrete (eg binary) idden nodes, exact

More information

Probabilistic Graphical Models Homework 1: Due January 29, 2014 at 4 pm

Probabilistic Graphical Models Homework 1: Due January 29, 2014 at 4 pm Probabilistic Grapical Models 10-708 Homework 1: Due January 29, 2014 at 4 pm Directions. Tis omework assignment covers te material presented in Lectures 1-3. You must complete all four problems to obtain

More information

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab

Te comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk

More information

Notes on Neural Networks

Notes on Neural Networks Artificial neurons otes on eural etwors Paulo Eduardo Rauber 205 Consider te data set D {(x i y i ) i { n} x i R m y i R d } Te tas of supervised learning consists on finding a function f : R m R d tat

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

Overdispersed Variational Autoencoders

Overdispersed Variational Autoencoders Overdispersed Variational Autoencoders Harsil Sa, David Barber and Aleksandar Botev Department of Computer Science, University College London Alan Turing Institute arsil.sa.15@ucl.ac.uk, david.barber@ucl.ac.uk,

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

Adaptive Neural Filters with Fixed Weights

Adaptive Neural Filters with Fixed Weights Adaptive Neural Filters wit Fixed Weigts James T. Lo and Justin Nave Department of Matematics and Statistics University of Maryland Baltimore County Baltimore, MD 150, U.S.A. e-mail: jameslo@umbc.edu Abstract

More information

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx. Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x

More information

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties

More information

New Distribution Theory for the Estimation of Structural Break Point in Mean

New Distribution Theory for the Estimation of Structural Break Point in Mean New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

Introduction to Machine Learning. Recitation 8. w 2, b 2. w 1, b 1. z 0 z 1. The function we want to minimize is the loss over all examples: f =

Introduction to Machine Learning. Recitation 8. w 2, b 2. w 1, b 1. z 0 z 1. The function we want to minimize is the loss over all examples: f = Introduction to Macine Learning Lecturer: Regev Scweiger Recitation 8 Fall Semester Scribe: Regev Scweiger 8.1 Backpropagation We will develop and review te backpropagation algoritm for neural networks.

More information

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Speaker Representation and Verification Part II. by Vasileios Vasilakakis Speaker Representation and Verification Part II by Vasileios Vasilakakis Outline -Approaches of Neural Networks in Speaker/Speech Recognition -Feed-Forward Neural Networks -Training with Back-propagation

More information

How to Find the Derivative of a Function: Calculus 1

How to Find the Derivative of a Function: Calculus 1 Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te

More information

Technology-Independent Design of Neurocomputers: The Universal Field Computer 1

Technology-Independent Design of Neurocomputers: The Universal Field Computer 1 Tecnology-Independent Design of Neurocomputers: Te Universal Field Computer 1 Abstract Bruce J. MacLennan Computer Science Department Naval Postgraduate Scool Monterey, CA 9393 We argue tat AI is moving

More information

Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks

Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks Learning to Reject Sequential Importance Steps for Continuous-Time Bayesian Networks Jeremy C. Weiss University of Wisconsin-Madison Madison, WI, US jcweiss@cs.wisc.edu Sriraam Natarajan Indiana University

More information

Bayesian ML Sequence Detection for ISI Channels

Bayesian ML Sequence Detection for ISI Channels Bayesian ML Sequence Detection for ISI Cannels Jill K. Nelson Department of Electrical and Computer Engineering George Mason University Fairfax, VA 030 Email: jnelson@gmu.edu Andrew C. Singer Department

More information

2.11 That s So Derivative

2.11 That s So Derivative 2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point

More information

Usually the estimation of the partition function is intractable and it becomes exponentially hard when the complexity of the model increases. However,

Usually the estimation of the partition function is intractable and it becomes exponentially hard when the complexity of the model increases. However, Odyssey 2012 The Speaker and Language Recognition Workshop 25-28 June 2012, Singapore First attempt of Boltzmann Machines for Speaker Verification Mohammed Senoussaoui 1,2, Najim Dehak 3, Patrick Kenny

More information

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

Lecture 16 Deep Neural Generative Models

Lecture 16 Deep Neural Generative Models Lecture 16 Deep Neural Generative Models CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago May 22, 2017 Approach so far: We have considered simple models and then constructed

More information

RECOGNITION of online handwriting aims at finding the

RECOGNITION of online handwriting aims at finding the SUBMITTED ON SEPTEMBER 2017 1 A General Framework for te Recognition of Online Handwritten Grapics Frank Julca-Aguilar, Harold Moucère, Cristian Viard-Gaudin, and Nina S. T. Hirata arxiv:1709.06389v1 [cs.cv]

More information

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS Po-Ceng Cang National Standard Time & Frequency Lab., TL, Taiwan 1, Lane 551, Min-Tsu Road, Sec. 5, Yang-Mei, Taoyuan, Taiwan 36 Tel: 886 3

More information

Efficient algorithms for for clone items detection

Efficient algorithms for for clone items detection Efficient algoritms for for clone items detection Raoul Medina, Caroline Noyer, and Olivier Raynaud Raoul Medina, Caroline Noyer and Olivier Raynaud LIMOS - Université Blaise Pascal, Campus universitaire

More information

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LAURA EVANS.. Introduction Not all differential equations can be explicitly solved for y. Tis can be problematic if we need to know te value of y

More information

1watt=1W=1kg m 2 /s 3

1watt=1W=1kg m 2 /s 3 Appendix A Matematics Appendix A.1 Units To measure a pysical quantity, you need a standard. Eac pysical quantity as certain units. A unit is just a standard we use to compare, e.g. a ruler. In tis laboratory

More information

Learning based super-resolution land cover mapping

Learning based super-resolution land cover mapping earning based super-resolution land cover mapping Feng ing, Yiang Zang, Giles M. Foody IEEE Fellow, Xiaodong Xiuua Zang, Siming Fang, Wenbo Yun Du is work was supported in part by te National Basic Researc

More information

Spike train entropy-rate estimation using hierarchical Dirichlet process priors

Spike train entropy-rate estimation using hierarchical Dirichlet process priors publised in: Advances in Neural Information Processing Systems 26 (23), 276 284. Spike train entropy-rate estimation using ierarcical Diriclet process priors Karin Knudson Department of Matematics kknudson@mat.utexas.edu

More information

EDML: A Method for Learning Parameters in Bayesian Networks

EDML: A Method for Learning Parameters in Bayesian Networks : A Metod for Learning Parameters in Bayesian Networks Artur Coi, Kaled S. Refaat and Adnan Darwice Computer Science Department University of California, Los Angeles {aycoi, krefaat, darwice}@cs.ucla.edu

More information

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL Georgeta Budura Politenica University of Timisoara, Faculty of Electronics and Telecommunications, Comm. Dep., georgeta.budura@etc.utt.ro Abstract:

More information

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h Lecture Numerical differentiation Introduction We can analytically calculate te derivative of any elementary function, so tere migt seem to be no motivation for calculating derivatives numerically. However

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

Parameter Fitted Scheme for Singularly Perturbed Delay Differential Equations

Parameter Fitted Scheme for Singularly Perturbed Delay Differential Equations International Journal of Applied Science and Engineering 2013. 11, 4: 361-373 Parameter Fitted Sceme for Singularly Perturbed Delay Differential Equations Awoke Andargiea* and Y. N. Reddyb a b Department

More information

Order of Accuracy. ũ h u Ch p, (1)

Order of Accuracy. ũ h u Ch p, (1) Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Chapter 5 FINITE DIFFERENCE METHOD (FDM)

Chapter 5 FINITE DIFFERENCE METHOD (FDM) MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential

More information

Homework 1 Due: Wednesday, September 28, 2016

Homework 1 Due: Wednesday, September 28, 2016 0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q

More information

Autoregressive Neural Models for Statistical Parametric Speech Synthesis

Autoregressive Neural Models for Statistical Parametric Speech Synthesis Autoregressive Neural Models for Statistical Parametric Speech Synthesis シンワン Xin WANG 2018-01-11 contact: wangxin@nii.ac.jp we welcome critical comments, suggestions, and discussion 1 https://www.slideshare.net/kotarotanahashi/deep-learning-library-coyotecnn

More information

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul

More information

Haplotyping. Biostatistics 666

Haplotyping. Biostatistics 666 Haplotyping Biostatistics 666 Previously Introduction to te E-M algoritm Approac for likeliood optimization Examples related to gene counting Allele frequency estimation recessive disorder Allele frequency

More information

Pre-Calculus Review Preemptive Strike

Pre-Calculus Review Preemptive Strike Pre-Calculus Review Preemptive Strike Attaced are some notes and one assignment wit tree parts. Tese are due on te day tat we start te pre-calculus review. I strongly suggest reading troug te notes torougly

More information

Deep unsupervised learning

Deep unsupervised learning Deep unsupervised learning Advanced data-mining Yongdai Kim Department of Statistics, Seoul National University, South Korea Unsupervised learning In machine learning, there are 3 kinds of learning paradigm.

More information

The Origin of Deep Learning. Lili Mou Jan, 2015

The Origin of Deep Learning. Lili Mou Jan, 2015 The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets

More information

Sin, Cos and All That

Sin, Cos and All That Sin, Cos and All Tat James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 9, 2017 Outline Sin, Cos and all tat! A New Power Rule Derivatives

More information

Math 312 Lecture Notes Modeling

Math 312 Lecture Notes Modeling Mat 3 Lecture Notes Modeling Warren Weckesser Department of Matematics Colgate University 5 7 January 006 Classifying Matematical Models An Example We consider te following scenario. During a storm, a

More information

A Theoretically Grounded Application of Dropout in Recurrent Neural Networks

A Theoretically Grounded Application of Dropout in Recurrent Neural Networks A Teoretically Grounded Application of Dropout in Recurrent Neural Networks Yarin Gal University of Cambridge {yg279,zg201}@cam.ac.uk oubin Garamani Abstract Recurrent neural networks (RNNs) stand at te

More information

Symmetry Labeling of Molecular Energies

Symmetry Labeling of Molecular Energies Capter 7. Symmetry Labeling of Molecular Energies Notes: Most of te material presented in tis capter is taken from Bunker and Jensen 1998, Cap. 6, and Bunker and Jensen 2005, Cap. 7. 7.1 Hamiltonian Symmetry

More information

arxiv: v3 [cs.lg] 14 Oct 2015

arxiv: v3 [cs.lg] 14 Oct 2015 Learning Deep Generative Models wit Doubly Stocastic MCMC Cao Du Jun Zu Bo Zang Dept. of Comp. Sci. & Tec., State Key Lab of Intell. Tec. & Sys., TNList Lab, Center for Bio-Inspired Computing Researc,

More information

Flavius Guiaş. X(t + h) = X(t) + F (X(s)) ds.

Flavius Guiaş. X(t + h) = X(t) + F (X(s)) ds. Numerical solvers for large systems of ordinary differential equations based on te stocastic direct simulation metod improved by te and Runge Kutta principles Flavius Guiaş Abstract We present a numerical

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Model development for the beveling of quartz crystal blanks

Model development for the beveling of quartz crystal blanks 9t International Congress on Modelling and Simulation, Pert, Australia, 6 December 0 ttp://mssanz.org.au/modsim0 Model development for te beveling of quartz crystal blanks C. Dong a a Department of Mecanical

More information

232 Calculus and Structures

232 Calculus and Structures 3 Calculus and Structures CHAPTER 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS FOR EVALUATING BEAMS Calculus and Structures 33 Copyrigt Capter 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS 17.1 THE

More information

LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION

LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION LAPLACIAN MATRIX LEARNING FOR SMOOTH GRAPH SIGNAL REPRESENTATION Xiaowen Dong, Dorina Tanou, Pascal Frossard and Pierre Vandergeynst Media Lab, MIT, USA xdong@mit.edu Signal Processing Laboratories, EPFL,

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

Introduction to Derivatives

Introduction to Derivatives Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))

More information

arxiv: v1 [cs.cl] 23 Sep 2013

arxiv: v1 [cs.cl] 23 Sep 2013 Feature Learning with Gaussian Restricted Boltzmann Machine for Robust Speech Recognition Xin Zheng 1,2, Zhiyong Wu 1,2,3, Helen Meng 1,3, Weifeng Li 1, Lianhong Cai 1,2 arxiv:1309.6176v1 [cs.cl] 23 Sep

More information

Nearly Perfect Detection of Continuous F 0 Contour and Frame Classification for TTS Synthesis. Thomas Ewender

Nearly Perfect Detection of Continuous F 0 Contour and Frame Classification for TTS Synthesis. Thomas Ewender Nearly Perfect Detection of Continuous F 0 Contour and Frame Classification for TTS Synthesis Thomas Ewender Outline Motivation Detection algorithm of continuous F 0 contour Frame classification algorithm

More information

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2

More information

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning Long Term Time Series Prediction wit Multi-Input Multi-Output Local Learning Gianluca Bontempi Macine Learning Group, Département d Informatique Faculté des Sciences, ULB, Université Libre de Bruxelles

More information

Online Learning: Bandit Setting

Online Learning: Bandit Setting Online Learning: Bandit Setting Daniel asabi Summer 04 Last Update: October 0, 06 Introduction [TODO Bandits. Stocastic setting Suppose tere exists unknown distributions ν,..., ν, suc tat te loss at eac

More information

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES Cristoper J. Roy Sandia National Laboratories* P. O. Box 5800, MS 085 Albuquerque, NM 8785-085 AIAA Paper 00-606 Abstract New developments

More information

3.1 Extreme Values of a Function

3.1 Extreme Values of a Function .1 Etreme Values of a Function Section.1 Notes Page 1 One application of te derivative is finding minimum and maimum values off a grap. In precalculus we were only able to do tis wit quadratics by find

More information

Chapter 2 Ising Model for Ferromagnetism

Chapter 2 Ising Model for Ferromagnetism Capter Ising Model for Ferromagnetism Abstract Tis capter presents te Ising model for ferromagnetism, wic is a standard simple model of a pase transition. Using te approximation of mean-field teory, te

More information

5.1 We will begin this section with the definition of a rational expression. We

5.1 We will begin this section with the definition of a rational expression. We Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go

More information

Artificial Neural Network Model Based Estimation of Finite Population Total

Artificial Neural Network Model Based Estimation of Finite Population Total International Journal of Science and Researc (IJSR), India Online ISSN: 2319-7064 Artificial Neural Network Model Based Estimation of Finite Population Total Robert Kasisi 1, Romanus O. Odiambo 2, Antony

More information

Quantum Numbers and Rules

Quantum Numbers and Rules OpenStax-CNX module: m42614 1 Quantum Numbers and Rules OpenStax College Tis work is produced by OpenStax-CNX and licensed under te Creative Commons Attribution License 3.0 Abstract Dene quantum number.

More information

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example, NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing

More information

An Autoregressive Recurrent Mixture Density Network for Parametric Speech Synthesis

An Autoregressive Recurrent Mixture Density Network for Parametric Speech Synthesis ICASSP 07 New Orleans, USA An Autoregressive Recurrent Mixture Density Network for Parametric Speech Synthesis Xin WANG, Shinji TAKAKI, Junichi YAMAGISHI National Institute of Informatics, Japan 07-03-07

More information

INTRODUCTION AND MATHEMATICAL CONCEPTS

INTRODUCTION AND MATHEMATICAL CONCEPTS Capter 1 INTRODUCTION ND MTHEMTICL CONCEPTS PREVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips

More information

arxiv: v3 [cs.ds] 4 Aug 2017

arxiv: v3 [cs.ds] 4 Aug 2017 Non-preemptive Sceduling in a Smart Grid Model and its Implications on Macine Minimization Fu-Hong Liu 1, Hsiang-Hsuan Liu 1,2, and Prudence W.H. Wong 2 1 Department of Computer Science, National Tsing

More information

Exact Maximum Margin Structure Learning of Bayesian Networks

Exact Maximum Margin Structure Learning of Bayesian Networks Robert Pearz robert.pearz@tugraz.at Franz Pernkopf pernkopf@tugraz.at Signal Processing and Speec Communication Laboratory, Graz University of Tecnology, Austria Abstract Recently, tere as been muc interest

More information

Section 15.6 Directional Derivatives and the Gradient Vector

Section 15.6 Directional Derivatives and the Gradient Vector Section 15.6 Directional Derivatives and te Gradient Vector Finding rates of cange in different directions Recall tat wen we first started considering derivatives of functions of more tan one variable,

More information

Exemplar-based voice conversion using non-negative spectrogram deconvolution

Exemplar-based voice conversion using non-negative spectrogram deconvolution Exemplar-based voice conversion using non-negative spectrogram deconvolution Zhizheng Wu 1, Tuomas Virtanen 2, Tomi Kinnunen 3, Eng Siong Chng 1, Haizhou Li 1,4 1 Nanyang Technological University, Singapore

More information

COMPARISON OF FUZZY LOGIC CONTROLLERS FOR A MULTIVARIABLE PROCESS

COMPARISON OF FUZZY LOGIC CONTROLLERS FOR A MULTIVARIABLE PROCESS COMPARISON OF FUZZY LOGIC CONTROLLERS FOR A MULTIVARIABLE PROCESS KARTHICK S, LAKSHMI P, DEEPA T 3 PG Student, DEEE, College of Engineering, Guindy, Anna University, Cennai Associate Professor, DEEE, College

More information

1. AB Calculus Introduction

1. AB Calculus Introduction 1. AB Calculus Introduction Before we get into wat calculus is, ere are several eamples of wat you could do BC (before calculus) and wat you will be able to do at te end of tis course. Eample 1: On April

More information

Greedy Layer-Wise Training of Deep Networks

Greedy Layer-Wise Training of Deep Networks Greedy Layer-Wise Training of Deep Networks Yoshua Bengio, Pascal Lamblin, Dan Popovici, Hugo Larochelle NIPS 2007 Presented by Ahmed Hefny Story so far Deep neural nets are more expressive: Can learn

More information

AN IMPROVED WEIGHTED TOTAL HARMONIC DISTORTION INDEX FOR INDUCTION MOTOR DRIVES

AN IMPROVED WEIGHTED TOTAL HARMONIC DISTORTION INDEX FOR INDUCTION MOTOR DRIVES AN IMPROVED WEIGHTED TOTA HARMONIC DISTORTION INDEX FOR INDUCTION MOTOR DRIVES Tomas A. IPO University of Wisconsin, 45 Engineering Drive, Madison WI, USA P: -(608)-6-087, Fax: -(608)-6-5559, lipo@engr.wisc.edu

More information

Comparing Recent Waveform Generation and Acoustic Modeling Methods for Neural-network-based Speech Synthesis

Comparing Recent Waveform Generation and Acoustic Modeling Methods for Neural-network-based Speech Synthesis ICASSP 2018 Calgary, Canada Comparing Recent Waveform Generation and Acoustic Modeling Methods for Neural-network-based Speech Synthesis Xin WANG, Jaime Lorenzo-Trueba, Shinji TAKAKI, Lauri Juvela, Junichi

More information

Financial Econometrics Prof. Massimo Guidolin

Financial Econometrics Prof. Massimo Guidolin CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis

More information

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means

More information

IEOR 165 Lecture 10 Distribution Estimation

IEOR 165 Lecture 10 Distribution Estimation IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat

More information

Termination Problems in Chemical Kinetics

Termination Problems in Chemical Kinetics Termination Problems in Cemical Kinetics Gianluigi Zavattaro and Luca Cardelli 2 Dip. Scienze dell Informazione, Università di Bologna, Italy 2 Microsoft Researc, Cambridge, UK Abstract. We consider nondeterministic

More information

= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c)

= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c) Paper 1: Pure Matematics 1 Mark Sceme 1(a) (i) (ii) d d y 3 1x 4x x M1 A1 d y dx 1.1b 1.1b 36x 48x A1ft 1.1b Substitutes x = into teir dx (3) 3 1 4 Sows d y 0 and states ''ence tere is a stationary point''

More information

Why gravity is not an entropic force

Why gravity is not an entropic force Wy gravity is not an entropic force San Gao Unit for History and Pilosopy of Science & Centre for Time, SOPHI, University of Sydney Email: sgao7319@uni.sydney.edu.au Te remarkable connections between gravity

More information

Knowledge Extraction from DBNs for Images

Knowledge Extraction from DBNs for Images Knowledge Extraction from DBNs for Images Son N. Tran and Artur d Avila Garcez Department of Computer Science City University London Contents 1 Introduction 2 Knowledge Extraction from DBNs 3 Experimental

More information

Deep Generative Models

Deep Generative Models Deep Generative Models Durk Kingma Max Welling Deep Probabilistic Models Worksop Wednesday, 1st of Oct, 2014 D.P. Kingma Deep generative models Transformations between Bayes nets and Neural nets Transformation

More information

Bob Brown Math 251 Calculus 1 Chapter 3, Section 1 Completed 1 CCBC Dundalk

Bob Brown Math 251 Calculus 1 Chapter 3, Section 1 Completed 1 CCBC Dundalk Bob Brown Mat 251 Calculus 1 Capter 3, Section 1 Completed 1 Te Tangent Line Problem Te idea of a tangent line first arises in geometry in te context of a circle. But before we jump into a discussion of

More information

HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS

HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS V Gosbell University of Wollongong Department of Electrical, Computer & Telecommunications Engineering, Wollongong, NSW 2522, Australia

More information

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5 THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION NEW MODULAR SCHEME introduced from te examinations in 009 MODULE 5 SOLUTIONS FOR SPECIMEN PAPER B THE QUESTIONS ARE CONTAINED IN A SEPARATE FILE

More information

Journal of Computational and Applied Mathematics

Journal of Computational and Applied Mathematics Journal of Computational and Applied Matematics 94 (6) 75 96 Contents lists available at ScienceDirect Journal of Computational and Applied Matematics journal omepage: www.elsevier.com/locate/cam Smootness-Increasing

More information

Excluded Volume Effects in Gene Stretching. Pui-Man Lam Physics Department, Southern University Baton Rouge, Louisiana

Excluded Volume Effects in Gene Stretching. Pui-Man Lam Physics Department, Southern University Baton Rouge, Louisiana Excluded Volume Effects in Gene Stretcing Pui-Man Lam Pysics Department, Soutern University Baton Rouge, Louisiana 7083 Abstract We investigate te effects excluded volume on te stretcing of a single DNA

More information

Handling Missing Data on Asymmetric Distribution

Handling Missing Data on Asymmetric Distribution International Matematical Forum, Vol. 8, 03, no. 4, 53-65 Handling Missing Data on Asymmetric Distribution Amad M. H. Al-Kazale Department of Matematics, Faculty of Science Al-albayt University, Al-Mafraq-Jordan

More information