Deep Learning for Speech Processing An NST Perspective

Similar documents
4F10: Deep Learning. Mark Gales. Michaelmas 2016

Segmental Recurrent Neural Networks for End-to-end Speech Recognition

Residual LSTM: Design of a Deep Recurrent Architecture for Distant Speech Recognition

Speech recognition. Lecture 14: Neural Networks. Andrew Senior December 12, Google NYC

Very Deep Convolutional Neural Networks for LVCSR

Modeling Time-Frequency Patterns with LSTM vs. Convolutional Architectures for LVCSR Tasks

Unfolded Recurrent Neural Networks for Speech Recognition

Deep Neural Networks

Why DNN Works for Acoustic Modeling in Speech Recognition?

An exploration of dropout with LSTMs

arxiv: v4 [cs.cl] 5 Jun 2017

Highway-LSTM and Recurrent Highway Networks for Speech Recognition

arxiv: v1 [cs.cl] 22 Feb 2018

SPEECH RECOGNITION WITH DEEP RECURRENT NEURAL NETWORKS. Alex Graves, Abdel-rahman Mohamed and Geoffrey Hinton

End-to-end Automatic Speech Recognition

Deep Recurrent Neural Networks

Gate Activation Signal Analysis for Gated Recurrent Neural Networks and Its Correlation with Phoneme Boundaries

Large Vocabulary Continuous Speech Recognition with Long Short-Term Recurrent Networks

MULTI-FRAME FACTORISATION FOR LONG-SPAN ACOUSTIC MODELLING. Liang Lu and Steve Renals

Introduction to Neural Networks

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

A Primal-Dual Method for Training Recurrent Neural Networks Constrained by the Echo-State Property

Generative Model-Based Text-to-Speech Synthesis. Heiga Zen (Google London) February

Augmented Statistical Models for Speech Recognition

Boundary Contraction Training for Acoustic Models based on Discrete Deep Neural Networks

WaveNet: A Generative Model for Raw Audio

Introduction to Deep Neural Networks

Recurrent Neural Network

Improved Learning through Augmenting the Loss

Deep Learning for Speech Recognition. Hung-yi Lee

Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks

Deep Learning for Automatic Speech Recognition Part II

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Recurrent Neural Networks (Part - 2) Sumit Chopra Facebook

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji

SPEECH recognition systems based on hidden Markov

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Speaker recognition by means of Deep Belief Networks

Language Modelling. Steve Renals. Automatic Speech Recognition ASR Lecture 11 6 March ASR Lecture 11 Language Modelling 1

Neural Architectures for Image, Language, and Speech Processing

Knowledge Extraction from DBNs for Images

From perceptrons to word embeddings. Simon Šuster University of Groningen

a) b) (Natural Language Processing; NLP) (Deep Learning) Bag of words White House RGB [1] IBM

Beyond Cross-entropy: Towards Better Frame-level Objective Functions For Deep Neural Network Training In Automatic Speech Recognition

Deep Learning Recurrent Networks 2/28/2018

Feature-space Speaker Adaptation for Probabilistic Linear Discriminant Analysis Acoustic Models

EE-559 Deep learning Recurrent Neural Networks

Faster Training of Very Deep Networks Via p-norm Gates

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Autoregressive Neural Models for Statistical Parametric Speech Synthesis

arxiv: v1 [cs.cl] 31 May 2015

Reformulating the HMM as a trajectory model by imposing explicit relationship between static and dynamic features

Google s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

arxiv: v1 [cs.cl] 21 May 2017

arxiv: v1 [cs.cl] 23 Sep 2013

Presented By: Omer Shmueli and Sivan Niv

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Generating Sequences with Recurrent Neural Networks

Deep Learning Sequence to Sequence models: Attention Models. 17 March 2018

GAUSSIAN PROCESS LSTM RECURRENT NEURAL NETWORK LANGUAGE MODELS FOR SPEECH RECOGNITION

arxiv: v2 [cs.sd] 7 Feb 2018

Sequence Transduction with Recurrent Neural Networks

Recurrent Neural Networks with Flexible Gates using Kernel Activation Functions

Brief Introduction of Machine Learning Techniques for Content Analysis

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Deep Learning Autoencoder Models

Lecture 14. Advanced Neural Networks. Michael Picheny, Bhuvana Ramabhadran, Stanley F. Chen, Markus Nussbaum-Thom

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Lecture 16 Deep Neural Generative Models

arxiv: v1 [cs.ne] 14 Nov 2012

Reading Group on Deep Learning Session 1

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

arxiv: v2 [cs.cl] 1 Jan 2019

Lecture 11 Recurrent Neural Networks I

STA 4273H: Statistical Machine Learning

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 6: Multi-Layer Perceptrons I

Deep Belief Networks are compact universal approximators

Long-Short Term Memory and Other Gated RNNs

arxiv: v3 [cs.lg] 14 Jan 2018

Gated Recurrent Neural Tensor Network

ADVANCING CONNECTIONIST TEMPORAL CLASSIFICATION WITH ATTENTION MODELING

Conditional Language modeling with attention

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Based on the original slides of Hung-yi Lee

STA 414/2104: Machine Learning

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Probabilistic Reasoning in Deep Learning

arxiv: v2 [cs.ne] 7 Apr 2015

UNSUPERVISED LEARNING

Usually the estimation of the partition function is intractable and it becomes exponentially hard when the complexity of the model increases. However,

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

Recent Advances in Bayesian Inference Techniques

Lecture 17: Neural Networks and Deep Learning

BIDIRECTIONAL LSTM-HMM HYBRID SYSTEM FOR POLYPHONIC SOUND EVENT DETECTION

Adapting Wavenet for Speech Enhancement DARIO RETHAGE JULY 12, 2017

Deep Learning & Neural Networks Lecture 2

Learning Deep Architectures for AI. Part II - Vijay Chakilam

Introduction to Convolutional Neural Networks 2018 / 02 / 23

Transcription:

Deep Learning for Speech Processing An NST Perspective Mark Gales University of Cambridge June 2016 September 2016

Natural Speech Technology (NST) EPSRC (UK Government) Programme Grant: collaboration significantly advance state-of-the-art in speech technology more natural, approaching human levels of reliability, adaptability and conversational richness ran from 2011 to 2016 - interesting times... 2 of 67

What is Deep Learning? From Wikipedia: Deep learning is a branch of machine learning based on a set of algorithms that attempt to model high-level abstractions in data by using multiple processing layers, with complex structures or otherwise, composed of multiple non-linear transformations. For NST: Deep learning allows both speech synthesis and speech recognition to use the same underlying, highly flexible, building blocks and adaptation techniques (and improved performance). 3 of 67

What is Deep Learning? From Wikipedia: Deep learning is a branch of machine learning based on a set of algorithms that attempt to model high-level abstractions in data by using multiple processing layers, with complex structures or otherwise, composed of multiple non-linear transformations. For NST: Deep learning allows both speech synthesis and speech recognition to use the same underlying, highly flexible, building blocks and adaptation techniques (and improved performance). 4 of 67

The Rise of Deep Learning (June 2016) NST 5 of 67

Overview Basic Building Blocks neural network architectures activation functions Sequence-to-Sequence Modelling generative models discriminative models encoder-decoder models Speech Processing Applications language modelling speech recognition speech synthesis Adaptation for Deep Learning 6 of 67

What I am Not Discussing History of development Optimisation essential aspect of all systems (major issue) Only work from NST though (of course) the talk is biased indicates papers/contributions from NST in area Experimental Results see NST publications and other papers (references at end) My personal opinion of DNNs... 7 of 67

Basic Building Blocks 8 of 67

Deep Neural Networks [19] x t y t General mapping process from input x t to output y t y t = F(x t ) deep refers to number of hidden layers Depending on nature of outputs (y t ) vary activation function softmax often for classification tanh or sigmoid common for hidden layers 9 of 67

Neural Network Layer/Node w i z i φ() 10 of 67

Activation Functions 1 0.8 0.6 0.4 0.2 0 0.2 0.4 Activation functions: step function (green) sigmoid function (red) tanh function (blue) 0.6 0.8 1 5 4 3 2 1 0 1 2 3 4 5 softmax, usual output layer (sum-to-one/positive) for classification φ i (x t ) = exp (w T i x t + b i ) Jj=1 exp (w T j x t + b j ) 11 of 67

Activation Functions - ReLUs [35, 54] Alternative activation function: Rectified Linear Units y i = max(0, z i ) Related activation function noisy ReLU: y i = max(0, z i + ɛ); ɛ N (0, σ 2 ) efficient, no exponential/division, rapid convergence in training Possible to train the activation function parameters e.g. gradient of slopes for ReLUs 12 of 67

Activation Functions - Residual Networks [17] x W φ() W + y ReLU Modify layer to model the residual y t = F(x t ) + x t allows deeper networks to be built deep residual learning Links to highway connections 13 of 67

Pooling/Max-Out Functions [27, 43] Possible to pool the output of a set of node reduces the number of weights to connect layers together ϕ() } A range of functions have been examined maxout φ(y 1, y 2, y 3 ) = max(y 1, y 2, y 3 ) soft-maxout φ(y 1, y 2, y 3 ) = log( 3 i=1 exp(y i)) p-norm φ(y 1, y 2, y 3 ) = ( 3 i=1 y i ) 1/p Has also been applied for unsupervised adaptation 14 of 67

Convolutional Neural Networks [26, 1] k frequencies n frames filter 1 filter n pooling layer 15 of 67

Mixture Density Neural Networks [4, 52] x t component means component (log) variances component priors Predict a mixture of experts multiple components M F m (c) (x t ): prior prediction F (µ) m F (σ) m (x t ): mean prediction (x t ): variance prediction Optimise using maximum likelihood M p(y t x t ) = F (c) m=1 m (x t )N (y t ; F m (µ) (x t ), F m (σ) (x t )) Form of output influences output activation function used 16 of 67

Recurrent Neural Networks [38, 37] x t h t 1 h t Time delay y t Introduce recurrent units h t = f h ( W f hx t + W r hh t 1 + b h ) y t = f f ( W y h t + W x x t + b y ) h t history vector at time t Two history weight matrices W f h forward W r h recursion Uses (general) approximation (no optional i/o link) 17 of 67 p(y t x 1:t ) = p(y t x t, x 1:t 1 ) p(y t x t, h t 1 ) p(y t h t ) network has (causal) memory encoded in history vector (h t )

RNN Variants [40, 10] x t x t+1 h t h t+1 y t y t+1 xt h t h t+1 y t y t+1 ~ h t x t+1 h ~ t+1 xt x t+1 z t z t+1 h t h t+1 y t y t+1 RNN (+i/o link) Bi-Directional RNN Variational RNN Bi-directional: use complete observation sequence - non-causal Variational: introduce a latent variable sequence z 1:T p(y t x 1:t ) p(y t x t, z t, h t 1 )p(z t h t 1 )dz t 18 of 67

Network Gating A flexible extension to activation function is gating standard form is (σ() sigmoid activation function) i = σ(w f x t + W r h t 1 + b) vector acts a a probabilistic gate on network values Gating can be applied at various levels features: impact of input/output features on nodes time: memory of the network layer: influence of a layer s activation function 19 of 67

Long-Short Term Memory Networks [20, 16] x t h t 1 x t h t 1 σ σ i i i o x t h t 1 f m c t f h time delay h t σ i f x t h t 1 20 of 67

Long-Short Term Memory Networks The operations can be written as (peephole config): Forget gate (i f ), Input gate (i i ), Output gate (i o ) i f = σ(w f fx t + W r fh t 1 + W m fc t 1 + b f ) i i = σ(w f ix t + W r ih t 1 + W m ic t 1 + b i ) i o = σ(w f ox t + W r oh t 1 + W m oc t + b o ) Memory Cell, history vector and gates are related by c t = i f c t 1 + i i f m (W f cx t + W r ch t 1 + b c ) h t = i o f h (c t ) is element-by-element multiplication memory cell weight matrices (W m f, W m i, W m o) diagonal can allow explicit analysis of individual cell elements 21 of 67

Highway Connections [41] x t h t 1 xt ht 1 xt ht 1 σ σ σ i h x t h t 1 f m ii c t f h time delay io + h t if σ xt ht 1 Gate the output of the node (example from LSTM) combine with output from previous layer (x t ) 22 of 67 i h = σ(w f hx t + W r hh t 1 + b h ) h t = i h (i o f h (c t )) + (1 i h ) x t x t

Sequence-to-Sequence Modelling 23 of 67

Sequence-to-Sequence Modelling Sequence-to-sequence modelling central to speech/language: speech synthesis: word sequence (discrete) waveform (continuous) speech recognition: waveform (continuous) word sequence (discrete) machine translation: word sequence (discrete) word sequence (discrete) The sequence lengths on either side can differ waveform sampled at 10ms/5ms frame-rate word sequences (are words...) Description focuses on ASR with RNNs (other models possible) 24 of 67

S2S: Generative Models [5, 6] Consider two sequences L T : input: x 1:T = {x 1, x 2,..., x T } output: y 1:L = {y 1, y 2,..., y L } Consider generative model (ASR language) p(y 1:L, x 1:T ) = p(y 1:L )p(x 1:T y 1:L ) = p(y 1:L ) p(x 1:T φ 1:T )P(φ 1:T y 1:L ) φ 1:T Φy 1:L p(y 1:L ): prior ( language ) model p(x 1:T φ 1:T ): (conditional) acoustic model P(φ 1:T y 1:L ): alignment model - handles variable length 25 of 67

Prior Model Approximations [21, 3, 33] Markovian: N-gram and Feed-Forward neural network L L p(y 1:L ) = p(y i y 1:i 1 ) p(y i y i 1,..., y i N+1 ) i=1 i=1 Non-Markovian: Recurrent neural network p(y 1:L ) L p(y i y i 1, h i 2 ) i=1 L p(y i h i 1 ) i=1 depends on complete unobserved word history y i y i+1 ~ ~ h i 1 h i 26 of 67

Hidden Markov Models [2, 13] Important sequence model is the hidden Markov model (HMM) an example of a dynamic Bayesian network (DBN) x t x t+1 φ t φt+1 discrete latent variables φ t describes discrete state-space conditional independence assumptions P(φ t φ 1:t 1, y 1:L ) = P(φ t φ t 1 ) p(x t x 1:t 1, φ 1:t ) = p(x t φ t ) The likelihood of the data is ( T ) p(x 1:T y 1:L ) = p(x t φ t )P(φ t φ t 1 ) φ t=1 1:T Φy 1:L 27 of 67

Acoustic Model Approximations Fully Markovian: HMM, simplest form of approximation T p(x 1:T φ 1:T ) p(x t φ t ) t=1 State Markovian: T T p(x 1:T φ 1:T ) p(x t φ t, x 1:t 1 ) p(x t φ t, h t 1 ) t=1 t=1 Feature Markovian: T T p(x 1:T φ 1:T ) p(x t φ 1:t ) p(x t h t ) t=1 t=1 28 of 67

Markovian Approximations and Inference h t 1 h t x t x t+1 x t x t+1 x t x t+1 h ~ ~ t h t+1 φ t φt+1 φ t φt+1 φ φ t t+1 Markovian (HMM) State Markovian Feature Markovian Tt=1 p(x t φ t ) Tt=1 p(x t φ t, h t 1 ) Tt=1 p(x t h t ) Inference costs significantly different: state Markovian: all past history observed - deterministic feature Markovian: past history unobserved - depends on path 29 of 67

Likelihoods [6] Deep learning can be used to estimate distributions - MDNN more often trained as a discriminative model need to convert to a likelihood Most common form (for RNN acoustic model): p(x t φ t, h t 1 ) P(φ t x t, h t 1 ) P(φ t ) P(φ t x t, h t 1 ): modelled by a standard RNN P(φ t ): state/phone prior probability Why use a generative sequence-to-sequence model? 30 of 67

S2S: Discriminative Models [5] Directly compute posterior of sequence p(y 1:L x 1:T ) = p(y 1 1:L φ 1:T )P(φ 1:T x 1:T ) φ 1:T Φy 1:L state Markovian RNNs used to model history/alignment P(φ 1:T x 1:T ) T P(φ t x 1:t ) t=1 T T P(φ t x t, h t 1 ) P(φ t h t ) t=1 t=1 Expression does not have alignment/language models 31 of 67

Connectionist Temporal Classification [15] CTC: discriminative model, no explicit alignment model introduces a blank output symbol (ɛ) ε /T/ ε /A/ ε /C/ Consider word: CAT Pronunciation: /C/ /A/ /T/ Observe 7 frames possible state transitions example path: /C/ ɛ /A/ /A/ ɛ /T/ ɛ Time 32 of 67

Extension to Non-Markovian x t x t+1 x t x t+1 x t x t+1 h t h t+1 h ~ ~ t h t+1 φ t φt+1 φ t φ t+1 φ φ t t+1 MEMM [25] CTC Non-Markovian Interesting to consider state dependencies (right) 33 of 67 T T P(φ 1:T x 1:T ) P(φ t x 1:t, φ 1:t 1 ) P(φ t h t ) t=1 t=1

Link of Pain - Non-Markovian φ t 1 x t ~ h t φ t Exact inference intractable complete history dependence see RNNLM ASR decoding ~ h t 1 Time delay 34 of 67

Discriminative Models and Priors [18] No language models in (this form of) discriminative model in CTC the word history captured in state (frame) history no explicit dependence on state (word) history Treat as a product of experts (log-linear model): for CTC ( ) 1 p(y 1:L x 1:T ) = Z(x 1:T ) exp α T log φ P(φ 1:T Φy 1:T x 1:T ) 1:L log (p(y 1:L )) α trainable parameter (related to LM scale) p(y 1:L ) standard prior (language) model Normalisation term not required in decoding α often empirically tuned 35 of 67

S2S: Encoder-Decoder Style Models Directly model relationship p(y 1:L x 1:T ) = L p(y i y 1:i 1, x 1:T ) i=1 L p(y i y i 1, h i 2, c) i=1 looks like an RNN LM with additional dependence on c c = φ(x 1:T ) c is a fixed length vector - like a sequence kernel 36 of 67

RNN Encoder-Decoder Model [14, 32] Decoder y i y i+1 ~ ~ h i 1 h i h t 1 h t h T 1 h T x t x t+1 x T Encoder Simplest form is to use hidden unit from acoustic RNN/LSTM c = φ(x 1:T ) = h T dependence on context is global via c - possibly limiting 37 of 67

Attention-Based Models [9, 7, 32] Decoder y i y i+1 ~ ~ h i 1 h i Attention c i c i+1 h t 1 h t h T 1 h T x t x t+1 x T Encoder 38 of 67

Attention-Based Models Introduce attention layer to system introduce dependence on locality i p(y 1:L x 1:T ) L p(y i y i 1, h i 2, c i ) i=1 L p(y i h i 1 ) i=1 c i = T exp(e iτ ) α iτ h τ ; α iτ = T k=1 exp(e ik), e iτ = f e ( hi 2 ), h τ τ=1 e iτ how well position i 1 in input matches position τ in output h τ is representation (RNN) for the input at position τ Attention can wander with large input size (T ) use a pyramidal network to reduce frame-rate for attention 39 of 67

S2S: Structured Discriminative Models [12, 31, 50] General form of structured discriminative model p(y 1:L x 1:T ) = 1 Z(x 1:T ) ( exp α T f (x 1:T, φ 1:T, y 1:L ) ) φ 1:T Φy 1:L f (x 1:T, φ 1:T, y 1:L ) extracts features from observations/states/words need to map variable length sequence to a fixed length (again) latent variables, state sequence φ 1:T, aid attention Integrate with deep learning to model segment features: RNN to map segments to a fixed length vector segment posterior outputs from multiple systems (joint decoding) 40 of 67

Speech Processing Applications 41 of 67

LM: Neural Network Language Models [3, 33] Neural networks extensively used for language modelling recurrent neural networks - complete word history P(ω 1:L ) = L P(ω i ω 1:i 1 ) i=1 L P(ω i ω i 1, h i 2 ) i=1 1-of-K ( one-hot ) coding for i th word, ω i, y i additional out-of-shortlist symbol may be added softmax activation function on output layer L P(ω i h i 1 ) i=1 Issues that need to be addressed 1. training: how to efficiently train on billions of words? 2. decoding: how to handle dependence on complete history? 42 of 67

LM: Cross-Entropy Training Criteria Standard training criterion for word sequence ω 1:L = ω 1,..., ω L F ce = 1 L L i=1 ( ) log P(ω i h i 1 ) GPU training makes this reasonable BUT Compute cost for softmax normalisation term Z( h i 1 ) P(ω i h i 1 ) = 1 ( ) Z( h i 1 ) exp w T f (ω i ) h i 1 required as unobserved sequence (contrast acoustic model) 43 of 67

LM: Alternative Training Criteria [8] Variance Regularisation: eliminate normalisation from decoding F vr = F ce + γ 2 1 L L i=1 ( ) 2 log(z( h i 1 )) log(z) log(z) average (log) history normalisation all normalisation terms tend to be the same Noise Contrastive Estimation: efficient decoding and training F nce = 1 L k log(p(y i = T ω i, h i 1 ) + log(p(y i = F ˆω ij, h i 1 ) L i=1 j=1 44 of 67 ˆω ij are competing samples for ω i - often sample from uni-gram LM

LM: ASR Decoding with RNNLMs w i 1 ~ h i 1 w i ~ ~ h i 1 h i ~ h i 2 Time delay w i w i+1 ASR decoding with RNNLM has a link of pain history vector depends on unobserved word sequence 45 of 67

LM: ASR Decoding with RNNLMs <s> there is a cat </s> hat <s> there is a cat </s> here is a mat cat here hat hat Lattice mat Prefix Tree mat Consider word-lattice on the left expands to prefix tree (right) if complete history taken into account significant increase in number of paths 46 of 67

LM: N-Gram History Approximation [29] <s> there is a cat hat mat </s> <s> ac= 10.0 lm= 3.0 rnn hist: <s> there ac= 9.0 lm= 2.3 <s> there is ac= 14.0 lm= 1.5 <s> there is rnn hist: <s> there is a a cat </s> here is a Prefix Tree cat hat mat ac= 13.0 lm= 3.4 rnn hist: <s> here ac= 11.0 lm= 1.5 <s> here is ac= 14.0 lm= 2.0 <s> here is hat mat N-Gram Approximation Use exact RNN LM value but merge paths based on N-gram history can also use history vector distance merging 47 of 67

ASR: Sequence Training [24] Cross-Entropy using fixed alignment standard criterion (RNN) F ce = T t=1 ( ) log P(ˆφ t x t, h t 1 ) criterion based on frame-by-frame classification Sequence training integrates sequence modelling into training F mbr = R P( ω x (r) 1:T )L( ω, ω(r) ) ω r=1 MBR described (various loss functions) - also CML used may be applied to generative and discriminative models 48 of 67

ASR: Sequence Training and CTC [36] Sequence-training is discriminative, so is CTC... let s ignote the blank symbol consider MMI as the training criterion F mmi = R r=1 ( ) log P(ω (r) x (r) 1:T ) lattice-free training uses some of the CTC-style approaches CTC has local (every frame) normalistion discriminatively-trained generative models use global normalisation CTC has no language-model use phone level language model P(ph i ph j ) (a 4-gram used) 49 of 67

ASR: Joint Decoding (Stacked Hybrid) [50] Pitch PLP Bottleneck Tandem HMM GMM Log Likelihoods Language Independent Language Dependent Speaker Dependent Bottleneck Layer FBank Pitch Fusion Score Stacked Hybrid Bottleneck PLP Pitch Log Posteriors DNN acoustic models and Tandem systems uses bottleneck features and stacking fusion: based on log-linear models - structured SVM 50 of 67

TTS: Deep Learning for Synthesis [53, 28] x t x t+1 x t x t+1 x t x t+1 ht h t+1 ht h t+1 ht h t+1 φ t φt+1 φ t φt+1 φ t φt+1 RNN/LSTM Output Recurrent Non-Markovian For statistical speech synthesis model p(x 1:T y 1:L ) = p(x 1:T φ 1:T )P(φ 1:T y 1:L ) p(x 1:T ˆφ 1:T ) φ 1:T Φy 1:L 51 of 67

TTS: Bottleneck Features [49] 5th MCC 1 0.5 0 Natural speech FE DNN MTE BN DNN 0.5 0 50 100 150 200 250 300 350 Frame Figure 1: Trajectories of the 5 th Mel-Cepstral Coefficient (MCC) of reference natural speech and those predicted by baseline FE-DNN and proposed MTE-BN-DNN systems. Can also include bottleneck features (similar to ASR) ing dynamic constraints at the output acoustic level. However, combining both gives the best overall performance: MTE-BN- DNN achieves FE-DNN: a 0.04 dbstandard reduction in MCDfeed-forward compared to FE- NN BN-DNN, with similar F0 RMSE performance. MTE-BN-DNN: feed-forward NN with MTE and BN features Subjectively better as well Table 1: Objective results. MCD = Mel Cepstral Distortion. Root mean squared error (RMSE) of F0 was computed in linear frequency. V/UV error means frame-level voiced/unvoiced 52 of 67 error. MCD F0 V/UV error preference score than FE-BN-DNN, although the difference is not significant. MTE DNN MTE BN DNN MTE BN DNN FE BN DNN FE DNN FE BN DNN MTE DNN MTE DNN

TTS: Minimum Trajectory Error [47] Smoothing is an important issue for generating trajectories in TTS predict experts for static and dynamics parameters (MLPG) use recursion on the output layer When using experts can minimise trajectory error (MTE) F mge = = R (ˆx (r) 1:T x(r) 1:T )T (ˆx (r) 1:T x(r) 1:T ) r=1 R (Rô (r) 1:T x(r) 1:T )T (Rô (r) 1:T x(r) 1:T ) r=1 ô (r) 1:T is the sequence of static/delta (means) R is the mapping matrix from static/deltas to trajectory 53 of 67

TTS: Conditional Language Modelling [44] x t x t+1 ht h t+1 φ φ t t+1 Make the system non-markovian 1-of-K coding for samples sample-level synthesis 8-bit = 256 output softmax output activation Recurrent units (shown): 240ms=3840 samples insufficient history memory Replace recurrent units by a sparse Markovian history extensive use of dilation to limit model parameters Search not an issue - simply sampling! 54 of 67

example, in Fig. 2 the receptive field is only 5 (= #layers + filter length - 1). In this paper we use dilated convolutions to increase the receptive field by orders of magnitude, without greatly increasing computational cost. TTS: CNN and RNN History Modelling [51, 44] A dilated convolution (also called à trous, or convolution with holes) is a convolution where the filter is applied over an area larger than its length by skipping input values with a certain step. It is equivalent to a convolution with a larger filter derived from the original filter by dilating it with zeros, but is significantly more efficient. A dilated convolution effectively allows the network to operate on a coarser scale than with a normal convolution. This is similar to pooling or strided convolutions, but here the output has the same size as the input. As a special case, dilated convolution with dilation 1 yields the standard convolution. Fig. 3 depicts dilated causal convolutions for dilations 1, 2, 4, and 8. Dilated convolutions have previously been used in various contexts, e.g. signal processing (Holschneider et al., 1989; Dutilleux, 1989), and image segmentation (Chen et al., 2015; Yu & Markovian history modelling limimted by parameter growth network parameters grows linearly with the length of history Non-Markovan (recurrent) approaches address parameter issue Koltun, 2016). but hard to train, and long-term representation (often) poor Output Dilation = 8 Hidden Layer Dilation = 4 Hidden Layer Dilation = 2 Hidden Layer Dilation = 1 Input CNN withfigure dilation 3: Visualization an alternative of a stack ofbalance dilated causal convolutional layers. exponential expansion of history, limited parameeter growth Stacked dilated convolutions enable networks to have very large receptive fields with just a few lay- 67 while preserving the input resolution throughout the network as well as computational efficiency. 55 ofers, In this paper, the dilation is doubled for every layer up to a limit and then repeated: e.g.

56 of 67 Neural Network Adaptation

Neural Network Adaptation a t ~ y t Auxiliary information Network adaptation Feature transformation y t x t Similar approaches used for TTS/ASR/LMs - broad classes auxiliary information: i-vectors, gender information, emotion ] network adaptation: weight adaptation, activation function adaptation feature transformation: linear transformation of the features Possible to combine multiple approaches together 57 of 67

Auxiliary Information [34, 39, 22, 23, 8] a t Auxiliary information Network adaptation Feature transformation y t x t Include information about speaker/topic/environment to network often vector representation used - ivectors, LDA topic spaces possible to adapt to speaker without hypothesis (e.g. ivectors) Applied to language modelling, speaker and noise adaptation 58 of 67

Network Parameter Adaptation Auxiliary information Network adaptation Feature transformation y t x t Seen a range of network adaptation approaches for ASR use well-trained network parameters as a prior updates all the parameters, weights, of the system Is it possible to reduce number of parameters to be adapted? 59 of 67

Structuring Networks for Adaptation [45] Hidden Layers x t Basis 1 Basis K Interpolation y t Structure network as a series of bases interpolation layer is speaker dependent h (ls) = K k=1 λ (s) k h(l) k few parameters per speaker - very rapid adaptation interpolation estimation convex optimisation problem 60 of 67

Activation Function Adaptation [42, 54, 43] Consider a layer of a network with 1000 1000 connections weights: 1,000,000 parameters to adjust activation functions: 2,000 functions (output and input) Take the example of a sigmoid activation function φ i (α (s) i, α (s), α (s) b ) = o ( 1 + exp α (s) i α (s) o wi Tx t + α (s) ) b α (s) i : scaling of the input α o (s) : scaling of the output α (s) b : offset on the activation train these (or subset) parameters to be speaker specific Exact form of parameter adaptation depends on activation function 61 of 67

Factored Weight Adaptation [30] Consider a speaker-specific linear transform of the weight matrix W (s) = A (s) W act on a low-dimensional compression layer Compact transform central to aspects like covariance modelling: W (s) = W + P i=1 λ (s) i A (i) W number of parameters is P - independent of weight matrix size A (i) low-rank, limits number of parameters can make a function of auxiliary information (e.g. ivectors) 62 of 67

Feature Transformation (TTS) [11, 48] Auxiliary information Network adaptation Feature transformation y~ t y t x t Train the network to predict experts for normalised features, ỹ t, transform normalised experts to target speaker experts ASR constrained to use global transform TTS can make use of regression class trees (CSMAPLR) 63 of 67

NST and Deep Learning Many other applications of deep learning under NST Speech Recognition audio data segmentation beamforming and channel combination network initialisation and regularisation acoustic feature dependency modelling paraphrastic language models Speech Synthesis post-processing influence/limitations of LSTM parameters on synthesis robust duration modelling unit selection 64 of 67

Is Deep Learning the Solution? Most research still uses a two-stage approach to training: 1. feature extraction: convert waveform to parametric form 2. modelling: given parameters train model Limitations in the feature extraction stage cannot be overcome... integrate feature extraction into process attempt to directly model/synthesise waveform (WaveNet) Both are interesting, active, areas of research links with integrated end-to-end systems: waveform-in words-out feasible as quantity of data increases BUT networks are difficult to optimise - tuning required hard to interpret networks to get insights sometimes difficult to learn from previous tasks... 65 of 67

Network Interpretation [46] Standard /ay/ Stimulated /ay/ Deep learning usually highly distributed - hard to interpret awkward to adapt/understand/regularise modify training - add stimulation regularisation (improves ASR!) 66 of 67

67 of 67 Thank-you!

[1] O. Abdel-Hamid, A.-R. Mohamed, H. Jiang, and G. Penn, Applying convolutional neural networks concepts to hybrid NN-HMM model for speech recognition, in 2012 IEEE international conference on Acoustics, speech and signal processing (ICASSP). IEEE, 2012, pp. 4277 4280. [2] L. Baum and J. Eagon, An Inequality with Applications to Statistical Estimation for Probabilistic Functions of Markov Processes and to a Model for Ecology, Bull Amer Math Soc, vol. 73, pp. 360 363, 1967. [3] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin, A neural probabilistic language model, journal of machine learning research, vol. 3, no. Feb, pp. 1137 1155, 2003. [4] C. Bishop, Mixture density networks, in Tech. Rep. NCRG/94/004, Neural Computing Research Group, Aston University, 1994. [5] C. M. Bishop, Pattern Recognition and Machine Learning. Springer Verlag, 2006. [6] H. Bourlard and N. Morgan, Connectionist speech recognition: A hybrid approach, 1994. [7] W. Chan, N. Jaitly, Q. V. Le, and O. Vinyals, Listen, attend and spell, CoRR, vol. abs/1508.01211, 2015. [Online]. Available: http://arxiv.org/abs/1508.01211 [8] X. Chen, T. Tan, X. Liu, P. Lanchantin, M. Wan, M. J. Gales, and P. C. Woodland, Recurrent neural network language model adaptation for multi-genre broadcast speech recognition, in Proceedings of InterSpeech, 2015. [9] J. Chorowski, D. Bahdanau, D. Serdyuk, K. Cho, and Y. Bengio, Attention-based models for speech recognition, CoRR, vol. abs/1506.07503, 2015. [Online]. Available: http://arxiv.org/abs/1506.07503 [10] J. Chung, K. Kastner, L. Dinh, K. Goel, A. C. Courville, and Y. Bengio, A recurrent latent variable model for sequential data, CoRR, vol. abs/1506.02216, 2015. [Online]. Available: http://arxiv.org/abs/1506.02216 [11] S. Deena, M. Hasan, M. Doulaty, O. Saz, and T. Hain, Combining feature and model-based adaptation of RNNLMs for multi-genre broadcast speech recognition, in Proceedings of the 17th Annual Conference of the International Speech Communication Association (Interspeech), 2016. [12] M. J. F. Gales, S. Watanabe, and E. Fosler-Lussier, Structured discriminative models for speech recognition: An overview, IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 70 81, 2012. 67 of 67

[13] M. Gales and S. Young, The application of hidden Markov models in speech recognition, Foundations and Trends in Signal Processing, vol. 1, no. 3, 2007. [14] A. Graves, Sequence transduction with recurrent neural networks, CoRR, vol. abs/1211.3711, 2012. [Online]. Available: http://arxiv.org/abs/1211.3711 [15] A. Graves, S. Fernández, F. Gomez, and J. Schmidhuber, Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks, in Proceedings of the 23rd international conference on Machine learning. ACM, 2006, pp. 369 376. [16] A. Graves, A.-R. Mohamed, and G. Hinton, Speech recognition with deep recurrent neural networks, in 2013 IEEE international conference on acoustics, speech and signal processing. IEEE, 2013, pp. 6645 6649. [17] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, CoRR, vol. abs/1512.03385, 2015. [Online]. Available: http://arxiv.org/abs/1512.03385 [18] G. E. Hinton, Products of experts, in Proceedings of the Ninth International Conference on Artificial Neural Networks (ICANN 99), 1999, pp. 1 6. [19] G. Hinton, L. Deng, D. Yu, G. E. Dahl, A.-R. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. N. Sainath, et al., Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups, IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 82 97, 2012. [20] S. Hochreiter and J. Schmidhuber, Long short-term memory, Neural Comput., vol. 9, no. 8, pp. 1735 1780, Nov. 1997. [21] F. Jelinek, Statistical methods for speech recognition, ser. Language, speech, and communication. Cambridge (Mass.), London: MIT Press, 1997. [22] P. Karanasou, M. Gales, and P. Woodland, I-vector estimation using informative priors for adaptation of deep neural networks, in Proc. of Interspeech, 2015. [23] P. Karanasou, Y. Wang, M. Gales, and P. Woodland, Adaptation of deep neural network acoustic models using factorised i-vectors, in Proceedings of Interspeech 14, 2014. [24] B. Kingsbury, Lattice-based optimization of sequence classification criteria for neural-network acoustic modeling, in 2009 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 2009, pp. 3761 3764. 67 of 67

[25] H.-K. Kuo and Y. Gao, Maximum entropy direct models for speech recognition, IEEE Transactions Audio Speech and Language Processing, 2006. [26] Y. LeCun and Y. Bengio, Convolutional networks for images, speech, and time series, The handbook of brain theory and neural networks, vol. 3361, no. 10, p. 1995, 1995. [27] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, Backpropagation applied to handwritten zip code recognition, Neural computation, vol. 1, no. 4, pp. 541 551, 1989. [28] Z.-H. Ling, S.-Y. Kang, H. Zen, A. Senior, M. Schuster, X.-J. Qian, H. M. Meng, and L. Deng, Deep learning for acoustic modeling in parametric speech generation: A systematic review of existing techniques and future trends, IEEE Signal Processing Magazine, vol. 32, no. 3, pp. 35 52, 2015. [29] X. Liu, Y. Wang, X. Chen, M. Gales, and P. Woodland, Efficient lattice rescoring using recurrent neural network language models, in IEEE ICASSP2014, IEEE ICASSP2014. IEEE ICASSP2014, 04/05/2014 2014. [30] Y. Liu, P. Karanasou, and T. Hain, An investigation into speaker informed dnn front-end for LVCSR, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), April 2015. [31] L. Lu, L. Kong, C. Dyer, N. A. Smith, and S. Renals, Segmental recurrent neural networks for end-to-end speech recognition, in Proc. INTERSPEECH, 2016. [32] L. Lu, X. Zhang, K. Cho, and S. Renals, A study of the recurrent neural network encoder-decoder for large vocabulary speech recognition, in Proc. INTERSPEECH, 2015. [33] T. Mikolov, M. Karafiát, L. Burget, J. Cernockỳ, and S. Khudanpur, Recurrent neural network based language model. in Interspeech, vol. 2, 2010, p. 3. [34] T. Mikolov and G. Zweig, Context dependent recurrent neural network language model. in SLT, 2012, pp. 234 239. [35] V. Nair and G. E. Hinton, Rectified linear units improve restricted Boltzmann machines, in Proceedings of the 27th International Conference on Machine Learning (ICML-10), 2010, pp. 807 814. [36] D. Povey, V. Peddinti, D. Galvez, P. Ghahrmani, V. Manohar, X. Na, Y. Wang, and S. Khudanpur, Purely sequence-trained neural networks for ASR based on lattice-free MMI, Interspeech, 2016. 67 of 67

[37] T. Robinson and F. Fallside, A recurrent error propagation network speech recognition system, Computer Speech & Language, vol. 5, no. 3, pp. 259 274, 1991. [38] D. E. Rumelhart, G. E. Hinton, and R. J. Williams, Parallel distributed processing: Explorations in the microstructure of cognition, vol. 1, D. E. Rumelhart, J. L. McClelland, and C. PDP Research Group, Eds. Cambridge, MA, USA: MIT Press, 1986, ch. Learning Internal Representations by Error Propagation, pp. 318 362. [39] G. Saon, H. Soltau, D. Nahamoo, and M. Picheny, Speaker adaptation of neural network acoustic models using i-vectors. in ASRU, 2013, pp. 55 59. [40] M. Schuster and K. K. Paliwal, Bidirectional recurrent neural networks, IEEE Transactions on Signal Processing, vol. 45, no. 11, pp. 2673 2681, 1997. [41] R. K. Srivastava, K. Greff, and J. Schmidhuber, Highway networks, CoRR, vol. abs/1505.00387, 2015. [Online]. Available: http://arxiv.org/abs/1505.00387 [42] P. Swietojanski and S. Renals, Learning hidden unit contributions for unsupervised speaker adaptation of neural network acoustic models, in Proc. IEEE Workshop on Spoken Language Technology, December 2014. [43], Differentiable pooling for unsupervised acoustic model adaptation, IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. PP, no. 99, pp. 1 1, 2016. [44] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, Wavenet: A generative model for raw auaio, CoRR, 2016. [Online]. Available: https://arxiv.org/pdf/1609.03499.pdf [45] C. Wu and M. Gales, Multi-basis adaptive neural network for rapid adaptation in speech recognition, in Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on, IEEE. IEEE, 2015. [46] C. Wu, P. Karanasou, M. Gales, and K. C. Sim, Stimulated deep neural network for speech recognition, in Proceedings interspeech, 2016. [47] Z. Wu and S. King, Minimum trajectory error training for deep neural networks, combined with stacked bottleneck features, in Proc. Interspeech, 2015, pp. 309 313. [48] Z. Wu, P. Swietojanski, C. Veaux, S. Renals, and S. King, A study of speaker adaptation for dnn-based speech synthesis, in Proceedings interspeech, 2015. 67 of 67

[49] Z. Wu, C. Valentini-Botinhao, O. Watts, and S. King, Deep neural networks employing multi-task learning and stacked bottleneck features for speech synthesis, in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2015. [50] J. Yang, C. Zhang, A. Ragni, M. Gales, and P. Woodland, System combiantion with log-linear models, in Proc. ICASSP 16, Shanghai, China, 2016. [51] F. Yu and V. Koltun, Multi-scale context aggregation by dilated convolutions, CoRR, vol. abs/1511.07122, 2015. [Online]. Available: http://arxiv.org/abs/1511.07122 [52] H. Zen and A. Senior, Deep mixture density networks for acoustic modeling in statistical parametric speech synthesis, in 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2014, pp. 3844 3848. [53] H. Zen, A. Senior, and M. Schuster, Statistical parametric speech synthesis using deep neural networks, in IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP). IEEE, 2013, pp. 7962 7966. [54] C. Zhang and P. Woodland, DNN speaker adaptation using parameterised sigmoid and ReLU hidden activation functions, in Proc. ICASSP 16, 2016. 67 of 67