Inference in Explicit Duration Hidden Markov Models

Size: px
Start display at page:

Download "Inference in Explicit Duration Hidden Markov Models"

Transcription

1 Inference in Explicit Duration Hidden Markov Models Frank Wood Joint work with Chris Wiggins, Mike Dewar Columbia University November, 2011 Wood (Columbia University) EDHMM Inference November, / 38

2 Hidden Markov Models Hidden Markov models (HMMs) [Rabiner, 1989] are an important tool for data exploration and engineering applications. Applications include Speech recognition [Jelinek, 1997, Juang and Rabiner, 1985] Natural language processing [Manning and Schütze, 1999] Hand-writing recognition [Nag et al., 1986] DNA and other biological sequence modeling applications [Krogh et al., 1994] Gesture recognition [Tanguay Jr, 1995, Wilson and Bobick, 1999] Financial data modeling [Rydén et al., 1998]... and many more. Wood (Columbia University) EDHMM Inference November, / 38

3 Graphical Model: Hidden Markov Model π m z 0 z 1 z 2 z 3 z T θ m y 1 y 2 y y K 3 T Wood (Columbia University) EDHMM Inference November, / 38

4 Notation: Hidden Markov Model z t z t 1 = m Discrete(π m ) y t z t = m F θ (θ m ) A =... π 1 π m π K... Wood (Columbia University) EDHMM Inference November, / 38

5 HMM: Typical Usage Scenario (Character Recognition) Training data: multiple observed y t = {v t, h t } sequences of stylus positions for each kind of character Task: train Σ different models, one for each character Latent states: sequences of strokes Usage: classify new stylus position sequences using trained models M σ = {A σ, Θ σ } P(M σ y 1,..., y T ) P(y 1,..., y T M σ )P(M σ ) Wood (Columbia University) EDHMM Inference November, / 38

6 Shortcomings of Original HMM Specification Latent state dwell times are not usually geometrically distributed P(z t = m,..., z t+l = m A) L = P(z t+l+1 = m z t+l = m, A) l=1 = Geometric(L; π m (m)) Prior knowledge often suggests structural constraints on allowable transitions The state cardinality of the latent Markov chain K is usually unknown Wood (Columbia University) EDHMM Inference November, / 38

7 Explicit Duration HMM / H. Semi-Markov Model [Mitchell et al., 1995, Murphy, 2002, Yu and Kobayashi, 2003, Yu, 2010] Latent state sequence z = ({s 1, r 1 },..., {s T, r T }) Latent state id sequence s = (s 1,..., s T ) Latent remaining duration sequence r = (r 1,..., r T ) State-specific duration distribution F r (λ m ) Other distributions the same An EDHMM transitions between states in a different way than does a typical HMM. Unless r t = 0 the current remaining duration is decremented and the state does not change. If r t = 0 then the EDHMM transitions to a state m s t according to the distribution defined by π st Problem: inference requires enumerating possible durations. Wood (Columbia University) EDHMM Inference November, / 38

8 EDHMM notation Latent state z t = {s t, r t } is tuple consisting of state identity and time left in state. s t s t 1, r t 1 r t s t, r t 1 y t s t F θ (θ st ) { I(s t = s t 1 ), r t 1 > 0 Discrete(π st 1 ), { r t 1 = 0 I(r t = r t 1 1), r t 1 > 0 F r (λ st ), r t 1 = 0 Wood (Columbia University) EDHMM Inference November, / 38

9 EDHMM: Graphical Model λ m r 0 r 1 r 2 r 3 r T π m s 0 s 1 s 2 s 3 s T θ m K y 1 y 2 y 3 y T Wood (Columbia University) EDHMM Inference November, / 38

10 Structured HMMs: i.e. left-to-right HMM [Rabiner, 1989] Example: Chicken pox Observations vital signs Latent states pre-infection, infected, post-infection a State transition structure can t go from infected to pre-infection a disregarding shingles Structured transitions also imply zeros in the transition matrix A, i.e. p(s t = m s t 1 = l) = 0 m < l Wood (Columbia University) EDHMM Inference November, / 38

11 Bayesian HMM We will put a prior on parameters so that we can effect a solution that conforms to our ideas about what the solution should look like Structured prior examples A i,j = 0 (hard constraints) A i,j P j A i,j (rich get richer) Regularization means that we can specify a model with more parameters than could possibly be needed infinite complexity (i.e. K ) avoids many model selection problems extra states can be thought of as auxiliary or nuisance variables inference requires sampling in a model with countably infinite support Posterior over latent variables encodes uncertainty about interpretation of data. Wood (Columbia University) EDHMM Inference November, / 38

12 Bayesian HMM H z π m z 0 z 1 z 2 z 3 z T H y θ m y 1 y 2 y y K 3 T π m H z θ m H y z t z t 1 = m Discrete(π m ) y t z t = m F θ (θ m ) Wood (Columbia University) EDHMM Inference November, / 38

13 Infinite HMMs (IHMM) [Beal et al., 2002, Teh et al., 2006] c 0 γ c 1 π m z 0 z 1 z 2 z 3 z T H y θ m y 1 y 2 y y 3 T K, Sticky Infinite HMM [Fox et al., 2011] Wood (Columbia University) EDHMM Inference November, / 38

14 Inference for the Explicit Duration HMM (EDHMM) Simple Idea Result Infinite HMMs and EDHMMs share fundamental characteristic: countable support Inference techniques for Bayesian nonparametric (infinite) HMMs can be used for EDHMM inference Approximation-free, efficient inference algorithm for EDHMM inference Wood (Columbia University) EDHMM Inference November, / 38

15 Bayesian EDHMM: Graphical Model H r λ m r 0 r 1 r 2 r 3 r T H z π m s 0 s 1 s 2 s 3 s T H y θ m K y 1 y 2 y 3 y T A choice of prior λ m H r Gamma(H r ) π m H z Dir(1/K, 1/K,..., 1/K, 0, 1/K,..., 1/K, 1/K) Wood (Columbia University) EDHMM Inference November, / 38

16 EDHMM Inference: Beam Sampling [Dewar, Wiggins and W, 2011] We employ the forward-filtering, backward slice-sampling approach for the IHMM of [Van Gael et al., 2008], in which the state and duration variables s and r are sampled conditioned on auxiliary slice variables u. Net result: efficient, always finite forward-backward procedure for sampling latent states and the amount of time spent in them. Wood (Columbia University) EDHMM Inference November, / 38

17 Auxiliary Variables for Sampling Objective: get samples of x. λ x Wood (Columbia University) EDHMM Inference November, / 38

18 Auxiliary Variables for Sampling Objective: get samples of x. Sometimes it is easier to introduce an auxiliary variable u and to Gibbs sample the joint P(x, u) (i.e. sample from P(x u; λ) then P(u x, λ), etc.) then discard the u values than it is to directly sample from p(x λ). λ x λ x u Wood (Columbia University) EDHMM Inference November, / 38

19 Auxiliary Variables for Sampling Objective: get samples of x. Sometimes it is easier to introduce an auxiliary variable u and to Gibbs sample the joint P(x, u) (i.e. sample from P(x u; λ) then P(u x, λ), etc.) then discard the u values than it is to directly sample from p(x λ). Useful when: p(x λ) does not have a known parametric form but adding u results in a parametric form and when x has countable support and sampling it requires enumerating all values. λ x λ x u Wood (Columbia University) EDHMM Inference November, / 38

20 Slice Sampling: A very useful auxiliary variable sampling trick Pedagogical Example: x λ Poisson(λ) (countable support) enumeration strategy for sampling x (impossible) 1 auxiliary variable u with P(u x, λ) = I(0 u P(x λ)) P(x λ) Note: Marginal distribution of x is P(x λ) = u P(x, u λ) = u = u = u P(x λ)p(u x, λ) I(0 u P(x λ)) P(x λ) P(x λ) I(0 u P(x λ)) = P(x λ) 1 Necessary in EDHMM Wood (Columbia University) EDHMM Inference November, / 38

21 Slice Sampling: A very useful auxiliary variable sampling trick This suggests a Gibbs sampling scheme: alternately sampling from P(x u, λ) I(u P(x λ)) finite support, uniform above slice, enumeration possible P(u x, λ) = I(0 u P(x λ)) P(x λ) uniform between 0 and y = P(x λ) then discarding the u values to arrive at x samples marginally distributed according to P(x λ). PoissonPDF(r t ; λ m = 5) u t = r t Wood (Columbia University) EDHMM Inference November, / 38

22 EDHMM Graphical Model with Auxiliary Variables r 0 r 1 r 2 r 3 r T u 1 u 2 u T s 0 s 1 s 2 s 3 s T y 1 y 2 y 3 y T Wood (Columbia University) EDHMM Inference November, / 38

23 EDHMM Inference: Beam Sampling With auxiliary variables defined as and p(u t z t, z t 1 ) = I(0 < u t < p(z t z t 1 )) p(z t z t 1 ) p(z t z t 1 ) = t, r t ) (s t 1, r t 1 )) p((s { = r t 1 > 0, I(s t = s t 1 )I(r t = r t 1 1) r t 1 = 0, π st 1 s t F r (r t ; λ st ). one can run standard forward-backward conditioned on u s. Forward-backward slice sampling only has to consider a finite number of successor states at each timestep. Wood (Columbia University) EDHMM Inference November, / 38

24 Forward recursion ˆα t (z t ) = p(z t, Y t 1, U t 1) = z t 1 p(z t, z t 1, Y t 1, U t 1) z t 1 p(u t z t, z t 1 )p(z t, z t 1, Y t 1, U t 1 1 ) = z t 1 I(0 < u t < p(z t z t 1 ))p(y t z t )ˆα t 1 (z t 1 ). Wood (Columbia University) EDHMM Inference November, / 38

25 Backward Sampling Recursively sample a state sequence from the distribution p(z t 1 z t, Y, U) which can expressed in terms of the forward variable: p(z t 1 z t, Y, U) p(z t, z t 1, Y, U) p(u t z t, z t 1 )p(z t z t 1 )ˆα t 1 (z t 1 ) I(0 < u t < p(z t z t 1 ))ˆα t 1 (z t 1 ). Wood (Columbia University) EDHMM Inference November, / 38

26 Results To illustrate EDHMM learning on synthetic data, five hundred datapoints were generated using a 4 state EDHMM with Poisson duration distributions λ = (10, 20, 3, 7) and Gaussian emission distributions with means all unit variance. µ = ( 6, 2, 2, 6) Wood (Columbia University) EDHMM Inference November, / 38

27 EDHMM: Synthetic Data Results 10 emission/mean time Wood (Columbia University) EDHMM Inference November, / 38

28 EDHMM: Synthetic Data, State Mean Posterior 600 count emission mean Wood (Columbia University) EDHMM Inference November, / 38

29 EDHMM: Synthetic Data, State Duration Parameter Posterior 300 count duration rate Wood (Columbia University) EDHMM Inference November, / 38

30 EDHMM: Nanoscale Transistor Spontaneous Voltage Fluctuation 2 1 emission/mean time Wood (Columbia University) EDHMM Inference November, / 38

31 EDHMM: States Not Distinguishable By Output Task: understand system with states that have identical observation distributions and differ only in duration distribution. Observation distributions have means µ 1 = 0, µ 2 = 0, and µ 3 = 3 and the duration distributions have rates λ 1 = 5, λ 2 = 15, λ 3 = 20. (Next slide) a) observations; b) true states overlaid with 20 randomly selected state traces produced by the sampler. Samples from the posterior observation distribution mean are shown in c), and samples from the posterior duration distribution rates are shown in d). Wood (Columbia University) EDHMM Inference November, / 38

32 EDHMM: States Not Distinguishable By Output yt time (a) xt time (b) p(µj Y) y (c) p(λj Y) time (d) Wood (Columbia University) EDHMM Inference November, / 38

33 EDHMM vs. IHMM: Modeling the Morse Code Cepstrum Wood (Columbia University) EDHMM Inference November, / 38

34 Wrap-Up Novel inference procedure for EDHMMs that doesn t require truncation and is more efficient than considering all possible durations. number of transitions visited iteration Figure: Mean number of transitions considered per time point by the beam sampler for 1000 post-burn-in iterations. Compare to (KT ) 2 = O(10 6 ) transitions that would need to be considered by standard forward backward without truncation, the only surely safe, truncation-free alternative. Wood (Columbia University) EDHMM Inference November, / 38

35 Future Work Novel Gamma process construction for dependent, structured, infinite dimensional HMM transition distributions. Generalize to spatial prior on HMM states ( location ) Simultaneous location and mapping Process diagram modeling for systems biology Applications; seeking users Wood (Columbia University) EDHMM Inference November, / 38

36 Questions? Thank you! Wood (Columbia University) EDHMM Inference November, / 38

37 Bibliography I M J Beal, Z Ghahramani, and C E Rasmussen. The Infinite Hidden Markov Model. In Advances in Neural Information Processing Systems, pages , March M Dewar, C Wiggins, and F Wood. Inference in hidden Markov models with explicit state duration distributions. In Submission, E B Fox, E B Sudderth, M I Jordan, and A S Willsky. A Sticky HDP-HMM with Application to Speaker Diarization. Annals of Applied Statistics, 5(2A): , F. Jelinek. Statistical Methods for Speech Recognition. MIT Press, B H Juang and L R Rabiner. Mixture Autoregressive Hidden Markov Models for Speech Signals. IEEE Transactions on Acoustics, Speech, and Signal Processing, 33(6): , Wood (Columbia University) EDHMM Inference November, / 38

38 Bibliography II A. Krogh, M. Brown, I. S. Mian, K. Sjolander, and D. Haussler. Hidden Markov models in computational biology: Applications to protein modelling. Journal of Molecular Biology, 235: , C. Manning and H. Schütze. Foundations of statistical natural language processing. MIT Press, Cambridge, MA, C Mitchell, M Harper, and L Jamieson. On the complexity of explicit duration HMM s. IEEE Transactions on Speech and Audio Processing, 3 (3): , K P Murphy. Hidden semi-markov models (HSMMs). Technical report, MIT, R. Nag, K. Wong, and F. Fallside. Script regonition using hidden Markov models. In ICASSP86, pages , L R Rabiner. A tutorial on hidden Markov models and selected applications in speech recognition. In Proceedings of the IEEE, pages , Wood (Columbia University) EDHMM Inference November, / 38

39 Bibliography III T. Rydén, T. Teräsvirta, and S. rasbrink. Stylized facts of daily return series and the hidden markov model. Journal of Applied Econometrics, 13(3): , D.O. Tanguay Jr. Hidden Markov models for gesture recognition. PhD thesis, Massachusetts Institute of Technology, Y W Teh, M I Jordan, M J Beal, and D M Blei. Hierarchical Dirichlet Processes. Journal of the American Statistical Association, 101(476): , J Van Gael, Y Saatci, Y W Teh, and Z Ghahramani. Beam sampling for the infinite hidden Markov model. In Proceedings of the 25th International Conference on Machine Learning, pages ACM, A.D. Wilson and A.F. Bobick. Parametric hidden Markov models for gesture recognition. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 21(9): , Wood (Columbia University) EDHMM Inference November, / 38

40 Bibliography IV S Yu and H Kobayashi. An Efficient Forward Backward Algorithm for an Explicit-Duration Hidden Markov Model. Signal Processing letters, 10 (1):11 14, S Z Yu. Hidden semi-markov models. Artificial Intelligence, 174(2): , Wood (Columbia University) EDHMM Inference November, / 38

Hidden Markov models: from the beginning to the state of the art

Hidden Markov models: from the beginning to the state of the art Hidden Markov models: from the beginning to the state of the art Frank Wood Columbia University November, 2011 Wood (Columbia University) HMMs November, 2011 1 / 44 Outline Overview of hidden Markov models

More information

Bayesian Hidden Markov Models and Extensions

Bayesian Hidden Markov Models and Extensions Bayesian Hidden Markov Models and Extensions Zoubin Ghahramani Department of Engineering University of Cambridge joint work with Matt Beal, Jurgen van Gael, Yunus Saatci, Tom Stepleton, Yee Whye Teh Modeling

More information

Bayesian Nonparametric Learning of Complex Dynamical Phenomena

Bayesian Nonparametric Learning of Complex Dynamical Phenomena Duke University Department of Statistical Science Bayesian Nonparametric Learning of Complex Dynamical Phenomena Emily Fox Joint work with Erik Sudderth (Brown University), Michael Jordan (UC Berkeley),

More information

Applied Bayesian Nonparametrics 3. Infinite Hidden Markov Models

Applied Bayesian Nonparametrics 3. Infinite Hidden Markov Models Applied Bayesian Nonparametrics 3. Infinite Hidden Markov Models Tutorial at CVPR 2012 Erik Sudderth Brown University Work by E. Fox, E. Sudderth, M. Jordan, & A. Willsky AOAS 2011: A Sticky HDP-HMM with

More information

A Brief Overview of Nonparametric Bayesian Models

A Brief Overview of Nonparametric Bayesian Models A Brief Overview of Nonparametric Bayesian Models Eurandom Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin Also at Machine

More information

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check

More information

Bayesian Nonparametrics for Speech and Signal Processing

Bayesian Nonparametrics for Speech and Signal Processing Bayesian Nonparametrics for Speech and Signal Processing Michael I. Jordan University of California, Berkeley June 28, 2011 Acknowledgments: Emily Fox, Erik Sudderth, Yee Whye Teh, and Romain Thibaux Computer

More information

Master 2 Informatique Probabilistic Learning and Data Analysis

Master 2 Informatique Probabilistic Learning and Data Analysis Master 2 Informatique Probabilistic Learning and Data Analysis Faicel Chamroukhi Maître de Conférences USTV, LSIS UMR CNRS 7296 email: chamroukhi@univ-tln.fr web: chamroukhi.univ-tln.fr 2013/2014 Faicel

More information

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning

More information

Sequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them

Sequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them HMM, MEMM and CRF 40-957 Special opics in Artificial Intelligence: Probabilistic Graphical Models Sharif University of echnology Soleymani Spring 2014 Sequence labeling aking collective a set of interrelated

More information

Bayesian Nonparametric Hidden Semi-Markov Models

Bayesian Nonparametric Hidden Semi-Markov Models Journal of Machine Learning Research 14 (2013) 673-701 Submitted 12/11; Revised 9/12; Published 2/13 Bayesian Nonparametric Hidden Semi-Markov Models Matthew J. Johnson Alan S. Willsky Laboratory for Information

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers

More information

Dynamic Approaches: The Hidden Markov Model

Dynamic Approaches: The Hidden Markov Model Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message

More information

Graphical Models for Query-driven Analysis of Multimodal Data

Graphical Models for Query-driven Analysis of Multimodal Data Graphical Models for Query-driven Analysis of Multimodal Data John Fisher Sensing, Learning, & Inference Group Computer Science & Artificial Intelligence Laboratory Massachusetts Institute of Technology

More information

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

INFINITE MIXTURES OF MULTIVARIATE GAUSSIAN PROCESSES

INFINITE MIXTURES OF MULTIVARIATE GAUSSIAN PROCESSES INFINITE MIXTURES OF MULTIVARIATE GAUSSIAN PROCESSES SHILIANG SUN Department of Computer Science and Technology, East China Normal University 500 Dongchuan Road, Shanghai 20024, China E-MAIL: slsun@cs.ecnu.edu.cn,

More information

Stochastic Variational Inference for the HDP-HMM

Stochastic Variational Inference for the HDP-HMM Stochastic Variational Inference for the HDP-HMM Aonan Zhang San Gultekin John Paisley Department of Electrical Engineering & Data Science Institute Columbia University, New York, NY Abstract We derive

More information

HMM part 1. Dr Philip Jackson

HMM part 1. Dr Philip Jackson Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. HMM part 1 Dr Philip Jackson Probability fundamentals Markov models State topology diagrams Hidden Markov models -

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models $ Technical Report, University of Toronto, CSRG-501, October 2004 Density Propagation for Continuous Temporal Chains Generative and Discriminative Models Cristian Sminchisescu and Allan Jepson Department

More information

The Infinite Factorial Hidden Markov Model

The Infinite Factorial Hidden Markov Model The Infinite Factorial Hidden Markov Model Jurgen Van Gael Department of Engineering University of Cambridge, UK jv279@cam.ac.uk Yee Whye Teh Gatsby Unit University College London, UK ywteh@gatsby.ucl.ac.uk

More information

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes (bilmes@cs.berkeley.edu) International Computer Science Institute

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Hidden Markov Models. Terminology, Representation and Basic Problems

Hidden Markov Models. Terminology, Representation and Basic Problems Hidden Markov Models Terminology, Representation and Basic Problems Data analysis? Machine learning? In bioinformatics, we analyze a lot of (sequential) data (biological sequences) to learn unknown parameters

More information

Data Analyzing and Daily Activity Learning with Hidden Markov Model

Data Analyzing and Daily Activity Learning with Hidden Markov Model Data Analyzing and Daily Activity Learning with Hidden Markov Model GuoQing Yin and Dietmar Bruckner Institute of Computer Technology Vienna University of Technology, Austria, Europe {yin, bruckner}@ict.tuwien.ac.at

More information

Learning from Sequential and Time-Series Data

Learning from Sequential and Time-Series Data Learning from Sequential and Time-Series Data Sridhar Mahadevan mahadeva@cs.umass.edu University of Massachusetts Sridhar Mahadevan: CMPSCI 689 p. 1/? Sequential and Time-Series Data Many real-world applications

More information

ECE521 Lecture 19 HMM cont. Inference in HMM

ECE521 Lecture 19 HMM cont. Inference in HMM ECE521 Lecture 19 HMM cont. Inference in HMM Outline Hidden Markov models Model definitions and notations Inference in HMMs Learning in HMMs 2 Formally, a hidden Markov model defines a generative process

More information

A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems

A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems Daniel Meyer-Delius 1, Christian Plagemann 1, Georg von Wichert 2, Wendelin Feiten 2, Gisbert Lawitzky 2, and

More information

Expectation Propagation in Dynamical Systems

Expectation Propagation in Dynamical Systems Expectation Propagation in Dynamical Systems Marc Peter Deisenroth Joint Work with Shakir Mohamed (UBC) August 10, 2012 Marc Deisenroth (TU Darmstadt) EP in Dynamical Systems 1 Motivation Figure : Complex

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Hidden Markov Models Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Additional References: David

More information

Conditional Random Field

Conditional Random Field Introduction Linear-Chain General Specific Implementations Conclusions Corso di Elaborazione del Linguaggio Naturale Pisa, May, 2011 Introduction Linear-Chain General Specific Implementations Conclusions

More information

Gaussian Mixture Model

Gaussian Mixture Model Case Study : Document Retrieval MAP EM, Latent Dirichlet Allocation, Gibbs Sampling Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 5 th,

More information

order is number of previous outputs

order is number of previous outputs Markov Models Lecture : Markov and Hidden Markov Models PSfrag Use past replacements as state. Next output depends on previous output(s): y t = f[y t, y t,...] order is number of previous outputs y t y

More information

An Introduction to Bayesian Machine Learning

An Introduction to Bayesian Machine Learning 1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems

More information

Note Set 5: Hidden Markov Models

Note Set 5: Hidden Markov Models Note Set 5: Hidden Markov Models Probabilistic Learning: Theory and Algorithms, CS 274A, Winter 2016 1 Hidden Markov Models (HMMs) 1.1 Introduction Consider observed data vectors x t that are d-dimensional

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010 Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Lecture 3a: Dirichlet processes

Lecture 3a: Dirichlet processes Lecture 3a: Dirichlet processes Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk Advanced Topics

More information

Bayesian Nonparametric Models

Bayesian Nonparametric Models Bayesian Nonparametric Models David M. Blei Columbia University December 15, 2015 Introduction We have been looking at models that posit latent structure in high dimensional data. We use the posterior

More information

Household Energy Disaggregation based on Difference Hidden Markov Model

Household Energy Disaggregation based on Difference Hidden Markov Model 1 Household Energy Disaggregation based on Difference Hidden Markov Model Ali Hemmatifar (ahemmati@stanford.edu) Manohar Mogadali (manoharm@stanford.edu) Abstract We aim to address the problem of energy

More information

Infinite Hierarchical Hidden Markov Models

Infinite Hierarchical Hidden Markov Models Katherine A. Heller Engineering Department University of Cambridge Cambridge, UK heller@gatsby.ucl.ac.uk Yee Whye Teh and Dilan Görür Gatsby Unit University College London London, UK {ywteh,dilan}@gatsby.ucl.ac.uk

More information

Nonparametric Bayesian Models for Sparse Matrices and Covariances

Nonparametric Bayesian Models for Sparse Matrices and Covariances Nonparametric Bayesian Models for Sparse Matrices and Covariances Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Bayes

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Lecture 3: Machine learning, classification, and generative models

Lecture 3: Machine learning, classification, and generative models EE E6820: Speech & Audio Processing & Recognition Lecture 3: Machine learning, classification, and generative models 1 Classification 2 Generative models 3 Gaussian models Michael Mandel

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Advanced Data Science

Advanced Data Science Advanced Data Science Dr. Kira Radinsky Slides Adapted from Tom M. Mitchell Agenda Topics Covered: Time series data Markov Models Hidden Markov Models Dynamic Bayes Nets Additional Reading: Bishop: Chapter

More information

Chapter 05: Hidden Markov Models

Chapter 05: Hidden Markov Models LEARNING AND INFERENCE IN GRAPHICAL MODELS Chapter 05: Hidden Markov Models Dr. Martin Lauer University of Freiburg Machine Learning Lab Karlsruhe Institute of Technology Institute of Measurement and Control

More information

IEEE TRANSACTION ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1. Variational Infinite Hidden Conditional Random Fields

IEEE TRANSACTION ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1. Variational Infinite Hidden Conditional Random Fields IEEE TRANSACTION ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 Variational Infinite Hidden Conditional Random Fields Konstantinos Bousmalis, Student Member, IEEE, Stefanos Zafeiriou, Member, IEEE, Louis-Philippe

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Lecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010

Lecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010 Hidden Lecture 4: Hidden : An Introduction to Dynamic Decision Making November 11, 2010 Special Meeting 1/26 Markov Model Hidden When a dynamical system is probabilistic it may be determined by the transition

More information

Course 495: Advanced Statistical Machine Learning/Pattern Recognition

Course 495: Advanced Statistical Machine Learning/Pattern Recognition Course 495: Advanced Statistical Machine Learning/Pattern Recognition Lecturer: Stefanos Zafeiriou Goal (Lectures): To present discrete and continuous valued probabilistic linear dynamical systems (HMMs

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Infinite Feature Models: The Indian Buffet Process Eric Xing Lecture 21, April 2, 214 Acknowledgement: slides first drafted by Sinead Williamson

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Nonparametrics: Dirichlet Process

Bayesian Nonparametrics: Dirichlet Process Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian

More information

Hidden Markov models

Hidden Markov models Hidden Markov models Charles Elkan November 26, 2012 Important: These lecture notes are based on notes written by Lawrence Saul. Also, these typeset notes lack illustrations. See the classroom lectures

More information

Non-parametric Clustering with Dirichlet Processes

Non-parametric Clustering with Dirichlet Processes Non-parametric Clustering with Dirichlet Processes Timothy Burns SUNY at Buffalo Mar. 31 2009 T. Burns (SUNY at Buffalo) Non-parametric Clustering with Dirichlet Processes Mar. 31 2009 1 / 24 Introduction

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

An Evolutionary Programming Based Algorithm for HMM training

An Evolutionary Programming Based Algorithm for HMM training An Evolutionary Programming Based Algorithm for HMM training Ewa Figielska,Wlodzimierz Kasprzak Institute of Control and Computation Engineering, Warsaw University of Technology ul. Nowowiejska 15/19,

More information

MAD-Bayes: MAP-based Asymptotic Derivations from Bayes

MAD-Bayes: MAP-based Asymptotic Derivations from Bayes MAD-Bayes: MAP-based Asymptotic Derivations from Bayes Tamara Broderick Brian Kulis Michael I. Jordan Cat Clusters Mouse clusters Dog 1 Cat Clusters Dog Mouse Lizard Sheep Picture 1 Picture 2 Picture 3

More information

Hidden Markov Models. Terminology and Basic Algorithms

Hidden Markov Models. Terminology and Basic Algorithms Hidden Markov Models Terminology and Basic Algorithms What is machine learning? From http://en.wikipedia.org/wiki/machine_learning Machine learning, a branch of artificial intelligence, is about the construction

More information

Hidden Markov Models Hamid R. Rabiee

Hidden Markov Models Hamid R. Rabiee Hidden Markov Models Hamid R. Rabiee 1 Hidden Markov Models (HMMs) In the previous slides, we have seen that in many cases the underlying behavior of nature could be modeled as a Markov process. However

More information

Probabilistic Time Series Classification

Probabilistic Time Series Classification Probabilistic Time Series Classification Y. Cem Sübakan Boğaziçi University 25.06.2013 Y. Cem Sübakan (Boğaziçi University) M.Sc. Thesis Defense 25.06.2013 1 / 54 Problem Statement The goal is to assign

More information

Infering the Number of State Clusters in Hidden Markov Model and its Extension

Infering the Number of State Clusters in Hidden Markov Model and its Extension Infering the Number of State Clusters in Hidden Markov Model and its Extension Xugang Ye Department of Applied Mathematics and Statistics, Johns Hopkins University Elements of a Hidden Markov Model (HMM)

More information

MACHINE LEARNING 2 UGM,HMMS Lecture 7

MACHINE LEARNING 2 UGM,HMMS Lecture 7 LOREM I P S U M Royal Institute of Technology MACHINE LEARNING 2 UGM,HMMS Lecture 7 THIS LECTURE DGM semantics UGM De-noising HMMs Applications (interesting probabilities) DP for generation probability

More information

Hidden Markov Models and Gaussian Mixture Models

Hidden Markov Models and Gaussian Mixture Models Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Gentle Introduction to Infinite Gaussian Mixture Modeling

Gentle Introduction to Infinite Gaussian Mixture Modeling Gentle Introduction to Infinite Gaussian Mixture Modeling with an application in neuroscience By Frank Wood Rasmussen, NIPS 1999 Neuroscience Application: Spike Sorting Important in neuroscience and for

More information

Introduction to Artificial Intelligence (AI)

Introduction to Artificial Intelligence (AI) Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 10 Oct, 13, 2011 CPSC 502, Lecture 10 Slide 1 Today Oct 13 Inference in HMMs More on Robot Localization CPSC 502, Lecture

More information

Basic math for biology

Basic math for biology Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood

More information

Bayesian Inference Course, WTCN, UCL, March 2013

Bayesian Inference Course, WTCN, UCL, March 2013 Bayesian Course, WTCN, UCL, March 2013 Shannon (1948) asked how much information is received when we observe a specific value of the variable x? If an unlikely event occurs then one would expect the information

More information

Bayesian Mixtures of Bernoulli Distributions

Bayesian Mixtures of Bernoulli Distributions Bayesian Mixtures of Bernoulli Distributions Laurens van der Maaten Department of Computer Science and Engineering University of California, San Diego Introduction The mixture of Bernoulli distributions

More information

CSEP 573: Artificial Intelligence

CSEP 573: Artificial Intelligence CSEP 573: Artificial Intelligence Hidden Markov Models Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell, Andrew Moore, Ali Farhadi, or Dan Weld 1 Outline Probabilistic

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling

Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling BMI/CS 776 www.biostat.wisc.edu/bmi776/ Spring 009 Mark Craven craven@biostat.wisc.edu Sequence Motifs what is a sequence

More information

Fast MCMC sampling for Markov jump processes and extensions

Fast MCMC sampling for Markov jump processes and extensions Fast MCMC sampling for Markov jump processes and extensions Vinayak Rao and Yee Whye Teh Rao: Department of Statistical Science, Duke University Teh: Department of Statistics, Oxford University Work done

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Hidden Markov Models Barnabás Póczos & Aarti Singh Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed

More information

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing Hidden Markov Models By Parisa Abedi Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed data Sequential (non i.i.d.) data Time-series data E.g. Speech

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Hidden Markov Models Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

CSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado

CSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado CSCI 5822 Probabilistic Model of Human and Machine Learning Mike Mozer University of Colorado Topics Language modeling Hierarchical processes Pitman-Yor processes Based on work of Teh (2006), A hierarchical

More information

Two Useful Bounds for Variational Inference

Two Useful Bounds for Variational Inference Two Useful Bounds for Variational Inference John Paisley Department of Computer Science Princeton University, Princeton, NJ jpaisley@princeton.edu Abstract We review and derive two lower bounds on the

More information

Fast Inference and Learning with Sparse Belief Propagation

Fast Inference and Learning with Sparse Belief Propagation Fast Inference and Learning with Sparse Belief Propagation Chris Pal, Charles Sutton and Andrew McCallum University of Massachusetts Department of Computer Science Amherst, MA 01003 {pal,casutton,mccallum}@cs.umass.edu

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Hidden Markov Models. Terminology and Basic Algorithms

Hidden Markov Models. Terminology and Basic Algorithms Hidden Markov Models Terminology and Basic Algorithms The next two weeks Hidden Markov models (HMMs): Wed 9/11: Terminology and basic algorithms Mon 14/11: Implementing the basic algorithms Wed 16/11:

More information

Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection

Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection Alireza Ahrabian 1 arxiv:1901.0545v1 [eess.sp] 16 Jan 019 Abstract This work addresses the problem of segmentation in time series

More information

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume

More information

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma COMS 4771 Probabilistic Reasoning via Graphical Models Nakul Verma Last time Dimensionality Reduction Linear vs non-linear Dimensionality Reduction Principal Component Analysis (PCA) Non-linear methods

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Topic Modelling and Latent Dirichlet Allocation

Topic Modelling and Latent Dirichlet Allocation Topic Modelling and Latent Dirichlet Allocation Stephen Clark (with thanks to Mark Gales for some of the slides) Lent 2013 Machine Learning for Language Processing: Lecture 7 MPhil in Advanced Computer

More information

Acoustic Unit Discovery (AUD) Models. Leda Sarı

Acoustic Unit Discovery (AUD) Models. Leda Sarı Acoustic Unit Discovery (AUD) Models Leda Sarı Lucas Ondel and Lukáš Burget A summary of AUD experiments from JHU Frederick Jelinek Summer Workshop 2016 lsari2@illinois.edu November 07, 2016 1 / 23 The

More information

Page 1. References. Hidden Markov models and multiple sequence alignment. Markov chains. Probability review. Example. Markovian sequence

Page 1. References. Hidden Markov models and multiple sequence alignment. Markov chains. Probability review. Example. Markovian sequence Page Hidden Markov models and multiple sequence alignment Russ B Altman BMI 4 CS 74 Some slides borrowed from Scott C Schmidler (BMI graduate student) References Bioinformatics Classic: Krogh et al (994)

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data Statistical Machine Learning from Data Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique Fédérale de Lausanne (EPFL),

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory

More information

Expectation propagation for signal detection in flat-fading channels

Expectation propagation for signal detection in flat-fading channels Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA

More information

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures 17th Europ. Conf. on Machine Learning, Berlin, Germany, 2006. Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures Shipeng Yu 1,2, Kai Yu 2, Volker Tresp 2, and Hans-Peter

More information

Efficient Sensitivity Analysis in Hidden Markov Models

Efficient Sensitivity Analysis in Hidden Markov Models Efficient Sensitivity Analysis in Hidden Markov Models Silja Renooij Department of Information and Computing Sciences, Utrecht University P.O. Box 80.089, 3508 TB Utrecht, The Netherlands silja@cs.uu.nl

More information