Monitoring and data filtering II. Dan Jensen IPH, KU
|
|
- Doreen Todd
- 5 years ago
- Views:
Transcription
1 Monitoring and data filtering II Dan Jensen IPH, KU
2 Outline Introduction to Dynamic Linear Models (DLM) - Conceptual introduction - Difference between the Classical methods and DLM - A very simple DLM and the Kalman filter - Break (5 minutes) Appliction examples using the simple DLM - Break (5 minutes) General form of the DLM Appliction examples using the general DLM Concluding remarks and Exercises
3 The basics of a DLM Dynamic i.e. non-static, adaptive TRUE VALUE Estimation Observed value Forecasted value Uncertainties
4 The basics of a DLM Linear Current value = Previous value + Trend
5 The basics of a DLM Model i.e. we can make forecasts! Alarm Decision Effect estimation: system: support: If tried I staythis the new course, feedhow mixture. will my production look over Doesthe it make nextmy fewproduction years? look better or worse, Therefore: after we strip away the observational noise? How much will it better? look if I change to a faster growing breed? IF everything is fine THEN things progress as expected IF things progress UN-expectedly THEN Something is wrong!
6 Classical methods compared to DLM In Chapter 7: Time series: k1,, kt Model: Moving average EWMA Control charts: test if θ = θ Fundemental assumption: θ is constant over time Here: Model: (Observation equation) Notice: the underlying mean,, can change over time! (System equation)
7 First Order Univariate Polynomial DLM A very simple DLM Updated with each time step!
8 Updating the (simple) DLM: the Kalman filter Prior information: 1-step forecast information: Adaptation information Filtered (posterior) information Prior mean Prior variance Forecast Forecast variance Forecast error Adaptive coefficient Filtered mean Filtered variance
9 Relation to the EWMA EWMA: Filtered mean: When and :
10
11 Incorporate external information Incorporate external information: intervention Types of external information: 1. Known effect, experienced before (ex: new breed with a known different performance) We want the model to adapt to the new known conditions 2. Unknown effect (ex: wave of heat, introduction of new animals in a group) We want the model to adapt to the new unknown conditions 3. Unknown effect we want to measure (ex: change of feed composition, new veterinary treatments) We want to measure the effect of a voluntary change
12 Incorporate external information Intervention - 1. Known effect We want the model to adapt to the new known conditions Ex. 8.4, p. 88: Productivity in broilers / reference weight 38 days V=10000 g 2 (i.e. SD of 333 g) Until batch 10: Ross 208; m 10 = 1883 g, W= 100 g 2 From batch 11: Ross 308,» N (µ, W ) where µ = 70 and W = 100 Revised prior: (θ t D t -1 )» N(m t -1 + µ, R t ), R t = C t -1 + W t + W m 11 = m = = 1953 And R 11 = C 10 + W 11 + W = = 1401
13 Incorporate external information Intervention - 2. Unknown effect We want the model to adapt to the new unknown conditions Ex 1: broiler example with no prior information ~ N (0, 20000) for t = 11 In practice we temporarily apply a larger system variance W
14 Incorporate external information Intervention - 2. Unknown effect Productivity in broilers / reference weight Intervention, prior knowlegde ~ N (70, 100) Intervention, no prior knowlegde ~ N (0, 20000) No intervention Change in production should be modeled as intervention If any prior information is available: use it!
15 Incorporate external information Intervention - 3. Estimation of effect Productivity in broilers / reference weight Smoothed mean + 75 g Intervention, no prior knowlegde ~ N (0, 20000) Smoothing Retrospective analysis
16 Specification of variance components Discount factor (δ)as an aid to choosing W t To run the model (assume with constant parameters) we need: m 0, C 0, V, W Discount factor δ can be used if W is unknown - 0 < δ < 1 we know that W is a fixed proportion of C R t = C t-1 + W R t = C t-1 / δ For a process in control we use the value of delta that minimize the sum of the squares of the forecast errors e t
17 Specification of variance components Daily gain example High value of delta: small system (evolution) variance W, slow adaptation to new information Low value of delta: very adaptive model NB: lower delta can be used for modeling intervention!
18
19 The general Dynamic Linear Model Definition In a general DLM, observations may be multivariate (i.e. vectors) Let Y t = (y 1,, y n ) be a vector of key figures observed at time t. Let θ t = (θ 1,, θ m ) be a vector of parameters describing the system at time t. General form of the DLM Observation Equation: Y t = F t θ t + ν t, ν t ~ N(0,V t ) System Equation: θ t = G t θ t-1 + ω t, ω t ~ N(0,W t ) F t is the design matrix: extracts expected observations from θ t G t is the system matrix: describe how θ t changes over time DLM combined with Kalman Filter: estimate the underlying state vector θ t by its mean vector m t and its variance-covariance matrix C t.
20 A linear growth model Parameter vector Design matrix Observation equation k t = F t θ t + ν t, ν t ~ N(0,V t ) System matrix Covariance matrix System equation θ t = G t θ t-1 + ω t, ω t ~ N(0,W t ) Modeling patterns with a DLM = 0 1 F t = t t θ θ θ = t G = t t t W W W
21 Modeling patterns with a DLM A linear growth model: example of daily feed intake Parameter vector System matrix θ t G t 600 = 25 1 = What is the feed intake at t+1? θ t+1 = G t θ t + ω t Design matrix F t 1 = 0 Observation Equation: Y t = F t θ t + ν t, ν t ~ N(0,V t ) System Equation: θ t = G t θ t-1 + ω t, ω t ~ N(0,W t )
22 Updating the general DLM Prior information: 1-step forecast information: Adaptation information Filtered (posterior) information Prior mean Prior variance Forecast Forecast variance Adaptive coefficient Forecast error Filtered mean Filtered variance
23 Modeling patterns with a DLM A linear growth model: example of daily feed intake Specification of the priors m 0 Observational variance: V = (i.e. 300 g standard deviation) Evolution variance: delta = = 25 C =
24 Monitoring Deviations from the model First of all: use standardized errors: Then: Control chart, with alarm limits Or V-mask (parameters d and Ψ) Applied on the cumulative sum (cusum) of the standardized error: C t t = ut = ut + ct t= 1 1 Or: Others, e.g. Bayesian networks, neural networks, etc.
25 Concluding remarks Differents Models were presented Simple local level model DLM in its general form Examples The general form of the model allows us to include patterns (e.g. cyclic patterns for for drinking activity, linear patterns for daily gain) Not necessarily as graphs automatic alarms (as V-mask, MPKF) Many handles to adjust dangerous Always combine with your knowledge on animal production
Monitoring and data filtering II. Dynamic Linear Models
Monitoring and data filtering II. Dynamic Linear Models Advanced Herd Management Cécile Cornou, IPH Dias 1 Program for monitoring and data filtering Friday 26 (morning) - Lecture for part I : use of control
More informationMonitoring and data filtering III. The Kalman Filter and its relation with the other methods
Monitoring and data filtering III. The Kalman Filter and its relation with the other methods Advanced Herd Management Cécile Cornou, IPH Dias 1 Gain (g) Before this part of the course Compare key figures
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationDynamic Matrix-Variate Graphical Models A Synopsis 1
Proc. Valencia / ISBA 8th World Meeting on Bayesian Statistics Benidorm (Alicante, Spain), June 1st 6th, 2006 Dynamic Matrix-Variate Graphical Models A Synopsis 1 Carlos M. Carvalho & Mike West ISDS, Duke
More informationCIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions
CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions December 14, 2016 Questions Throughout the following questions we will assume that x t is the state vector at time t, z t is the
More informationModel Probability Var = 1 Var = 2.25 Var = 25
1 Background Detection of Oestrus by Monitoring Boar Visits Tage Ostersen 2 Data 3 Duration of Visits 4 Break 5 Frequency of Visits 6 Visits 7 Conclusion Advanced Herd Management 22/9 29 Slide 1/3 Slide
More informationL06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms
L06. LINEAR KALMAN FILTERS NA568 Mobile Robotics: Methods & Algorithms 2 PS2 is out! Landmark-based Localization: EKF, UKF, PF Today s Lecture Minimum Mean Square Error (MMSE) Linear Kalman Filter Gaussian
More informationDetection of Oestrus by Monitoring Boar Visits Tage Ostersen Advanced Herd Management 22/ Slide 1/30 1 Background 2 Data 3 Duration of Visits 4
Detection of Oestrus by Monitoring Boar Visits Tage Ostersen Advanced Herd Management 22/9 2009 Slide 1/30 1 Background 2 Data 3 Duration of Visits 4 Break 5 Frequency of Visits 6 Combining Frequency and
More information2D Image Processing. Bayes filter implementation: Kalman filter
2D Image Processing Bayes filter implementation: Kalman filter Prof. Didier Stricker Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
More informationIntroduction to Probabilistic Graphical Models: Exercises
Introduction to Probabilistic Graphical Models: Exercises Cédric Archambeau Xerox Research Centre Europe cedric.archambeau@xrce.xerox.com Pascal Bootcamp Marseille, France, July 2010 Exercise 1: basics
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationBasics: Definitions and Notation. Stationarity. A More Formal Definition
Basics: Definitions and Notation A Univariate is a sequence of measurements of the same variable collected over (usually regular intervals of) time. Usual assumption in many time series techniques is that
More informationDynamic monitoring of mortality rate for sows and piglets
Dynamic monitoring of mortality rate for sows and piglets Claudia Bono C. Cornou A. R. Kristensen Dias 1 Background Problem Existing Management Information Systems (MIS) are static and typically computed
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More information2D Image Processing. Bayes filter implementation: Kalman filter
2D Image Processing Bayes filter implementation: Kalman filter Prof. Didier Stricker Dr. Gabriele Bleser Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationEKF and SLAM. McGill COMP 765 Sept 18 th, 2017
EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/term
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t.
ECO 513 Fall 2008 C.Sims KALMAN FILTER Model in the form 1. THE KALMAN FILTER Plant equation : s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. Var(ε t ) = Ω, Var(ν t ) = Ξ. ε t ν t and (ε t,
More informationStatistical Methods for Forecasting
Statistical Methods for Forecasting BOVAS ABRAHAM University of Waterloo JOHANNES LEDOLTER University of Iowa John Wiley & Sons New York Chichester Brisbane Toronto Singapore Contents 1 INTRODUCTION AND
More informationFactor Analysis and Kalman Filtering (11/2/04)
CS281A/Stat241A: Statistical Learning Theory Factor Analysis and Kalman Filtering (11/2/04) Lecturer: Michael I. Jordan Scribes: Byung-Gon Chun and Sunghoon Kim 1 Factor Analysis Factor analysis is used
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.2: Kalman Filter Jürgen Sturm Technische Universität München Motivation Bayes filter is a useful tool for state
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationLecture 16: State Space Model and Kalman Filter Bus 41910, Time Series Analysis, Mr. R. Tsay
Lecture 6: State Space Model and Kalman Filter Bus 490, Time Series Analysis, Mr R Tsay A state space model consists of two equations: S t+ F S t + Ge t+, () Z t HS t + ɛ t (2) where S t is a state vector
More informationProbabilistic numerics for deep learning
Presenter: Shijia Wang Department of Engineering Science, University of Oxford rning (RLSS) Summer School, Montreal 2017 Outline 1 Introduction Probabilistic Numerics 2 Components Probabilistic modeling
More informationDynamic chain graphs for high-dimensional time series: an application to real-time traffic flow forecasting
Dynamic chain graphs for high-dimensional time series: an application to real-time traffic flow forecasting Catriona Queen Joint work with Osvaldo Anacleto 1 Department of Mathematics and Statistics The
More informationKalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University
Kalman Filter 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Examples up to now have been discrete (binary) random variables Kalman filtering can be seen as a special case of a temporal
More informationMarkov localization uses an explicit, discrete representation for the probability of all position in the state space.
Markov Kalman Filter Localization Markov localization localization starting from any unknown position recovers from ambiguous situation. However, to update the probability of all positions within the whole
More informationPerformance of Conventional X-bar Chart for Autocorrelated Data Using Smaller Sample Sizes
, 23-25 October, 2013, San Francisco, USA Performance of Conventional X-bar Chart for Autocorrelated Data Using Smaller Sample Sizes D. R. Prajapati Abstract Control charts are used to determine whether
More informationContents. Preface to Second Edition Preface to First Edition Abbreviations PART I PRINCIPLES OF STATISTICAL THINKING AND ANALYSIS 1
Contents Preface to Second Edition Preface to First Edition Abbreviations xv xvii xix PART I PRINCIPLES OF STATISTICAL THINKING AND ANALYSIS 1 1 The Role of Statistical Methods in Modern Industry and Services
More informationMiscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier
Miscellaneous Regarding reading materials Reading materials will be provided as needed If no assigned reading, it means I think the material from class is sufficient Should be enough for you to do your
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationXXV ENCONTRO BRASILEIRO DE ECONOMETRIA Porto Seguro - BA, 2003 REVISITING DISTRIBUTED LAG MODELS THROUGH A BAYESIAN PERSPECTIVE
XXV ENCONTRO BRASILEIRO DE ECONOMETRIA Porto Seguro - BA, 2003 REVISITING DISTRIBUTED LAG MODELS THROUGH A BAYESIAN PERSPECTIVE Romy R. Ravines, Alexandra M. Schmidt and Helio S. Migon 1 Instituto de Matemática
More informationIntroduction to Time Series (I)
Introduction to Time Series (I) ZHANG RONG Department of Social Networking Operations Social Networking Group Tencent Company November 20, 2017 ZHANG RONG Introduction to Time Series (I) 1/69 Outline 1
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49
State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing
More informationHierarchical Modelling for Univariate Spatial Data
Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationBayesian non-parametric model to longitudinally predict churn
Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics
More informationPart I State space models
Part I State space models 1 Introduction to state space time series analysis James Durbin Department of Statistics, London School of Economics and Political Science Abstract The paper presents a broad
More informationMaximum Likelihood Estimation. only training data is available to design a classifier
Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture State space models, 1st part: Model: Sec. 10.1 The
More informationState Observers and the Kalman filter
Modelling and Control of Dynamic Systems State Observers and the Kalman filter Prof. Oreste S. Bursi University of Trento Page 1 Feedback System State variable feedback system: Control feedback law:u =
More informationGaussians. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics
Gaussians Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Univariate Gaussian Multivariate Gaussian Law of Total Probability Conditioning
More informationCS491/691: Introduction to Aerial Robotics
CS491/691: Introduction to Aerial Robotics Topic: State Estimation Dr. Kostas Alexis (CSE) World state (or system state) Belief state: Our belief/estimate of the world state World state: Real state of
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationThe Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance
The Kalman Filter Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience Sarah Dance School of Mathematical and Physical Sciences, University of Reading s.l.dance@reading.ac.uk July
More informationMultivariate Gaussian Random Fields with SPDEs
Multivariate Gaussian Random Fields with SPDEs Xiangping Hu Daniel Simpson, Finn Lindgren and Håvard Rue Department of Mathematics, University of Oslo PASI, 214 Outline The Matérn covariance function and
More informationTime Series Prediction by Kalman Smoother with Cross-Validated Noise Density
Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density Simo Särkkä E-mail: simo.sarkka@hut.fi Aki Vehtari E-mail: aki.vehtari@hut.fi Jouko Lampinen E-mail: jouko.lampinen@hut.fi Abstract
More informationDependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.
Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,
More informationTIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.
TIME SERIES ANALYSIS Forecasting and Control Fifth Edition GEORGE E. P. BOX GWILYM M. JENKINS GREGORY C. REINSEL GRETA M. LJUNG Wiley CONTENTS PREFACE TO THE FIFTH EDITION PREFACE TO THE FOURTH EDITION
More informationKalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q
Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I
More informationBayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.
Kalman Filter Localization Bayes Filter Reminder Prediction Correction Gaussians p(x) ~ N(µ,σ 2 ) : Properties of Gaussians Univariate p(x) = 1 1 2πσ e 2 (x µ) 2 σ 2 µ Univariate -σ σ Multivariate µ Multivariate
More informationSection II: Assessing Chart Performance. (Jim Benneyan)
Section II: Assessing Chart Performance (Jim Benneyan) 1 Learning Objectives Understand concepts of chart performance Two types of errors o Type 1: Call an in-control process out-of-control o Type 2: Call
More informationStochastic Models, Estimation and Control Peter S. Maybeck Volumes 1, 2 & 3 Tables of Contents
Navtech Part #s Volume 1 #1277 Volume 2 #1278 Volume 3 #1279 3 Volume Set #1280 Stochastic Models, Estimation and Control Peter S. Maybeck Volumes 1, 2 & 3 Tables of Contents Volume 1 Preface Contents
More information9 Multi-Model State Estimation
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State
More informationData assimilation in high dimensions
Data assimilation in high dimensions David Kelly Courant Institute New York University New York NY www.dtbkelly.com February 12, 2015 Graduate seminar, CIMS David Kelly (CIMS) Data assimilation February
More informationTMA4285 December 2015 Time series models, solution.
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 5 TMA4285 December 205 Time series models, solution. Problem a) (i) The slow decay of the ACF of z t suggest that
More informationModels for models. Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research
Models for models Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research Outline Statistical models and tools Spatial fields (Wavelets) Climate regimes (Regression and clustering)
More informationCondensed Table of Contents for Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control by J. C.
Condensed Table of Contents for Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control by J. C. Spall John Wiley and Sons, Inc., 2003 Preface... xiii 1. Stochastic Search
More informationBAYESIAN DYNAMIC MODELLING
BAYESIAN DYNAMIC MODELLING Mike West Department of Statistical Science, Duke University in honour of Professor Sir Adrian F.M. Smith in Bayesian Theory and Applications (eds: P. Damien and P. Dellaportas
More informationCHAPTER 3 DYNAMIC PRODUCTION MONITORING IN PIG HERDS I: MODELING AND MONITORING LITTER SIZE AT HERD AND SOW LEVEL
CHAPTER 3 DYNAMIC PRODUCTION MONITORING IN PIG HERDS I: MODELING AND MONITORING LITTER SIZE AT HERD AND SOW LEVEL Claudia Bono, Cécile Cornou and Anders Ringgaard Kristensen Published in Livestock Science
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationCS260: Machine Learning Algorithms
CS260: Machine Learning Algorithms Lecture 4: Stochastic Gradient Descent Cho-Jui Hsieh UCLA Jan 16, 2019 Large-scale Problems Machine learning: usually minimizing the training loss min w { 1 N min w {
More informationSoftware for non-stationary time series analysis and decompositions via TVAR models. Raquel Prado July 26, Overview This software implements ti
Software for non-stationary time series analysis and decompositions via TVAR models. Raquel Prado July 26, 2000 1 Overview This software implements time-varying autoregressions or TVAR models and allows
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning
More informationECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form
ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER KRISTOFFER P. NIMARK The Kalman Filter We will be concerned with state space systems of the form X t = A t X t 1 + C t u t 0.1 Z t
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationStatistics Introduction to Probability
Statistics 110 - Introduction to Probability Mark E. Irwin Department of Statistics Harvard University Summer Term Monday, June 26, 2006 - Wednesday, August 16, 2006 Copyright 2006 by Mark E. Irwin Personnel
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationHypothesis Testing. ) the hypothesis that suggests no change from previous experience
Hypothesis Testing Definitions Hypothesis a claim about something Null hypothesis ( H 0 ) the hypothesis that suggests no change from previous experience Alternative hypothesis ( H 1 ) the hypothesis that
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Centre for Computational Statistics and Machine Learning University College London c.archambeau@cs.ucl.ac.uk CSML
More informationLecture 9. Time series prediction
Lecture 9 Time series prediction Prediction is about function fitting To predict we need to model There are a bewildering number of models for data we look at some of the major approaches in this lecture
More information+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing
Linear recurrent networks Simpler, much more amenable to analytic treatment E.g. by choosing + ( + ) = Firing rates can be negative Approximates dynamics around fixed point Approximation often reasonable
More informationLecture 4: Dynamic models
linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationRobotics 2 Target Tracking. Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard
Robotics 2 Target Tracking Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard Linear Dynamical System (LDS) Stochastic process governed by is the state vector is the input vector is the process
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationMultivariate Discount Weighted Regression and Local Level Models
Multivariate Discount Weighted Regression and Local Level Models Kostas Triantafyllopoulos a a University of Newcastle, Newcastle upon Tyne, UK. Abstract In this paper we propose a multivariate discount
More informationSequential Bayesian Updating
BS2 Statistical Inference, Lectures 14 and 15, Hilary Term 2009 May 28, 2009 We consider data arriving sequentially X 1,..., X n,... and wish to update inference on an unknown parameter θ online. In a
More informationDimension Reduction. David M. Blei. April 23, 2012
Dimension Reduction David M. Blei April 23, 2012 1 Basic idea Goal: Compute a reduced representation of data from p -dimensional to q-dimensional, where q < p. x 1,...,x p z 1,...,z q (1) We want to do
More informationLong-Run Covariability
Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationMaximum Likelihood (ML), Expectation Maximization (EM) Pieter Abbeel UC Berkeley EECS
Maximum Likelihood (ML), Expectation Maximization (EM) Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Outline Maximum likelihood (ML) Priors, and
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationMachine Learning Lecture 3
Announcements Machine Learning Lecture 3 Eam dates We re in the process of fiing the first eam date Probability Density Estimation II 9.0.207 Eercises The first eercise sheet is available on L2P now First
More informationCOM336: Neural Computing
COM336: Neural Computing http://www.dcs.shef.ac.uk/ sjr/com336/ Lecture 2: Density Estimation Steve Renals Department of Computer Science University of Sheffield Sheffield S1 4DP UK email: s.renals@dcs.shef.ac.uk
More informationIntroduction to Mobile Robotics Probabilistic Robotics
Introduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1 Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception Action
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationEstimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio
Estimation of reliability parameters from Experimental data (Parte 2) This lecture Life test (t 1,t 2,...,t n ) Estimate θ of f T t θ For example: λ of f T (t)= λe - λt Classical approach (frequentist
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationGaussian Filtering Strategies for Nonlinear Systems
Gaussian Filtering Strategies for Nonlinear Systems Canonical Nonlinear Filtering Problem ~u m+1 = ~ f (~u m )+~ m+1 ~v m+1 = ~g(~u m+1 )+~ o m+1 I ~ f and ~g are nonlinear & deterministic I Noise/Errors
More informationExpectation Propagation for Approximate Bayesian Inference
Expectation Propagation for Approximate Bayesian Inference José Miguel Hernández Lobato Universidad Autónoma de Madrid, Computer Science Department February 5, 2007 1/ 24 Bayesian Inference Inference Given
More informationUncertainty in energy system models
Uncertainty in energy system models Amy Wilson Durham University May 2015 Table of Contents 1 Model uncertainty 2 3 Example - generation investment 4 Conclusion Model uncertainty Contents 1 Model uncertainty
More informationParameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets
Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Matthias Katzfuß Advisor: Dr. Noel Cressie Department of Statistics The Ohio State University
More information1.1. VARs, Wold representations and their limits
1. Shocks Nr. 1 1.1. VARs, Wold representations and their limits A brief review of VARs. Assume a true model, in MA form: X = A 0 e + A 1 e( 1) + A 2 e( 2) +...; E(ee ) = I = (A 0 + A 1 L + A 2 L 2 +...)
More information