Statistical Inference and Methods

Size: px
Start display at page:

Download "Statistical Inference and Methods"

Transcription

1 Department of Mathematics Imperial College London das01/ 31st January 2006

2 Part VI Session 6: Filtering and Time to Event Data

3 Session 6: Filtering and 1/ 85 Filtering State-space Models Kalman Filter Non-linear/Non-Gaussian Filtering Particle Filters Discrete & Continuous Time Models Non-parametric analysis Kaplan-Meier plots Proportional Hazards/Accelerated Life Multivariate Reliability & Competing Risks Multi-State Models

4 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 2/ 85 State-Space Models The state-space form of time series models has received considerable attention, as they can better represent the actual dynamics of a data generation process. Let y t denote the observation from a (possibly multivariate) time series at time t, related to a vector α t, called the state vector, which is possibly unobserved and whose dimension, m say, is independent of the dimension, n, of y t.

5 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 3/ 85 The general form of a linear state-space model, is given by the following two equations y t = Z t α t + d t + G t ε t, t = 1,..., T (1a) α t+1 = T t α t + c t + H t ε t. (1b) Equation (1a) is the observation or measurement equation, while (1b) is transition equation.

6 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 4/ 85 In (1) Z t is an (n m) matrix, d t is an (n 1) vector, G t an (n (n + m)) matrix, T t is (m m), c t is (m 1) and H t is an (m (n + m)) matrix. All of the latter matrices are referred to as the system matrices and are assumed to be non-stochastic.

7 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 5/ 85 Process ε t is an ((n + m) 1) vector of serially independent, identically distributed disturbances with E (ε t ) = 0 and Var (ε t ) = I, the identity matrix. The formulation is completed by the assumption that the initial state vector α 1 is independent of ε t at all times t and has unconditional mean and variance a 1 0 and P 1 0 respectively.

8 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 6/ 85 The state-space model is characterized by the properties that the system matrices, the disturbance terms ε t and initial state vector α 1 possess. If the system matrices do not evolve with time, the state-space model is called time-invariant or time-homogeneous. If the disturbances, ε t, and initial state vector, α 1, are assumed to have a normal distribution, then the model is termed Gaussian. Finally, it should be noted that if G t H t = 0 for all t, then the measurement and transition equations are uncorrelated.

9 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 7/ 85 Example AR(2) Model Let {X t } be a zero-mean, Gaussian AR(2) process given by X t+1 = φ 1 X t + φ 2 X t 1 + u 2t where u 2t NID ( 0, σ 2 2). Suppose also that we observe the process with noise, so that Y t = X t + u 1t with u 1t NID ( 0, σ 2 1) and independent of u2t.

10 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 8/ 85 Example A state-space representation of this model can be formed as follows Y t = [ 1 0 ] [ ] X t + [ σ ] [ ] ε 1t [ Xt+1 X t ] = [ φ1 φ X t 1 ] [ Xt X t 1 ] + [ 0 σ ε 2t ] [ ε1t where ε t = (ε 1t, ε 2t ) is NID (0, I). Clearly the latter state-space model is time-homogeneous and Gaussian and the measurement and transition equations are uncorrelated. ε 2t ]

11 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 9/ 85 Kalman Filter The Kalman filter is used for prediction, filtering and smoothing. In particular, if we let Ψ t denote the information set up to time t, i.e. Ψ t = {y 1,..., y t }, then the problem of prediction is to compute E (α t+1 Ψ t ). Filtering is concerned with calculating E (α t Ψ t ), while (fixed-interval) smoothing is concerned with estimating E (α t Ψ T ), for all t < T. The Kalman filter consists of a set of simple recursions that compute the latter quantities.

12 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 10/ 85 Linear Gaussian State-Space Model : Assume that G t H t = 0 Moreover, we drop the (exogenous) terms d t and c t from the observation and transition equations for simplicity and write G t G t = Σ t and H t H t = Ω t.

13 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 11/ 85 The Kalman filter recursively computes the quantities a t t = E (α t Ψ t ), a t+1 t = E (α t+1 Ψ t ), P t t = MSE (α t Ψ t 1 ), P t+1 t = MSE (α t+1 Ψ t ). where MSE is the mean-square error or one-step ahead prediction variance.

14 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 12/ 85 Then, starting with a 1 0 and P 1 0, a t t, a t+1 t and their MSEs are obtained by running for t = 1,..., T, the recursions v t = y t Z t a t t 1, F t = Z t P t t 1 Z t + Σ t (2a) a t t = a t t 1 + P t t 1 Z tf 1 t v t, (2b) P t t = P t t 1 P t t 1 Z tf t 1 Z t P t t 1, (2c) a t+1 t = T t a t t, (2d) P t+1 t = T t P t t T t + Ω t. (2e)

15 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 13/ 85 Notes: quantities v t and F t in (2a) respectively denote the one-step-ahead error in forecasting y t conditional on the information set at time t 1 and its MSE. This facilitates computation of the likelihood function. the quantities a t t and a t t 1 are optimal estimators of α t conditional on the available information, in terms of minimum mean square. the latter property holds under Gaussianity; if this is not assumed, a t t and a t t 1 are optimal only within the class of linear estimators, that is, they are the minimum mean square linear estimators of α t conditional on Ψ t and Ψ t 1.

16 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 14/ 85 Finally, equations (2b) and (2d ) can be combined together to yield recursions that only compute the one-step-ahead prediction estimates of α t+1 given Ψ t. Similarly, by taking together (2c) and (2e), a single set of recursions for the MSE is obtained, which goes directly from P t t 1 to P t+1 t.

17 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 15/ 85 The resulting recursions, along with (2a), for t = 1,..., T 1, are as follows: a t+1 t = T t a t t 1 + K t v t (3a) K t = T t P t t 1 Z tf t 1 (3b) P t+1 t = T t P t+1 t L t + Ω t (3c) L t = T t K t Z t. (3d)

18 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 16/ 85 Parameter Estimation Another application of the Kalman filter is the estimation of any unknown parameters θ that appear in the system matrices. The likelihood for data y = (y 1,..., y T ) can be constructed as p (y 1,..., y T ) = p (y T Ψ T 1 ) p (y 2 Ψ 1 ) p (y 1 ) = T p (y t Ψ t 1 ). t=1

19 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 17/ 85 Assuming that the state-space model is Gaussian, by taking conditional expectations on both sides of the observation equation, with d t 0, we deduce that for t = 1,..., T, E (y t Ψ t 1 ) = Z t a t t 1, Var (y t Ψ t 1 ) = F t. Crucially, the one-step-ahead prediction density p (y t Ψ t 1 ) is the density of a multivariate normal random variable with mean Z t a t t 1 and covariance matrix F t.

20 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 18/ 85 Thus, the log-likelihood function is given by log L = nt 2 log (2π) 1 2 where T log (det F t ) 1 2 t=1 v t = ( y t Z t a t t 1 ) T t=1 v tf 1 t v t. Typically, numerical procedures are used in order to maximize the log-likelihood to obtain the ML estimates of the parameters θ which are consistent and asymptotically normal.

21 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 19/ 85 Non-Linear Non-Gaussian State-Space Models : If the state-space model is not Gaussian, the likelihood can still be constructed in the same way using the minimum mean square linear estimators of the state vector. However, the estimators θ that maximize the likelihood are the quasi-maximum likelihood (QML) estimators of the parameters. It can be shown that the QML estimators are also consistent and asymptotically normal.

22 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 20/ 85 Another algorithm can be applied to a state-space model; given a fixed set of data, estimates of the state vector are computed at each time t in the sample period taking into account the full information set available. This algorithm computes a t T = E (α t Ψ T ) along with its MSE, P t T, for all t = 1,..., T 1.

23 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 21/ 85 The latter quantities are computed via a set of backward recursions and use the quantities a t+1 t, a t t and the MSE matrices P t t, P t+1 t, which are obtained from (2). In particular, to obtain a t T and P t T, we start with a T T and P T T and we run backwards for t = T 1,..., 0 a t T = a t t + Pt ( ) at+1 T a t+1 t P t T = P t t + Pt ( ) Pt+1 T P t+1 t P t, Pt = P t T tp t+1 t.

24 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 22/ 85 Markov chain Monte Carlo Applications During the last decade, the extensive use of the MCMC, in particular the Gibbs sampler, has given rise to another smoothing algorithm called the simulation smoother and is also closely related to the Kalman filter. In contrast, to the fixed interval smoother, which computes the conditional mean and variance of the state vector at each time t in the sample, a simulation smoother is used for drawing samples from the density p (α 0,..., α T Y T ).

25 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 23/ 85 The first simulation smoother is based on the identity p (α 0,..., α T Ψ T ) = p (α T Ψ T ) T 1 t=0 p (α t Ψ t, α t+1 ) and a draw from p (α 0,..., α T Ψ T ) is recursively constructed in terms of α t.

26 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 24/ 85 Starting with a draw α T N ( α T T, P T T ), the main idea is that for a Gaussian state space model p (α t Ψ t, α t+1 ) is a multivariate normal density and hence it is completely characterized by its first and second moments. In order to compute these moments, the usual Kalman filter recursions (2) are run, so that α t t is initially obtained. Then, the draw α t+1 p (α t+1 Ψ t, α t+2 ) is treated as m additional observations and a second set of m Kalman filter recursions is run for each element of the state vector α t+1.

27 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 25/ 85 However, the latter procedure involves the inversion of system matrices, which are not necessarily non-singular. To overcome this problem another simulation smoother algorithm has been devised which is computationally more efficient. The main idea of the algorithm is to obtain a sample from the joint density of the transition equation disturbances, i.e. p (H 0 ε 0,..., H T ε T Ψ T ), by sampling η t p (H t ε t Ψ T, H t+1 ε t+1,..., H T ε T ).

28 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 26/ 85 This allows the state vector to be reconstructed, using the transition equation (1b) and substituting the simulated disturbances. In particular, the new simulation smoother requires only the prediction equations (3) to be initially run and the quantities v t, F t and L t to be stored. Then, starting with r T = 0 and U T = 0, the following backward recursions are run for t = T, T 1,..., 1. C t = Ω t Ω t U t Ω t, κ t N (0, C t ), r t 1 = Z tf t 1 v t + L ( t rt U t Ω t Ct 1 ) κ t, U t 1 = Z tf t 1 Z t + L ( t Ut + U t Ω t Ct 1 ) Ω t U t Lt.

29 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 27/ 85 Then, for t = 1,..., T, η t = Ω t r t + κ t is a draw from p (H t ε t Ψ T, H t+1 ε t+1,..., H T ε T ). For the initial disturbance term, η 0 = P 1 0 r 0 + κ 0, where κ 0 N ( 0, P 1 0 P 1 0 U 0 P 1 0 ), is a draw from p (H 0 ε 0 Ψ T, H 1 ε 1,..., H T ε T ). The state vector is simulated by starting with α 0 = 0 and running for t = 0,..., T the transition equation recursion with H t ε t substituted by η t, i.e. α t+1 = T t α t + η t.

30 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 28/ 85 Inference in General State-Space Models To recap, inference goals for state-space models are filtering/tracking: estimate state at time t from all observations up to time t p(α t Ψ t ) smoothing: estimate state at time t from all past and possibly some future observations p(α t Ψ t+s ) prediction: estimate state at time t from observations up to a previous time point t s p(α t Ψ t s )

31 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 29/ 85 The general (first-order Markov) state equation takes the form α t = f (α t 1, θ t 1 ) + η t 1 and the general observation equation takes the form y t = h(α t, θ t ) + ɛ t. For simplicity, we assume that the error processes {η t } and {ɛ t } are independent.

32 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 30/ 85 The fundamental inference mechanism is Bayesian; we compute the posterior quantities of interest sequentially in the following recursive calculation: Prediction p(α t Ψ t 1 ) = p(α t α t 1 )p(α t 1 Ψ t 1 ) dα t 1 Updating p(α t Ψ t ) = p(y t α t )p(α t Ψ t 1 ) p(y t Ψ t 1 )

33 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 31/ 85 The above equations only provide an analytical solution if all densities in the state and observation equation are Gaussian, and both the state and the observation equation are linear. If these conditions are met, the Kalman filter provides the optimal Bayesian solution to the tracking problem. If these conditions are not met, we require approximations: Extended Kalman filter (EKF): requires Gaussian densities, approximates non-linearities (using a first-order Taylor expansion) Particle filter (PF): approximates non-gaussian densities and non-linear equations

34 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 32/ 85 Particle Filter The particle filter uses Monte Carlo methods, in particular Importance sampling, to construct the approximations. We approximate the integral E f [g(x )] = g(x)f (x) dx by writing E f [g(x )] E f0 [ f (X )g(x ) f 0 (X ) for some importance sampling density f 0. ] g(x)f (x) = f 0 (x) dx f 0 (x)

35 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 33/ 85 We then approximate the integral by sampling x 1,..., x N from f 0 for large N, then constructing the approximation Ê fn = 1 N N i=1 g(x i )f (x i ) f 0 (x i ) = 1 N N w i g(x i ), i=1 say, where, if f 0 is chosen carefully as N Ê fn E f [g(x )]

36 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 34/ 85 If the function f is not known exactly, but only up to proportionality, the same procedure can work if the weights w i are suitably normalized to sum to one, that is, if above is replaced by w i = w i W w i = f (x i) f 0 (x i ) as W approximates the integral f (x) dx = W = n w i. i=1 f (x) f 0 (x) f 0(x) dx

37 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 35/ 85 Given a non-linear and non-gaussian state space model, the goal is to compute the posterior density p(α t Ψ t ) at each time t exactly. At each time point t, perform the following Sequential Importance Sampling (SIS) procedure: Draw random samples ( particles ) from a chosen importance density that can be sampled directly. Compute a normalized weight for each particle Approximate the true posterior density at time t by the weighted sum of the particles. Given sufficiently large numbers of particles, this IS characterization converges to the true posterior density.

38 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 36/ 85 In the filtering problem, at time t, we approximate p(α t Ψ t ) by particles x t = (x t1,..., x tn ), a sample from some importance sampling density f 0. normalized weights w t = (w t1,..., w tn ) via the discrete distribution {(w ti, x ti ), i = 1..., N}

39 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 37/ 85 Two major questions: What importance density f 0 can or should be chosen? How do we compute the weights efficiently, i.e. in a recursive fashion such that for a given weight wti, we just need a new observation y t to get an approximation to the posterior p(α t Ψ t ).

40 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 38/ 85 A simple choice for the importance sampling density at time t is the transition prior, that is as then f 0 (α t x t 1, Ψ t ) p(α t x t 1 ) w ti w t 1,ip(y t x ti). This procedure works well in general.

41 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 39/ 85 Particle Degeneracy: A problem with sequential importance sampling is that typically, after a few iterations, all but one particle have negligible weights. This phenomenon is termed degeneracy. The degree of degeneracy is represented by the effective sample size ( N ) 1 N t = (wti) 2 i=1 Possible solutions to overcome degeneracy effects: brute force: very large N optimized choice of importance density resampling

42 State-Space Models Kalman Filter MCMC General State-Space Models Session 6: State-space Models & Filtering 40/ 85 Resampling: Goal is to eliminate particles with small weights and replace by new particles drawn from the vicinity of particles with larger weights concentrate on particles with large weights model important parts of the posterior more precisely Various efficient algorithms exist for resampling which operate in O(N) time (e.g. Sampling Importance Resampling (SIR))

43 Session 6: 41/ 85 In many data sets the outcome of interest is the time to an event time between trades time to a credit default time until stock value threshold excedance actuarial survival warranty exposure to insurance risk

44 Session 6: 42/ 85 The distinguishing feature of such data is that at the end of the follow up period the event may not have be observed, and thus the occurrence time is censored. Censoring may occur as truncation of study period loss to follow up from non-specific cause loss to a competing event unrelated to the cause of failure being studied

45 Session 6: 43/ 85 Typically, components are installed over a period and followed up to a fixed date beyond the end of study, and the last components installed will thus be studied for a shorter period than those installed first, and will be less likely to undergo failure. We often can legitimately re-align installation times, and work in relative rather than calendar time.

46 Session 6: 44/ 85 Mathematical Notation: The principal difference between time-to-event analysis and conventional regression is that account is taken of potential censoring in the response variable we may observe some actual responses (survival, failure) times, censored responses where we do not observe an actual failure but rather only that the failure occurs after a censoring time (the end of study) this is called right-censoring occasionally, we observe left-censoring or interval-censoring

47 Session 6: 45/ 85 the response data is thus bivariate (Y, Z) where Y is the time at which the response is measured, and { 1 Failure is observed Z = 0 Censored

48 Session 6: 46/ 85 The potential presence of censoring fundamentally changes how we view the modelling process - previously we have looked at probability densities and expected responses etc. we have previously only dealt with data y for which we need to specify P [Y = y]; we now need to think about P [Y > y] for right censoring P [Y y] for left censoring We now take an alternative view, and examine reliability/survivor and hazard functions.

49 Session 6: 47/ 85 Survival In Discrete Time The probability mass function for response variable Y is f Y, f Y (y) = P [Y = y] y = 0, 1, 2,... The distribution function F Y is F Y (y) = P [Y y] = y f Y (t) = P [Y = 0] P [Y = y] t=0 that is a cumulative probability function. Note that the function F Y (y) is a non-decreasing function.

50 Session 6: 48/ 85 In conventional regression modelling, the probability contribution for data point i with response y i is f Y (y i ). For right-censored data with censoring at y i, however, we only observe the event Y > y i that is, death/failure has not occurred before y i time units. This event has probability P [Y > y i ] = 1 F Y (y i ) This motivates consideration of the survivor (reliability) function S Y (y) = 1 F Y (y) Note that S Y (y) is a non-increasing function.

51 Session 6: 49/ 85 The likelihood function (via which inference and testing will be done) is thus f Y (y i ) S Y (y i ) that is i:z i =1 i:z i =0 LIKELIHOOD FOR UNCENSORED DATA LIKELIHOOD FOR CENSORED DATA and the role of the predictors can be introduced via the parameters of f Y and F Y.

52 Session 6: 50/ 85 Let f Y (y) = P [Y = y] y = 0, 1, 2,... define a discrete failure distribution. Then S Y (y) = P [Y > y] = 1 F Y (y) = f Y (j) j=y+1

53 Session 6: 51/ 85 Example Geometric Model For some probability π) f Y (y) = (1 π) y π y = 0, 1, 2,... and S Y (y) = (1 π) y+1 y = 0, 1, 2,...

54 Session 6: 52/ 85 The Discrete Hazard Function As an alternative method of specification, we consider the discrete hazard function h Y (y) = P [Failure at y Survival y] = and the integrated hazard H Y (y) = y h Y (t). t=0 f Y (y) S Y (y 1)

55 Session 6: 53/ 85 Thus and y 1 f Y (y) = (1 h Y (j)) h Y (y) j=0 y S Y (y) = (1 h Y (j)) j=0

56 Session 6: 54/ 85 Example Constant Hazard If f Y (y) = (1 π) y π y = 0, 1, 2,... then h Y (y) = (1 π)y π (1 π) y+1 = π 1 π that is, a constant, independent of y.

57 Session 6: 55/ 85 The Continuous Time Model The probability density function for continuous response variable Y is f Y, and the expectation, likelihood function and so on that are required for regression modelling are formed from f Y. The distribution function F Y is F Y (y) = P [Y y] = y 0 f Y (t) dt In conventional regression modelling, the likelihood contribution for data point i with response y i is f Y (y i ). For right-censored data with censoring at y i, we have again the reliability function S Y (y) = 1 F Y (y)

58 Session 6: 56/ 85 Continuous Hazards As a further alternative method of specification, we consider the continuous hazard function h Y (y) = lim δy 0 P [Failure in (y, y + δy) Survival y] = f Y (y) S Y (y) and the integrated hazard and it can be shown that H Y (y) = y 0 h Y (t) dt S Y (y) = exp { H Y (y)}

59 Session 6: 57/ 85 The Kaplan-Meier Curve The Kaplan-Meier curve (or product-limit estimate) is a non-parametric estimate of the reliability function; it is a decreasing step-function, where the downward steps take place at the times of the failures The estimated reliability function at the jth failure/censoring time as j ( ) z i Ŝ j = 1 (4) n i + 1 i=1 Standard errors are also available.

60 Session 6: 58/ 85 Construction: Let sample size n comprise observed and censored failure times 0 < y (1) < y (2) <... < y (m), be the distinct failure times, sorted into ascending order d j be the number of number of failures observed at time y (j) usually d j = 1 certainly d j 1 (d j > 1 implies tied failure times) n j be the number of patients at risk of failure at time t (j), that is, the number of patients who have failure/censoring time greater than or equal to t (j).

61 Session 6: 59/ 85 Then the observed probability of surviving beyond t (j) (conditional on having survived that long) is p j = n j d j n j = 1 q j say, where q j = d j /n j is the estimated conditional probability of failure at time t (j). Using the chain rule for probabilities, the estimated probability of surviving at least until time t is J t J t ( P (t) = p j = 1 d ) j = n Ŝ KM (t) (5) j j=1 where J t = max { j : t (j) t }. j=1

62 Session 6: 60/ 85 Standard Errors: A number of possibilities have been suggested. Let P j = P ( ) t (j). Then Greenwood s Formula Peto s Formula ( ) s.e. Pj ( ) s.e. Pj = P j 1 d i j n i d i i=0 = P j 1 P j n j where n j is an adjusted or effective sample size, the number of survivors at the beginning of the interval ( tj, t (j+1) ).

63 Session 6: 61/ 85 The Nelson-Aalen Curve The Nelson-Aalen estimate is a non-parametric estimate of the cumulative hazard function; it takes the form J t ( ) dj Ĥ (t) = (6) j=1 where J t = max { j : t (j) t }. From this, we can construct another estimate of the reliability function } Ŝ FH (t) = exp { Ĥ (t) this is the Fleming-Harrington estimate of the reliability function. n j

64 Session 6: 62/ 85 Standard Errors: If Ĥj = Ĥ(t (j)), can use Greenwood Tsiatis Klein ) s.e. (Ĥj = j ) s.e. (Ĥj = ) s.e. (Ĥj = i=0 j d i n i (n i d i ) j i=0 d i n 2 i=0 i d i (n i d i ) n 3 i

65 Session 6: 63/ 85 The Cox Regression Model The Cox (or Proportional Hazards) model provides a simple way of introducing exogenous variables into the survival model. The basic components are a baseline hazard function, h 0 and a linear predictor and (positive) link function g. Then for observed predictor values X 1 = x 1, X 2 = x 2,..., X K = x K, the hazard function takes the form h Y (y; x) = g(x T β)h 0 (y) that is, the hazard is modified in a multiplicative fashion by the linked-linear predictor. Typically, g is the exponential function.

66 Session 6: 64/ 85 From the previously established relationships, { y } { S Y (y; x) = exp h Y (t) dt = exp 0 y 0 } g(x T β)h 0 (y) dt If a coefficient β k is positive, the hazard is increased, and the expected failure time decreased. The significance of a particular predictor is based on the magnitude of β t = ) s.e. ( β If t > 2, then the hypothesis that β = 0 can be rejected.

67 Session 6: 65/ 85 Discrete Time Proportional Modelling: In the discrete time case, proportional hazards modelling needs to be modified to respect constraints on the hazards to be probabilities. Recall that h Y (y) = P [Failure at y Survival y] = Then, by construction, 0 h Y (y) 1. f Y (y) S Y (y 1) If this hazard is a baseline hazard, and wish to recognize modification of the hazard by exogenous variables, these constraints have to be respected.

68 Session 6: 66/ 85 Specifically, h Y (y; x) = g(x T β)h 0 (y) 1 may not be guaranteed after the multiplicative modification by g. One construction that guarantees the constraints are met is to model on the transformed scale, that is ( ) ( ) hy (y; x) = g(x T h0 (y) β). 1 h Y (y; x) 1 h 0 (y) This is the proportional odds model.

69 Session 6: 67/ 85 But, for general hazard probabilities h, h(y) f (y)/s(y 1) = 1 h(y) 1 f (y)/s(y 1) = f (y) S(y 1) f (y) = f (y) S(y) so proportional odds modelling has a sensible interpretation.

70 Session 6: 68/ 85 The Accelerated Life Model The Accelerated Life model provides another way of introducing the influence of predictors into the survival model. The basic components now are a baseline reliability function, S 0 and a linear predictor and (positive) link function g.. Then for an experimental unit with observed predictor values X 1 = x 1, X 2 = x 2,..., X K = x K, the reliability function takes the form S Y (y; x) = S 0 (g(x T β)y) that is, the time scale is modified in a multiplicative fashion by the linked-linear predictor; this allows direct modelling of the influence of predictors on survival.

71 Session 6: 69/ 85 Frailty Modelling The idea of frailty modelling is to introduce random effects terms into the linear predictor that appears in the proportional hazards and accelerated life models. For example, we extend x T i β = β 0 + β 1 x i1 + β 1 x id β D x id to include a random component that is specific to the individual observational unit (bond, company etc.) concerned, that is, we have β = β 0 + β 1 x i1 + β 1 x i β D x id + L i x T i where L i is some (usually zero mean) random variable.

72 Session 6: 70/ 85 The Log-Rank Test : The log-rank test is a standard test for significant differences between two (or more) reliability functions that differ because of the influence of the different levels of a discrete predictor. H 0 : S 1 = S 2 H 1 : S 1 S 2 It is a non-parametric test based on ranks of samples for the two or more subgroups. Asymptotic or exact versions of the test can be carried out.

73 Session 6: 71/ 85 It is possible to fit and compare parametric survival models to such data. Parametric densities, reliability functions, hazards etc. can be readily used in the formation of a likelihood, potentially within the proportional hazards/accelerated life framework. Typical models used are Weibull Gamma Log-Logistic Log-Normal Pareto

74 Session 6: 72/ 85 Weibull : for y > 0, f (y) = α λ α y α 1 exp { ( y α } λ) { ( y α } F (y) = 1 exp λ) { ( y α } = S(y) = exp λ) h(y) = α λ α y α 1 H(y) = ( y λ) α for parameters α, λ > 0 (the shape and scale parameters respectively).

75 Session 6: 73/ 85 The Weibull pdf for different Alpha (Lambda=5) f(y) Alpha=1 Alpha=2 Alpha=3 Alpha=4 Alpha= y

76 Session 6: 74/ 85 The Weibull survivor function for different Alpha (Lambda=5) S(y) Alpha=1 Alpha=2 Alpha=3 Alpha=4 Alpha= y

77 Session 6: 75/ 85 The Weibull hazard function for different Alpha S(y) Alpha=1 Alpha=2 Alpha=3 Alpha=4 Alpha= y

78 Session 6: 76/ 85 The Weibull log-hazard S(y) Alpha=1 Alpha=2 Alpha=3 Alpha=4 Alpha= y

79 Session 6: 77/ 85 Multivariate Models & Competing Risks An important generalization extends event time Y to be a vector quantity, that is, we have say K different aspects of failure, with random variable Y = (Y 1,...Y K ) requiring a joint probability model. A model that captures joint structure of time-to-event observations must respect marginal and conditional coherence requirements, and if possible, capture a full range of dependency structures. Typically, such models are difficult to construct.

80 Session 6: 78/ 85 Some joint models exist Multivariate Exponential Multivariate Weibull Multivariate Lognormal Hougaard etc.

81 Session 6: 79/ 85 Some joint models can be constructed Copula (latent uniform) Latent Multivariate Normal Typically, inference for these models is quite complex, but can be achieved using numerical procedures.

82 Session 6: 80/ 85 Some models use the idea of a frailty or factor-type structure. Y 1,..., Y K positive random variables, potential censoring, corresponding to different items (defaultable bonds etc.) modelling achieved through hazard rates h 1,..., h K set h i (t) = h(t) + η i (t) i = 1,..., K. h(t) is some market-level factor determining global hazard rates, induces dependence across the individual items η i (t) is some item specific hazard.

83 Session 6: 81/ 85 Another common experimental situation is one of competing risks; that is, there are K potential causes of failure, but at most one is observed for each individual in the study. Then the failure time, T, is defined by T = min {Y 1,..., Y K }

84 Session 6: 82/ 85 If the cause of failure, C, is recorded as C = k, we observe Y 1 > t,..., Y k 1 > t, Y k = t, Y k+1 > t,..., Y K > t whereas if the observation is censored, we observe Y 1 > t,..., Y k 1 > t, Y k > t, Y k+1 > t,..., Y K > t A joint model is again often difficult to construct, and in addition there are issues to do with identifiability of the marginal failure processes for the components of Y. i.e. without sufficient data, there are problems in estimating the models for Y 1,..., Y K considered on their own

85 Session 6: 83/ 85 Multi-State Modelling In multi-state modelling, rather than just having the standard failed/not failed (dead/alive) dichotomy, with { 0 Censored Z = 1 Failure is observed we have an extension to polytomy, where 0 Censored at time t 1 State 1 at time t Z(t) =.. M State M at time t

86 Session 6: 84/ 85 In such a model, we attempt to estimate the probability π ij (t i, t j ) = P [State j at time t j State i at time t i ] for t i < t j, or rate, λ ij, of transition from one state to another. In a discrete time framework, homogeneous Markov Models are typically used, characterized by a transition matrix P, with (i, j) th entry π ij, independent of t, with M π ij = 1 j=0

87 Session 6: 85/ 85 A multi-state process is a random process {Z(t)} t 0 describing the state within which the individual lies at time t. This kind of modelling is very useful for modelling progression to credit default; different states could correspond to different credit ratings.

Statistics 262: Intermediate Biostatistics Non-parametric Survival Analysis

Statistics 262: Intermediate Biostatistics Non-parametric Survival Analysis Statistics 262: Intermediate Biostatistics Non-parametric Survival Analysis Jonathan Taylor & Kristin Cobb Statistics 262: Intermediate Biostatistics p.1/?? Overview of today s class Kaplan-Meier Curve

More information

Multistate Modeling and Applications

Multistate Modeling and Applications Multistate Modeling and Applications Yang Yang Department of Statistics University of Michigan, Ann Arbor IBM Research Graduate Student Workshop: Statistics for a Smarter Planet Yang Yang (UM, Ann Arbor)

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

Lecture 5 Models and methods for recurrent event data

Lecture 5 Models and methods for recurrent event data Lecture 5 Models and methods for recurrent event data Recurrent and multiple events are commonly encountered in longitudinal studies. In this chapter we consider ordered recurrent and multiple events.

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Estimation for Modified Data

Estimation for Modified Data Definition. Estimation for Modified Data 1. Empirical distribution for complete individual data (section 11.) An observation X is truncated from below ( left truncated) at d if when it is at or below d

More information

Introduction to Smoothing spline ANOVA models (metamodelling)

Introduction to Smoothing spline ANOVA models (metamodelling) Introduction to Smoothing spline ANOVA models (metamodelling) M. Ratto DYNARE Summer School, Paris, June 215. Joint Research Centre www.jrc.ec.europa.eu Serving society Stimulating innovation Supporting

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

ECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t.

ECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t. ECO 513 Fall 2008 C.Sims KALMAN FILTER Model in the form 1. THE KALMAN FILTER Plant equation : s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. Var(ε t ) = Ω, Var(ν t ) = Ξ. ε t ν t and (ε t,

More information

Survival Analysis Math 434 Fall 2011

Survival Analysis Math 434 Fall 2011 Survival Analysis Math 434 Fall 2011 Part IV: Chap. 8,9.2,9.3,11: Semiparametric Proportional Hazards Regression Jimin Ding Math Dept. www.math.wustl.edu/ jmding/math434/fall09/index.html Basic Model Setup

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Markov Chain Monte Carlo Methods for Stochastic Optimization

Markov Chain Monte Carlo Methods for Stochastic Optimization Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,

More information

Particle Filtering Approaches for Dynamic Stochastic Optimization

Particle Filtering Approaches for Dynamic Stochastic Optimization Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,

More information

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Research Division Federal Reserve Bank of St. Louis Working Paper Series Research Division Federal Reserve Bank of St Louis Working Paper Series Kalman Filtering with Truncated Normal State Variables for Bayesian Estimation of Macroeconomic Models Michael Dueker Working Paper

More information

Markov Chain Monte Carlo Methods for Stochastic

Markov Chain Monte Carlo Methods for Stochastic Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Sequential Bayesian Updating

Sequential Bayesian Updating BS2 Statistical Inference, Lectures 14 and 15, Hilary Term 2009 May 28, 2009 We consider data arriving sequentially X 1,..., X n,... and wish to update inference on an unknown parameter θ online. In a

More information

Lecture 6: Bayesian Inference in SDE Models

Lecture 6: Bayesian Inference in SDE Models Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:

More information

ST745: Survival Analysis: Nonparametric methods

ST745: Survival Analysis: Nonparametric methods ST745: Survival Analysis: Nonparametric methods Eric B. Laber Department of Statistics, North Carolina State University February 5, 2015 The KM estimator is used ubiquitously in medical studies to estimate

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Other Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model

Other Survival Models. (1) Non-PH models. We briefly discussed the non-proportional hazards (non-ph) model Other Survival Models (1) Non-PH models We briefly discussed the non-proportional hazards (non-ph) model λ(t Z) = λ 0 (t) exp{β(t) Z}, where β(t) can be estimated by: piecewise constants (recall how);

More information

Lecture 3. Truncation, length-bias and prevalence sampling

Lecture 3. Truncation, length-bias and prevalence sampling Lecture 3. Truncation, length-bias and prevalence sampling 3.1 Prevalent sampling Statistical techniques for truncated data have been integrated into survival analysis in last two decades. Truncation in

More information

Step-Stress Models and Associated Inference

Step-Stress Models and Associated Inference Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated

More information

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania Nonlinear and/or Non-normal Filtering Jesús Fernández-Villaverde University of Pennsylvania 1 Motivation Nonlinear and/or non-gaussian filtering, smoothing, and forecasting (NLGF) problems are pervasive

More information

Cox s proportional hazards model and Cox s partial likelihood

Cox s proportional hazards model and Cox s partial likelihood Cox s proportional hazards model and Cox s partial likelihood Rasmus Waagepetersen October 12, 2018 1 / 27 Non-parametric vs. parametric Suppose we want to estimate unknown function, e.g. survival function.

More information

Switching Regime Estimation

Switching Regime Estimation Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms

More information

Survival Analysis. Stat 526. April 13, 2018

Survival Analysis. Stat 526. April 13, 2018 Survival Analysis Stat 526 April 13, 2018 1 Functions of Survival Time Let T be the survival time for a subject Then P [T < 0] = 0 and T is a continuous random variable The Survival function is defined

More information

MAS3301 / MAS8311 Biostatistics Part II: Survival

MAS3301 / MAS8311 Biostatistics Part II: Survival MAS3301 / MAS8311 Biostatistics Part II: Survival M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2009-10 1 13 The Cox proportional hazards model 13.1 Introduction In the

More information

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA Kasun Rathnayake ; A/Prof Jun Ma Department of Statistics Faculty of Science and Engineering Macquarie University

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

Sequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007

Sequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007 Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember

More information

Towards a Bayesian model for Cyber Security

Towards a Bayesian model for Cyber Security Towards a Bayesian model for Cyber Security Mark Briers (mbriers@turing.ac.uk) Joint work with Henry Clausen and Prof. Niall Adams (Imperial College London) 27 September 2017 The Alan Turing Institute

More information

CTDL-Positive Stable Frailty Model

CTDL-Positive Stable Frailty Model CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Particle Filters. Outline

Particle Filters. Outline Particle Filters M. Sami Fadali Professor of EE University of Nevada Outline Monte Carlo integration. Particle filter. Importance sampling. Degeneracy Resampling Example. 1 2 Monte Carlo Integration Numerical

More information

Exercises. (a) Prove that m(t) =

Exercises. (a) Prove that m(t) = Exercises 1. Lack of memory. Verify that the exponential distribution has the lack of memory property, that is, if T is exponentially distributed with parameter λ > then so is T t given that T > t for

More information

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Lecture 8: Bayesian Estimation of Parameters in State Space Models in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space

More information

Basic math for biology

Basic math for biology Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood

More information

Sensor Tasking and Control

Sensor Tasking and Control Sensor Tasking and Control Sensing Networking Leonidas Guibas Stanford University Computation CS428 Sensor systems are about sensing, after all... System State Continuous and Discrete Variables The quantities

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53 State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State

More information

Generalized Autoregressive Score Models

Generalized Autoregressive Score Models Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be

More information

4 Testing Hypotheses. 4.1 Tests in the regression setting. 4.2 Non-parametric testing of survival between groups

4 Testing Hypotheses. 4.1 Tests in the regression setting. 4.2 Non-parametric testing of survival between groups 4 Testing Hypotheses The next lectures will look at tests, some in an actuarial setting, and in the last subsection we will also consider tests applied to graduation 4 Tests in the regression setting )

More information

Generalized Autoregressive Score Smoothers

Generalized Autoregressive Score Smoothers Generalized Autoregressive Score Smoothers Giuseppe Buccheri 1, Giacomo Bormetti 2, Fulvio Corsi 3,4, and Fabrizio Lillo 2 1 Scuola Normale Superiore, Italy 2 University of Bologna, Italy 3 University

More information

Gaussian Process Approximations of Stochastic Differential Equations

Gaussian Process Approximations of Stochastic Differential Equations Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run

More information

Approximate Bayesian Computation and Particle Filters

Approximate Bayesian Computation and Particle Filters Approximate Bayesian Computation and Particle Filters Dennis Prangle Reading University 5th February 2014 Introduction Talk is mostly a literature review A few comments on my own ongoing research See Jasra

More information

You know I m not goin diss you on the internet Cause my mama taught me better than that I m a survivor (What?) I m not goin give up (What?

You know I m not goin diss you on the internet Cause my mama taught me better than that I m a survivor (What?) I m not goin give up (What? You know I m not goin diss you on the internet Cause my mama taught me better than that I m a survivor (What?) I m not goin give up (What?) I m not goin stop (What?) I m goin work harder (What?) Sir David

More information

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω ECO 513 Spring 2015 TAKEHOME FINAL EXAM (1) Suppose the univariate stochastic process y is ARMA(2,2) of the following form: y t = 1.6974y t 1.9604y t 2 + ε t 1.6628ε t 1 +.9216ε t 2, (1) where ε is i.i.d.

More information

Bivariate Degradation Modeling Based on Gamma Process

Bivariate Degradation Modeling Based on Gamma Process Bivariate Degradation Modeling Based on Gamma Process Jinglun Zhou Zhengqiang Pan Member IAENG and Quan Sun Abstract Many highly reliable products have two or more performance characteristics (PCs). The

More information

Cox regression: Estimation

Cox regression: Estimation Cox regression: Estimation Patrick Breheny October 27 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/19 Introduction The Cox Partial Likelihood In our last lecture, we introduced the Cox partial

More information

Basic Sampling Methods

Basic Sampling Methods Basic Sampling Methods Sargur Srihari srihari@cedar.buffalo.edu 1 1. Motivation Topics Intractability in ML How sampling can help 2. Ancestral Sampling Using BNs 3. Transforming a Uniform Distribution

More information

[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements

[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements [Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements Aasthaa Bansal PhD Pharmaceutical Outcomes Research & Policy Program University of Washington 69 Biomarkers

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

CIMAT Taller de Modelos de Capture y Recaptura Known Fate Survival Analysis

CIMAT Taller de Modelos de Capture y Recaptura Known Fate Survival Analysis CIMAT Taller de Modelos de Capture y Recaptura 2010 Known Fate urvival Analysis B D BALANCE MODEL implest population model N = λ t+ 1 N t Deeper understanding of dynamics can be gained by identifying variation

More information

L09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms

L09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state

More information

STA 216, GLM, Lecture 16. October 29, 2007

STA 216, GLM, Lecture 16. October 29, 2007 STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Survival Analysis for Case-Cohort Studies

Survival Analysis for Case-Cohort Studies Survival Analysis for ase-ohort Studies Petr Klášterecký Dept. of Probability and Mathematical Statistics, Faculty of Mathematics and Physics, harles University, Prague, zech Republic e-mail: petr.klasterecky@matfyz.cz

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Analysing geoadditive regression data: a mixed model approach

Analysing geoadditive regression data: a mixed model approach Analysing geoadditive regression data: a mixed model approach Institut für Statistik, Ludwig-Maximilians-Universität München Joint work with Ludwig Fahrmeir & Stefan Lang 25.11.2005 Spatio-temporal regression

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Hidden Markov Models Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Additional References: David

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

STAT331. Cox s Proportional Hazards Model

STAT331. Cox s Proportional Hazards Model STAT331 Cox s Proportional Hazards Model In this unit we introduce Cox s proportional hazards (Cox s PH) model, give a heuristic development of the partial likelihood function, and discuss adaptations

More information

The Unscented Particle Filter

The Unscented Particle Filter The Unscented Particle Filter Rudolph van der Merwe (OGI) Nando de Freitas (UC Bereley) Arnaud Doucet (Cambridge University) Eric Wan (OGI) Outline Optimal Estimation & Filtering Optimal Recursive Bayesian

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations Department of Biomedical Engineering and Computational Science Aalto University April 28, 2010 Contents 1 Multiple Model

More information

Particle Filtering for Data-Driven Simulation and Optimization

Particle Filtering for Data-Driven Simulation and Optimization Particle Filtering for Data-Driven Simulation and Optimization John R. Birge The University of Chicago Booth School of Business Includes joint work with Nicholas Polson. JRBirge INFORMS Phoenix, October

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features Yangxin Huang Department of Epidemiology and Biostatistics, COPH, USF, Tampa, FL yhuang@health.usf.edu January

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Survival Regression Models

Survival Regression Models Survival Regression Models David M. Rocke May 18, 2017 David M. Rocke Survival Regression Models May 18, 2017 1 / 32 Background on the Proportional Hazards Model The exponential distribution has constant

More information

In contrast, parametric techniques (fitting exponential or Weibull, for example) are more focussed, can handle general covariates, but require

In contrast, parametric techniques (fitting exponential or Weibull, for example) are more focussed, can handle general covariates, but require Chapter 5 modelling Semi parametric We have considered parametric and nonparametric techniques for comparing survival distributions between different treatment groups. Nonparametric techniques, such as

More information

Modeling conditional distributions with mixture models: Theory and Inference

Modeling conditional distributions with mixture models: Theory and Inference Modeling conditional distributions with mixture models: Theory and Inference John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università di Venezia Italia June 2, 2005

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Lecture 4: Dynamic models

Lecture 4: Dynamic models linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu

More information

Inferring biological dynamics Iterated filtering (IF)

Inferring biological dynamics Iterated filtering (IF) Inferring biological dynamics 101 3. Iterated filtering (IF) IF originated in 2006 [6]. For plug-and-play likelihood-based inference on POMP models, there are not many alternatives. Directly estimating

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49 State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing

More information

TMA 4275 Lifetime Analysis June 2004 Solution

TMA 4275 Lifetime Analysis June 2004 Solution TMA 4275 Lifetime Analysis June 2004 Solution Problem 1 a) Observation of the outcome is censored, if the time of the outcome is not known exactly and only the last time when it was observed being intact,

More information

Introduction to BGPhazard

Introduction to BGPhazard Introduction to BGPhazard José A. García-Bueno and Luis E. Nieto-Barajas February 11, 2016 Abstract We present BGPhazard, an R package which computes hazard rates from a Bayesian nonparametric view. This

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

X t = a t + r t, (7.1)

X t = a t + r t, (7.1) Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Introduction to Reliability Theory (part 2)

Introduction to Reliability Theory (part 2) Introduction to Reliability Theory (part 2) Frank Coolen UTOPIAE Training School II, Durham University 3 July 2018 (UTOPIAE) Introduction to Reliability Theory 1 / 21 Outline Statistical issues Software

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Evaluating the value of structural heath monitoring with longitudinal performance indicators and hazard functions using Bayesian dynamic predictions

Evaluating the value of structural heath monitoring with longitudinal performance indicators and hazard functions using Bayesian dynamic predictions Evaluating the value of structural heath monitoring with longitudinal performance indicators and hazard functions using Bayesian dynamic predictions C. Xing, R. Caspeele, L. Taerwe Ghent University, Department

More information

The Kalman Filter ImPr Talk

The Kalman Filter ImPr Talk The Kalman Filter ImPr Talk Ged Ridgway Centre for Medical Image Computing November, 2006 Outline What is the Kalman Filter? State Space Models Kalman Filter Overview Bayesian Updating of Estimates Kalman

More information

Sensor Fusion: Particle Filter

Sensor Fusion: Particle Filter Sensor Fusion: Particle Filter By: Gordana Stojceska stojcesk@in.tum.de Outline Motivation Applications Fundamentals Tracking People Advantages and disadvantages Summary June 05 JASS '05, St.Petersburg,

More information

Content.

Content. Content Fundamentals of Bayesian Techniques (E. Sucar) Bayesian Filters (O. Aycard) Definition & interests Implementations Hidden Markov models Discrete Bayesian Filters or Markov localization Kalman filters

More information

Multi-state Models: An Overview

Multi-state Models: An Overview Multi-state Models: An Overview Andrew Titman Lancaster University 14 April 2016 Overview Introduction to multi-state modelling Examples of applications Continuously observed processes Intermittently observed

More information