A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy.
|
|
- Morgan Garrett
- 5 years ago
- Views:
Transcription
1 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1. A Student-t based sparsity enforcing hierarchical prior for linear inverse problems and its efficient Bayesian computation for 2D and 3D Computed Tomography Ali Mohammad-Djafari, Li Wang, Nicolas Gac and Folker Bleichrodt Laboratoire des Signaux et Systèmes (L2S) UMR8506 CNRS-CentraleSupélec-UNIV PARIS SUD SUPELEC, Gif-sur-Yvette, France djafari@lss.supelec.fr itwist2016, Aug , 2016, Aalborg, Denemark
2 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 Contents 1. Computed Tomography in 2D and 3D 2. Main classical methods 3. Basic Bayesian approach 4. Sparsity enforcing models through Student-t and IGSM 5. Computational tools: JMAP, EM, VBA 6. Implementation issues Main GPU implementation steps: Forward and Back Projections Multi-Resolution implementation 7. Some results 8. Conclusions
3 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 3 Computed Tomography: Seeing inside of a body f (x, y) a section of a real 3D body f (x, y, z) g φ (r) a line of observed radiography g φ (r, z) Forward model: Line integrals or Radon Transform g φ (r) = f (x, y) dl + ɛ φ (r) L r,φ = f (x, y) δ(r x cos φ y sin φ) dx dy + ɛ φ (r) Inverse problem: Image reconstruction Given the forward model H (Radon Transform) and a set of data g φi (r), i = 1,, M find f (x, y)
4 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 4 2D and 3D Computed Tomography 3D 2D g φ (r 1, r 2 ) = f (x, y, z) dl L r1,r 2,φ g φ (r) = f (x, y) dl L r,φ Forward probelm: f (x, y) or f (x, y, z) g φ (r) or g φ (r 1, r 2 ) Inverse problem: g φ (r) or g φ (r 1, r 2 ) f (x, y) or f (x, y, z)
5 Algebraic methods: Discretization y S r H ij f (x, y) f 1 f j g i φ f N x D f (x, y) = j f j b j (x, y) { 1 if (x, y) pixel j b j (x, y) = g(r, φ) 0 else N g(r, φ) = f (x, y) dl g i = H ij f j + ɛ i L j=1 g k = H k f + ɛ k, k = 1,, K g = Hf + ɛ g k projection at angle φ k, g all the projections. A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 5
6 Algebraic methods g 1 g =. g K, H = H 1. H K g k = H k f+ɛ k g = k H k f+ɛ k = Hf+ɛ H is huge dimensional: 2D: , 3D: Hf corresponds to forward projection H t g corresponds to Back projection (BP) H may not be invertible and even not square H is, in general, ill-conditioned In limited angle tomography H is under determined, si the problem has infinite number of solutions Minimum Norm Solution f = H t (HH t ) 1 g = k H t k (H kh t k ) 1 g k can be interpreted as the Filtered Back Projection solution. A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 6
7 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 7 Prior information or constraints Positivity: f j > 0 or f j IR + Boundedness: 1 > f j ) > 0 or f j [0, 1] Smoothness: Sparsity: many f j are zeros. f j depends on the neighborhoods. Sparsity in a transform domain: f = Dz and many z j are zeros. Discrete valued (DV): f j {0, 1,..., K} Binary valued (BV): f j {0, 1} Compactness: f (r) is non zero in one or few non-overlapping compact regions Combination of the above mentioned constraints Main mathematical questions: Which combination results to unique solution? How to apply them?
8 Algebraic methods: Regularization Minimum Norm Solution: minimize f 2 2 s.t. Hf = g f = H t (HH t ) 1 g Least square Solution: f = arg minf { J(f) = g Hf 2 } f = (H t H) 1 H t g Quadratic Regularization: J(f) = g Hf 2 + λ f 2 2 f = (H t H + λi) 1 H t g L1 Regularization: J(f) = g Hf 2 + λ f 1 Lpq Regularization: J(f) = g Hf p p + λ Df q q More general Regularization: J(f) = i φ(g i [Hf] i ) + λ j ψ(df] j ) with convex potential functions φ and Ψ or J(f) = 1 (g, Hf) + λ 2 (f, f 0 ) with 1 and 2 any distances (L2, L2,..) or divergence (KL) A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 8
9 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 9 Deterministic approaches Iterative methods: SIRT, ART, Quadratic or L1 regularization, Bloc Coordinate Descent, Multiplicative ART,... Criteria: J(f) = g Hf 2 + λ Df 2 Gradient based algorithms: J(f) = 2H t (g Hf) + 2λD t Df Simplest algorithm: f(k+1) = f(k) + α (k) [ H t (g H f (k) ) + 2λD t D f (k)] More criteria: J(f) = i φ(g i [Hf] i ) + λ j ψ((df] j) with φ(t) and ψ(t) = {t 2, t, t p,...} or non convex ones. Imposing constraints in each iteration (example: DART) Mathematical studies of uniqueness and convergence of these algorithms are necessary Many specialized algorithms (ISTA, FISTA, ADMM, AMP, GAMP,...) are developped for L1 regularization.
10 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Bayesian estimation approach M : g = Hf + ɛ Observation model M + Hypothesis on the noise ɛ p(g f; M) = pɛ(g Hf) A priori information p(f M) p(g f; M) p(f M) Bayes : p(f g; M) = p(g M) Gaussian priors: Prior knowledge on the noise: ɛ N (0, 2 vɛ I) Prior knowledge on f: f N (0, v 2 f (D D) 1 ) ] A posteriori: p(f g) exp [ 1 2v 2 ɛ g Hf 2 1 Df 2 2vf 2 MAP : f = arg maxf {p(f g)} = arg min f {J(f)} with J(f) = g Hf 2 + λ Df 2, λ = vɛ2 vf 2 Advantage : characterization of the solution p(f g) = N ( f, Σ) with f = ( H H + λd D ) 1 H g, Σ = vɛ ( H H + λd D ) 1
11 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Sparsity enforcing models 3 classes of models: 1- Generalized Gaussian (Double Exp. as particular case) 2- Mixture models (2 Gaussians mixture, BG,..) 3- Heavy tailed (Student-t and Cauchy) Student-t model [ St(f ν) exp ν + 1 log ( 1 + f 2 /ν )] 2 Infinite Gausian Scaled Mixture (IGSM) equivalence St(f ν) N (f, 0, 1/z) G(z α, β) dz, with α = β = ν/2 0 p(f z) = j p(f j z j ) = [ ] j N (f j 0, 1/z j ) exp 1 2 j z jfj 2 p(z α, β) = j G(z j α, β) [ j z(α 1) j exp [ βz j ] ] exp j (α 1) ln z j βz j [ ] p(f, z α, β) exp 1 2 j z jfj 2 + (α 1) ln z j βz j
12 Sparse model in a Transform domain f piecewise continuous gradient of f Sparse z =Haar Transform of f Sparse I Analysis: J(f) = kg Hfk22 + λkdfk1 I Synthesis: f = Dz J(z) = kg HDzk2 + λkzk1 I Explicit modelling f = Dz + ξ, z sparse, ξ Gaussian with unknown variance A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1
13 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Sparse model in a Transform domain g = Hf + ɛ, f = Dz + ξ, z sparse p(g f, v ɛ ) = N (g Hf, V ɛ ) V ɛ = diag [v ɛ ] α ξ0, β ξ0 α z0, β p(f z) = N (f Dz, V z0 ξ I) V ξ = diag [v ξ ] p(z v z ) = N (z 0, V z ), V z = diag [v z ] v ξ v z p(v ɛ ) = i α ɛ0, β IG(v ɛ i α ɛ0, β ɛ0 ) ɛ0 p(v z ) = j IG(v z j α z0, β z0 ) p(v ξ z v ξ ) = j IG(v ξ j α ξ0, β ξ0 ) ɛ p(f, z, v ɛ, v z, v ξ g) p(g f, v ɛ ) p(f z f ) p(z v z ) D p(v ɛ ) p(v z ) p(v ξ ) f ɛ JMAP: H ( f, ẑ, ˆv ɛ, v z, v ξ ) = arg max {p(f, z, v ɛ, v z, v ξ g)} (f,z,v ɛ,v z,v ξ ) g Alternate optimization. VBA: Approximate p(f, z, v ɛ, v z, v ξ g) by q 1 (f) q 2 (z) q 3 (v ɛ ) q 4 (v z ) q 5 (v ξ ) Alternate optimization.
14 JMAP Algorithm [ f, ẑ, v z, v ɛ, v ξ ] = arg max (f,z,v z,v ɛ,v ξ ) {p (f, z, v z, v ɛ, v ξ g)} iter : f (k+1) = f (k) γ (k) f J ( f (k) ) iter : ẑ (k+1) = ẑ (k) γ z (k) J (ẑ (k) ) v zj where = βz ẑ2 j α z0 +3/2, ˆv ɛ i = βɛ 0 + J (f) = 1 2 (g Hf) V 1 ɛ J (z) = 1 2 (f Dz) V 1 γ (k) f = γ (k) z = 1 2 ( g i [ H f ] i ) 2 ξ J ( f (k) 2 ) ŶɛH J ( f (k) 2 ) + Ŷξ J ( f (k) ) J (ẑ (k) ) 2 ŶξD J (ẑ (k) ) 2 + Ŷz J (ẑ(k) ) 2 and J ( ) is the gradient of J ( ). α ɛ0 +3/2, v ξj = β ξ (g Hf) (f Dz) V 1 ξ (f Dz) z V 1 z z ( f j [Dẑ] j ) 2 α ξ0 +3/2 (f Dz) 2 where Ŷɛ = V 1 2 ɛ and Ŷξ = V 1 2 where Ŷz = V 1 2 z A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 ξ
15 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Implementation issues In almost all the algorithms, the step of computation of f needs an optimization algorithm. The criterion to optimize is often in the form of J(f) = g Hf λ Df 2 2 or J(z) = g HDz 2 + λ z 1 Very often, we use the gradient based algorithms which need to compute J(f) = 2H t (g Hf) + 2λD t Df and so, the simplest case, in each step, we have f(k+1) = f(k) + α (k) [H t (g H f (k) ) + 2λD t D f (k)] 1. Compute ĝ = H f (Forward projection) 2. Compute δg = g ĝ (Error or residual) 3. Compute δf 1 = H δg (Backprojection of error) 4. Compute δf 2 = D D f and update f (k+1) = f (k) + [δf 1 + δf 2 ] Steps 1 and 3 need great computational cost and have been implemented on GPU. In this work, we used ASTRA Toolbox.
16 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Multi-Resolution Implementation Sacle 1: black g (1) = H (1) f (1) ( N N ) Sacle 2: green g (2) = H (2) f (2) (N/2 N/2) Sacle 3: red g (3) = H (3) f (3) (N/4 N/4)
17 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Results Phantom: 128 x 128 x 128 Projections: 128, 64 and 32 SNR=40 db 128 projections 64 projections 32 projections
18 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Results Phantom: 128 x 128 x 128 Projections: 128 SNR=20 db QR TV HHBM
19 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 1 Results with High SNR=30dB δf = f f f δg = H f g g
20 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 Results: High SNR=30dB and Low SNR=20dB δf = f f f High SNR=30dB δf = f f f Low SNR=20dB
21 Variational Bayesian Approximation (VBA) Depending on cases, we have to handle p(f, θ g), p(f, z, θ g), p(f, w, θ g) or p(f, w, z, θ g). Let consider the simplest case: Approximate p(f, θ g) by q(f, θ g) = q 1 (f g) q 2 (θ g) and then continue computations. Criterion KL(q(f, θ g) : p(f, θ g)) KL(q : p) = q ln q/p = q 1 q 2 ln q 1q 2 p = q 1 ln q 1 + q2 ln q 2 q ln p = H(q 1 ) H(q 2 ) < ln p > q Iterative algorithm q 1 q 2 q 1 q 2, q 1 (f) q 2 (θ) p(f, θ g) [ ] exp ln p(g, f, θ; M) [ q2 (θ)] exp ln p(g, f, θ; M) q1 (f) Variational Bayesian Approximation q 1 (f) f q 2 (θ) θ A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2
22 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 JMAP, Marginalization, Sampling and exploration, VBA JMAP: p(f, θ g) optimization Marginalization p(f, θ g) Joint Posterior f θ Alternate Optimization p(θ g) Marginalize over f { f = arg maxf {p(f θ, g)} θ = arg max θ {p(θ fg)} θ p(f θ, g) f Sampling and Exploration Gibbs sampling: f p(f θ, g) θ p(θ fg) Other sampling methods: IS, MH, Slice sampling,... Variational Bayesian Approximation p(f, θ g) Variational Bayesian Approximation q 1 (f) f q 2 (θ) θ
23 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 VBA: Choice of family of laws q 1 and q 2 Case 1 : Joint MAP { q 1 (f f) = δ(f f) f = arg maxf {p(f, θ g; } M) } (1) q 2 (θ θ) = δ(θ θ) θ= arg max θ {p( f, θ g; M) Case 2 : EM { q1 (f) p(f θ, g) Q(θ, θ)= ln p(f, θ g; M) q1 (f θ) q 2 (θ θ) = δ(θ θ) θ = arg max θ {Q(θ, θ) } (2) { Appropriate choice for inverse problems { q 1 (f) p(f θ, g; M) Accounts for the uncertainties of q 2 (θ) p(θ f, g; M) θ for f and vise versa. (3) Exponential families, Conjugate priors
24 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 JMAP, EM and VBA JMAP Alternate optimization Algorithm: θ (0) θ f = arg max f {p(f, θ g) } θ θ θ } = arg max θ {p( f, θ g) f f f EM: θ (0) θ q 1 (f) = p(f θ, g) q 1 (f) f θ θ Q(θ, θ) = ln p(f, θ g) q1(f) θ = arg max θ {Q(θ, θ) } q 1 (f) VBA: θ (0) q 2 (θ) q 1 (f) exp θ q 2 (θ) q 2 (θ) exp [ ] ln p(f, θ g) q2(θ) [ ] ln p(f, θ g) q1(f) q 1 (f) f q 1 (f)
25 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 VBA Approximate p(f, z, v z, v ξ, v ɛ g) by q 1 (f) q 2 (z) q 3 (v z ) q 4 (v ξ ) q 5 (v ɛ ) Alternate optimization: q 1 (f) = N (f µ f, Σ f ) q 2 (z) = N (z µ z, Σ z ) q 3 (v z ) = j IG(v z j α zj, β zj ) q 4 (v ξ ) = j IG(v ξ i α ξj, β ξj ) q 5 (v ɛ ) = i IG(v ɛ i α ɛi, β ɛi ) ( ) 1 µ f = H V 1 ɛ H + v 1 ξ H V 1 ɛ (g Dẑ) ( ) 1 Σ f = H V 1 ɛ H + v 1 ξ ( ) 1 µ z = D V 1 ξ D + v 1 z D V 1 f ξ ( ) 1 Σ z = D V 1 ξ D + v 1 z
26 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 JMAP Algorithm [ f, ẑ, v z, v ɛ, v ξ ] = arg max (f,z,v z,v ɛ,v ξ ) {p (f, z, v z, v ɛ, v ξ g)} iter : f (k+1) = f (k) γ (k) f J ( f (k) ) iter : ẑ (k+1) = ẑ (k) γ z (k) J (ẑ (k) ) v zj = βz ẑ2 j α z0 +3/2, ˆv ɛ i = βɛ ( g i [ H f ] i ) 2 α ɛ0 +3/2, v ξj = β ξ ( f j [Dẑ] j ) 2 α ξ0 +3/2 The expressions of v zj, ˆv ɛi and v ξj are more complex and needs the computation of the diagonal elements of the posterior covariances. Specific techniques are needed to compute them efficiently. ( [ ] ) 2 ) 2 v zj = βz <ẑ2 j > β α z0 +3/2, ˆv ɛ0 + 1 g 2 i H f β ξ0 + 1 i 2 ( f j [Dẑ] j ɛ i = α ɛ0 +3/2, v ξj = α ξ0 +3/2
27 A. Mohammad-Djafari, Bayesian Discrete Tomography from a few number of projections, Mars 21-23, 2016, Polytechnico de Milan, Italy. 2 Conclusions Computed Tomography is an ill-posed Inverse problem Algebraic methods and Regularization methods push further the limitations. Bayesian approach has many potentials for real applications where we need to estimate the hyper parameters (semi-supervised) and to quantify the uncertainties. Hierarchical prior model with hidden variables are very powerful tools for Bayesian approach to inverse problems. Main Bayesian computation tools: JMAP, VBA, AMP and MCMC Application in different other imaging systems: Microwaves, PET, Ultrasound, Optical Diffusion Tomography (ODT),.. Current Projects: Efficient implementation in 3D cases to reconstruct 1024 x 1024 x 1024 volumes using GPU.
Hierarchical prior models for scalable Bayesian Tomography
A. Mohammad-Djafari, Hierarchical prior models for scalable Bayesian Tomography, WSTL, January 25-27, 2016, Szeged, Hungary. 1/45. Hierarchical prior models for scalable Bayesian Tomography Ali Mohammad-Djafari
More informationBayesian X-ray Computed Tomography using a Three-level Hierarchical Prior Model
L. Wang, A. Mohammad-Djafari, N. Gac, MaxEnt 16, Ghent, Belgium. 1/26 Bayesian X-ray Computed Tomography using a Three-level Hierarchical Prior Model Li Wang, Ali Mohammad-Djafari, Nicolas Gac Laboratoire
More informationApproximate Bayesian Computation for Big Data
A. Mohammad-Djafari, Approximate Bayesian Computation for Big Data, Tutorial at MaxEnt 2016, July 10-15, Gent, Belgium. 1/63. Approximate Bayesian Computation for Big Data Ali Mohammad-Djafari Laboratoire
More informationBayesian inference for Machine Learning, Inverse Problems and Big Data: from Basics to Computational algorithms
A. Mohammad-Djafari, Approximate Bayesian Computation for Big Data, Seminar 2016, Nov. 21-25, Aix-Marseille Univ. 1/76. Bayesian inference for Machine Learning, Inverse Problems and Big Data: from Basics
More informationInverse Problems: From Regularization to Bayesian Inference
1 / 42 Inverse Problems: From Regularization to Bayesian Inference An Overview on Prior Modeling and Bayesian Computation Application to Computed Tomography Ali Mohammad-Djafari Groupe Problèmes Inverses
More informationBayesian inference methods for sources separation
A. Mohammad-Djafari, BeBec2012, February 22-23, 2012, Berlin, Germany, 1/12. Bayesian inference methods for sources separation Ali Mohammad-Djafari Laboratoire des Signaux et Systèmes, UMR8506 CNRS-SUPELEC-UNIV
More informationBayesian Microwave Breast Imaging
Bayesian Microwave Breast Imaging Leila Gharsalli 1 Hacheme Ayasso 2 Bernard Duchêne 1 Ali Mohammad-Djafari 1 1 Laboratoire des Signaux et Systèmes (L2S) (UMR8506: CNRS-SUPELEC-Univ Paris-Sud) Plateau
More informationBayesian Approach for Inverse Problem in Microwave Tomography
Bayesian Approach for Inverse Problem in Microwave Tomography Hacheme Ayasso Ali Mohammad-Djafari Bernard Duchêne Laboratoire des Signaux et Systèmes, UMRS 08506 (CNRS-SUPELEC-UNIV PARIS SUD 11), 3 rue
More informationEntropy, Information Theory, Information Geometry and Bayesian Inference in Data, Signal and Image Processing and Inverse Problems
Entropy 2015, 17, 3989-4027; doi:10.3390/e17063989 OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Article Entropy, Information Theory, Information Geometry and Bayesian Inference in Data,
More informationMAXIMUM ENTROPIES COPULAS
MAXIMUM ENTROPIES COPULAS Doriano-Boris Pougaza & Ali Mohammad-Djafari Groupe Problèmes Inverses Laboratoire des Signaux et Systèmes (UMR 8506 CNRS - SUPELEC - UNIV PARIS SUD) Supélec, Plateau de Moulon,
More informationInverse Problems: From Regularization Theory to Probabilistic and Bayesian Inference
A. Mohammad-Djafari, Inverse Problems: From Regularization to Bayesian Inference, Pôle signaux, 1/11. Inverse Problems: From Regularization Theory to Probabilistic and Bayesian Inference Ali Mohammad-Djafari
More informationVariational inference
Simon Leglaive Télécom ParisTech, CNRS LTCI, Université Paris Saclay November 18, 2016, Télécom ParisTech, Paris, France. Outline Introduction Probabilistic model Problem Log-likelihood decomposition EM
More informationInverse problems in imaging and computer vision
1 / 62. Inverse problems in imaging and computer vision From deterministic methods to probabilistic Bayesian inference Ali Mohammad-Djafari Groupe Problèmes Inverses Laboratoire des Signaux et Systèmes
More informationVARIATIONAL BAYES WITH GAUSS-MARKOV-POTTS PRIOR MODELS FOR JOINT IMAGE RESTORATION AND SEGMENTATION
VARIATIONAL BAYES WITH GAUSS-MARKOV-POTTS PRIOR MODELS FOR JOINT IMAGE RESTORATION AND SEGMENTATION Hacheme AYASSO and Ali MOHAMMAD-DJAFARI Laboratoire des Signaux et Systèmes, UMR 8506 (CNRS-SUPELEC-UPS)
More informationProbabilistic & Unsupervised Learning
Probabilistic & Unsupervised Learning Week 2: Latent Variable Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationInverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology
Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 27 Introduction Fredholm first kind integral equation of convolution type in one space dimension: g(x) = 1 k(x x )f(x
More informationVariable selection for model-based clustering
Variable selection for model-based clustering Matthieu Marbac (Ensai - Crest) Joint works with: M. Sedki (Univ. Paris-sud) and V. Vandewalle (Univ. Lille 2) The problem Objective: Estimation of a partition
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul
More informationThe Bayesian approach to inverse problems
The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu
More informationSparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference
Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationCovariance Matrix Simplification For Efficient Uncertainty Management
PASEO MaxEnt 2007 Covariance Matrix Simplification For Efficient Uncertainty Management André Jalobeanu, Jorge A. Gutiérrez PASEO Research Group LSIIT (CNRS/ Univ. Strasbourg) - Illkirch, France *part
More informationGraphical Models for Collaborative Filtering
Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More information2 Regularized Image Reconstruction for Compressive Imaging and Beyond
EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationHierarchical Sparse Bayesian Learning. Pierre Garrigues UC Berkeley
Hierarchical Sparse Bayesian Learning Pierre Garrigues UC Berkeley Outline Motivation Description of the model Inference Learning Results Motivation Assumption: sensory systems are adapted to the statistical
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationEE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)
EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in
More informationComputer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression
Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Expectation Maximization Mark Schmidt University of British Columbia Winter 2018 Last Time: Learning with MAR Values We discussed learning with missing at random values in data:
More informationInverse problem and optimization
Inverse problem and optimization Laurent Condat, Nelly Pustelnik CNRS, Gipsa-lab CNRS, Laboratoire de Physique de l ENS de Lyon Decembre, 15th 2016 Inverse problem and optimization 2/36 Plan 1. Examples
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ and Center for Automated Learning and
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationComputer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression
Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationAlgorithms for Variational Learning of Mixture of Gaussians
Algorithms for Variational Learning of Mixture of Gaussians Instructors: Tapani Raiko and Antti Honkela Bayes Group Adaptive Informatics Research Center 28.08.2008 Variational Bayesian Inference Mixture
More informationVariational Methods in Bayesian Deconvolution
PHYSTAT, SLAC, Stanford, California, September 8-, Variational Methods in Bayesian Deconvolution K. Zarb Adami Cavendish Laboratory, University of Cambridge, UK This paper gives an introduction to the
More information2D X-Ray Tomographic Reconstruction From Few Projections
2D X-Ray Tomographic Reconstruction From Few Projections Application of Compressed Sensing Theory CEA-LID, Thalès, UJF 6 octobre 2009 Introduction Plan 1 Introduction 2 Overview of Compressed Sensing Theory
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationAlgorithmisches Lernen/Machine Learning
Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationLecture 1b: Linear Models for Regression
Lecture 1b: Linear Models for Regression Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London c.archambeau@cs.ucl.ac.uk Advanced
More information(4D) Variational Models Preserving Sharp Edges. Martin Burger. Institute for Computational and Applied Mathematics
(4D) Variational Models Preserving Sharp Edges Institute for Computational and Applied Mathematics Intensity (cnt) Mathematical Imaging Workgroup @WWU 2 0.65 0.60 DNA Akrosom Flagellum Glass 0.55 0.50
More informationBayesian Paradigm. Maximum A Posteriori Estimation
Bayesian Paradigm Maximum A Posteriori Estimation Simple acquisition model noise + degradation Constraint minimization or Equivalent formulation Constraint minimization Lagrangian (unconstraint minimization)
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationWavelet Based Image Restoration Using Cross-Band Operators
1 Wavelet Based Image Restoration Using Cross-Band Operators Erez Cohen Electrical Engineering Department Technion - Israel Institute of Technology Supervised by Prof. Israel Cohen 2 Layout Introduction
More informationSUPREME: a SUPer REsolution Mapmaker for Extended emission
SUPREME: a SUPer REsolution Mapmaker for Extended emission Hacheme Ayasso 1,2 Alain Abergel 2 Thomas Rodet 3 François Orieux 2,3,4 Jean-François Giovannelli 5 Karin Dassas 2 Boualam Hasnoun 1 1 GIPSA-LAB
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate
More informationRigorous Dynamics and Consistent Estimation in Arbitrarily Conditioned Linear Systems
1 Rigorous Dynamics and Consistent Estimation in Arbitrarily Conditioned Linear Systems Alyson K. Fletcher, Mojtaba Sahraee-Ardakan, Philip Schniter, and Sundeep Rangan Abstract arxiv:1706.06054v1 cs.it
More informationExpectation propagation for symbol detection in large-scale MIMO communications
Expectation propagation for symbol detection in large-scale MIMO communications Pablo M. Olmos olmos@tsc.uc3m.es Joint work with Javier Céspedes (UC3M) Matilde Sánchez-Fernández (UC3M) and Fernando Pérez-Cruz
More informationData assimilation with and without a model
Data assimilation with and without a model Tyrus Berry George Mason University NJIT Feb. 28, 2017 Postdoc supported by NSF This work is in collaboration with: Tim Sauer, GMU Franz Hamilton, Postdoc, NCSU
More informationLeast Absolute Shrinkage is Equivalent to Quadratic Penalization
Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr
More informationRecovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm
Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm J. K. Pant, W.-S. Lu, and A. Antoniou University of Victoria August 25, 2011 Compressive Sensing 1 University
More informationA hierarchical Krylov-Bayes iterative inverse solver for MEG with physiological preconditioning
A hierarchical Krylov-Bayes iterative inverse solver for MEG with physiological preconditioning D Calvetti 1 A Pascarella 3 F Pitolli 2 E Somersalo 1 B Vantaggi 2 1 Case Western Reserve University Department
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationCheng Soon Ong & Christian Walder. Canberra February June 2017
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2017 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 679 Part XIX
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationMark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.
University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationPattern Recognition and Machine Learning. Bishop Chapter 9: Mixture Models and EM
Pattern Recognition and Machine Learning Chapter 9: Mixture Models and EM Thomas Mensink Jakob Verbeek October 11, 27 Le Menu 9.1 K-means clustering Getting the idea with a simple example 9.2 Mixtures
More informationDeep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationCOMPRESSIVE OPTICAL DEFLECTOMETRIC TOMOGRAPHY
COMPRESSIVE OPTICAL DEFLECTOMETRIC TOMOGRAPHY Adriana González ICTEAM/UCL March 26th, 214 1 ISPGroup - ICTEAM - UCL http://sites.uclouvain.be/ispgroup Université catholique de Louvain, Louvain-la-Neuve,
More informationOutline Lecture 2 2(32)
Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationBayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models
Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationLinear Inverse Problems
Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?
More informationRecent Advances in Bayesian Inference for Inverse Problems
Recent Advances in Bayesian Inference for Inverse Problems Felix Lucka University College London, UK f.lucka@ucl.ac.uk Applied Inverse Problems Helsinki, May 25, 2015 Bayesian Inference for Inverse Problems
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationFaster Stochastic Variational Inference using Proximal-Gradient Methods with General Divergence Functions
Faster Stochastic Variational Inference using Proximal-Gradient Methods with General Divergence Functions Mohammad Emtiyaz Khan, Reza Babanezhad, Wu Lin, Mark Schmidt, Masashi Sugiyama Conference on Uncertainty
More informationNumerical Methods. Rafał Zdunek Underdetermined problems (2h.) Applications) (FOCUSS, M-FOCUSS,
Numerical Methods Rafał Zdunek Underdetermined problems (h.) (FOCUSS, M-FOCUSS, M Applications) Introduction Solutions to underdetermined linear systems, Morphological constraints, FOCUSS algorithm, M-FOCUSS
More informationChapter 20. Deep Generative Models
Peng et al.: Deep Learning and Practice 1 Chapter 20 Deep Generative Models Peng et al.: Deep Learning and Practice 2 Generative Models Models that are able to Provide an estimate of the probability distribution
More informationGaussian Processes (10/16/13)
STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs
More informationOutline Lecture 3 2(40)
Outline Lecture 3 4 Lecture 3 Expectation aximization E and Clustering Thomas Schön Division of Automatic Control Linöping University Linöping Sweden. Email: schon@isy.liu.se Phone: 3-8373 Office: House
More informationProbabilistic Reasoning in Deep Learning
Probabilistic Reasoning in Deep Learning Dr Konstantina Palla, PhD palla@stats.ox.ac.uk September 2017 Deep Learning Indaba, Johannesburgh Konstantina Palla 1 / 39 OVERVIEW OF THE TALK Basics of Bayesian
More informationSparse Linear Models (10/7/13)
STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationLikelihood NIPS July 30, Gaussian Process Regression with Student-t. Likelihood. Jarno Vanhatalo, Pasi Jylanki and Aki Vehtari NIPS-2009
with with July 30, 2010 with 1 2 3 Representation Representation for Distribution Inference for the Augmented Model 4 Approximate Laplacian Approximation Introduction to Laplacian Approximation Laplacian
More informationPattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods
Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs
More information13 : Variational Inference: Loopy Belief Propagation and Mean Field
10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationMixtures of Gaussians. Sargur Srihari
Mixtures of Gaussians Sargur srihari@cedar.buffalo.edu 1 9. Mixture Models and EM 0. Mixture Models Overview 1. K-Means Clustering 2. Mixtures of Gaussians 3. An Alternative View of EM 4. The EM Algorithm
More informationBayesian inference J. Daunizeau
Bayesian inference J. Daunizeau Brain and Spine Institute, Paris, France Wellcome Trust Centre for Neuroimaging, London, UK Overview of the talk 1 Probabilistic modelling and representation of uncertainty
More informationBayesian Knowledge Fusion in Prognostics and Health Management A Case Study
Bayesian Knowledge Fusion in Prognostics and Health Management A Case Study Masoud Rabiei Mohammad Modarres Center for Risk and Reliability University of Maryland-College Park Ali Moahmamd-Djafari Laboratoire
More informationEfficient Bayesian Multivariate Surface Regression
Efficient Bayesian Multivariate Surface Regression Feng Li (joint with Mattias Villani) Department of Statistics, Stockholm University October, 211 Outline of the talk 1 Flexible regression models 2 The
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More informationMarginal density. If the unknown is of the form x = (x 1, x 2 ) in which the target of investigation is x 1, a marginal posterior density
Marginal density If the unknown is of the form x = x 1, x 2 ) in which the target of investigation is x 1, a marginal posterior density πx 1 y) = πx 1, x 2 y)dx 2 = πx 2 )πx 1 y, x 2 )dx 2 needs to be
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationSTA414/2104. Lecture 11: Gaussian Processes. Department of Statistics
STA414/2104 Lecture 11: Gaussian Processes Department of Statistics www.utstat.utoronto.ca Delivered by Mark Ebden with thanks to Russ Salakhutdinov Outline Gaussian Processes Exam review Course evaluations
More informationNonparametric Drift Estimation for Stochastic Differential Equations
Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,
More information