Augmented Tikhonov Regularization
|
|
- Anastasia Hutchinson
- 5 years ago
- Views:
Transcription
1 Augmented Tikhonov Regularization Bangti JIN Universität Bremen, Zentrum für Technomathematik Seminar, November 14, 2008
2 Outline 1 Background 2 Bayesian inference 3 Augmented Tikhonov regularization 4 Numerical results
3 Outline 1 Background 2 Bayesian inference 3 Augmented Tikhonov regularization 4 Numerical results
4 Linear inverse problem Hm = d, with H R n m ill-conditioned and the noisy data d Example d = d + ω noise vector Fredholm integral equations of the first kind (Tikhonov; Phillips) b a k(s, t)f (t)dt = g(s) inverse Laplace transform: k(s, t) = e st image deblurring
5 Linear inverse problem Hm = d, with H R n m ill-conditioned and the noisy data d d = d + ω noise vector Example Cauchy problem for Laplace equation (Hadamard, 1923) Ω: open bdd. domain with boundary Ω = Γ i Γ c Cauchy data: Dirichlet and Neumann b.c. on Γ c Cauchy problem: recover Dirichlet b.c. on Γ i applications: cardiography, i.h.c.p.
6 Linear inverse problem Hm = d, with H R n m ill-conditioned and the noisy data d d = d + ω noise vector Example Robin inverse problem for Laplace equation (Ingles 1997) Cauchy data: Dirichlet and Neumann b.c. on Γ c Robin inverse problem: recover Robin coefficient on Γ i applications: quenching process, corrosion detection, MOSFET semiconductor device
7 Ill-posedness (Hadamard, 1923) Existence of a solution not ensured Uniqueness of the solution violated Stable dependence on the data missing Tikhonov regularization (Tikhonov, 1960s) { } m η = arg min Hm d 2 m 2 + η Lm 2 2 ker(h) ker(l) = {0} uniqueness L: (often) discretized diff. operator smoothness Regularization parameter η plays an essential role! Data-fidelity term and regularization term can be rather general: impulsive noise, sparsity,...
8 Regularization parameter choice rules Discrepancy principle (Morozov, 1967) Hm η d 2 cδ estimated noise level mathematically rigorous: convergence and convergence rates require an estimate of the noise level very sensitive to under-estimation Quasi-optimality criterion (Tikhonov et al, 1979) η = arg min η η dm η dη no theoretical justification computationally expensive: sampling 2
9 Regularization parameter choice rules Generalized cross-validation (Golub-Heath-Wahba, 1979) G(η) = Hm η d 2 2 tr(i n HH # η ) 2 with H # η = (H T H + ηl T L) 1 H T existence of a minimizer not guaranteed often suffer from severe under-regularization computationally expensive
10 Regularization parameter choice rules L-curve criterion (Hansen, 1992) corner of L-curve L = { (log Hm η d 2, log Lm η 2 ) : η R +} existence of a corner not ensured nonconvergence results (Hanke, 1996; Vogel, 1996) computationally expensive, challenging to locate corner
11 Outline 1 Background 2 Bayesian inference 3 Augmented Tikhonov regularization 4 Numerical results
12 Rev. Thomas Bayes ( ) An essay towards solving a problem in the doctrine of chances, salient features uncertainty quantification flexible regularization
13 Fundamentals m and d are random variables. Goal: deduce the conditional distribution p(m d) (PPDF) of m on d. Bayes rule asserts p(m d) = p(d m)p(m) p(d m)p(m)dm p(d m)p(m) unimportant normalizing constant (a.k.a. Bayes factor) p(d) = p(d m)p(m)dm neglected building blocks likelihood: p(d m) data plays the role! prior distribution: p(m) Feature: The information is a prob. distr., and uncertainties of a specific inverse solution can be quantified!
14 Numerical exploration The PPDF lives in high-dim. space direct vis. impossible summarizing statistics of the PPDF posterior mean ˆm pm ˆm pm = E p(m d) [m] = mp(m d)dm R m maximum a posteriori (MAP) ˆm map marginal distribution p(m i d) ˆm map = arg max p(m d) m Caution: point estimates may not be representative!
15 Numerical strategies integration problems: no smoothness req., curse-of-dim. optimization methods: iterative, gradient type methods... Monte Carlo method Basic idea drawing i.i.d. samples {m i } N i=1 (nontrivial) sample approximation of the mean ˆm pm 1 N N m i, i=1 Traditional: importance sampling, accept-reject Modern: Markov chain Monte Carlo Metropolis-Hastings sampler, Gibbs sampler
16 Noise model additive noise model d i = (Hm) i + ξ i, i = 1, 2,..., n with ξ i being i.i.d. additive Gaussian random variables with mean zero and variance σ 2 = 1 τ, i.e. p(ξ i) = τ 2π e τ 2 ξ2 i p(d m) τ n 2 exp ( τ 2 d Hm 2 2 Remark: this noise model is not always suitable, e.g. multiplicative noise, impulsive noise, counting processes )
17 Prior model most challenging modeling task versatile tools: Markov random field, Gaussian process Markov random field A neighborhood system N is a collection of index sets verifying: i / N i ; j N i iff i N j m discrete MRF w.r.t. N if p(m i m i ) = p(m i m j, j N i ) Hammersley-Clifford theorem: p(m) exp( m i=1 V i(m)) V i (m) m i, m j, j N i German & German (1984, IEEE PAMI paper)
18 Markov random field Example: m lives on one-dim. equi. distr. lattice N i = {i 1, i + 1} V i (m) = λ 2 j N i (m j m i ) 2 p(m) λ m 2 exp ( λ 2 mt Wm ) with W = [w ij ] w ij = { 1, j Ni, 2, j = i. W = L T L with L being first-order difference operator p(m) λ m 2 exp ( λ 2 Lm 2 2) smoothness Remark: MRF can devise TV, L 1,...
19 PPDF Bayes rule PPDF p(m d) (with known τ and λ) ( p(m d) exp τ ) ( 2 d Hm 2 2 exp λ ) 2 Lm 2 2 Tikhonov regularization (MAP ˆm map ) ˆm map arg max m p(m d) = arg min m { } Hm d λτ 1 Lm 2 2 statistical basis: additive Gaussian noise with smoothness prior Regularization parameter η = λτ 1 plays the role of a regularization parameter parameters λ and τ nontrivial to determine!
20 Hierarchical modeling idea: let the data determine λ and τ! conjugate prior (Gamma distr.) for λ and τ p(λ) λ α 0 1 e β 0λ and p(τ) τ α 1 1 e β 1τ PPDF p(m, λ, τ d) p(m, λ, τ d) τ n 2 exp ( τ ) 2 d Hm 2 2 λ m 2 exp ( λ ) 2 Lm 2 2 λ α0 1 e β 0λ τ α1 1 e β 1τ
21 Outline 1 Background 2 Bayesian inference 3 Augmented Tikhonov regularization 4 Numerical results
22 Augmented Tikhonov functional MAP (m, λ, τ) map (m, λ, τ) map = arg max p(m, λ, τ d) = arg min J (m, λ, τ) (m,λ,τ) (m,λ,τ) with functional J (m, λ, τ) being J (m, λ, τ) = τ 2 Hm d λ 2 Lm β 0λ α 0 ln λ α 1 ln τ + β 1τ resembles Tikhonov method, but determines λ and τ automatically augmented Tikhonov regularization strictly biconvex w.r.t. m and (λ, τ) Theorem The functional J (m, λ, τ) has a finite minimizer.
23 Optimality system (with η = λτ 1 ) GSVD ( H T H + ηl T L ) m H T d = 0, 1 2 Lm β 0 α 0 1 λ = 0, 1 2 Hm d β 1 α 1 1 τ = 0. For matrix pair H R n m and L R p m, there holds ( ) Σ 0 H = U X 1, L = V(M 0 0 I p (m p) )X 1, m p where Σ = diag(σ 1,..., σ p ) and M = diag(µ 1,..., µ p ). The ratios γ i = σ i µ i are generalized singular values.
24 Variance estimate a-tikhonov estimate σ 2 (η) = Hmη d β 1 n+2α 1 2 GCV estimate V(η) = Hmη d 2 2 T (η) with T (η) = tr(i n HH # η ) α 1 1, β 1 0 σ 2 (η) V(η) T (η) n Lemma (Hansen,1993) If generalized s.v.s γ i relate as γ 2 i = c γ γ 2 i+1 (0 < c γ < 1), then n m + κ η 1 1 c γ T (η) n m + κ η c γ, where κ η is the number of γ 2 i less than η, i.e. γ 2 k η < γ γ 2 κ η+1.
25 The GCV estimate is only numerically verified. Theorem Let ȳ i = u T i d. Assume that there exist two constants 0 < c d < c γ < 1 such that γi 2 = c γ γi+1 2 and ȳ i 2 = c d ȳi+1 2. Moreover, PU d = 0, then V(η) satisfies [ ] ( ) K η 2 1 c γ σ c κη d 1 c d 2 1 c γc d ȳκ 2 η K η c γ where K η = n m + κ η. E[V(η)] [ K η c γ ] σ 2 0 +
26 Fix τ value at σ 2 0 η satisfies ( ) η Lm η β 0 = 2α 0 σ2 0. existence of a solution ensured! Lemma Assume the random vector ω satisfies ω c ω σ 0. There exists two constants c η,0 and c η,1 dependent on α 0 such that Theorem c η,0 σ 2 0 η c η,1 σ 2 0. For fixed β 0 and α 0 increasing like σ d 0 with 0 < d < 2, the a-tikhonov solution m converges to the generalized least-squares solution as the noise level σ 0 tends to zero.
27 Implications Hierarchical formulations with fixed α 0 and β 0 might fail for arbitrarily varying noise. Strategies to adapt α 0 are necessary. Choice of parameters variance estimate: α 1 1, β 1 0 convergence analysis: α 0 O(σ d 0 )(0 < d < 2), β 0 O( Lm )
28 biconvexity of J (m, λ, τ): alternating iterative algorithm (i) Set k = 0 and choose η 0. (ii) Solve for m k+1 by the Tikhonov regularization method { } m k+1 = arg min Hm d 2 m 2 + η k Lm 2 2. (iii) Update the parameter λ k+1 and τ k+1 by λ k+1 = α Lm k β, τ k+1 = 0 α Hm k+1 d β 1 and set η k+1 = λ k+1 τ 1 k+1. (iv) Check the stopping criterion. If not converged, set k = k + 1 and repeat from (ii).,
29 Theorem Let {(m k, λ k, τ k )} k be the sequence generated by the above algorithm. Then the sequence {J (m k, λ k, τ k )} k converges monotonically. Lemma For any η 0, the sequence {η k } k converges monotonically. Theorem The sequence {(m k, λ k, τ k )} k converges to a critical point (m, λ, τ ) of the functional J (m, λ, τ). Moreover, the convergence of the sequences {λ k } k and {τ k } k is monotonic.
30 Outline 1 Background 2 Bayesian inference 3 Augmented Tikhonov regularization 4 Numerical results
31 Phillips problem (FIE of 1st kind): Let φ(t) = [ 1 + cos( πt 3 )] χ t <3. The functions k, f, and g are respectively given by k(s, t) = φ(s t), f (t) = φ(t) and g(s) = (6 s ) ( ) πs 2 cos π sin( π 3 s ). The integration interval is [ 6, 6]. The problem is mildly ill-posed, with Cond(H) = The probability densities are calculated from 1000 Monte Carlo simulations.
32 Comparison of the regularization parameter and accuracy error. Observations AT works as well as QO, LC, and robust GCV is sensitive to noise realization
33 a-tikhonov estimate agrees well with GCV estimate, albeit slightly smaller GCV occasionally fails due to too small regularization parameter variance estimates
34 Convergence of the algorithm. Observations Variance estimate varies slowly during the iteration Converged value of λ is relatively independent of the noise level, and thus η σ 2 0 under-regularization
35 Convergence of the algorithm. Observations η k converges monotonically! AT under-regularizes for 1% noise!
36 Sensitivity analysis in case of 5% noise. β 0 σ λ η e 1e2 1.34e e e e0 1e0 1.36e e0 1.74e e-1 1e e e1 2.16e e-1 1e e e1 2.17e e-1 1e e e1 2.17e e-1 1e e e1 2.17e e-1 Observation Small β 0 yields practically identical results. Large β 0 under-regularizes.
37 Convergence of the algorithm (inverse Laplace transform). Observations The variance estimate converges within one iteration! The functional value increases with the noise level.
38 Summary Summary brief introduction to Bayesian inference propose an augmented Tikhonov functional mathematical analysis of a-tikhonov functional propose an efficient iterative algorithm Future research problems convergence rate analysis of the algorithm adaptive strategies
39 Further reading list 1 Gelman A, Carlin JB, Stern HS, Rubin DB. Bayesian Data Analysis (2nd edn). CRC, Vogel CR. Computational Methods for Inverse Problems. SIAM, Engl HW, Hanke M, Neubauer A. Regularization of Inverse Problems. Kluwer, Kaipio J, Somersalo E. Statistical and Computational Inverse Problems. Springer, Jin B, Zou J. Augmented Tikhonov regularization, in press. 6 Ito K, Jin B, Zou J. A new choice rule of regularization parameters in Tikhonov regularization, submitted.
MCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationUncertainty Quantification for Inverse Problems. November 7, 2011
Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance
More informationDeblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.
Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical
More informationRegularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution
Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationBayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems
Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.
More informationBagging During Markov Chain Monte Carlo for Smoother Predictions
Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationBayesian Inference. Chapter 1. Introduction and basic concepts
Bayesian Inference Chapter 1. Introduction and basic concepts M. Concepción Ausín Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master
More informationStatistical Estimation of the Parameters of a PDE
PIMS-MITACS 2001 1 Statistical Estimation of the Parameters of a PDE Colin Fox, Geoff Nicholls (University of Auckland) Nomenclature for image recovery Statistical model for inverse problems Traditional
More informationBayesian Inference for the Multivariate Normal
Bayesian Inference for the Multivariate Normal Will Penny Wellcome Trust Centre for Neuroimaging, University College, London WC1N 3BG, UK. November 28, 2014 Abstract Bayesian inference for the multivariate
More informationBayesian data analysis in practice: Three simple examples
Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to
More informationUnbiased Risk Estimation as Parameter Choice Rule for Filter-based Regularization Methods
Unbiased Risk Estimation as Parameter Choice Rule for Filter-based Regularization Methods Frank Werner 1 Statistical Inverse Problems in Biophysics Group Max Planck Institute for Biophysical Chemistry,
More informationLaplace-distributed increments, the Laplace prior, and edge-preserving regularization
J. Inverse Ill-Posed Probl.? (????), 1 15 DOI 1.1515/JIIP.????.??? de Gruyter???? Laplace-distributed increments, the Laplace prior, and edge-preserving regularization Johnathan M. Bardsley Abstract. For
More informationBayesian Inference in Astronomy & Astrophysics A Short Course
Bayesian Inference in Astronomy & Astrophysics A Short Course Tom Loredo Dept. of Astronomy, Cornell University p.1/37 Five Lectures Overview of Bayesian Inference From Gaussians to Periodograms Learning
More informationRegularization methods for large-scale, ill-posed, linear, discrete, inverse problems
Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationOn a Data Assimilation Method coupling Kalman Filtering, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model
On a Data Assimilation Method coupling, MCRE Concept and PGD Model Reduction for Real-Time Updating of Structural Mechanics Model 2016 SIAM Conference on Uncertainty Quantification Basile Marchand 1, Ludovic
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationBayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models
Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology
More informationRecent Advances in Bayesian Inference for Inverse Problems
Recent Advances in Bayesian Inference for Inverse Problems Felix Lucka University College London, UK f.lucka@ucl.ac.uk Applied Inverse Problems Helsinki, May 25, 2015 Bayesian Inference for Inverse Problems
More informationSupplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements
Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationDoing Bayesian Integrals
ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to
More informationNewton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve
Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationEfficient MCMC Sampling for Hierarchical Bayesian Inverse Problems
Efficient MCMC Sampling for Hierarchical Bayesian Inverse Problems Andrew Brown 1,2, Arvind Saibaba 3, Sarah Vallélian 2,3 CCNS Transition Workshop SAMSI May 5, 2016 Supported by SAMSI Visiting Research
More informationMH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution
MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous
More informationHierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,
More informationA model function method in total least squares
www.oeaw.ac.at A model function method in total least squares S. Lu, S. Pereverzyev, U. Tautenhahn RICAM-Report 2008-18 www.ricam.oeaw.ac.at A MODEL FUNCTION METHOD IN TOTAL LEAST SQUARES SHUAI LU, SERGEI
More informationwissen leben WWU Münster
MÜNSTER Sparsity Constraints in Bayesian Inversion Inverse Days conference in Jyväskylä, Finland. Felix Lucka 19.12.2012 MÜNSTER 2 /41 Sparsity Constraints in Inverse Problems Current trend in high dimensional
More informationBayesian Inverse Problems
Bayesian Inverse Problems Jonas Latz Input/Output: www.latz.io Technical University of Munich Department of Mathematics, Chair for Numerical Analysis Email: jonas.latz@tum.de Garching, July 10 2018 Guest
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationDiscrete ill posed problems
Discrete ill posed problems Gérard MEURANT October, 2008 1 Introduction to ill posed problems 2 Tikhonov regularization 3 The L curve criterion 4 Generalized cross validation 5 Comparisons of methods Introduction
More informationBayesian System Identification based on Hierarchical Sparse Bayesian Learning and Gibbs Sampling with Application to Structural Damage Assessment
Bayesian System Identification based on Hierarchical Sparse Bayesian Learning and Gibbs Sampling with Application to Structural Damage Assessment Yong Huang a,b, James L. Beck b,* and Hui Li a a Key Lab
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationTwo-parameter regularization method for determining the heat source
Global Journal of Pure and Applied Mathematics. ISSN 0973-1768 Volume 13, Number 8 (017), pp. 3937-3950 Research India Publications http://www.ripublication.com Two-parameter regularization method for
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationParameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1
Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationDiscretization-invariant Bayesian inversion and Besov space priors. Samuli Siltanen RICAM Tampere University of Technology
Discretization-invariant Bayesian inversion and Besov space priors Samuli Siltanen RICAM 28.10.2008 Tampere University of Technology http://math.tkk.fi/inverse-coe/ This is a joint work with Matti Lassas
More informationParameter estimation: A new approach to weighting a priori information
c de Gruyter 27 J. Inv. Ill-Posed Problems 5 (27), 2 DOI.55 / JIP.27.xxx Parameter estimation: A new approach to weighting a priori information J.L. Mead Communicated by Abstract. We propose a new approach
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationOptimization with nonnegativity constraints
Optimization with nonnegativity constraints Arie Verhoeven averhoev@win.tue.nl CASA Seminar, May 30, 2007 Seminar: Inverse problems 1 Introduction Yves van Gennip February 21 2 Regularization strategies
More informationPenalized Loss functions for Bayesian Model Choice
Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented
More informationVariational Methods in Bayesian Deconvolution
PHYSTAT, SLAC, Stanford, California, September 8-, Variational Methods in Bayesian Deconvolution K. Zarb Adami Cavendish Laboratory, University of Cambridge, UK This paper gives an introduction to the
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationStrong Lens Modeling (II): Statistical Methods
Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationREGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE
Int. J. Appl. Math. Comput. Sci., 007, Vol. 17, No., 157 164 DOI: 10.478/v10006-007-0014-3 REGULARIZATION PARAMETER SELECTION IN DISCRETE ILL POSED PROBLEMS THE USE OF THE U CURVE DOROTA KRAWCZYK-STAŃDO,
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationSolving discrete ill posed problems with Tikhonov regularization and generalized cross validation
Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation Gérard MEURANT November 2010 1 Introduction to ill posed problems 2 Examples of ill-posed problems 3 Tikhonov
More informationEfficient Variational Inference in Large-Scale Bayesian Compressed Sensing
Efficient Variational Inference in Large-Scale Bayesian Compressed Sensing George Papandreou and Alan Yuille Department of Statistics University of California, Los Angeles ICCV Workshop on Information
More informationHierarchical Bayesian Modeling
Hierarchical Bayesian Modeling Making scientific inferences about a population based on many individuals Angie Wolfgang NSF Postdoctoral Fellow, Penn State Astronomical Populations Once we discover an
More informationMultilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems
Jonas Latz 1 Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems Jonas Latz Technische Universität München Fakultät für Mathematik Lehrstuhl für Numerische Mathematik jonas.latz@tum.de November
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationSAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software
SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:
More informationStatistical Methods in Particle Physics Lecture 1: Bayesian methods
Statistical Methods in Particle Physics Lecture 1: Bayesian methods SUSSP65 St Andrews 16 29 August 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationMONTE CARLO METHODS. Hedibert Freitas Lopes
MONTE CARLO METHODS Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationStat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline
More informationRegularization parameter estimation for underdetermined problems by the χ 2 principle with application to 2D focusing gravity inversion
Regularization parameter estimation for underdetermined problems by the χ principle with application to D focusing gravity inversion Saeed Vatankhah, Rosemary A Renaut and Vahid E Ardestani, Institute
More informationMinimum Message Length Analysis of the Behrens Fisher Problem
Analysis of the Behrens Fisher Problem Enes Makalic and Daniel F Schmidt Centre for MEGA Epidemiology The University of Melbourne Solomonoff 85th Memorial Conference, 2011 Outline Introduction 1 Introduction
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationEmpirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods
Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods Frank Werner 1 Statistical Inverse Problems in Biophysics Group Max Planck Institute for Biophysical Chemistry,
More informationST 740: Markov Chain Monte Carlo
ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:
More informationKarl-Rudolf Koch Introduction to Bayesian Statistics Second Edition
Karl-Rudolf Koch Introduction to Bayesian Statistics Second Edition Karl-Rudolf Koch Introduction to Bayesian Statistics Second, updated and enlarged Edition With 17 Figures Professor Dr.-Ing., Dr.-Ing.
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationSIGNAL AND IMAGE RESTORATION: SOLVING
1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationThe Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.
Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface
More informationMarkov Chain Monte Carlo
1 Motivation 1.1 Bayesian Learning Markov Chain Monte Carlo Yale Chang In Bayesian learning, given data X, we make assumptions on the generative process of X by introducing hidden variables Z: p(z): prior
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian
More informationBayesian Inverse problem, Data assimilation and Localization
Bayesian Inverse problem, Data assimilation and Localization Xin T Tong National University of Singapore ICIP, Singapore 2018 X.Tong Localization 1 / 37 Content What is Bayesian inverse problem? What is
More informationUse of probability gradients in hybrid MCMC and a new convergence test
Use of probability gradients in hybrid MCMC and a new convergence test Ken Hanson Methods for Advanced Scientific Simulations Group This presentation available under http://www.lanl.gov/home/kmh/ June
More informationBayesian Approach 2. CSC412 Probabilistic Learning & Reasoning
CSC412 Probabilistic Learning & Reasoning Lecture 12: Bayesian Parameter Estimation February 27, 2006 Sam Roweis Bayesian Approach 2 The Bayesian programme (after Rev. Thomas Bayes) treats all unnown quantities
More informationAn introduction to Bayesian statistics and model calibration and a host of related topics
An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the
More informationA MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS
A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS SILVIA NOSCHESE AND LOTHAR REICHEL Abstract. Truncated singular value decomposition (TSVD) is a popular method for solving linear discrete ill-posed
More informationMarkov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationSimultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors
Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,
More informationSTAT 425: Introduction to Bayesian Analysis
STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte
More informationA Double Regularization Approach for Inverse Problems with Noisy Data and Inexact Operator
A Double Regularization Approach for Inverse Problems with Noisy Data and Inexact Operator Ismael Rodrigo Bleyer Prof. Dr. Ronny Ramlau Johannes Kepler Universität - Linz Florianópolis - September, 2011.
More informationMachine Learning Lecture Notes
Machine Learning Lecture Notes Predrag Radivojac January 25, 205 Basic Principles of Parameter Estimation In probabilistic modeling, we are typically presented with a set of observations and the objective
More informationBayesian Sparse Correlated Factor Analysis
Bayesian Sparse Correlated Factor Analysis 1 Abstract In this paper, we propose a new sparse correlated factor model under a Bayesian framework that intended to model transcription factor regulation in
More informationFitting Narrow Emission Lines in X-ray Spectra
Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:
More informationHierarchical Modelling for Univariate Spatial Data
Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul
More information