Bayesian Registration of Functions with a Gaussian Process Prior

Similar documents
Alignment and Analysis of Proteomics Data using Square Root Slope Function Framework

The Square Root Velocity Framework for Curves in a Homogeneous Space

Registration of Functional Data Using Fisher-Rao Metric

CPSC 540: Machine Learning

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010

Supervised Dimension Reduction:

Introduction to Machine Learning

Pattern Recognition and Machine Learning

Bayesian Nonparametric Regression for Diabetes Deaths

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Contents. Part I: Fundamentals of Bayesian Inference 1

A Bayesian Nonparametric Model for Predicting Disease Status Using Longitudinal Profiles

Bayesian Methods for Machine Learning

Numerical Analysis for Statisticians

Monte Carlo in Bayesian Statistics

FLORIDA STATE UNIVERSITY COLLEGE OF ARTS AND SCIENCES FUNCTIONAL COMPONENT ANALYSIS AND REGRESSION USING ELASTIC METHODS J.

Bayesian Inference of Multiple Gaussian Graphical Models

Lecture 16: Mixtures of Generalized Linear Models

Nonparametric Bayesian Methods (Gaussian Processes)

Reducing The Computational Cost of Bayesian Indoor Positioning Systems

A Bayesian Approach to Phylogenetics

Durham Research Online

Information geometry for bivariate distribution control

Mixed effect model for the spatiotemporal analysis of longitudinal manifold value data

Bayesian Principal Geodesic Analysis in Diffeomorphic Image Registration

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

CSC 2541: Bayesian Methods for Machine Learning

NONPARAMETRIC BAYESIAN INFERENCE ON PLANAR SHAPES

Manifold Monte Carlo Methods

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Small-variance Asymptotics for Dirichlet Process Mixtures of SVMs

Efficient Bayesian Multivariate Surface Regression

Bayesian model selection in graphs by using BDgraph package

Forward Problems and their Inverse Solutions

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

Gaussian Mixture Model

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Algorithms for Variational Learning of Mixture of Gaussians

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Part IV: Monte Carlo and nonparametric Bayes

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Markov Chain Monte Carlo Lecture 6

Monte Carlo methods for sampling-based Stochastic Optimization

David B. Dahl. Department of Statistics, and Department of Biostatistics & Medical Informatics University of Wisconsin Madison

ECE 5984: Introduction to Machine Learning

Designing Kernel Functions Using the Karhunen-Loève Expansion

Markov chain Monte Carlo methods in atmospheric remote sensing

Lecture Note 1: Background

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm

Markov Chain Monte Carlo Algorithms for Gaussian Processes

Computer Practical: Metropolis-Hastings-based MCMC

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Dimension-Independent likelihood-informed (DILI) MCMC

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

Markov Chain Monte Carlo (MCMC)

Bayesian Analysis. Bayesian Analysis: Bayesian methods concern one s belief about θ. [Current Belief (Posterior)] (Prior Belief) x (Data) Outline

Distributed Estimation, Information Loss and Exponential Families. Qiang Liu Department of Computer Science Dartmouth College

Machine Learning Linear Classification. Prof. Matteo Matteucci

Stat 516, Homework 1

Non-convex optimization. Issam Laradji

Statistical techniques for data analysis in Cosmology

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California

Elastic Handling of Predictor Phase in Functional Regression Models

STA 4273H: Statistical Machine Learning

Bayesian Estimation of Input Output Tables for Russia

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Disk Diffusion Breakpoint Determination Using a Bayesian Nonparametric Variation of the Errors-in-Variables Model

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Default Priors and Effcient Posterior Computation in Bayesian

Introduction to Bayesian methods in inverse problems

Non-Parametric Bayes

CSC 2541: Bayesian Methods for Machine Learning

Consistent Downscaling of Seismic Inversions to Cornerpoint Flow Models SPE

Advanced Statistical Computing

Blind Equalization via Particle Filtering

Spatial Normalized Gamma Process

Kernel adaptive Sequential Monte Carlo

Principles of Bayesian Inference

Lecture 16 Deep Neural Generative Models

Consistent high-dimensional Bayesian variable selection via penalized credible regions

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Patterns of Scalable Bayesian Inference Background (Session 1)

Bagging During Markov Chain Monte Carlo for Smoother Predictions

A Process over all Stationary Covariance Kernels

Riemannian Stein Variational Gradient Descent for Bayesian Inference

Gentle Introduction to Infinite Gaussian Mixture Modeling

MCMC Sampling for Bayesian Inference using L1-type Priors

Afternoon Meeting on Bayesian Computation 2018 University of Reading

Metamorphic Geodesic Regression

CPSC 540: Machine Learning

Bayesian model selection: methodology, computation and applications

Can we do statistical inference in a non-asymptotic way? 1

Probability for Statistics and Machine Learning

CSC 446 Notes: Lecture 13

Transcription:

Bayesian Registration of Functions with a Gaussian Process Prior

RELATED PROBLEMS Functions Curves Images Surfaces

FUNCTIONAL DATA Electrocardiogram Signals (Heart Disease) Gait Pressure Measurements (Parkinson s) Gene Expression Data Respiration Strain During 4D-CT Scan Signature Acceleration Curves (Fraud Detection) Hand-grip Strength (Arthritis) Spike Trains Growth Curves

ALIGNMENT PROBLEM DESCRIPTION

MAIN GOALS The function alignment problem has been referred to by many names including registration and warping. Note: I will use these terms interchangeably throughout this talk. Main Goals: 1. Function alignment: principled approach to separate the amplitude and phase variabilities of functional datasets for subsequent statistical analysis. 2. Alignment uncertainty: framework that allows computation of confidence values associated with alignment estimates. 3. Multiple plausible alignments: approach that provides a recipe for identifying multiple good registrations. 4. Statistical analysis of warping functions: tools for computing statistics on the space of warping functions including the mean, covariance, PCA, cluster analysis, etc.

RELATED WORK Bayesian alignment of functions and curves: B-splines: Telesca and Inoue, 2008 (JASA) Multiresolution approach or warplets : Claeskens et al., 2010 (JRSSB) SRVF with Dirichlet prior: Cheng et al., 2016 (Bayesian Analysis) Bayesian deformable models: Allassonniere et al., 2010 (Bernoulli) Computational anatomy: Ma et al., 2008 (NeuroImage); Tang et al., 2013 (PLoS ONE) Other works: Srivastava and Jermyn, 2009 (PAMI); Zhang and Fletcher, 2016 (Algorithmic Advances in Riemannian Geometry and Applications); Wassermann et al., 2014 (WBIR); Simpson et al., 2015 (Medical Image Analysis); etc.

PAIRWISE ALIGNMENT PROBLEM Observation space : set of absolutely continuous functions on [0,1]. Parameter space of warping functions: acts on the function space by composition:. PQRST complex f 1 and f 2 γ f 1, f 2 γ

PAIRWISE ALIGNMENT PROBLEM Most common optimization approach (Dynamic Time Warping): Known issues: (1) lack of symmetry, (2) pinching effect: (one can pinch the entire function f 2 to get arbitrarily close to f 1 in norm), (3) not a proper metric on the quotient space : cannot be used for statistical analysis. Another approach based on the norm: Adding a regularization penalty on γ eliminates the pinching effect, but the other drawbacks remain. The choice of the regularization parameter λ is not obvious in general cases.

SOLUTION BASED ON SRSF Square-root slope function (SRSF):. Inverse-mapping is well defined and easy to compute. Space of SRSFs:. Action of warping group on SRSF space:. Important property:. Equivalence class of SRSF under warping:. Distance can be efficiently computed using Dynamic Programming (DP): This approach overcomes all previously mentioned issues. Separate issue: this is a purely optimization-based approach, which does not allow the user to easily assess alignment uncertainty or to discover multiple plausible alignments.

OUTLINE OF PROPOSED APPROACH Bayesian Model + Riemannian Geometry + Markov Chain Monte Carlo (1) Bayesian Alignment Model: Allows comprehensive exploration of the variable space. Provides credible intervals of warping function estimates. Allows discovering multiple registration solutions through multimodal posteriors. (2) Riemannian Geometry of Warping Group: Enables efficient Riemannian computation on the space of warping functions: tools for computing statistics and implementation of a k-means clustering algorithm to discover multiple modes in the posterior. Allows a geometric prior distribution on the space of warpings. (3) Markov Chain Monte Carlo: Allows sampling from the posterior distribution. We use a specialized preconditioned Crank Nicolson algorithm for functional parameters.

RIEMANNIAN GEOMETRY OF WARPINGS Riemannian metric on (, Fisher-Rao metric): This results in a non-trivial Riemannian geometry of. Simplification: Definition: Define the mapping. Then, given an element, define a new representation using the square-root of its derivative as. We refer to ψ as the square-root density (SRD). Main Result: simplifies to the positive orthant of the Hilbert sphere and the Fisher-Rao metric simplifies to the metric. The Riemannian geometry of the Hilbert sphere is well known providing efficient tools for statistical analysis on the warping group.

GEOMETRIC TOOLS Identity mapping, i.e. no warping:. denotes the inner product. Tangent space:. Geodesic distance:. Exponential and inverse-exponential maps:

STATISTICS OF WARPING FUNCTIONS Given: a sample of warping functions and their corresponding SRDs. Karcher Mean:. Geometric Median (Fletcher et al., 2009):. Both quantities found using a standard gradient descent algorithm. K-means clustering: utilize the Karcher mean and Fisher-Rao distance on the warping group.

PAIRWISE ALIGNMENT MODEL Function discretization:. Pairwise alignment model: Prior truncated to Ψ. C g : pre-specified covariance operator. a and b: pre-specified constants. Sampling from posterior via Metropolis within Gibbs: Update sequentially from full conditionals of g (Metropolis step using pcn) and (direct draw).

Z-MIXTURE pcn MCMC ALGORITHM We use the Z-mixture preconditioned Crank Nicolson (pcn) to update the function g (Cotter et al., 2013): 1. Propose. 2. Accept g with probability. (prior) via Karhunen-Loeve expansion: 1. Assume the eigenpairs of C g are (λ i2, b i (t); i 1). We use a Fourier basis. 2. Sample and let.

Z-MIXTURE pcn MCMC ALGORITHM β is a tuning parameter: 1. Large (small) values imply large (small) jumps in the parameter space. 2. We use a mixture to improve MCMC convergence. Advantages of mixture pcn MCMC algorithm: 1. The acceptance ratio is independent of the dimension of the discretized g. 2. The tuning parameter β is random; flexible to control size of jump.

MULTIPLE ALIGNMENT MODEL Multiple alignment model: Sampling from posterior by updating sequentially: g 1,,g C, q* (Metropolis steps using pcn) and (direct draw).

EXAMPLES: PAIRWISE SIMULATIONS Data: Three simulation settings: (1) randomly generate γ using the first pair of Fourier basis (fourier1), (2) randomly generate γ using the first 20 pairs of Fourier basis (fourier20), (3) generate γ by smoothing a random step function (random) Each simulation setting is repeated ten times. Evaluation: Fisher-Rao distance between posterior mean and true warping.

EXAMPLES: PAIRWISE SIMULATIONS

EXAMPLES: MULTIPLE ALIGNMENT Data: right knee flexion functions for 12 individuals

EXAMPLES: MULTIPLE ALIGNMENT Original Data Aligned Data Cross-sectional Averages Estimated Warps Simulated Gait Data: Pelvis Right Roll Spike Trains Pinch Force

SUMMARY AND FUTURE WORK Contributions: 1. We proposed a novel Bayesian model for pairwise and multiple alignment of functional data. 2. We use the Riemannian geometry of the space of warping functions for: defining a Gaussian prior; efficient sampling from the posterior distribution; computation of statistics of warping functions. 3. The proposed model is very good at exploring the entire variable space. In some cases, it performs better than the optimization-based DP algorithm. Future Work: 1. Theoretical properties of the model. 2. Sensitivity analysis. 3. Alternative priors and posterior sampling strategies for warping functions. 4. Extension to higher-dimensional curves, images and surfaces.

CBMS CONFERENCE CBMS Conference: Elastic Functional and Shape Data Analysis (EFSDA) Where: The Ohio State University, Columbus, OH When: July 16-20, 2018 Primary Lecturer: Prof. Anuj Srivastava Organizers: Sebastian Kurtek (Statistics, OSU), Facundo Memoli (Mathematics, OSU), Yusu Wang (Computer Science and Engineering, OSU), Tingting Zhang (Statistics, University of Virginia), Hongtu Zhu (Biostatistics, MD Anderson Cancer Center)

THANK YOU! QUESTIONS?