Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 26, 2013
|
|
- Posy Morris
- 5 years ago
- Views:
Transcription
1 Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 6, 13 SLIM University of British Columbia
2 Motivation 3D seismic experiments - 5D data expensive to acquire, store sample at sub- Nyquist rates Data exhibits low- rank structure exploit structure for interpolation Fully sampled data simultaneous sources in wave- equation based inversion mitigating multiples
3 [1] Kreimer, N, and Sacchi, M. D. "A tensor higher- order singular value decomposivon for prestack seismic data noise reducvon and interpolavon." (1) [] Gao, Jianjun, Vicente Oropeza, and Mauricio D. Sacchi. "EvaluaVon of a fast algorithm for the eigen- decomposivon of large block Toeplitz matrices with applicavon to 5D seismic data interpolavon." (11) Context Low- rank matrix/tensor completion via nuclear norm projection [1] Require SVDs on huge data matrices Not scalable to large problem sizes Data completion via Toeplitz embedding [] Problem size - (# data points)
4 Goals Generalization of Compressive Sensing to multiple dimensions what can we learn from 1D/D recovery? Randomized source/receiver acquisition reduce acquisition financial/time costs Efficient solver SVD- free, parallelizable # parameters << # data points
5 7.3 Hz - 75% missing receivers Common source gather x 1 6 x Receiver y 1 Receiver y Receiver x Receiver x True data Subsampled data
6 7.3 Hz - 75% missing receivers Common source gather x 1 6 x Receiver y 1 Receiver y Receiver x Receiver x 6 True data Recovered data - SNR 17. db
7 Compressive sensing with sparsity promotion Successful reconstruction scheme Signal structure sparsity Sampling subsampling decreases sparsity Optimization look for sparsest solution
8 Multidimensional interpolation with Hierarchical Tucker Successful reconstruction scheme Signal structure Hierarchical Tucker Sampling subsampling increases hierarchical rank Optimization fit data in the Hierarchical Tucker format
9 Matricization The matricization of a tensor X with dimensions 1,...,d along the dimensions t =(t 1,...,t r ) dimensions t along the rows and dimensions is the matrix formed by placing the t c along the columns Denoted X (t)
10 Example in Matlab n1 = ; n = ; n3 = ; n = ; % Tensor x = randn(n1,n,n3,n); X (1,) X (3,) X (1,3) % Matricization along dimensions 1 and x1 = reshape(x,n1 * n, n3 * n); % Matricization along dimensions 3 and y3 = permute(x,[3 1 ]); x3 = reshape(x, n3 * n, n1 * n); % Matricization along dimensions 1 and 3 y13 = permute(x,[1 3 ]); x13 = reshape(x,n1 * n3, n * n);
11 Hierarchical Tucker format X n 1 n n 3 n tensor B 13 k U T 3 3 n1n X (1,) = n1n U 1 n 3 n n 3 n k 1 SVD -like decomposition
12 Hierarchical Tucker format X n 1 n n 3 n tensor n1n U 1 n 1! U 1 n k 1 k 1
13 Hierarchical Tucker format X n 1 n n 3 n tensor B 1 n1n U 1 n 1! U 1 n 1! U 1 U T n k 1 k 1 n k k 1
14 Hierarchical Tucker format Intermediate matrices don t need to be stored U t,b t - small parameter matrices specify the tensor completely Separating groups of dimensions from each other dimension tree
15 A. Uschmajew, B. Vandereycken. The geometry of algorithms using hierarchical tensors. Linear algebra and its applications, 13. Example B 135 {1,, 3,, 5} = t r U 13 B 13 U 5 B 5 {1,, 3} {, 5} U 1 U 3 B 3 U U 5 {1} {, 3} = t {} {5} U U 3 {} = t 1 {3} = t
16 Hierarchical Tucker format Storage apple dnk +(d )K 3 + K Compare to N d storage for the full tensor Effectively breaking the curse of dimensionality when K N d Low frequency data compresses in HT
17 Hierarchical Tucker example For a cube with max rank N = 1,d=,K = Full storage: N d = 1 8 values HTucker storage: values Compression of a factor of 99.97%
18 Seismic Hierarchical Tucker We consider a 3D seismic survey with coordinates (src x, src y, rec x, rec y, time) We take a Fourier transform in time and restrict ourselves to a single frequency slice
19 Seismic Hierarchical Tucker For a frequency slice with coordinates (src x, src y, rec x, rec y), there are essentially two choices of dimension splitting (by reciprocity) {src x, srcy, rec x, rec y} {src x, rec x, src y, rec y} {src x, src y} {rec x, rec y} {src x, rec x} {src y, rec y} {src x} {src y} {rec x} {rec y} {src x} {rec x} {src y} {rec y} Canonical Decomposition Non- canonical Decomposition
20 Matricizations 1 x rec, y rec Full data Normalized singular value x src, y src 1 6 (Rec x, Rec y) matricization - Canonical ordering
21 Matricizations 1 x src,x rec (Rx Ry) matricization (Sx Rx) matricization Normalized singular value y src,y rec 1 6 (Src x, Rec x) matricization - Noncanonical ordering
22 Multidimensional interpolation with Hierarchical Tucker Successful reconstruction scheme Signal structure Hierarchical Tucker Sampling subsampling increases hierarchical rank Optimization fit data in the Hierarchical Tucker format
23 Matrix Completion X Normalized singular value
24 Matrix Completion A(X) Normalized singular value
25 Tensor Completion Structure - recover a tensor Well represented in HT X which has low hierarchical rank Sampling - random removal of points increases rank Poorly represented in HT Idealized sampling
26 Idealized recovery 75% random entries removed Common receiver gather x 1 3 x Source x 5 Source x Source y Source y 1 True data Subsampled data
27 Idealized recovery 75% random entries removed Common receiver gather x x Source x 5 Source x Source y Source y True data Subsampled data
28 Idealized recovery 75% random entries removed Common receiver gather x x Source x 5 Source x Source y Source y True data Recovered data - SNR 19.3 db
29 Sampling Sampling (x src,y src,x rec,y rec ) points from the data idealized recovery impossible to physically implement Sampling (x rec,y rec ) points from the data less idealized possible to acquire data - e.g. ocean bottom nodes
30 Realistic recovery 5% random receivers removed x rec, y rec No subsampling 5% missing sources Normalized singular value x src, y src 1 6 (Rec x, Rec y) matricization - Canonical ordering
31 Realistic recovery 5% random receivers removed x src,x rec No subsampling 5% missing sources y,y src rec Normalized singular value (Src x, Rec x) matricization - Noncanonical ordering
32 Data organization In summary: (rec x, rec y) organization High rank Missing sources operator - removes columns Poor recovery scenario (src x, rec x) organization Low rank Missing sources operator - removes blocks Closer to ideal recovery scenario
33 Multidimensional interpolation with Hierarchical Tucker Successful reconstruction scheme Signal structure Hierarchical Tucker Sampling subsampling increases hierarchical rank Optimization fit data in the Hierarchical Tucker format
34 Optimization Given data b with missing sources and/or receivers, subsampling operator A, full tensor expansion operator :(U t,b t )! C n 1...n d solve min x=(u t,b t ) 1 ka (x) bk
35 A. Uschmajew, B. Vandereycken. The geometry of algorithms using hierarchical tensors. Linear algebra and its applications, 13 C. Da Silva and F. J. Herrmann, OpFmizaFon on the Hierarchical Tucker manifold - applicafons to tensor complefon, 13 Differential geometry HT tensors parametrize a submanifold of full tensor space C n 1...n d Nonlinear, nonconvex space Generalization of curved surfaces Steepest Descent, Conjugate gradient, Gauss- Newton without SVDs in the full tensor space
36 Optimization program (x) Full- tensor space Parameter space
37 Optimization program
38 Derivatives Derivatives of a particular node with respect to its children can be computed efficiently, i.e. via (I U tl U H t l )hu T t r Z, B t i (,3),(,3) (I U tr U H t r )hu T t l 1 Z, B t i (1,3),(1,3) X P i Q multiplies Q by P in dimension i, hx, Y i (1,3),(1,3) = X i1,:,i Y 3 i1,:,i 3 i 1,i 3 The chain rule gives the gradient of the function
39 Derivatives Only involves matrix- matrix multiplications of small matrices compared to the full- tensor space Parallelizable - multilinear product can be done in parallel SVD- free - no large- scale SVDs, unlike nuclear norm- based methods
40 Results
41 Synthetic BG Group data Unknown model 68 x 68 sources with 1 x 1 receivers, data at.68hz, 7.3 Hz Receivers subsampled to 1 x 1 Recovered with Gauss- Newton
42 .68 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Subsampled data
43 .68 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Recovered data - SNR 19.7 db
44 .68 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Difference
45 .68 Hz - 75% missing receivers Common receiver gather - no data initially x 1 8 x Source y 3 Source y Source x Source x 8 True data Recovered data - SNR 3 db
46 .68 Hz - 75% missing receivers Common receiver gather - no data initially x 1 8 x Source y 3 Source y Source x Source x 8 True data Difference
47 7.3 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Subsampled data
48 7.3 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Recovered data - SNR 17.6 db
49 7.3 Hz - 75% missing receivers Common source gather x 1 x Receiver y 1 Receiver y Receiver x Receiver x True data Difference
50 7.3 Hz - 75% missing receivers Common receiver gather - no data initially x 1 x Source y 3 1 Source y Source x Source x True data Recovered data - SNR 16. db
51 7.3 Hz - 75% missing receivers Common receiver gather - no data initially x 1 x Source y 3 Source y Source x Source x True data Difference
52 7.3 Hz - simultaneous receivers - 9% data reduction Common source gather x 1 x Receiver y Receiver y Receiver x Receiver x True data Input data - A T Ab A - subsampling operator b - full data
53 7.3 Hz - simultaneous receivers - 9% data reduction Common source gather x 1 x Receiver y Receiver y Receiver x Receiver x True data Recovered data - SNR 16. db
54 7.3 Hz - simultaneous receivers - 9% data reduction Common source gather x 1 x Receiver y Receiver y Receiver x Receiver x True data Difference
55 Conclusion 3D seismic data has an underlying structure that we can exploit for interpolation (Hierarchical Tucker format) Different schemes for organizing data - important for recovery
56 Conclusion We can interpolate HT tensors with missing entries using the Riemannian manifold structure of the HT format Achieve good results from largely subsampled data (75% missing receivers) Can use this method to create full volumes from subsampled data Migration, multiple removal, etc.
57 Acknowledgements Thank you for your attention This work was in part financially supported by the Natural Sciences and Engineering Research Council of Canada Discovery Grant (R815) and the Collaborative Research and Development Grant DNOISE II (3751-8). This research was carried out as part of the SINBAD II project with support from the following organizations: BG Group, BGP, BP, CGG, Chevron, ConocoPhillips, ION, Petrobras, PGS, Total SA, WesternGeco, Statoil, and Woodside.
Low-rank Promoting Transformations and Tensor Interpolation - Applications to Seismic Data Denoising
Low-rank Promoting Transformations and Tensor Interpolation - Applications to Seismic Data Denoising Curt Da Silva and Felix J. Herrmann 2 Dept. of Mathematics 2 Dept. of Earth and Ocean Sciences, University
More informationSeismic data interpolation and denoising using SVD-free low-rank matrix factorization
Seismic data interpolation and denoising using SVD-free low-rank matrix factorization R. Kumar, A.Y. Aravkin,, H. Mansour,, B. Recht and F.J. Herrmann Dept. of Earth and Ocean sciences, University of British
More informationMitigating data gaps in the estimation of primaries by sparse inversion without data reconstruction
Mitigating data gaps in the estimation of primaries by sparse inversion without data reconstruction Tim T.Y. Lin SLIM University of British Columbia Talk outline Brief review of REPSI Data reconstruction
More informationApplication of matrix square root and its inverse to downward wavefield extrapolation
Application of matrix square root and its inverse to downward wavefield extrapolation Polina Zheglova and Felix J. Herrmann Department of Earth and Ocean sciences, University of British Columbia, Vancouver,
More informationWavefield Reconstruction Inversion (WRI) a new take on wave-equation based inversion Felix J. Herrmann
Wavefield Reconstruction Inversion (WRI) a new take on wave-equation based inversion Felix J. Herrmann SLIM University of British Columbia van Leeuwen, T and Herrmann, F J (2013). Mitigating local minima
More informationSparsity-promoting migration with multiples
Sparsity-promoting migration with multiples Tim Lin, Ning Tu and Felix Herrmann SLIM Seismic Laboratory for Imaging and Modeling the University of British Columbia Courtesy of Verschuur, 29 SLIM Motivation..
More informationSimultaneous estimation of wavefields & medium parameters
Simultaneous estimation of wavefields & medium parameters reduced-space versus full-space waveform inversion Bas Peters, Felix J. Herrmann Workshop W- 2, The limit of FWI in subsurface parameter recovery.
More informationSource estimation for frequency-domain FWI with robust penalties
Source estimation for frequency-domain FWI with robust penalties Aleksandr Y. Aravkin, Tristan van Leeuwen, Henri Calandra, and Felix J. Herrmann Dept. of Earth and Ocean sciences University of British
More informationCompressive Sensing Applied to Full-wave Form Inversion
Compressive Sensing Applied to Full-wave Form Inversion Felix J. Herrmann* fherrmann@eos.ubc.ca Joint work with Yogi Erlangga, and Tim Lin *Seismic Laboratory for Imaging & Modeling Department of Earth
More informationFWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke
Consortium 2010 FWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke SLIM University of British Columbia Full Waveform Inversion The Full Waveform
More informationSeismic wavefield inversion with curvelet-domain sparsity promotion
Seismic wavefield inversion with curvelet-domain sparsity promotion Felix J. Herrmann* fherrmann@eos.ubc.ca Deli Wang** wangdeli@email.jlu.edu.cn *Seismic Laboratory for Imaging & Modeling Department of
More informationCompressive sampling meets seismic imaging
Compressive sampling meets seismic imaging Felix J. Herrmann fherrmann@eos.ubc.ca http://slim.eos.ubc.ca joint work with Tim Lin and Yogi Erlangga Seismic Laboratory for Imaging & Modeling Department of
More information3D INTERPOLATION USING HANKEL TENSOR COMPLETION BY ORTHOGONAL MATCHING PURSUIT A. Adamo, P. Mazzucchelli Aresys, Milano, Italy
3D INTERPOLATION USING HANKEL TENSOR COMPLETION BY ORTHOGONAL MATCHING PURSUIT A. Adamo, P. Mazzucchelli Aresys, Milano, Italy Introduction. Seismic data are often sparsely or irregularly sampled along
More informationFrequency-Domain Rank Reduction in Seismic Processing An Overview
Frequency-Domain Rank Reduction in Seismic Processing An Overview Stewart Trickett Absolute Imaging Inc. Summary Over the last fifteen years, a new family of matrix rank reduction methods has been developed
More informationMigration with Implicit Solvers for the Time-harmonic Helmholtz
Migration with Implicit Solvers for the Time-harmonic Helmholtz Yogi A. Erlangga, Felix J. Herrmann Seismic Laboratory for Imaging and Modeling, The University of British Columbia {yerlangga,fherrmann}@eos.ubc.ca
More informationFull-Waveform Inversion with Gauss- Newton-Krylov Method
Full-Waveform Inversion with Gauss- Newton-Krylov Method Yogi A. Erlangga and Felix J. Herrmann {yerlangga,fherrmann}@eos.ubc.ca Seismic Laboratory for Imaging and Modeling The University of British Columbia
More informationSLIM. University of British Columbia
Accelerating an Iterative Helmholtz Solver Using Reconfigurable Hardware Art Petrenko M.Sc. Defence, April 9, 2014 Seismic Laboratory for Imaging and Modelling Department of Earth, Ocean and Atmospheric
More informationUncertainty quantification for Wavefield Reconstruction Inversion
Uncertainty quantification for Wavefield Reconstruction Inversion Zhilong Fang *, Chia Ying Lee, Curt Da Silva *, Felix J. Herrmann *, and Rachel Kuske * Seismic Laboratory for Imaging and Modeling (SLIM),
More informationMatrix Probing and Simultaneous Sources: A New Approach for Preconditioning the Hessian
Matrix Probing and Simultaneous Sources: A New Approach for Preconditioning the Hessian Curt Da Silva 1 and Felix J. Herrmann 2 1 Dept. of Mathematics 2 Dept. of Earth and Ocean SciencesUniversity of British
More informationTime domain sparsity promoting LSRTM with source estimation
Time domain sparsity promoting LSRTM with source estimation Mengmeng Yang, Philipp Witte, Zhilong Fang & Felix J. Herrmann SLIM University of British Columbia Motivation Features of RTM: pros - no dip
More informationMath 671: Tensor Train decomposition methods
Math 671: Eduardo Corona 1 1 University of Michigan at Ann Arbor December 8, 2016 Table of Contents 1 Preliminaries and goal 2 Unfolding matrices for tensorized arrays The Tensor Train decomposition 3
More informationPostgraduate Course Signal Processing for Big Data (MSc)
Postgraduate Course Signal Processing for Big Data (MSc) Jesús Gustavo Cuevas del Río E-mail: gustavo.cuevas@upm.es Work Phone: +34 91 549 57 00 Ext: 4039 Course Description Instructor Information Course
More informationA randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors
A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017 2 Classification of hazardous
More informationIntroduction How it works Theory behind Compressed Sensing. Compressed Sensing. Huichao Xue. CS3750 Fall 2011
Compressed Sensing Huichao Xue CS3750 Fall 2011 Table of Contents Introduction From News Reports Abstract Definition How it works A review of L 1 norm The Algorithm Backgrounds for underdetermined linear
More informationSelf-Calibration and Biconvex Compressive Sensing
Self-Calibration and Biconvex Compressive Sensing Shuyang Ling Department of Mathematics, UC Davis July 12, 2017 Shuyang Ling (UC Davis) SIAM Annual Meeting, 2017, Pittsburgh July 12, 2017 1 / 22 Acknowledgements
More informationRecent results in curvelet-based primary-multiple separation: application to real data
Recent results in curvelet-based primary-multiple separation: application to real data Deli Wang 1,2, Rayan Saab 3, Ozgur Yilmaz 4, Felix J. Herrmann 2 1.College of Geoexploration Science and Technology,
More informationTensor-Based Dictionary Learning for Multidimensional Sparse Recovery. Florian Römer and Giovanni Del Galdo
Tensor-Based Dictionary Learning for Multidimensional Sparse Recovery Florian Römer and Giovanni Del Galdo 2nd CoSeRa, Bonn, 17-19 Sept. 2013 Ilmenau University of Technology Institute for Information
More informationSUMMARY. H (ω)u(ω,x s ;x) :=
Interpolating solutions of the Helmholtz equation with compressed sensing Tim TY Lin*, Evgeniy Lebed, Yogi A Erlangga, and Felix J Herrmann, University of British Columbia, EOS SUMMARY We present an algorithm
More informationMatrix-Product-States/ Tensor-Trains
/ Tensor-Trains November 22, 2016 / Tensor-Trains 1 Matrices What Can We Do With Matrices? Tensors What Can We Do With Tensors? Diagrammatic Notation 2 Singular-Value-Decomposition 3 Curse of Dimensionality
More informationSparsity Matters. Robert J. Vanderbei September 20. IDA: Center for Communications Research Princeton NJ.
Sparsity Matters Robert J. Vanderbei 2017 September 20 http://www.princeton.edu/ rvdb IDA: Center for Communications Research Princeton NJ The simplex method is 200 times faster... The simplex method is
More informationDealing with curse and blessing of dimensionality through tensor decompositions
Dealing with curse and blessing of dimensionality through tensor decompositions Lieven De Lathauwer Joint work with Nico Vervliet, Martijn Boussé and Otto Debals June 26, 2017 2 Overview Curse of dimensionality
More informationRecovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm
Recovery of Sparse Signals from Noisy Measurements Using an l p -Regularized Least-Squares Algorithm J. K. Pant, W.-S. Lu, and A. Antoniou University of Victoria August 25, 2011 Compressive Sensing 1 University
More informationTensor data analysis Part 2
Tensor data analysis Part 2 Mariya Ishteva Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Outline Last Dme:! MoDvaDon! Basic concepts! Basic tensor decomposidons Today:! Other useful decomposidons!
More informationTENSOR APPROXIMATION TOOLS FREE OF THE CURSE OF DIMENSIONALITY
TENSOR APPROXIMATION TOOLS FREE OF THE CURSE OF DIMENSIONALITY Eugene Tyrtyshnikov Institute of Numerical Mathematics Russian Academy of Sciences (joint work with Ivan Oseledets) WHAT ARE TENSORS? Tensors
More informationSparse Solutions of an Undetermined Linear System
1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationIntroduction to the Tensor Train Decomposition and Its Applications in Machine Learning
Introduction to the Tensor Train Decomposition and Its Applications in Machine Learning Anton Rodomanov Higher School of Economics, Russia Bayesian methods research group (http://bayesgroup.ru) 14 March
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationCombining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation
UIUC CSL Mar. 24 Combining Sparsity with Physically-Meaningful Constraints in Sparse Parameter Estimation Yuejie Chi Department of ECE and BMI Ohio State University Joint work with Yuxin Chen (Stanford).
More informationsublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)
sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU) 0 overview Our Contributions: 1 overview Our Contributions: A near optimal low-rank
More informationAvailable Ph.D position in Big data processing using sparse tensor representations
Available Ph.D position in Big data processing using sparse tensor representations Research area: advanced mathematical methods applied to big data processing. Keywords: compressed sensing, tensor models,
More informationDynamical low-rank approximation
Dynamical low-rank approximation Christian Lubich Univ. Tübingen Genève, Swiss Numerical Analysis Day, 17 April 2015 Coauthors Othmar Koch 2007, 2010 Achim Nonnenmacher 2008 Dajana Conte 2010 Thorsten
More informationAccelerated large-scale inversion with message passing Felix J. Herrmann, the University of British Columbia, Canada
Accelerated large-scale inversion with message passing Felix J. Herrmann, the University of British Columbia, Canada SUMMARY To meet current-day challenges, exploration seismology increasingly relies on
More informationTensors and graphical models
Tensors and graphical models Mariya Ishteva with Haesun Park, Le Song Dept. ELEC, VUB Georgia Tech, USA INMA Seminar, May 7, 2013, LLN Outline Tensors Random variables and graphical models Tractable representations
More informationAdaptive linear prediction filtering for random noise attenuation Mauricio D. Sacchi* and Mostafa Naghizadeh, University of Alberta
Adaptive linear prediction filtering for random noise attenuation Mauricio D. Sacchi* and Mostafa Naghizadeh, University of Alberta SUMMARY We propose an algorithm to compute time and space variant prediction
More informationA Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem
A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem Morteza Ashraphijuo Columbia University ashraphijuo@ee.columbia.edu Xiaodong Wang Columbia University wangx@ee.columbia.edu
More informationDynamical low rank approximation in hierarchical tensor format
Dynamical low rank approximation in hierarchical tensor format R. Schneider (TUB Matheon) John von Neumann Lecture TU Munich, 2012 Motivation Equations describing complex systems with multi-variate solution
More informationOptimization for Compressed Sensing
Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve
More informationSample ECE275A Midterm Exam Questions
Sample ECE275A Midterm Exam Questions The questions given below are actual problems taken from exams given in in the past few years. Solutions to these problems will NOT be provided. These problems and
More informationSUMMARY. The main contribution of this paper is threefold. First, we show that FWI based on the Helmholtz equation has the advantage
Randomized full-waveform inversion: a dimenstionality-reduction approach Peyman P. Moghaddam and Felix J. Herrmann, University of British Columbia, Canada SUMMARY Full-waveform inversion relies on the
More informationMachine Learning for Signal Processing Sparse and Overcomplete Representations. Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013
Machine Learning for Signal Processing Sparse and Overcomplete Representations Bhiksha Raj (slides from Sourish Chaudhuri) Oct 22, 2013 1 Key Topics in this Lecture Basics Component-based representations
More informationAdaptive low-rank approximation in hierarchical tensor format using least-squares method
Workshop on Challenges in HD Analysis and Computation, San Servolo 4/5/2016 Adaptive low-rank approximation in hierarchical tensor format using least-squares method Anthony Nouy Ecole Centrale Nantes,
More informationHierarchical Matrices. Jon Cockayne April 18, 2017
Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications
More informationLecture Notes: Geometric Considerations in Unconstrained Optimization
Lecture Notes: Geometric Considerations in Unconstrained Optimization James T. Allison February 15, 2006 The primary objectives of this lecture on unconstrained optimization are to: Establish connections
More informationLearning goals: students learn to use the SVD to find good approximations to matrices and to compute the pseudoinverse.
Application of the SVD: Compression and Pseudoinverse Learning goals: students learn to use the SVD to find good approximations to matrices and to compute the pseudoinverse. Low rank approximation One
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationRecitation 1. Gradients and Directional Derivatives. Brett Bernstein. CDS at NYU. January 21, 2018
Gradients and Directional Derivatives Brett Bernstein CDS at NYU January 21, 2018 Brett Bernstein (CDS at NYU) Recitation 1 January 21, 2018 1 / 23 Initial Question Intro Question Question We are given
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationRanking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization
Ranking from Crowdsourced Pairwise Comparisons via Matrix Manifold Optimization Jialin Dong ShanghaiTech University 1 Outline Introduction FourVignettes: System Model and Problem Formulation Problem Analysis
More informationAn Introduction to Hierachical (H ) Rank and TT Rank of Tensors with Examples
An Introduction to Hierachical (H ) Rank and TT Rank of Tensors with Examples Lars Grasedyck and Wolfgang Hackbusch Bericht Nr. 329 August 2011 Key words: MSC: hierarchical Tucker tensor rank tensor approximation
More informationProvable Alternating Minimization Methods for Non-convex Optimization
Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish
More informationDictionary Learning for photo-z estimation
Dictionary Learning for photo-z estimation Joana Frontera-Pons, Florent Sureau, Jérôme Bobin 5th September 2017 - Workshop Dictionary Learning on Manifolds MOTIVATION! Goal : Measure the radial positions
More informationMobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti
Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes
More information10-725/36-725: Convex Optimization Prerequisite Topics
10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the
More informationThe multiple-vector tensor-vector product
I TD MTVP C KU Leuven August 29, 2013 In collaboration with: N Vanbaelen, K Meerbergen, and R Vandebril Overview I TD MTVP C 1 Introduction Inspiring example Notation 2 Tensor decompositions The CP decomposition
More informationA fast randomized algorithm for approximating an SVD of a matrix
A fast randomized algorithm for approximating an SVD of a matrix Joint work with Franco Woolfe, Edo Liberty, and Vladimir Rokhlin Mark Tygert Program in Applied Mathematics Yale University Place July 17,
More informationNUMERICAL METHODS WITH TENSOR REPRESENTATIONS OF DATA
NUMERICAL METHODS WITH TENSOR REPRESENTATIONS OF DATA Institute of Numerical Mathematics of Russian Academy of Sciences eugene.tyrtyshnikov@gmail.com 2 June 2012 COLLABORATION MOSCOW: I.Oseledets, D.Savostyanov
More informationRobust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds
Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof.
More informationCompressive Sensing of Sparse Tensor
Shmuel Friedland Univ. Illinois at Chicago Matheon Workshop CSA2013 December 12, 2013 joint work with Q. Li and D. Schonfeld, UIC Abstract Conventional Compressive sensing (CS) theory relies on data representation
More informationCompressive Imaging by Generalized Total Variation Minimization
1 / 23 Compressive Imaging by Generalized Total Variation Minimization Jie Yan and Wu-Sheng Lu Department of Electrical and Computer Engineering University of Victoria, Victoria, BC, Canada APCCAS 2014,
More informationNumerical Linear and Multilinear Algebra in Quantum Tensor Networks
Numerical Linear and Multilinear Algebra in Quantum Tensor Networks Konrad Waldherr October 20, 2013 Joint work with Thomas Huckle QCCC 2013, Prien, October 20, 2013 1 Outline Numerical (Multi-) Linear
More informationIterative Methods. Splitting Methods
Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition
More informationCS60021: Scalable Data Mining. Dimensionality Reduction
J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org 1 CS60021: Scalable Data Mining Dimensionality Reduction Sourangshu Bhattacharya Assumption: Data lies on or near a
More informationThe convex algebraic geometry of rank minimization
The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationUncertainty quantification for Wavefield Reconstruction Inversion
using a PDE free semidefinite Hessian and randomize-then-optimize method Zhilong Fang *, Chia Ying Lee, Curt Da Silva, Tristan van Leeuwen and Felix J. Herrmann Seismic Laboratory for Imaging and Modeling
More informationTensor networks, TT (Matrix Product States) and Hierarchical Tucker decomposition
Tensor networks, TT (Matrix Product States) and Hierarchical Tucker decomposition R. Schneider (TUB Matheon) John von Neumann Lecture TU Munich, 2012 Setting - Tensors V ν := R n, H d = H := d ν=1 V ν
More informationSolving linear equations with Gaussian Elimination (I)
Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian
More informationA Randomized Approach for Crowdsourcing in the Presence of Multiple Views
A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion
More informationOn the convergence of higher-order orthogonality iteration and its extension
On the convergence of higher-order orthogonality iteration and its extension Yangyang Xu IMA, University of Minnesota SIAM Conference LA15, Atlanta October 27, 2015 Best low-multilinear-rank approximation
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationUncertainty quantification for inverse problems with a weak wave-equation constraint
Uncertainty quantification for inverse problems with a weak wave-equation constraint Zhilong Fang*, Curt Da Silva*, Rachel Kuske** and Felix J. Herrmann* *Seismic Laboratory for Imaging and Modeling (SLIM),
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationLinear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4
Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix
More informationMinimum entropy deconvolution with frequency-domain constraints
GEOPHYSICS, VOL. 59, NO. 6 (JUNE 1994); P. 938-945, 9 FIGS., 1 TABLE. Minimum entropy deconvolution with frequency-domain constraints Mauricio D. Sacchi*, Danilo R. Velis*, and Alberto H. Cominguez ABSTRACT
More informationSparse & Redundant Signal Representation, and its Role in Image Processing
Sparse & Redundant Signal Representation, and its Role in Michael Elad The CS Department The Technion Israel Institute of technology Haifa 3000, Israel Wave 006 Wavelet and Applications Ecole Polytechnique
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationPractical Linear Algebra: A Geometry Toolbox
Practical Linear Algebra: A Geometry Toolbox Third edition Chapter 12: Gauss for Linear Systems Gerald Farin & Dianne Hansford CRC Press, Taylor & Francis Group, An A K Peters Book www.farinhansford.com/books/pla
More informationTensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia
Network Feature s Decompositions for Machine Learning 1 1 Department of Computer Science University of British Columbia UBC Machine Learning Group, June 15 2016 1/30 Contact information Network Feature
More informationA Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices
A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices Ryota Tomioka 1, Taiji Suzuki 1, Masashi Sugiyama 2, Hisashi Kashima 1 1 The University of Tokyo 2 Tokyo Institute of Technology 2010-06-22
More information/16/$ IEEE 1728
Extension of the Semi-Algebraic Framework for Approximate CP Decompositions via Simultaneous Matrix Diagonalization to the Efficient Calculation of Coupled CP Decompositions Kristina Naskovska and Martin
More informationSection 1.1: Systems of Linear Equations. A linear equation: a 1 x 1 a 2 x 2 a n x n b. EXAMPLE: 4x 1 5x 2 2 x 1 and x x 1 x 3
Section 1.1: Systems of Linear Equations A linear equation: a 1 x 1 a 2 x 2 a n x n b EXAMPLE: 4x 1 5x 2 2 x 1 and x 2 2 6 x 1 x 3 rearranged rearranged 3x 1 5x 2 2 2x 1 x 2 x 3 2 6 Not linear: 4x 1 6x
More informationCS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares
CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search
More informationCompressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles
Or: the equation Ax = b, revisited University of California, Los Angeles Mahler Lecture Series Acquiring signals Many types of real-world signals (e.g. sound, images, video) can be viewed as an n-dimensional
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More information