Experiences with Model Reduction and Interpolation
|
|
- Kristian Hopkins
- 6 years ago
- Views:
Transcription
1 Experiences with Model Reduction and Interpolation Paul Constantine Stanford University, Sandia National Laboratories Qiqi Wang (MIT) David Gleich (Purdue) Emory University March 7, 2012 Joe Ruthruff (SNL) Jeremy Templeton (SNL)
2 PA = LU = PAU 1 L 1 = I Education & Positions B.A. Math (U North Texas) M.S. & Ph.D. CME (Stanford) von Neumann Fellow (Sandia) Postdoc (Stanford, ME) Recent work Sparse Fourier transforms via Smolyak s algorithm Input dimension reduction for multivariate approximation Polynomial approximation for composite functions (via Lanczos) Stanford President John Hennessy Tony The Tiger me Application areas Turbulent flows Climate models Coupled multiphysics systems Paul G. Constantine
3 (inspiration for) Just ask a good question! Paul G. Constantine
4 Paul G. Constantine Motivation Uncertainty quantification Parameterized models and model reduction Two approaches: Optimization Interpolation A few examples
5 How can we test systems without testing them? Simulation! How do we get measures of confidence from simulations? Uncertainty quantification! 1996 Nuclear Test Ban Treaty Paul G. Constantine 1
6 differential operators / forcings / boundaries space L(u; x, t, ω) =0 randomness solution time Can we compute statistics of the solution with respect to the randomness? Paul G. Constantine 2
7 differential operators / forcings / boundaries space L(u; x, t, s) =0 parameters solution time Can we compute statistics of the solution with respect to the parameters? Each realization corresponds to a point in the parameter space, where we solve the differential equation. Paul G. Constantine 3
8 space time input parameters solution u(x, t, s) Given input parameters, a physical simulation approximates a space/time dependent solution. Each solution is computationally expensive, so exhaustive parameter studies (e.g., UQ) are infeasible. Cheaper reduced order models enable guesstimates for parameter inquiries. Paul G. Constantine 4
9 discretized differential operator discretized solution A(s)x(s) =b(s) forcing term Paul G. Constantine 5
10 discretized differential operator discretized solution A(s)x(s) =b(s) forcing term Suppose we have solved this system for some parameter values. Can we use those runs to approximate the solution at other points in the parameter space (more cheaply or efficiently)? For each we compute x j = x(s j ) that solves A(s j )x j = b(s j ). Then we approximate x = x(s) n a j x j j=1 How do we compute the coefficients? = Xa Paul G. Constantine 5
11 minimize a Xa x 2 = X T Xa = X T x It s a lot like Kriging Gaussian process emulation Radial basis functions where X T X approximates the correlation model. But what s the problem here? Paul G. Constantine 6
12 Paul G. Constantine Motivation Uncertainty quantification Parameterized models and model reduction Two approaches: Optimization Interpolation A few examples
13 For A = A(s) and b = b(s) minimize a AXa b 2 Funny story about this And then there s Galerkin X T AXa = X T b *Ahem* What about nonlinear problems? Paul G. Constantine 7
14 Full system: f(x,s)=0 Interpolatory system: minimize a f(xa,s) 2 A nonlinear least-squares problem! And time dependence? Paul G. Constantine 8
15 Full system: dx dt = f(x,s) minimize a minimize a Interpolatory system: T 2 d (Xa) f(xa,s) 0 dt dt 2 2 d dt (X ia) f(x i a,s) i 2 A bigger nonlinear least-squares problem (when discretized in time)! Constantine and Wang. Residual Minimizing Model Interpolation for Parameterized Nonlinear Dynamical Systems. In review. Paul G. Constantine 9
16 Point-wise optimality by construction. Norm of residual is natural measure of point-wise misfit. Implementation is independent of dimension of parameter space. Interpolates data with continuous basis functions. Yields a straightforward strategy for adaptively adding bases. Full reuse of previously computed solutions. Each additional basis enhances the approximation globally. Monotonic decrease of the maximum residual (over parameter space). Exact point-wise convergence (in exact arithmetic) after finite model evaluations (although this is a little ridiculous). A rank-revealing property. Easy to implement using only residuals. Paul G. Constantine 10
17 Sandia Research Project Questions: What is the probability of failure? Which input values cause failure? Challenges: Each simulation is expensive (30 minutes on 16 cores) Large scale nonlinear, time dependent heat transfer problem nodes, 10 3 time steps Approach: Construct an interpolating reduced order model from a budgetconstrained ensemble of runs for uncertainty and optimization studies. Paul G. Constantine 11
18 Glory: 4,608 cores 32 GB RAM/core MapReduce cluster: 256 cores 2 TB storage/core Researcher: Infinite curiosity! Each multi-day HPC simulation generates gigabytes of data. A MapReduce cluster can hold hundreds or thousands of simulations enabling engineers to query and analyze simulation data for statistical studies and uncertainty quantification. Paul G. Constantine 12
19 1. Getting the residual out of existing codes was no easy task. 2. For some simple numerical examples, solving the nonlinear least squares problem was as costly as solving the full model. Approach: Reformulate the interpolation scheme as a non-intrusive method, sacrificing accuracy for an increase in performance. Paul G. Constantine 13
20 Paul G. Constantine Motivation Uncertainty quantification Parameterized models and model reduction Two approaches: Optimization Interpolation A few examples
21 solution space/time bases u(x, t, s) = singular values φ i (x, t) σ i v i (s) parameter bases i=1 Exponential decay of singular values for smooth functions. Uniform convergence. See the work of P.C. Hansen on integral operator kernels. See also the Karhunen-Loeve expansion. Paul G. Constantine 14
22 fixed space/time discretization x(s) r u i σ i v i (s) i=1 Paul G. Constantine 15
23 fixed space/time discretization x(s) r u i σ i v i (s) i=1 experimental design SVD x(s1 ) x(s n ) = X = UΣV T Paul G. Constantine 15
24 fixed space/time discretization x(s) r u i σ i v i (s) i=1 experimental design SVD x(s1 ) x(s n ) = X = UΣV T Paul G. Constantine 15
25 fixed space/time discretization x(s) r u i σ i v i (s) i=1 experimental design SVD x(s1 ) x(s n ) = X = UΣV T Paul G. Constantine 15
26 fixed space/time discretization x(s) r u i σ i v i (s) i=1 experimental design SVD x(s1 ) x(s n ) = X = UΣV T Paul G. Constantine 15
27 fixed space/time discretization x(s) r u i σ i v i (s) i=1 experimental design SVD x(s1 ) x(s n ) = X = UΣV T treat each right singular vector as samples of the unknown basis functions! V ij = v i (s j ) Paul G. Constantine 15
28 vi (s j ) ṽ i (s) polynomials radial basis functions gaussian processes piecewise linear splines Paul G. Constantine 16
29 Which section would you rather try to interpolate, A or B? A B Paul G. Constantine 17
30 Folk Theorem (O Leary, 2011) The singular vectors of a matrix (with smooth data) become more oscillatory as the index increases. Paul G. Constantine 18
31 Folk Theorem (O Leary, 2011) The singular vectors of a matrix (with smooth data) become more oscillatory as the index increases. This implies The gradient of the singular vectors increases as the index increases. Paul G. Constantine 18
32 Folk Theorem (O Leary, 2011) The singular vectors of a matrix (with smooth data) become more oscillatory as the index increases. This implies The gradient of the singular vectors increases as the index increases. PREDICTABLE v 1 (s),..., v t (s) UNPREDICTABLE v t+1 (s),..., v r (s) Paul G. Constantine 18
33 predictable unpredictable x(s) t(s) u i σ i ṽ i (s) + r u i σ i η i t(s) i=1 i=t(s)+1 η N(0, 1) Space/time dependent prediction variance at each interpolation: var[x] = diag r u i σi 2 u T i i=t(s)+1 Paul G. Constantine 19
34 1 σ 1 t(s) τ i=1 is the largest σ i v i s τ such that < threshold This local truncation can be used as an importance sampling measure to adaptively refine the approximation. Paul G. Constantine 20
35 Paul G. Constantine Motivation Parameterized models and model reduction Two approaches: Optimization Interpolation A few examples
36 Model: d a(x, s) du dx dx Boundary conditions: u(0) = u(1) = 1 =1 Coefficients: a(x, s) =1+4s (x 2 x) Parameter space: s [0, 1 γ], γ > 0 There s a singularity at x =0.5 s =1 Solution: u(x, s) = 1 8s log 1+4s (x 2 x) Paul G. Constantine 21
37 x(s1 ) x(s n ) = UΣV T Paul G. Constantine 22
38 Paul G. Constantine 23
39 Paul G. Constantine 24
40 Paul G. Constantine 25
41 Heat equation: ρ t (c pt )= (k T ) Material properties: Paul G. Constantine 26
42 Heat equation: ρ t (c pt )= (k T ) Material properties: input parameters Paul G. Constantine 26
43 Heat equation: ρ t (c pt )= (k T ) Material properties: input parameters Paul G. Constantine 26
44 x(s1 ) x(s n ) = UΣV T 20 point Latin hypercube design Paul G. Constantine 27
45 Model Reduction x(s1 ) x(s n ) = UΣV T 20 point Latin hypercube design Make sure to ask me about higher dimensions. Paul G. Constantine 27
46 Model Reduction x(s1 ) x(s n ) = UΣV T Paul G. Constantine 28
47 1 σ 1 t i=1 σ i v i ξ There is probably a better way to do this. 10,000 samples of the weighted sum of the norm of the gradient Paul G. Constantine 29
48 1 σ 1 t i=1 σ i v i ξ There is probably a better way to do this. Paul G. Constantine 29
49 1 σ 1 t i=1 σ i v i ξ < threshold There is probably a better way to do this. Paul G. Constantine 29
50 1,585 randomly sampled cross-validation tests Paul G. Constantine 30
51 truth Where is the error the worst? ROM Paul G. Constantine 31
52 prediction variance Where is the error the worst? error Paul G. Constantine 31
53 Qualitatively compare the parameter distribution of the crossvalidation error to the metric that determines truncation. Paul G. Constantine 32
54 Use the truncation metric as a measure for importance sampling to refine the ROM in areas of large error. Paul G. Constantine 32
55 1,585 Cross-validation Tests Original Design (20 LHS) Refined Design (20 LHS Importance) Mean Relative Error Max Relative Error Paul G. Constantine 33
56 Full Model Reduced Model Prediction Variance Nonlinear heat transfer model 80k nodes, 300 time-steps 104 basis runs SVD of 24m x 104 data matrix 500x reduction in wall clock time Paul G. Constantine 34
57 A framework for model reduction Construct the time/space bases with SVD of sampled outputs. Interpolate the parameter bases. Use the gradient to determine if the interpolant is predictive. Use unpredictability as an importance sampling measure for refinement. Some interesting questions When will this work? (Rapid decay of singular values.) What s the relationship between the predictability measure and the error? When do the singular vectors behave like that? Spatial/temporal refinement? How the heck do you scale the SVD computation? Paul G. Constantine 35
58 Exascale computing requires exascale data management! Supercomputer MapReduce cluster Researcher Paul G. Constantine 36
59 QUESTIONS? Paul G. Constantine 37
60 We have implemented a QR factorization for tall and skinny matrices in MapReduce. A 1 A 2 A 3 A 4 = Q 1 Q 2 Q 4 8n 4n =Q R 1 R 2 R 3 R 4 4n n Q 1 = Q 2 Q 3 Q R github.com/dgleich/mrtsqr! Q 4n n n n 4 8n 4n The Tall, Skinny QR gives us the R factor. We compute Q = XR 1 with another MapReduce call. Demmel, et al. Communication-avoiding parallel and sequential QR factorizations Constantine and Gleich. Tall and Skinny QR factorizations in MapReduce Architectures. MAPREDUCE, '11. Paul G. Constantine 38
61 Reality Prediction/ Validation Modeling Computer Simulation Mathematical Model Verification Paul G. Constantine
Tall-and-skinny! QRs and SVDs in MapReduce
A 1 A 2 A 3 Tall-and-skinny! QRs and SVDs in MapReduce Yangyang Hou " Purdue, CS Austin Benson " Stanford University Paul G. Constantine Col. School. Mines " Joe Nichols U. of Minn James Demmel " UC Berkeley
More informationConcepts in Global Sensitivity Analysis IMA UQ Short Course, June 23, 2015
Concepts in Global Sensitivity Analysis IMA UQ Short Course, June 23, 2015 A good reference is Global Sensitivity Analysis: The Primer. Saltelli, et al. (2008) Paul Constantine Colorado School of Mines
More informationKarhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques
Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,
More informationACTIVE SUBSPACE METHODS IN THEORY AND PRACTICE: APPLICATIONS TO KRIGING SURFACES
ACTIVE SUBSPACE METHODS IN THEORY AND PRACTICE: APPLICATIONS TO KRIGING SURFACES PAUL G. CONSTANTINE, ERIC DOW, AND QIQI WANG Abstract. Many multivariate functions encountered in practical engineering
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationAdaptive Collocation with Kernel Density Estimation
Examples of with Kernel Density Estimation Howard C. Elman Department of Computer Science University of Maryland at College Park Christopher W. Miller Applied Mathematics and Scientific Computing Program
More informationNONLOCALITY AND STOCHASTICITY TWO EMERGENT DIRECTIONS FOR APPLIED MATHEMATICS. Max Gunzburger
NONLOCALITY AND STOCHASTICITY TWO EMERGENT DIRECTIONS FOR APPLIED MATHEMATICS Max Gunzburger Department of Scientific Computing Florida State University North Carolina State University, March 10, 2011
More informationUncertainty Quantification in Computational Science
DTU 2010 - Lecture I Uncertainty Quantification in Computational Science Jan S Hesthaven Brown University Jan.Hesthaven@Brown.edu Objective of lectures The main objective of these lectures are To offer
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationPolynomial Chaos and Karhunen-Loeve Expansion
Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the
More informationUncertainty Quantification in Viscous Hypersonic Flows using Gradient Information and Surrogate Modeling
Uncertainty Quantification in Viscous Hypersonic Flows using Gradient Information and Surrogate Modeling Brian A. Lockwood, Markus P. Rumpfkeil, Wataru Yamazaki and Dimitri J. Mavriplis Mechanical Engineering
More informationSimultaneous Robust Design and Tolerancing of Compressor Blades
Simultaneous Robust Design and Tolerancing of Compressor Blades Eric Dow Aerospace Computational Design Laboratory Department of Aeronautics and Astronautics Massachusetts Institute of Technology GTL Seminar
More informationAn Empirical Chaos Expansion Method for Uncertainty Quantification
An Empirical Chaos Expansion Method for Uncertainty Quantification Melvin Leok and Gautam Wilkins Abstract. Uncertainty quantification seeks to provide a quantitative means to understand complex systems
More informationMatrix assembly by low rank tensor approximation
Matrix assembly by low rank tensor approximation Felix Scholz 13.02.2017 References Angelos Mantzaflaris, Bert Juettler, Boris Khoromskij, and Ulrich Langer. Matrix generation in isogeometric analysis
More informationPredictive Engineering and Computational Sciences. Local Sensitivity Derivative Enhanced Monte Carlo Methods. Roy H. Stogner, Vikram Garg
PECOS Predictive Engineering and Computational Sciences Local Sensitivity Derivative Enhanced Monte Carlo Methods Roy H. Stogner, Vikram Garg Institute for Computational Engineering and Sciences The University
More informationStochastic Collocation Methods for Polynomial Chaos: Analysis and Applications
Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Taylor s Theorem Can often approximate a function by a polynomial The error in the approximation
More informationCS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation
Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80
More informationLecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko
tifica Lecture 1: Introduction to low-rank tensor representation/approximation Alexander Litvinenko http://sri-uq.kaust.edu.sa/ KAUST Figure : KAUST campus, 5 years old, approx. 7000 people (include 1400
More informationUtilizing Adjoint-Based Techniques to Improve the Accuracy and Reliability in Uncertainty Quantification
Utilizing Adjoint-Based Techniques to Improve the Accuracy and Reliability in Uncertainty Quantification Tim Wildey Sandia National Laboratories Center for Computing Research (CCR) Collaborators: E. Cyr,
More informationComputer simulation of multiscale problems
Progress in the SSF project CutFEM, Geometry, and Optimal design Computer simulation of multiscale problems Axel Målqvist and Daniel Elfverson University of Gothenburg and Uppsala University Umeå 2015-05-20
More informationHyperbolic Polynomial Chaos Expansion (HPCE) and its Application to Statistical Analysis of Nonlinear Circuits
Hyperbolic Polynomial Chaos Expansion HPCE and its Application to Statistical Analysis of Nonlinear Circuits Majid Ahadi, Aditi Krishna Prasad, Sourajeet Roy High Speed System Simulations Laboratory Department
More informationH 2 -matrices with adaptive bases
1 H 2 -matrices with adaptive bases Steffen Börm MPI für Mathematik in den Naturwissenschaften Inselstraße 22 26, 04103 Leipzig http://www.mis.mpg.de/ Problem 2 Goal: Treat certain large dense matrices
More informationSTOCHASTIC SAMPLING METHODS
STOCHASTIC SAMPLING METHODS APPROXIMATING QUANTITIES OF INTEREST USING SAMPLING METHODS Recall that quantities of interest often require the evaluation of stochastic integrals of functions of the solutions
More informationDiscrete Ill Posed and Rank Deficient Problems. Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1
Discrete Ill Posed and Rank Deficient Problems Alistair Boyle, Feb 2009, SYS5906: Directed Studies Inverse Problems 1 Definitions Overview Inversion, SVD, Picard Condition, Rank Deficient, Ill-Posed Classical
More informationEvaluation of Non-Intrusive Approaches for Wiener-Askey Generalized Polynomial Chaos
49th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference 6t 7 - April 28, Schaumburg, IL AIAA 28-892 Evaluation of Non-Intrusive Approaches for Wiener-Askey Generalized
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationarxiv: v2 [math.na] 5 Dec 2013
ACTIVE SUBSPACE METHODS IN THEORY AND PRACTICE: APPLICATIONS TO KRIGING SURFACES PAUL G. CONSTANTINE, ERIC DOW, AND QIQI WANG arxiv:1304.2070v2 [math.na 5 Dec 2013 Abstract. Many multivariate functions
More informationSensitivity Analysis of a Nuclear Reactor System Finite Element Model
Westinghouse Non-Proprietary Class 3 Sensitivity Analysis of a Nuclear Reactor System Finite Element Model 2018 ASME Verification & Validation Symposium VVS2018-9306 Gregory A. Banyay, P.E. Stephen D.
More informationLinear Models for Regression. Sargur Srihari
Linear Models for Regression Sargur srihari@cedar.buffalo.edu 1 Topics in Linear Regression What is regression? Polynomial Curve Fitting with Scalar input Linear Basis Function Models Maximum Likelihood
More informationSampling and Low-Rank Tensor Approximations
Sampling and Low-Rank Tensor Approximations Hermann G. Matthies Alexander Litvinenko, Tarek A. El-Moshely +, Brunswick, Germany + MIT, Cambridge, MA, USA wire@tu-bs.de http://www.wire.tu-bs.de $Id: 2_Sydney-MCQMC.tex,v.3
More informationProblem 1: Toolbox (25 pts) For all of the parts of this problem, you are limited to the following sets of tools:
CS 322 Final Exam Friday 18 May 2007 150 minutes Problem 1: Toolbox (25 pts) For all of the parts of this problem, you are limited to the following sets of tools: (A) Runge-Kutta 4/5 Method (B) Condition
More informationLeast Squares Approximation
Chapter 6 Least Squares Approximation As we saw in Chapter 5 we can interpret radial basis function interpolation as a constrained optimization problem. We now take this point of view again, but start
More informationGeostatistical Modeling for Large Data Sets: Low-rank methods
Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank
More informationSparse polynomial chaos expansions in engineering applications
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Sparse polynomial chaos expansions in engineering applications B. Sudret G. Blatman (EDF R&D,
More informationPolynomial chaos expansions for structural reliability analysis
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for structural reliability analysis B. Sudret & S. Marelli Incl.
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationSampling and low-rank tensor approximation of the response surface
Sampling and low-rank tensor approximation of the response surface tifica Alexander Litvinenko 1,2 (joint work with Hermann G. Matthies 3 ) 1 Group of Raul Tempone, SRI UQ, and 2 Group of David Keyes,
More informationA Stochastic Collocation based. for Data Assimilation
A Stochastic Collocation based Kalman Filter (SCKF) for Data Assimilation Lingzao Zeng and Dongxiao Zhang University of Southern California August 11, 2009 Los Angeles Outline Introduction SCKF Algorithm
More informationEstimating functional uncertainty using polynomial chaos and adjoint equations
0. Estimating functional uncertainty using polynomial chaos and adjoint equations February 24, 2011 1 Florida State University, Tallahassee, Florida, Usa 2 Moscow Institute of Physics and Technology, Moscow,
More informationLeast Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. under-determined
More informationResidual Minimizing Model Interpolation for Parameterized Nonlinear Dynamical Systems
Residual Minimizing Model Interpolation for Parameterized Nonlinear Dynamical Systems The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters.
More informationAccuracy, Precision and Efficiency in Sparse Grids
John, Information Technology Department, Virginia Tech.... http://people.sc.fsu.edu/ jburkardt/presentations/ sandia 2009.pdf... Computer Science Research Institute, Sandia National Laboratory, 23 July
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationCME 345: MODEL REDUCTION
CME 345: MODEL REDUCTION Proper Orthogonal Decomposition (POD) Charbel Farhat & David Amsallem Stanford University cfarhat@stanford.edu 1 / 43 Outline 1 Time-continuous Formulation 2 Method of Snapshots
More informationLecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.
Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference
More informationCSC 576: Linear System
CSC 576: Linear System Ji Liu Department of Computer Science, University of Rochester September 3, 206 Linear Equations Consider solving linear equations where A R m n and b R n m and n could be extremely
More informationA Statistical Look at Spectral Graph Analysis. Deep Mukhopadhyay
A Statistical Look at Spectral Graph Analysis Deep Mukhopadhyay Department of Statistics, Temple University Office: Speakman 335 deep@temple.edu http://sites.temple.edu/deepstat/ Graph Signal Processing
More informationAPPLIED NUMERICAL LINEAR ALGEBRA
APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation
More informationLeast-squares data fitting
EE263 Autumn 2015 S. Boyd and S. Lall Least-squares data fitting 1 Least-squares data fitting we are given: functions f 1,..., f n : S R, called regressors or basis functions data or measurements (s i,
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 34: Improving the Condition Number of the Interpolation Matrix Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu
More information. Frobenius-Perron Operator ACC Workshop on Uncertainty Analysis & Estimation. Raktim Bhattacharya
.. Frobenius-Perron Operator 2014 ACC Workshop on Uncertainty Analysis & Estimation Raktim Bhattacharya Laboratory For Uncertainty Quantification Aerospace Engineering, Texas A&M University. uq.tamu.edu
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationLecture 4: Numerical solution of ordinary differential equations
Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor
More informationSpatial Process Estimates as Smoothers: A Review
Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed
More informationKernel B Splines and Interpolation
Kernel B Splines and Interpolation M. Bozzini, L. Lenarduzzi and R. Schaback February 6, 5 Abstract This paper applies divided differences to conditionally positive definite kernels in order to generate
More informationCME 345: MODEL REDUCTION
CME 345: MODEL REDUCTION Methods for Nonlinear Systems Charbel Farhat & David Amsallem Stanford University cfarhat@stanford.edu 1 / 65 Outline 1 Nested Approximations 2 Trajectory PieceWise Linear (TPWL)
More informationSchwarz Preconditioner for the Stochastic Finite Element Method
Schwarz Preconditioner for the Stochastic Finite Element Method Waad Subber 1 and Sébastien Loisel 2 Preprint submitted to DD22 conference 1 Introduction The intrusive polynomial chaos approach for uncertainty
More informationSlide05 Haykin Chapter 5: Radial-Basis Function Networks
Slide5 Haykin Chapter 5: Radial-Basis Function Networks CPSC 636-6 Instructor: Yoonsuck Choe Spring Learning in MLP Supervised learning in multilayer perceptrons: Recursive technique of stochastic approximation,
More informationModeling uncertainty in metric space. Jef Caers Stanford University Stanford, California, USA
Modeling uncertainty in metric space Jef Caers Stanford University Stanford, California, USA Contributors Celine Scheidt (Research Associate) Kwangwon Park (PhD student) Motivation Modeling uncertainty
More informationOne Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017
One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary
More informationMeshfree Approximation Methods with MATLAB
Interdisciplinary Mathematical Sc Meshfree Approximation Methods with MATLAB Gregory E. Fasshauer Illinois Institute of Technology, USA Y f? World Scientific NEW JERSEY LONDON SINGAPORE BEIJING SHANGHAI
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationBig Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang Tutorial@ACML 2015 Hong Kong Department of Computer Science, The University of Iowa, IA, USA Nov. 20, 2015 Yang Tutorial for ACML 15 Nov.
More informationLearning gradients: prescriptive models
Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationApplied Numerical Analysis Quiz #2
Applied Numerical Analysis Quiz #2 Modules 3 and 4 Name: Student number: DO NOT OPEN UNTIL ASKED Instructions: Make sure you have a machine-readable answer form. Write your name and student number in the
More informationTowards Reduced Order Modeling (ROM) for Gust Simulations
Towards Reduced Order Modeling (ROM) for Gust Simulations S. Görtz, M. Ripepi DLR, Institute of Aerodynamics and Flow Technology, Braunschweig, Germany Deutscher Luft und Raumfahrtkongress 2017 5. 7. September
More informationarxiv: v1 [cs.lg] 26 Jul 2017
Updating Singular Value Decomposition for Rank One Matrix Perturbation Ratnik Gandhi, Amoli Rajgor School of Engineering & Applied Science, Ahmedabad University, Ahmedabad-380009, India arxiv:70708369v
More informationPART II : Least-Squares Approximation
PART II : Least-Squares Approximation Basic theory Let U be an inner product space. Let V be a subspace of U. For any g U, we look for a least-squares approximation of g in the subspace V min f V f g 2,
More informationFast Numerical Methods for Stochastic Computations
Fast AreviewbyDongbinXiu May 16 th,2013 Outline Motivation 1 Motivation 2 3 4 5 Example: Burgers Equation Let us consider the Burger s equation: u t + uu x = νu xx, x [ 1, 1] u( 1) =1 u(1) = 1 Example:
More informationFast Direct Methods for Gaussian Processes
Fast Direct Methods for Gaussian Processes Mike O Neil Departments of Mathematics New York University oneil@cims.nyu.edu December 12, 2015 1 Collaborators This is joint work with: Siva Ambikasaran Dan
More informationNONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition
NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationLow-Rank Approximations, Random Sampling and Subspace Iteration
Modified from Talk in 2012 For RNLA study group, March 2015 Ming Gu, UC Berkeley mgu@math.berkeley.edu Low-Rank Approximations, Random Sampling and Subspace Iteration Content! Approaches for low-rank matrix
More informationUncertainty Quantification in Computational Models
Uncertainty Quantification in Computational Models Habib N. Najm Sandia National Laboratories, Livermore, CA, USA Workshop on Understanding Climate Change from Data (UCC11) University of Minnesota, Minneapolis,
More information[2] (a) Develop and describe the piecewise linear Galerkin finite element approximation of,
269 C, Vese Practice problems [1] Write the differential equation u + u = f(x, y), (x, y) Ω u = 1 (x, y) Ω 1 n + u = x (x, y) Ω 2, Ω = {(x, y) x 2 + y 2 < 1}, Ω 1 = {(x, y) x 2 + y 2 = 1, x 0}, Ω 2 = {(x,
More informationNumerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018
Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationIntegration, differentiation, and root finding. Phys 420/580 Lecture 7
Integration, differentiation, and root finding Phys 420/580 Lecture 7 Numerical integration Compute an approximation to the definite integral I = b Find area under the curve in the interval Trapezoid Rule:
More informationFinal Exam Practice 3, May 8, 2018 Math 21b, Spring Name:
Final Exam Practice 3, May 8, 8 Math b, Spring 8 Name: MWF 9 Oliver Knill MWF Jeremy Hahn MWF Hunter Spink MWF Matt Demers MWF Yu-Wen Hsu MWF Ben Knudsen MWF Sander Kupers MWF Hakim Walker TTH Ana Balibanu
More informationBorel Summability in PDE initial value problems
Borel Summability in PDE initial value problems Saleh Tanveer (Ohio State University) Collaborator Ovidiu Costin & Guo Luo Research supported in part by Institute for Math Sciences (IC), EPSRC & NSF. Main
More informationMulti-Objective Stochastic Optimization of Magnetic Fields. David S Bindel, Cornell University
Multi-Objective Stochastic Optimization of Magnetic Fields David S Bindel, Cornell University Current State of the Art (STELLOPT) Optimizer Calculate! " (physics + engineering targets) Adjust plasma boundary
More information256 Summary. D n f(x j ) = f j+n f j n 2n x. j n=1. α m n = 2( 1) n (m!) 2 (m n)!(m + n)!. PPW = 2π k x 2 N + 1. i=0?d i,j. N/2} N + 1-dim.
56 Summary High order FD Finite-order finite differences: Points per Wavelength: Number of passes: D n f(x j ) = f j+n f j n n x df xj = m α m dx n D n f j j n= α m n = ( ) n (m!) (m n)!(m + n)!. PPW =
More information( nonlinear constraints)
Wavelet Design & Applications Basic requirements: Admissibility (single constraint) Orthogonality ( nonlinear constraints) Sparse Representation Smooth functions well approx. by Fourier High-frequency
More informationFixed point iteration and root finding
Fixed point iteration and root finding The sign function is defined as x > 0 sign(x) = 0 x = 0 x < 0. It can be evaluated via an iteration which is useful for some problems. One such iteration is given
More informationIntroduction to Uncertainty Quantification in Computational Science Handout #3
Introduction to Uncertainty Quantification in Computational Science Handout #3 Gianluca Iaccarino Department of Mechanical Engineering Stanford University June 29 - July 1, 2009 Scuola di Dottorato di
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationLecture 8: February 9
0-725/36-725: Convex Optimiation Spring 205 Lecturer: Ryan Tibshirani Lecture 8: February 9 Scribes: Kartikeya Bhardwaj, Sangwon Hyun, Irina Caan 8 Proximal Gradient Descent In the previous lecture, we
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Lecture 9: Sensitivity Analysis ST 2018 Tobias Neckel Scientific Computing in Computer Science TUM Repetition of Previous Lecture Sparse grids in Uncertainty Quantification
More informationLinear Least-Squares Data Fitting
CHAPTER 6 Linear Least-Squares Data Fitting 61 Introduction Recall that in chapter 3 we were discussing linear systems of equations, written in shorthand in the form Ax = b In chapter 3, we just considered
More informationAN EFFICIENT COMPUTATIONAL FRAMEWORK FOR UNCERTAINTY QUANTIFICATION IN MULTISCALE SYSTEMS
AN EFFICIENT COMPUTATIONAL FRAMEWORK FOR UNCERTAINTY QUANTIFICATION IN MULTISCALE SYSTEMS A Dissertation Presented to the Faculty of the Graduate School of Cornell University in Partial Fulfillment of
More informationAProofoftheStabilityoftheSpectral Difference Method For All Orders of Accuracy
AProofoftheStabilityoftheSpectral Difference Method For All Orders of Accuracy Antony Jameson 1 1 Thomas V. Jones Professor of Engineering Department of Aeronautics and Astronautics Stanford University
More informationParameter Selection Techniques and Surrogate Models
Parameter Selection Techniques and Surrogate Models Model Reduction: Will discuss two forms Parameter space reduction Surrogate models to reduce model complexity Input Representation Local Sensitivity
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple
More informationRegularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution
Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationPascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids
Covering : Accuracy, Precision and Efficiency in Sparse Grids https://people.sc.fsu.edu/ jburkardt/presentations/auburn 2009.pdf... John Interdisciplinary Center for Applied Mathematics & Information Technology
More informationGradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering
Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering Brian A. Lockwood 1 and Mihai Anitescu 2 1 Department of Mechanical Engineering University of Wyoming 2 Mathematics
More information8 A pseudo-spectral solution to the Stokes Problem
8 A pseudo-spectral solution to the Stokes Problem 8.1 The Method 8.1.1 Generalities We are interested in setting up a pseudo-spectral method for the following Stokes Problem u σu p = f in Ω u = 0 in Ω,
More information