Thomas Santner Ohio State University Columbus OH. Joint Statistical Meetings Seattle, WA August 2006
|
|
- Dorcas Morris
- 5 years ago
- Views:
Transcription
1 Introductory Overview Lecture on Computer Experiments - The Modeling and Analysis of Data from Computer Experiments Thomas Santner Ohio State University Columbus OH Joint Statistical Meetings Seattle, WA August 2006
2 Outline 1. Experiments 2. Experimentation using Computer Codes 3. A Taxonomy of Problems 4. Gaussian Stochastic Process (GaSP) Models &Þ Prediction based on the GaSP Model 6. Conclusions-Take Home Messages
3 1. Experiments Physical Experiments Gold standard for establishing cause and effect relationships Mainstay of Agriculture, Industry, Medicine Principles of randomization, blocking, choice of sample size, and stochastic modeling of response variables all developed in response to needs of physical experiments
4 1. Experiments Physical Experiments Gold standard for establishing cause and effect relationships Mainstay of Agriculture, Industry, Medicine Principles of randomization, blocking, choice of sample size, and stochastic modeling of response variables all developed in response to needs of physical experiments Simulation Experiments Complex physical system each of whose parts interact in a known stochastic manner but whose ensemble is not understood analytically. Heavily used in Industrial Engineering--e.g., compare job shop set ups
5 1. Experiments Physical Experiments Gold standard for establishing cause and effect relationships Mainstay of Agriculture, Industry, Medicine Principles of randomization, blocking, choice of sample size, and stochastic modeling of response variables all developed in response to needs of physical experiments Simulation Experiments Complex physical system each of whose parts interact in a known stochastic manner but whose ensemble is not understood analytically. Heavily used in Industrial Engineering--e.g., compare job shop set ups Computer Experiments relatively new (below)
6 1. Experiments Physical Experiments Gold standard for establishing cause and effect relationships Mainstay of Agriculture, Industry, Medicine Principles of randomization, blocking, choice of sample size, and stochastic modeling of response variables all developed in response to needs of physical experiments Simulation Experiments Complex physical system each of whose parts interact in a known stochastic manner but whose ensemble is not understood analytically. Heavily used in Industrial Engineering--e.g., compare job shop set ups Computer Experiments relatively new (below) Combinations of the above
7 2. Experimentation using Computer Codes In some situations performing a physical experiment is not feasible 1. Physical process is technically too difficult to study 2. Number of variables is too large 3. Too expensive to study directly (it's all money) 4. Ethical considerations
8 2. Experimentation using Computer Codes In some situations performing a physical experiment is not feasible 1. Physical process is technically too difficult to study 2. Number of variables is too large 3. Too expensive to study directly (it's all money) 4. Ethical considerations When physical experiments are not possible, it may still be feasible to conduct a computer experiment
9 2. Experimentation using Computer Codes In some situations performing a physical experiment is not feasible When physical experiments are not possible, it may still be feasible to conduct a computer experiment IF the physical process relating the inputs B to the response(s) a. Can be described by a mathematical model relating the output, CÐBÑ, to B b. Numerical methods exist for solving the mathematical model c. The numerical methods can be implemented with computer code (in finite time!) THEN one can run the computer code to produce a "response'' CÐBÑ at any input B k., i.e., one can conduct a computer experiment B Ò Code Ò CÐBÑ The computer code is a proxy for the physical process.
10 Features of Computer Experiments B Ò Code Ò CÐBÑ CÐBÑ is deterministic
11 Features of Computer Experiments B Ò Code Ò CÐBÑ CÐBÑ is deterministic CÐB Ñ may be biased (hence the need for "calibration")
12 Features of Computer Experiments B Ò Code Ò CÐBÑ CÐBÑ is deterministic CÐBÑ may be biased Traditional principles used in designing physical experiments to balance the effects of non-identical experimental units (randomization, blocking, etc) are irrelevant.
13 Features of Computer Experiments B Ò Code Ò CÐBÑ CÐBÑ is deterministic CÐBÑ may be biased Traditional DoE principles are irrelevant Sometimes output from a physical experiments is also available. Usual philosophy physical experiment is a noisy measurement of the true input- output relationship ÐB Ò. X ÐBÑÑÞ Model : X c X C B œ. ÐBÑ % ÐBÑ and C B œ. ÐBÑ $ ÐBÑ e% ÐBÑf B are measurement errors (usually, white noise) e$ ÐBÑfB is computer model bias Warning Sometimes physical experiments are available only for components of the ensemble process, eg, code that emulates an auto crash test. In other cases, only experiments that approximate reality are available, e.g., a knee simulator
14 Features of Computer Experiments B Ò Code Ò CÐBÑ CÐBÑ is deterministic CÐBÑ may be biased Traditional DoE principles are irrelevant Sometimes output from a physical experiments is also available. Our interest in settings where Few computer runs are possible - complex codes Ðfine-grid FEA codesñ Highqdimensional input B
15 (1) Design of VLSI circuits Examples of Computer Experiments (2) Design automobile engines, and other components Kai-Tai Fang, Runze Li, and Agus Sudjianto (2005) Design and Modeling for Computer Experiments (3) Determine optimum operating conditions for a compression molding process (4) Determine the performance of controlled nuclear fusion devices (5) Describe the temporal evolution of contained and wild fires (6) Design of jet engines, helicopter rotor blades (7) Biomechanics Explain behavior of (or even design) prosthetic devices
16 Example Zone Computer Models used to predict the evolution of a fire in an enclosed room (eg, the computer code ASET = Available Safe Egress Time) In particular, ASET-B describes the of a fire in a single room temporal evolution with closed doors and windows that contains an object at some point below the ceiling that has been ignited.
17
18
19
20 Mathematical Model Inputs to ASET-B Room ceiling height ÐB Ñ " Room floor area ÐB Ñ # Height of the fire source (the burning object) above the floor ÐB Ñ $ Heat loss fraction for the room (depends on room insulation) ÐB Ñ % (etc, material-specific heat release rate) Computed Response CÐB", B# ßB$ ßB% Ñ œ time required by smoke layer to reach 5 ft above the ground
21 Objective Predict the time for the fire plume to reach 5 ft above the ground for untried combinations
22 3. A Taxonomy of Problems Setup 1. Inputs BœÐBßB - /, B7Ñwhere B - œ control ( manufacturing, engineering design) variables B / œ noise ( field, enviromental) variables B7 œ model variables (Not all types of inputs need be present in every application Þ)
23 3. A Taxonomy of Problems Setup 1. Inputs BœÐBßB - /, B7Ñwhere B - œ control ( manufacturing, engineering design) variables B / œ noise ( field, enviromental) variables B7 œ model and calibration variables (Not all types of inputs need be present in every application Þ) 2. Outputs Real-valued : CÐBÑ or Multivariate: C" ÐBÑß C# ÐBÑß á ß C5ÐBÑ or Functional: (>ß CÐ>ß BÑÑ
24 3. A Taxonomy of Problems Setup 1. Inputs BœÐBßB - /, B7Ñwhere B - œ control ( manufacturing, engineering design) variables B / œ noise ( field, enviromental) variables B7 œ model variables (Not all types of inputs need be present in every application Þ) 2. Outputs Real-valued : CÐBÑ or Multivariate: C" ÐBÑß C# ÐBÑß á ß C5ÐBÑ or Functional: (>ß CÐ>ß BÑÑ 3. X / µjðñ describes "target field conditions'' X 7 µ 1( Ñ describes prior knowledge about the model variables
25 3. A Taxonomy of Problems Setup 1. Inputs BœÐBßB - /, B7Ñwhere B - œ control ( manufacturing, engineering design) variables B / œ noise ( field, environmental) variables B7 œ model variables (Not all types of inputs need be present in every application Þ) 2. Outputs Real-valued : CÐBÑ or Multivariate: C" ÐBÑß C# ÐBÑß á ß C5ÐBÑ or Functional: (>ß CÐ>ß BÑÑ 3. X / µjðñ describes "target field conditions'' X 7 µ 1( Ñ describes prior knowledge about the model specifications 4. Summary of the CÐBß\ Ñ distribution (assuming no X ) - / 7.ÐB-Ñ œ IJeCÐB- ß\ / Ñfà 0( B-): P JeCÐB- ß\ / Ñ Ÿ 0( B-) f œ α (median) à ( B ) œ Var ÐCÐB ß\ ÑÑ 5 # - J - /
26 3. A Taxonomy of Problems Setup 1. Inputs BœÐBßB - /, B7Ñwhere B - œ control ( manufacturing, engineering design) variables B / œ noise ( field, environmental) variables B7 œ B7ß7+>2ß B7ß8?7 œ model variables (Not all types of inputs need be present in every application Þ) 2. Outputs Real-valued : CÐBÑ or Multivariate: C" ÐBÑß C# ÐBÑß á ß C5ÐBÑ or Functional: (>ß CÐ>ß BÑÑ 3. X / µjðñ describes "target field conditions'' X 7 µ 1( Ñ describes prior knowledge about the model specifications 4. Summary (assuming no X7 ) of CÐBß\ - / Ñ distribution.ðb-ñ œ IJeCÐB- ß\ / Ñfà 0( B-): P JeCÐB- ß\ / Ñ Ÿ 0( B-) f œ α (median) à ( B ) œ Var ÐCÐB ß\ ÑÑ 5 # - J - / > > 5. "Design" of a computer experiment choice of B" ßá, B8 at which to > evaluate computer code, whereß eg, B œ ˆ > > B ßB, 3 = ", á, 8 3 -ß3 /ß3
27 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Given computer code output at a set of training inputs, ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ " " 8 8 predict CÐ Ñ at a new input B! (predictor metamodel)
28 3. Taxonomy of Problems Problem 1 inputs, Interpolation/Prediction Given computer code output at a set of "training" ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ " " 8 8 predict CÐ Ñ at a new input B! (predictor metamodel) Problem 2 Experimental design Determine a set of inputs at which to carry out the sequence of code runs (a "good" design of a physical or computer experiment depends on the scientific objective of the research) Exploratory Designs ("space-filling") Prediction-based Designs Optimization-based Designs (e.g., find B 9:> - œ argmin CÐBÑ)
29 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Problem 2 Experimental design Exploratory Designs ("space-filling") Prediction-based Designs Optimization-based Designs (e.g., find B 9:> - œ argmin CÐBÑ) Problem 3 Uncertainty/Output Analysis Determine the distribution of the random variable CÐB -, \/ Ñ. (Determine the variability in the performance measure C( ) for design B - when applied to the population defined by the distribution of \ /, eg, patient specific variables (patient weight or bone material properties) or surgeon specific variables (measuring surgical skill)
30 Example In his Cornell PhD thesis, Kevin Ong conducted an uncertainty analysis of the effect of Engineering Cup design, Surgical, Patient, and Fluid Effects on the Stability of Uncemented Acetabular Components
31 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Problem 2 Experimental design Problem 3 Uncertainty/Output Analysis Problem 4 Sensitivity Analysis Determine how variation in CÐBÑ can be apportioned to the different inputs of B (which inputs is CÐBÑ not sensitive to? which sets of inputs is CÐBÑ most sensitive to?) Philosophy Inputs that have relatively little effect on the output can be set to some nominal value; additional investigation can be restricted to determining how the output depends on the active inputs
32 <œroom ceiling height +œ Room floor area 0œHeight of the fire source above the floor 2œHeat loss fraction for the room
33 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Problem 2 Experimental design Problem 3 Uncertainty/Output Analysis Problem 4 Sensitivity Analysis Problem 5 Calibrate the computer code Use outputs from both a physical experiment and an associated computer code that represents the physical process to set the computer code calibration variables. Calibration variables are those B 7 variables that can be used to eliminate the bias in a computer code's representation of a physical input-output relationship Typical Numerical Variables (1) FEA Mesh Density œ? ß (2) Load Discretization œ? ß (3) Solution tolerances œ?, etc
34 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Problem 2 Experimental design Problem 3 Uncertainty/Output Analysis Problem 4 Sensitivity Analysis Problem 5 Calibrate the computer code Problem 6 Prediction Accuracy Using the data from both a physical experimental data and a (calibrated) computer experiment, give predictions (including uncertainty bounds) for an associated physical system.
35 3. Taxonomy of Problems Problem 1 Interpolation/Prediction Problem 2 Experimental design Problem 3 Uncertainty/Output Analysis Problem 4 Sensitivity Analysis Problem 5 Calibration Problem 6 Prediction Problem 7 Find Robust Inputs In experiments with engineering design and patient-specific environmental variables, determine robust choices of the engineering design variables. If B -. B œi ecðb, \ Ñf - J - / then a robust set of inputs B - is an engineering "design" whose output is minimally sensitive to the assumed distribution JÐ Ñ of \ /
36 Bottom Line Many of the problems above have "natural" solutions obtained by approximating CÐB-, B/ Ñ by a fast (linear in the training data) predictorß a metamodel
37 4. Gaussian Stochastic Process (GaSP) Models (used as basis for both prediction and some design choices) Idea Regard CÐBÑ as a realization, a "draw ß " of a random function ] (Bayesian Thinking) The simplest possible (prior) model for ] B is ]ÐBÑœ " ÐBÑ î^ B ðóóñóóò 4 " large scale trends" " smooth deviations" œ " ð 0ÐBÑ ^ B where 0Ð " BÑáß0Ð, 5 BÑare known regression functions, " is an unknown regression vector, and ^ B is a stationary Gaussian Stochastic Process (GaSP) B
38 ^ÐBÑß B k satisfies Ee^ÐBÑf œ!ðzero meanñ ð ð Ê I e]ðbñf œ " 0ÐBÑ! œ " 0ÐBÑ Var ^ÐBÑ œ 5 #^ Correlation Function: symmetric VÐ Ñwith VÐ!Ñ œ ", # Cov( ^ÐB" Ñ, ^ÐB# Ñ) œ 5^ VÐB" B# Ñ 5 #^ œ Var ^Ð B Ñ Typically VÐÑœVÐl 0Ñis a function of a finite number of unknown parameters GaSP: For any B", á, B= ß ^ÐB" Ñá,, ^ÐB= Ñ has the multivariate normal distribution Usually, taking " ð 0ÐB Ñ œ "! with a data-selected parametric correlation function R l0
39 4. GaSP Models GaSP Models Are Flexible Four draws from ^ÐBÑß a zero mean, unit variance GaSP with input B [0,1] and having correlation function for ) e%ß "'ß "!! f VÐ2Ñ œ /B: ˆ ) 2 #
40 Some draws from a GaSP with inputs B Ò!ß "Ó #
41 4. GaSP Models GaSP Models Are Flexible GaSP Models are computationally "simple"
42 4. GaSP Models GaSP Models Are Flexible GaSP Models are computationally "simple" GaSP Models are Bayesian in character
43 &Þ Prediction based on the GaSP Model Given ( training) data ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ " " 8 8 predict C B ß where B!! is an untried new input.
44 &Þ Prediction based on the GaSP Model Given ( training) data predict C B ß where ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ B!! " " 8 8 is an untried new input. 8 > > 8 > > " 8 " 8 Notation ] œð]ðb Ñßáß]ÐB ÑÑ and C œðcðb Ñß áßcðb ÑÑ
45 &Þ Prediction based on the GaSP Model Given ( training) data predict C B ß where ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ B!! " " 8 8 is an untried new input. 8 > > 8 > > " 8 " 8 Notation ] œð]ðb Ñßáß]ÐB ÑÑ and C œðcðb Ñß áßcðb ÑÑ The minimum MSPE predictor of CÐ Ñ is B! scðb Ñ œ I e]ðb Ñl] œ C!! 8 8 f
46 &Þ Prediction based on the GaSP Model Given ( training) data predict C B ß where ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ B!! " " 8 8 is an untried new input. 8 > > 8 > > " 8 " 8 Notation ] œð]ðb Ñßáß]ÐB ÑÑ and C œðcðb Ñß áßcðb ÑÑ The minimum MSPE predictor of CÐ Ñ is B! scðb Ñ œ I e]ðb Ñl] œ C!! 8 8 Example If ]ÐB Ñfollows the GaSP( " ß5 ßVÐ Ñß ) then scðb Ñ I ]ÐB Ñl] 8 œ C 8 œ s ð " 8 e f " < ÐB ÑV Š C " s "!!!!!! 8 where R œ ˆ > > VÐB B Ñ 3 4 is 8 8 ð > > ð <ÐBÑœ! ÐVÐB! B" Ñß á ß VÐB! B 8 ÑÑ is " 8 " s " œ ð " " ð " 8 WLSE of " R " " R C!! # ^ f
47 &Þ Prediction based on the GaSP Model Given ( training) data predict C B ß where ˆ > > B B ˆ > > ßCÐ Ñ ßáß B ßCÐB Ñ B!! " " 8 8 is an untried new input. 8 > > 8 > > " 8 " 8 Notation ] œð]ðb Ñßáß]ÐB ÑÑ and C œðcðb Ñß áßcðb ÑÑ The minimum MSPE predictor of CÐ Ñ is B! scðb Ñ œ I e]ðb Ñl] œ C!! 8 8 Example If ]ÐB Ñfollows the GaSP( " ß5 ßVÐ Ñß ) then scðb Ñ I ]ÐB Ñl] 8 œ C 8 œ s ð " 8 e f " < ÐB ÑV Š C " s "!!!!!! 8 IF only the moment assumptions holds, CÐ Ñ Best Linear Unbiased Predictor of B! s B! ]Ð Ñ # ^ f
48 &Þ Prediction based on the GaSP Model Model Prediction uncertainty at B! 5 # ( B! ) œ E Ð]ÐB! Ñ CÐ s B! ÑÑ # l] 8 œ C 8
49 &Þ Prediction based on the GaSP Model Model Prediction uncertainty at B! 5 # ( B! ) œ E Ð]ÐB! Ñ CÐ s B! ÑÑ # l] 8 œ C 8 Empirical BLUP If the correlation is unknown (the usual case Ñ, Ê R, <ÐBÑ ð, and " s are also unknown : (!! ß VÐ Ñ œ VÐ l0ñis parametric, and we estimate 0 by If further, we can predict using the corresponding empirical BLUP s 0, say, 8 8 scðb Ñ I ]ÐB Ñl] œ C ß s œ s s< ÐB ÑVs 8 C s! š 0 " Š " " ð "!!!! 8 s0 œ MLE, REML, penalized likelihood ß or other estimator of 0
50 &Þ Prediction based on the GaSP Model Model Prediction uncertainty at B! 5 # ( B! ) œ E Ð]ÐB! Ñ CÐ s B! ÑÑ # l] 8 œ C 8 Empirical BLUP If the correlation is unknown (the usual case Ñ, X Ê R, < ÐB Ñ, and " s are also unknown : (!! Ifß further, VÐ Ñ œ VÐ l0ñis parametric, and we estimate 0 by s0, say, we can predict using the corresponding empirical BLUP "!!!!! 8 scðb Ñ I ]ÐB Ñl] 8 œ C 8 ß s œ s s<ðb ÑVs 8 š 0 " Š C s " " s0 œ MLE, REML, penalized likelihood ß or other estimator of 0 Fully Bayesian Predictor (giving all parameters priors) 8 8 e B ] C f œðóóóóóóóóóóóñóóóóóóóóóóóò # 8! B! ßC scðb! Ñ œ I ]Ð Ñl œ œ I I ]Ð Ñl"! ß ß0 5^ 8 Need Ò"! ß5 #^ ß0lC Ó known
51 Properties of s " 8 CÐB! Ñ œ " s ð < ÐB ÑV Š C " s " Simple to compute (linear in C 8 )!!! 8 > scðbñ œ - B - ÐBÑCÐB Ñ 8! 4 4œ" 4
52 Properties of s " 8 CÐB! Ñ œ " s ð < ÐB ÑV Š C " s " Simple to compute (linear in C 8 )!!! 8 > scðb Ñ œ - B - ÐB ÑCÐB Ñ!!! 4! 4œ" but, unfortunately, not the Empirical BLUP :-( GASP (W. Welch) PErK (B. J. Williams) SAS Proc Mixed BACCO (Hankin) and others... R -1 can be computationally demanding 8 4
53 Properties of s " 8 CÐB! Ñ œ " s ð < ÐB ÑV Š C " s " Simple to compute (linear in C 8 ) Viewed as a function of B!,!!! 8 > scðb Ñ œ - B - ÐB ÑCÐB Ñ!!! 4! 4œ" scðb Ñ œ.. Vˆ > B B!! 4! 4œ"
54 Properties of s " 8 CÐB! Ñ œ " s ð < ÐB ÑV Š C " s " Simple to compute (linear in C 8 ) Viewed as a function of B, scðb Ñ interpolates data, i.e.,!!! 8 > scðb Ñ œ - B - ÐB ÑCÐB Ñ!!! 4! 4œ" scðb Ñ œ.. Vˆ > B B!! 4! 4œ" 4 > > 4 8 scðb Ñ œ CÐB Ñß 4 œ "ßáß
55 Properties of s " 8 CÐB! Ñ œ " s ð < ÐB ÑV Š C " s " Simple to compute (linear in C 8 ) Viewed as a function of B, scðbñ interpolates data, i.e.,!!! 8 > scðb Ñ œ - B - ÐB ÑCÐB Ñ!!! 4! 4œ" scðb Ñ œ.. Vˆ > B B!! 4! 4œ" 4 > > 4 8 scð B Ñ œ CÐB Ñß 4 œ "ß á ß 8 Splines, neural networks and other well-known interpolators correspond to specific choices of regressors and correlation function VÐ Ñ 8 4 4
56 Illustration True Curve ( solid); five training data points ( diamonds); REML-EBLUP with exponential correlation function ( dotted)
57
58 5. Conclusions-Take Home Messages ". An increasing number of phenomenon that could previously be studied only by physical experiments, can now be investigated using "computer experiments'' or combinations of computer and physical experiments #. Modeling the responses from computer experiments must account for the (highly) correlated nature of the output CÐBÑ over the input space. $. Prediction of the output function CÐBÑ based on Gaussian (or other) stochastic processes can be used to interpolate training data %. GaSP models are the basis for Determining interpolating predictors of CÐBÑ Assessing error bands of the predictor due to model uncertainty, Targeted experimental design, and Solving calibration problems
59 Related Sessions 335 Multivariate Functional ANOVA for Kriging Model in Computer Experiments (Zhe Zhang) 345 Statistical Design of Computer Experiments for a 3D Chemical Microanalysis Imaging System (Juan Soto, James J. Filliben, and John H. Scott) 363 Design and Analysis of Experiments for Complex Computer Simulators (Aug 9 Session) Designs for Integrated Computer and Physical Experiments (C. Shane Reese, Derek Bingham, and Wilson Lu) Sequential Experiment Design for Contour Estimation from Complex Computer Codes (Pritam Ranjan) Uncertainty Quantification for Combining Experimental Data and Computer Simulations from Multiple Data Sources (Brian J. Williams, Dave Higdon,Jim Gattiker) 474 Network Analysis and Spatial Applications-Contributed (Aug 9 Session) Calibration and Prediction for Computer Experiment Output Having Qualitative and Quantitative Input Variables (Gang Han, Thomas Santner, and William Notz) Validity of Likelihood and Bayesian Inference for Gaussian Process Regression (Bela Nagy, Jason Loeppky, and William J. Welch) Questions?
Tutorial on. The Design of Computer Experiments for Optimization. Thomas Santner Department of Statistics The Ohio State University
Tutorial on The Design of Computer Experiments for Optimization Thomas Santner Department of Statistics The Ohio State University The A-C-NW Optimization Tutorials June 13, 2005 Outline 1. Three Types
More informationIntroduction to Design and Analysis of Computer Experiments
Introduction to Design and Analysis of Thomas Santner Department of Statistics The Ohio State University October 2010 Outline Empirical Experimentation Empirical Experimentation Physical Experiments (agriculture,
More informationSequential adaptive designs in computer experiments for response surface model fit
Statistics and Applications Volume 6, Nos. &, 8 (New Series), pp.7-33 Sequential adaptive designs in computer experiments for response surface model fit Chen Quin Lam and William I. Notz Department of
More informationOutline. 1. The Screening Philosophy 2. Robust Products/Processes 3. A Statistical Screening Procedure 4. Some Caveats 5.
Screening to Identify Robust Products Based on Fractional Factorial Experiments Thomas J. Santner Ohio State University Guohua Pan Novartis Pharmaceuticals Corp Outline 1. The Screening Philosophy 2. Robust
More informationSEQUENTIAL ADAPTIVE DESIGNS IN COMPUTER EXPERIMENTS FOR RESPONSE SURFACE MODEL FIT
SEQUENTIAL ADAPTIVE DESIGNS IN COMPUTER EXPERIMENTS FOR RESPONSE SURFACE MODEL FIT DISSERTATION Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate
More informationSome general observations.
Modeling and analyzing data from computer experiments. Some general observations. 1. For simplicity, I assume that all factors (inputs) x1, x2,, xd are quantitative. 2. Because the code always produces
More informationSIMULTANEOUS CONFIDENCE BANDS FOR THE PTH PERCENTILE AND THE MEAN LIFETIME IN EXPONENTIAL AND WEIBULL REGRESSION MODELS. Ping Sa and S.J.
SIMULTANEOUS CONFIDENCE BANDS FOR THE PTH PERCENTILE AND THE MEAN LIFETIME IN EXPONENTIAL AND WEIBULL REGRESSION MODELS " # Ping Sa and S.J. Lee " Dept. of Mathematics and Statistics, U. of North Florida,
More informationMonitoring Wafer Geometric Quality using Additive Gaussian Process
Monitoring Wafer Geometric Quality using Additive Gaussian Process Linmiao Zhang 1 Kaibo Wang 2 Nan Chen 1 1 Department of Industrial and Systems Engineering, National University of Singapore 2 Department
More informationCovariance function estimation in Gaussian process regression
Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian
More informationOptimal Designs for Gaussian Process Models via Spectral Decomposition. Ofir Harari
Optimal Designs for Gaussian Process Models via Spectral Decomposition Ofir Harari Department of Statistics & Actuarial Sciences, Simon Fraser University September 2014 Dynamic Computer Experiments, 2014
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationAn investigation into the use of maximum posterior probability estimates for assessing the accuracy of large maps
An investigation into the use of maximum posterior probability estimates for assessing the accuracy of large maps Brian Steele 1 and Dave Patterson2 Dept. of Mathematical Sciences University of Montana
More informationAn R # Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM
An R Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM Lloyd J. Edwards, Ph.D. UNC-CH Department of Biostatistics email: Lloyd_Edwards@unc.edu Presented to the Department
More informationAn introduction to Bayesian statistics and model calibration and a host of related topics
An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the
More informationUncertainty Quantification for Inverse Problems. November 7, 2011
Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance
More informationIt's Only Fitting. Fitting model to data parameterizing model estimating unknown parameters in the model
It's Only Fitting Fitting model to data parameterizing model estimating unknown parameters in the model Likelihood: an example Cohort of 8! individuals observe survivors at times >œ 1, 2, 3,..., : 8",
More informationKriging Luc Anselin, All Rights Reserved
Kriging Luc Anselin Spatial Analysis Laboratory Dept. Agricultural and Consumer Economics University of Illinois, Urbana-Champaign http://sal.agecon.uiuc.edu Outline Principles Kriging Models Spatial Interpolation
More informationLA PRISE DE CALAIS. çoys, çoys, har - dis. çoys, dis. tons, mantz, tons, Gas. c est. à ce. C est à ce. coup, c est à ce
> ƒ? @ Z [ \ _ ' µ `. l 1 2 3 z Æ Ñ 6 = Ð l sl (~131 1606) rn % & +, l r s s, r 7 nr ss r r s s s, r s, r! " # $ s s ( ) r * s, / 0 s, r 4 r r 9;: < 10 r mnz, rz, r ns, 1 s ; j;k ns, q r s { } ~ l r mnz,
More informationMultiple Linear Regression for the Supervisor Data
for the Supervisor Data Rating 40 50 60 70 80 90 40 50 60 70 50 60 70 80 90 40 60 80 40 60 80 Complaints Privileges 30 50 70 40 60 Learn Raises 50 70 50 70 90 Critical 40 50 60 70 80 30 40 50 60 70 80
More informationSensitivity analysis in linear and nonlinear models: A review. Introduction
Sensitivity analysis in linear and nonlinear models: A review Caren Marzban Applied Physics Lab. and Department of Statistics Univ. of Washington, Seattle, WA, USA 98195 Consider: Introduction Question:
More informationPART I. Multiple choice. 1. Find the slope of the line shown here. 2. Find the slope of the line with equation $ÐB CÑœ(B &.
Math 1301 - College Algebra Final Exam Review Sheet Version X This review, while fairly comprehensive, should not be the only material used to study for the final exam. It should not be considered a preview
More informationAsymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands
Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department
More informationModels for models. Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research
Models for models Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research Outline Statistical models and tools Spatial fields (Wavelets) Climate regimes (Regression and clustering)
More informationSEQUENTIAL DESIGN OF COMPUTER EXPERIMENTS TO MINIMIZE INTEGRATED RESPONSE FUNCTIONS
Statistica Sinica 10(2000), 1133-1152 SEQUENTIAL DESIGN OF COMPUTER EXPERIMENTS TO MINIMIZE INTEGRATED RESPONSE FUNCTIONS Brian J. Williams, Thomas J. Santner and William I. Notz The Ohio State University
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationInner Product Spaces
Inner Product Spaces In 8 X, we defined an inner product? @? @?@ ÞÞÞ? 8@ 8. Another notation sometimes used is? @? ß@. The inner product in 8 has several important properties ( see Theorem, p. 33) that
More informationSimulation optimization via bootstrapped Kriging: Survey
Simulation optimization via bootstrapped Kriging: Survey Jack P.C. Kleijnen Department of Information Management / Center for Economic Research (CentER) Tilburg School of Economics & Management (TiSEM)
More informationE.G., data mining, prediction. Function approximation problem: How >9 best estimate the function 0 from partial information (examples)
Can Neural Network and Statistical Learning Theory be Formulated in terms of Continuous Complexity Theory? 1. Neural Networks and Statistical Learning Theory Many areas of mathematics, statistics, and
More informationRandom and mixed effects models
Random and mixed effects models Fixed effect: Three fields were available for an agricultural yield experiment. The experiment is conducted on those fields. The mean yield of this particular strain of
More informationIntegrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University
Integrated Likelihood Estimation in Semiparametric Regression Models Thomas A. Severini Department of Statistics Northwestern University Joint work with Heping He, University of York Introduction Let Y
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationConfidence Estimation Methods for Neural Networks: A Practical Comparison
, 6-8 000, Confidence Estimation Methods for : A Practical Comparison G. Papadopoulos, P.J. Edwards, A.F. Murray Department of Electronics and Electrical Engineering, University of Edinburgh Abstract.
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationBayesian Calibration of Simulators with Structured Discretization Uncertainty
Bayesian Calibration of Simulators with Structured Discretization Uncertainty Oksana A. Chkrebtii Department of Statistics, The Ohio State University Joint work with Matthew T. Pratola (Statistics, The
More informationUncertainty quantification and calibration of computer models. February 5th, 2014
Uncertainty quantification and calibration of computer models February 5th, 2014 Physical model Physical model is defined by a set of differential equations. Now, they are usually described by computer
More informationMAP METHODS FOR MACHINE LEARNING. Mark A. Kon, Boston University Leszek Plaskota, Warsaw University, Warsaw Andrzej Przybyszewski, McGill University
MAP METHODS FOR MACHINE LEARNING Mark A. Kon, Boston University Leszek Plaskota, Warsaw University, Warsaw Andrzej Przybyszewski, McGill University Example: Control System 1. Paradigm: Machine learning
More informationModeling of Atmospheric Effects on InSAR Measurements With the Method of Stochastic Simulation
Modeling of Atmospheric Effects on InSAR Measurements With the Method of Stochastic Simulation Z. W. LI, X. L. DING Department of Land Surveying and Geo-Informatics, Hong Kong Polytechnic University, Hung
More informationSensitivity analysis using the Metamodel of Optimal Prognosis. Lectures. Thomas Most & Johannes Will
Lectures Sensitivity analysis using the Metamodel of Optimal Prognosis Thomas Most & Johannes Will presented at the Weimar Optimization and Stochastic Days 2011 Source: www.dynardo.de/en/library Sensitivity
More informationGWAS V: Gaussian processes
GWAS V: Gaussian processes Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS V: Gaussian processes Summer 2011
More informationForecasting Data Streams: Next Generation Flow Field Forecasting
Forecasting Data Streams: Next Generation Flow Field Forecasting Kyle Caudle South Dakota School of Mines & Technology (SDSMT) kyle.caudle@sdsmt.edu Joint work with Michael Frey (Bucknell University) and
More informationSequential Importance Sampling for Rare Event Estimation with Computer Experiments
Sequential Importance Sampling for Rare Event Estimation with Computer Experiments Brian Williams and Rick Picard LA-UR-12-22467 Statistical Sciences Group, Los Alamos National Laboratory Abstract Importance
More informationData are collected along transect lines, with dense data along. Spatial modelling using GMRFs. 200m. Today? Spatial modelling using GMRFs
Centre for Mathematical Sciences Lund University Engineering geology Lund University Results A non-stationary extension The model Estimation Gaussian Markov random fields Basics Approximating Mate rn covariances
More informationDesign and Analysis of Computer Experiments for Screening Input Variables. Dissertation. Hyejung Moon, M.S. Graduate Program in Statistics
Design and Analysis of Computer Experiments for Screening Input Variables Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in the Graduate School of
More informationMustafa H. Tongarlak Bruce E. Ankenman Barry L. Nelson
Proceedings of the 0 Winter Simulation Conference S. Jain, R. R. Creasey, J. Himmelspach, K. P. White, and M. Fu, eds. RELATIVE ERROR STOCHASTIC KRIGING Mustafa H. Tongarlak Bruce E. Ankenman Barry L.
More informationData-Driven Forecasting Algorithms for Building Energy Consumption
Data-Driven Forecasting Algorithms for Building Energy Consumption Hae Young Noh a and Ram Rajagopal b a Department of Civil and Environmental Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213,
More informationarxiv: v1 [stat.me] 24 May 2010
The role of the nugget term in the Gaussian process method Andrey Pepelyshev arxiv:1005.4385v1 [stat.me] 24 May 2010 Abstract The maximum likelihood estimate of the correlation parameter of a Gaussian
More informationQUALIFYING EXAMINATION II IN MATHEMATICAL STATISTICS SATURDAY, MAY 9, Examiners: Drs. K. M. Ramachandran and G. S. Ladde
QUALIFYING EXAMINATION II IN MATHEMATICAL STATISTICS SATURDAY, MAY 9, 2015 Examiners: Drs. K. M. Ramachandran and G. S. Ladde INSTRUCTIONS: a. The quality is more important than the quantity. b. Attempt
More informationGene expression experiments. Infinite Dimensional Vector Spaces. 1. Motivation: Statistical machine learning and reproducing kernel Hilbert Spaces
MA 751 Part 3 Gene expression experiments Infinite Dimensional Vector Spaces 1. Motivation: Statistical machine learning and reproducing kernel Hilbert Spaces Gene expression experiments Question: Gene
More informationIntroduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones
Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 Last week... supervised and unsupervised methods need adaptive
More informationA Framework for Daily Spatio-Temporal Stochastic Weather Simulation
A Framework for Daily Spatio-Temporal Stochastic Weather Simulation, Rick Katz, Balaji Rajagopalan Geophysical Statistics Project Institute for Mathematics Applied to Geosciences National Center for Atmospheric
More informationmlegp: an R package for Gaussian process modeling and sensitivity analysis
mlegp: an R package for Gaussian process modeling and sensitivity analysis Garrett Dancik January 30, 2018 1 mlegp: an overview Gaussian processes (GPs) are commonly used as surrogate statistical models
More informationA Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,
A Course in Applied Econometrics Lecture 18: Missing Data Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. When Can Missing Data be Ignored? 2. Inverse Probability Weighting 3. Imputation 4. Heckman-Type
More informationStatistical modelling: Theory and practice
Statistical modelling: Theory and practice Introduction Gilles Guillot gigu@dtu.dk August 27, 2013 Gilles Guillot (gigu@dtu.dk) Stat. modelling August 27, 2013 1 / 6 Schedule 13 weeks weekly time slot:
More informationSVM example: cancer classification Support Vector Machines
SVM example: cancer classification Support Vector Machines 1. Cancer genomics: TCGA The cancer genome atlas (TCGA) will provide high-quality cancer data for large scale analysis by many groups: SVM example:
More informationarxiv: v1 [stat.me] 10 Jul 2009
6th St.Petersburg Workshop on Simulation (2009) 1091-1096 Improvement of random LHD for high dimensions arxiv:0907.1823v1 [stat.me] 10 Jul 2009 Andrey Pepelyshev 1 Abstract Designs of experiments for multivariate
More informationNonparametric Bayesian Methods - Lecture I
Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics
More informationCONTINUOUS SPATIAL DATA ANALYSIS
CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s
More informationWays to make neural networks generalize better
Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights
More informationBayesian Dynamic Linear Modelling for. Complex Computer Models
Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer
More informationContents 1 Introduction 2 Statistical Tools and Concepts
1 Introduction... 1 1.1 Objectives and Approach... 1 1.2 Scope of Resource Modeling... 2 1.3 Critical Aspects... 2 1.3.1 Data Assembly and Data Quality... 2 1.3.2 Geologic Model and Definition of Estimation
More informationRegression, Ridge Regression, Lasso
Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.
More information12 - Nonparametric Density Estimation
ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6
More informationGeneralized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science.
Texts in Statistical Science Generalized Linear Mixed Models Modern Concepts, Methods and Applications Walter W. Stroup CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint
More informationDesign of Computer Experiments for Optimization, Estimation of Function Contours, and Related Objectives
7 Design of Computer Experiments for Optimization, Estimation of Function Contours, and Related Objectives Derek Bingham Simon Fraser University, Burnaby, BC Pritam Ranjan Acadia University, Wolfville,
More informationNonconcave Penalized Likelihood with A Diverging Number of Parameters
Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized
More informationResampling techniques for statistical modeling
Resampling techniques for statistical modeling Gianluca Bontempi Département d Informatique Boulevard de Triomphe - CP 212 http://www.ulb.ac.be/di Resampling techniques p.1/33 Beyond the empirical error
More informationIntroduction to Machine Learning Midterm, Tues April 8
Introduction to Machine Learning 10-701 Midterm, Tues April 8 [1 point] Name: Andrew ID: Instructions: You are allowed a (two-sided) sheet of notes. Exam ends at 2:45pm Take a deep breath and don t spend
More informationBasics of Uncertainty Analysis
Basics of Uncertainty Analysis Chapter Six Basics of Uncertainty Analysis 6.1 Introduction As shown in Fig. 6.1, analysis models are used to predict the performances or behaviors of a product under design.
More informationSTAT 518 Intro Student Presentation
STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationCSE446: non-parametric methods Spring 2017
CSE446: non-parametric methods Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin and Luke Zettlemoyer Linear Regression: What can go wrong? What do we do if the bias is too strong? Might want
More informationOC330C. Wiring Diagram. Recommended PKH- P35 / P50 GALH PKA- RP35 / RP50. Remarks (Drawing No.) No. Parts No. Parts Name Specifications
G G " # $ % & " ' ( ) $ * " # $ % & " ( + ) $ * " # C % " ' ( ) $ * C " # C % " ( + ) $ * C D ; E @ F @ 9 = H I J ; @ = : @ A > B ; : K 9 L 9 M N O D K P D N O Q P D R S > T ; U V > = : W X Y J > E ; Z
More informationAUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET
The Problem Identification of Linear and onlinear Dynamical Systems Theme : Curve Fitting Division of Automatic Control Linköping University Sweden Data from Gripen Questions How do the control surface
More informationLecture 3: More on regularization. Bayesian vs maximum likelihood learning
Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationInfinite Dimensional Vector Spaces. 1. Motivation: Statistical machine learning and reproducing kernel Hilbert Spaces
MA 751 Part 3 Infinite Dimensional Vector Spaces 1. Motivation: Statistical machine learning and reproducing kernel Hilbert Spaces Microarray experiment: Question: Gene expression - when is the DNA in
More informationIntroduction. Chapter 1
Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics
More informationLimit Kriging. Abstract
Limit Kriging V. Roshan Joseph School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, GA 30332-0205, USA roshan@isye.gatech.edu Abstract A new kriging predictor is proposed.
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More informationSolving SVM: Quadratic Programming
MA 751 Part 7 Solving SVM: Quadratic Programming 1. Quadratic programming (QP): Introducing Lagrange multipliers α 4 and. 4 (can be justified in QP for inequality as well as equality constraints) we define
More informationThe Design and Analysis of Computer Experiments
Thomas J. Santner Brian J. Williams William I. Notz The Design and Analysis of Computer Experiments February 18, 2014 Springer Use the template dedic.tex together with the Springer document class SVMono
More informationAdaptive Data Assimilation and Multi-Model Fusion
Adaptive Data Assimilation and Multi-Model Fusion Pierre F.J. Lermusiaux, Oleg G. Logoutov and Patrick J. Haley Jr. Mechanical Engineering and Ocean Science and Engineering, MIT We thank: Allan R. Robinson
More informationIntroduction to Gaussian-process based Kriging models for metamodeling and validation of computer codes
Introduction to Gaussian-process based Kriging models for metamodeling and validation of computer codes François Bachoc Department of Statistics and Operations Research, University of Vienna (Former PhD
More informationLecture 9. Time series prediction
Lecture 9 Time series prediction Prediction is about function fitting To predict we need to model There are a bewildering number of models for data we look at some of the major approaches in this lecture
More informationGaussian Processes (10/16/13)
STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs
More informationCalibrating Environmental Engineering Models and Uncertainty Analysis
Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now
More informationEfficient sensitivity analysis for virtual prototyping. Lectures. Thomas Most & Johannes Will
Lectures Efficient sensitivity analysis for virtual prototyping Thomas Most & Johannes Will presented at the ECCOMAS Conference 2012 Source: www.dynardo.de/en/library European Congress on Computational
More information9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures
FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationConsistent high-dimensional Bayesian variable selection via penalized credible regions
Consistent high-dimensional Bayesian variable selection via penalized credible regions Howard Bondell bondell@stat.ncsu.edu Joint work with Brian Reich Howard Bondell p. 1 Outline High-Dimensional Variable
More informationBayesian Inference of Noise Levels in Regression
Bayesian Inference of Noise Levels in Regression Christopher M. Bishop Microsoft Research, 7 J. J. Thomson Avenue, Cambridge, CB FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationSIMULATION - PROBLEM SET 1
SIMULATION - PROBLEM SET 1 " if! Ÿ B Ÿ 1. The random variable X has probability density function 0ÐBÑ œ " $ if Ÿ B Ÿ.! otherwise Using the inverse transform method of simulation, find the random observation
More informationprobability of k samples out of J fall in R.
Nonparametric Techniques for Density Estimation (DHS Ch. 4) n Introduction n Estimation Procedure n Parzen Window Estimation n Parzen Window Example n K n -Nearest Neighbor Estimation Introduction Suppose
More informationStochastic Analogues to Deterministic Optimizers
Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured
More informationApplied Time Series Notes ( 1) Dates. ñ Internal: # days since Jan 1, ñ Need format for reading, one for writing
Applied Time Series Notes ( 1) Dates ñ Internal: # days since Jan 1, 1960 ñ Need format for reading, one for writing ñ Often DATE is ID variable (extrapolates) ñ Program has lots of examples: options ls=76
More informationUse of Design Sensitivity Information in Response Surface and Kriging Metamodels
Optimization and Engineering, 2, 469 484, 2001 c 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. Use of Design Sensitivity Information in Response Surface and Kriging Metamodels J. J.
More information