A Robust Strategy for Joint Data Reconciliation and Parameter Estimation

Size: px
Start display at page:

Download "A Robust Strategy for Joint Data Reconciliation and Parameter Estimation"

Transcription

1 A Robust Strategy for Joint Data Reconciliation and Parameter Estimation Yen Yen Joe 1) 3), David Wang ), Chi Bun Ching 3), Arthur Tay 1), Weng Khuen Ho 1) and Jose Romagnoli ) * 1) Dept. of Electrical & Computer Engineering, The National University of Singapore, 1 Kent Ridge Cres., Singapore 1196 ) Dept. of Chemical Engineering, The University of Sydney, NSW 6, Australia 3) Institute of Chemical and Engineering Sciences, Ayer Rajah Cres., Block 8, Unit #-8, Singapore Abstract In this work, the generalized T (GT) distribution is used to develop a statistically robust joint data reconciliation parameter estimation (DRPE) strategy. The robustness feature is provided by the GT distribution, which includes Normal, Laplacian and Cauchy distribution as special cases. We use historical data to first estimate the parameters of the GT distribution, so that the resulting estimator is efficient when the error is in the GT family. The strategy is implemented in a simulation of a practical chemical engineering plant. The results confirm the robustness and efficiency of the estimator. Keywords: parameter estimation, data reconciliation, error-in-all-variables, robust, estimators 1. Introduction A more efficient approach than the sequential data reconciliation (DR) parameter estimation (PE) that is common in practice is to jointly perform DR and PE (DRPE), such that the resulting reconciled data and model parameters are consistent with respect to both the process model and DR constraints. The DRPE can also be viewed as the error-in-all-variables-measured (EVM) formulation, which is the generalization of the conventional PE: in EVM, all measurements are s.t. errors, such that the distinction between independent and dependent variables is no longer clear (Romagnoli et al, ). The three main aspects of EVM discussed in the literature are the EVM algorithm (Valko et al, 1987), the optimization strategy (Kim et al, 199; Tjoa et al, 1991) and the robustness of the EVM estimation (Albuquerque et al, 1996; Arora et al, 1). In this paper we will focus mainly on the robustness of the EVM estimation. Various robust estimation approaches such as the M-estimators have been proposed, but most assumed, in a priori, some forms of error distribution, which, although robust, might not be representative of the actual distribution. On the other hand, the nonparametric methods such as the kernel function (Wang et al, 3) are free of such assumptions and fully flexible, but are also complex and computationally demanding. * Author to whom correspondence should be addressed: jose@chem.eng.usyd.edu.au

2 An alternative is to strike a balance between the simplicity of the parametric approach and the flexibility of the non-parametric approach, i.e. by adopting a specific objective function that covers a wide variety of common distributions. This corresponds to the generalized T (GT) distribution. The parameters of the GT distribution can be estimated posteriori to ensure its suitability to the data. In this work, we extend the robust DR strategy using GT distribution (Wang et al, 3) to incorporate parameter estimation. This results in a statistically robust EVM strategy that is also efficient. The paper is organised as follows. The next section discusses the incorporation of the robustness feature into DRPE within a probabilistic framework. Section 3 describes the DRPE strategy using the GT distribution, which is then applied to a case study of a general purpose chemical engineering plant in Section 4. Finally, Section 5 concludes the paper.. The Robustness of DRPE Estimator Within a probabilistic framework, by maximum-likelihood principle, the DRPE can be formulated as: max f ( ε ) = min -log(f( ε)) = min ρ( ε ) x, u, θ x, u, θ x, u, θ () s.t. model and bounds where ε = y x is the measurement error, and f (ε ) is the probability density of the error. As the efficiency of the estimator depends on how well f (ε ) characterizes the actual error, the estimator can be made robust by reducing the sensitivity of ρ(ε ) to large values of ε. This corresponds to the robust M-estimator. The robustness of the M- estimators can be explained by the influence function (IF), defined by ψ ( ε)= ρ( ε) ε. Essentially, IF gives a rough measure of how much influence a particular residual has in the estimation (McDonald et al, 1988; Hampel et al, 1986), so it is desirable to have an IF that is bounded for large residuals in order for them to have limited influence on the estimation. It should be pointed out at this point that the conventional WLS, where ρ( ε ) ε, has the IF that is a straight line, which is why large residuals have unlimited influence on and can dominate the estimation, resulting in biased estimates. Common choices of robust M-estimator such as the contaminated normal (Tjoa et al, 1991), the combination of Laplace and Normal distribution (Wang et al, 3), the fair function (Albuquerque et al, 1996) and the redescending estimator (Arora et al, 1) depend on parameters which either are assigned to them a priori or do not have meaningful association with the error distribution. As a result, the underlying error distribution may not be well characterised and the estimators may not be efficient in the MLE sense (Wang et al, 3). We therefore propose the use of GT distribution as robust estimator for DRPE problems, as the GT has the advantages that enable it to be robust while not sacrificing efficiency. This will be elaborated in the next section. 3. Robust DRPE Using the Generalized T (GT) Distribution The use of the GT distribution in estimation is first proposed by McDonald et al (1988) due to its flexibility to accommodate various distributional shapes. The density function is given by:

3 p f GT ( ε; σ, p, q) = ; - < ε < (3) q+ 1 p p 1/ p ε σq B(1 p, q) 1 + p qσ Depending on the values taken by { p, q,, can take the shape of any distribution within the family defined by the GT distribution. As illustrated in Figure 1, it covers most of the important distributions that are commonly encountered in practice. f GT GT p = q σ =α (a) q=5, sigma=1 p= -.- p= Power exponential/ Box-Tiao T-distribution, df = q p =1 p = σ =α q q = (b) p=, sigma=1 q=5 -.- q=4 Double exponential/ Laplacian Normal Figure 1: GT Distribution Tree Cauchy Figure : Influence Functions of GT with different parameter settings The IFs of f GT with different settings of { p, q, is shown Figure. σ only affects the distribution spread, while p and q determine the shape. It is seen that the robustness criteria are satisfied as the IFs are bounded and actually descending when the residuals get large. The two properties of GT demonstrated above: flexibility and robustness, enable us to achieve a robust yet efficient estimator in the MLE sense. This justifies our main motivation in selecting GT over other M-estimators, which as mentioned in Section, may not be efficient in the MLE sense as they may not characterize the error distribution well. The GT, on the other hand, can take a wide range of distributional shapes depending on its parameter values. When these parameters are estimated from the data, it is able to adapt its shape to the data. We therefore use a set of historical data to estimate the distribution parameters. To ensure robustness, however, care must be taken in estimating the distribution parameters. We see that as p increases (Figure a), the IF value for large residuals increases, and as q increases (Figure b), the IF becomes less bounded. It is thus necessary to impose bounds on p and q if robustness is to be preserved (in this work, p 5, q 5 ). We underline that although the bounds exclude a part of the GT family, by estimating p and q from the data, we effectively fit the data with the GT distribution within the parameter bounds. This ensures that the exclusion will have little effect on the asymptotic efficiency of the estimator (McDonald et al, 1988).

4 To estimate the GT distribution parameters, a preliminary reconciliation of the historical data is performed to obtain the residuals ε, which is then fed to the maximum likelihood estimator, given by (Wang et al, 3): max log f GT ( ε ; p, q, σ ) p, q, } (4) { σ The estimates of { p, q, are then obtained as the parameters of a GT member from which the data are most likely sampled. In DRPE, some of the measurement variables are non-redundant, which complicates the estimation of the distribution parameters. Since this work deals with steady-state data, we take the median as the estimated value of the non-redundant measurements. Taking the median as the estimated values corresponds to the use of robust L-estimator (Albuquerque et al, 1996; Hampel et al, 1986); however, it can only be used when the variables have repeated measurements or are known to be constant over the time horizon considered. In the case where these conditions do not hold, a robust preliminary parameter estimation or DRPE has to be performed to obtain the residuals for the nonredundant measurements. This is more complex and computationally expensive, especially if we would like to update the distribution parameters online, as the estimation of the distribution parameters may need a number of iterations. Another alternative is to assign fixed values of { p, q, that are sufficiently robust. In this case, the efficiency is traded off for convenience. Figure 3: Case Study Plant Flowsheet 4. Application Case Study The proposed robust DRPE strategy is applied to a case study of a pilot-scale setting containing two CSTRs, a mixer and a number of heat exchangers (Figure 3). Material feed from the feed tank is heated before being fed to the first reactor and the mixer. The effluent from first reactor is then mixed with the material feed in the mixer, and then fed to the second reactor. The effluent from the second reactor is, in turn, fed back to the feed tank and the cycle continues. Steady-state analysis of the system structure results in seven redundant equations involving 14 redundant variables. The model parameters estimated are the product of the heat transfer coefficient with the effective heat transfer area of the steam jacket (P1)

5 and the cooling coil (P) of the first reactor. For parameter estimation, two more model equations with five non-redundant variables are included. Associated with the pilot-scale plant, a virtual environment has been developed within the Matlab/Simulink framework which mimics the actual plant behaviour and will be used in this paper while the plant is being commissioned. Simulation data are generated with several different distributions: Normal, Laplacian and Cauchy distribution. The different distributions are considered as outliers. A data set having Normal distribution and with large random shifts as gross error is also generated. We then perform DRPE using three different methods, the conventional WLS, the contaminated (bivariate) Gaussian distribution with gross error probability p =. and gross error ratio b = (Tjoa et al, 1991), and the GT distribution with distribution parameters { p, q, estimated from a historical data set (n=1) having similar distribution as the current data set. The performance criterion used to compare the efficiencies of the different methods is the mean-squared error (MSE): MSE = 1 mk K m j= 1 i= 1 ) ( x i, j xi, j ) σ i where m is the number of measured variables and K is the number of data sets used for ) the DRPE (m=19, K=1 in this study). and are the estimates of the reconciled x i, j data and the actual value of the variable, respectively, while σ i is the standard deviation of the Gaussian noise on sensor i. The MSE results are shown in Figure 4, while Table 1 lists the estimated model parameters P1 and P. The fact that the bivariate Gaussian and GT method are more efficient (lower MSE and % discrepancy of parameter estimates) than WLS for distributions other than Normal, and for Normal noise with gross error, proves the robustness of the two M-estimators. Compared to the bivariate Gaussian method, the GT method is more efficient for the Laplacian and Cauchy error distributions, which are special cases of the GT distribution. The GT distribution parameters for some variables for the case of Laplacian and Cauchy noises are listed in Table. The reader can refer back to the distribution tree in Figure 1 to see that the values of the estimated p and q are close to the ideal p and q for the respective distributions. For example, for Laplacian noise, ideally p=1 and q ; the estimated p are close to one, while q are large or close to the upper bound, i.e. q=5. Figure 5 plots the relative frequency distribution of the noise, the estimated GT density with estimated{ p, q,, the bivariate density (p=.,q=), and the normal distribution (N(, σ )) corresponding to the actual noise, GT, bivariate and WLS estimator, respectively, for a temperature variable with Laplacian noise. It is seen that the GT estimator characterizes the data best, which explains its lowest MSE and %discrepancy for Laplacian noise in Figure 4 and Table 1. The same can be concluded for Cauchy noise. 5. Conclusion The DRPE based on GT distribution is robust and efficient, especially when the underlying error distribution is within the GT family. Since the GT family encompasses x i, j (5)

6 a wide variety and many important statistical distributions, the GT-based estimator is a very viable choice of estimator considering its simplicity. MSE WLS Bivariate GT Normal Normal + Gross Error Laplacian Cauchy *-. : relative freq : GT relative freq - - GT : Bivariate bivariate WLS Figure 4: Performance Comparison for Different Noise Profiles Figure 5: Distribution Plots 6 Table 1: Model Parameter Estimates and Their Accuracies %discrepancy %discrepancy P1 P P1 P P1 P P1 P Actual value 37,5 56, ,5 56, Normal WLS 39,17 57, Laplace 39,5 57, Biv. 36,6 58, ,17 57, GT 37,96 57, ,545 57, Normal+ WLS 4,331 57, Cauchy 41,34 51, Gross Biv. 39,45 57, ,1 58, Error GT 36,3 57, ,1 58, Table : GT Distribution Parameter Estimates Cauchy Noise Laplacian Noise Variable p=1 q =.5 p=1 q inf T T T T Trx Tmx References Albuquerque, J. S., Biegler, L.T., AIChE J., Vol. 4, No. 1, pp Arora, N., Biegler, L.T., 1. Comp. Chem. Eng., Vol. 5, pp Hampel, F.R., Ronchetti, E.M., Rousseeuw, P.J., Stahel, W.A., 1986, Robust Statistics: The Approach Based on Influence Functions, Wiley. Kim, I.W., Liebman, M.J., and Edgar, T.F., 199. AIChE, Vol. 36, pp McDonald, J.B., Newey, W.K., 1988, Partially Adaptive Estimation of Regression Models via the Generalized T Distribution, Econometric Theory, Vol. 4, pp Romagnoli, J.A., Sanchez, M.C.,, Data Processing and Reconciliation for Chemical Process Operations, Academic Press. Tjoa, I.B., Biegler, L.T., Comp. Chem. Eng., Vol. 15 No. 1, pp Valko, P., Vadja, S., Comp. Chem. Eng., Vol. 11, pp Wang, D., Romagnoli, J.A., 3. Ind.Eng.Chem.Res.,Vol.4, No.13, pp

ROBUST AND EFFICIENT JOINT DATA RECONCILIATION PARAMETER ESTIMATION USING A GENERALIZED OBJECTIVE FUNCTION

ROBUST AND EFFICIENT JOINT DATA RECONCILIATION PARAMETER ESTIMATION USING A GENERALIZED OBJECTIVE FUNCTION ROBUST AND EFFICIENT JOINT DATA RECONCILIATION PARAMETER ESTIMATION USING A GENERALIZED OBJECTIVE FUNCTION Yen Yen Joe i), iii), David Wang ii), Jose Romagnoli ii) and Arthur Tay i) i) Dept. of Electrical

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information

ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY

ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY ROBUST ESTIMATION OF A CORRELATION COEFFICIENT: AN ATTEMPT OF SURVEY G.L. Shevlyakov, P.O. Smirnov St. Petersburg State Polytechnic University St.Petersburg, RUSSIA E-mail: Georgy.Shevlyakov@gmail.com

More information

ON THE CALCULATION OF A ROBUST S-ESTIMATOR OF A COVARIANCE MATRIX

ON THE CALCULATION OF A ROBUST S-ESTIMATOR OF A COVARIANCE MATRIX STATISTICS IN MEDICINE Statist. Med. 17, 2685 2695 (1998) ON THE CALCULATION OF A ROBUST S-ESTIMATOR OF A COVARIANCE MATRIX N. A. CAMPBELL *, H. P. LOPUHAA AND P. J. ROUSSEEUW CSIRO Mathematical and Information

More information

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Kyriaki Kitikidou, Elias Milios, Lazaros Iliadis, and Minas Kaymakis Democritus University of Thrace,

More information

By 3DYHOýtåHN, Wolfgang Härdle

By 3DYHOýtåHN, Wolfgang Härdle No. 2005 31 ROBUST ESTIMATION OF DIMENSION REDUCTION SPACE By 3DYHOýtåHN, Wolfgang Härdle February 2005 ISSN 0924-7815 Robust estimation of dimension reduction space P. Čížek a and W. Härdle b a Department

More information

MODULE -4 BAYEIAN LEARNING

MODULE -4 BAYEIAN LEARNING MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION (REFEREED RESEARCH) COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION Hakan S. Sazak 1, *, Hülya Yılmaz 2 1 Ege University, Department

More information

Robust model selection criteria for robust S and LT S estimators

Robust model selection criteria for robust S and LT S estimators Hacettepe Journal of Mathematics and Statistics Volume 45 (1) (2016), 153 164 Robust model selection criteria for robust S and LT S estimators Meral Çetin Abstract Outliers and multi-collinearity often

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Introduction to Robust Statistics. Elvezio Ronchetti. Department of Econometrics University of Geneva Switzerland.

Introduction to Robust Statistics. Elvezio Ronchetti. Department of Econometrics University of Geneva Switzerland. Introduction to Robust Statistics Elvezio Ronchetti Department of Econometrics University of Geneva Switzerland Elvezio.Ronchetti@metri.unige.ch http://www.unige.ch/ses/metri/ronchetti/ 1 Outline Introduction

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Lecture 3: Statistical Decision Theory (Part II)

Lecture 3: Statistical Decision Theory (Part II) Lecture 3: Statistical Decision Theory (Part II) Hao Helen Zhang Hao Helen Zhang Lecture 3: Statistical Decision Theory (Part II) 1 / 27 Outline of This Note Part I: Statistics Decision Theory (Classical

More information

Robust high-dimensional linear regression: A statistical perspective

Robust high-dimensional linear regression: A statistical perspective Robust high-dimensional linear regression: A statistical perspective Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics STOC workshop on robustness and nonconvexity Montreal,

More information

Indian Statistical Institute

Indian Statistical Institute Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating

More information

A Modified M-estimator for the Detection of Outliers

A Modified M-estimator for the Detection of Outliers A Modified M-estimator for the Detection of Outliers Asad Ali Department of Statistics, University of Peshawar NWFP, Pakistan Email: asad_yousafzay@yahoo.com Muhammad F. Qadir Department of Statistics,

More information

Introduction Robust regression Examples Conclusion. Robust regression. Jiří Franc

Introduction Robust regression Examples Conclusion. Robust regression. Jiří Franc Robust regression Robust estimation of regression coefficients in linear regression model Jiří Franc Czech Technical University Faculty of Nuclear Sciences and Physical Engineering Department of Mathematics

More information

Testing for Regime Switching in Singaporean Business Cycles

Testing for Regime Switching in Singaporean Business Cycles Testing for Regime Switching in Singaporean Business Cycles Robert Breunig School of Economics Faculty of Economics and Commerce Australian National University and Alison Stegman Research School of Pacific

More information

Parameter estimation by anfis where dependent variable has outlier

Parameter estimation by anfis where dependent variable has outlier Hacettepe Journal of Mathematics and Statistics Volume 43 (2) (2014), 309 322 Parameter estimation by anfis where dependent variable has outlier Türkan Erbay Dalkılıç a, Kamile Şanlı Kula b, and Ayşen

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity

Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity John C. Chao, Department of Economics, University of Maryland, chao@econ.umd.edu. Jerry A. Hausman, Department of Economics,

More information

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Journal of Modern Applied Statistical Methods Volume 15 Issue 2 Article 23 11-1-2016 Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Ashok Vithoba

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS

ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS ROBUST TESTS BASED ON MINIMUM DENSITY POWER DIVERGENCE ESTIMATORS AND SADDLEPOINT APPROXIMATIONS AIDA TOMA The nonrobustness of classical tests for parametric models is a well known problem and various

More information

Regression Clustering

Regression Clustering Regression Clustering In regression clustering, we assume a model of the form y = f g (x, θ g ) + ɛ g for observations y and x in the g th group. Usually, of course, we assume linear models of the form

More information

J. W. LEE (Kumoh Institute of Technology, Kumi, South Korea) V. I. SHIN (Gwangju Institute of Science and Technology, Gwangju, South Korea)

J. W. LEE (Kumoh Institute of Technology, Kumi, South Korea) V. I. SHIN (Gwangju Institute of Science and Technology, Gwangju, South Korea) J. W. LEE (Kumoh Institute of Technology, Kumi, South Korea) V. I. SHIN (Gwangju Institute of Science and Technology, Gwangju, South Korea) G. L. SHEVLYAKOV (Gwangju Institute of Science and Technology,

More information

A Two-Stage Algorithm for Multi-Scenario Dynamic Optimization Problem

A Two-Stage Algorithm for Multi-Scenario Dynamic Optimization Problem A Two-Stage Algorithm for Multi-Scenario Dynamic Optimization Problem Weijie Lin, Lorenz T Biegler, Annette M. Jacobson March 8, 2011 EWO Annual Meeting Outline Project review and problem introduction

More information

Scale Mixture Modeling of Priors for Sparse Signal Recovery

Scale Mixture Modeling of Priors for Sparse Signal Recovery Scale Mixture Modeling of Priors for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Jason Palmer, Zhilin Zhang and Ritwik Giri Outline Outline Sparse

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

An algorithm for robust fitting of autoregressive models Dimitris N. Politis An algorithm for robust fitting of autoregressive models Dimitris N. Politis Abstract: An algorithm for robust fitting of AR models is given, based on a linear regression idea. The new method appears to

More information

More on Unsupervised Learning

More on Unsupervised Learning More on Unsupervised Learning Two types of problems are to find association rules for occurrences in common in observations (market basket analysis), and finding the groups of values of observational data

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

Discussion of Sensitivity and Informativeness under Local Misspecification

Discussion of Sensitivity and Informativeness under Local Misspecification Discussion of Sensitivity and Informativeness under Local Misspecification Jinyong Hahn April 4, 2019 Jinyong Hahn () Discussion of Sensitivity and Informativeness under Local Misspecification April 4,

More information

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression BSTT523: Kutner et al., Chapter 1 1 Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression Introduction: Functional relation between

More information

Lecture 12 Robust Estimation

Lecture 12 Robust Estimation Lecture 12 Robust Estimation Prof. Dr. Svetlozar Rachev Institute for Statistics and Mathematical Economics University of Karlsruhe Financial Econometrics, Summer Semester 2007 Copyright These lecture-notes

More information

Breakdown points of Cauchy regression-scale estimators

Breakdown points of Cauchy regression-scale estimators Breadown points of Cauchy regression-scale estimators Ivan Mizera University of Alberta 1 and Christine H. Müller Carl von Ossietzy University of Oldenburg Abstract. The lower bounds for the explosion

More information

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion 13 Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion Pedro Galeano and Daniel Peña CONTENTS 13.1 Introduction... 317 13.2 Formulation of the Outlier Detection

More information

Regression Analysis for Data Containing Outliers and High Leverage Points

Regression Analysis for Data Containing Outliers and High Leverage Points Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain

More information

A Brief Overview of Robust Statistics

A Brief Overview of Robust Statistics A Brief Overview of Robust Statistics Olfa Nasraoui Department of Computer Engineering & Computer Science University of Louisville, olfa.nasraoui_at_louisville.edu Robust Statistical Estimators Robust

More information

Figure 1. Sketch of various properties of an influence function. Rejection point

Figure 1. Sketch of various properties of an influence function. Rejection point Robust Filtering of NMR Images Petr Hotmar and Jarom r Kukal Prague Institute of Chemical Technology, Faculty of Chemical Engineering, Department of Computing and Control Engineering Introduction The development

More information

OPTIMAL B-ROBUST ESTIMATORS FOR THE PARAMETERS OF THE GENERALIZED HALF-NORMAL DISTRIBUTION

OPTIMAL B-ROBUST ESTIMATORS FOR THE PARAMETERS OF THE GENERALIZED HALF-NORMAL DISTRIBUTION REVSTAT Statistical Journal Volume 15, Number 3, July 2017, 455 471 OPTIMAL B-ROBUST ESTIMATORS FOR THE PARAMETERS OF THE GENERALIZED HALF-NORMAL DISTRIBUTION Authors: Fatma Zehra Doğru Department of Econometrics,

More information

Optimal Implementation of On-Line Optimization

Optimal Implementation of On-Line Optimization Optimal Implementation of On-Line Optimization Xueyu Chen and Ralph W. Pike Louisiana State University Baton Rouge, Louisiana USA Thomas A. Hertwig IMC Agrico Company Convent, Louisiana USA Jack R. Hopper

More information

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL B. N. MANDAL Abstract: Yearly sugarcane production data for the period of - to - of India were analyzed by time-series methods. Autocorrelation

More information

Robust regression in R. Eva Cantoni

Robust regression in R. Eva Cantoni Robust regression in R Eva Cantoni Research Center for Statistics and Geneva School of Economics and Management, University of Geneva, Switzerland April 4th, 2017 1 Robust statistics philosopy 2 Robust

More information

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Statistics 203: Introduction to Regression and Analysis of Variance Course review Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying

More information

9. Robust regression

9. Robust regression 9. Robust regression Least squares regression........................................................ 2 Problems with LS regression..................................................... 3 Robust regression............................................................

More information

Data Reconciliation: Measurement Variance, Robust Objectives and Optimization Techniques

Data Reconciliation: Measurement Variance, Robust Objectives and Optimization Techniques Data Reconciliation: Measurement Variance, Robust Objectives and Optimization Techniques Technical Report Olli Suominen November 2015, Tampere, Finland Version 1.0 CLEEN MMEA program, WP2. DL 2.1.2, FP4

More information

AC : A MODULE FOR TEACHING BATCH OPERATIONS

AC : A MODULE FOR TEACHING BATCH OPERATIONS AC 2010-726: A MODULE FOR TEACHING BATCH OPERATIONS Richard Turton, West Virginia University Richard Turton received his B.S. degree from the University of Nottingham and his M.S. and Ph.D. degrees from

More information

Uncertainty Quantification for Inverse Problems. November 7, 2011

Uncertainty Quantification for Inverse Problems. November 7, 2011 Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance

More information

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania Nonlinear and/or Non-normal Filtering Jesús Fernández-Villaverde University of Pennsylvania 1 Motivation Nonlinear and/or non-gaussian filtering, smoothing, and forecasting (NLGF) problems are pervasive

More information

Probabilistic Machine Learning. Industrial AI Lab.

Probabilistic Machine Learning. Industrial AI Lab. Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear

More information

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science. rofessor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Review of robability For additional review material, see http://eecs.mines.edu/courses/csci507/schedule/

More information

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning Tobias Scheffer, Niels Landwehr Remember: Normal Distribution Distribution over x. Density function with parameters

More information

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart 1 Motivation and Problem In Lecture 1 we briefly saw how histograms

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008 A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. Linear-in-Parameters Models: IV versus Control Functions 2. Correlated

More information

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model 204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:

More information

On a Strategy of Serial Identification with Collective Compensation for Multiple Gross Error Estimation in Linear Steady-State Reconciliation

On a Strategy of Serial Identification with Collective Compensation for Multiple Gross Error Estimation in Linear Steady-State Reconciliation Ind. Eng. Chem. Res. 1999, 38, 2119-2128 2119 On a Strategy of Serial Identification with Collective Compensation for Multiple Gross Error Estimation in Linear Steady-State Reconciliation Qiyou Jiang and

More information

The Bayesian approach to inverse problems

The Bayesian approach to inverse problems The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu

More information

WEIGHTED LIKELIHOOD NEGATIVE BINOMIAL REGRESSION

WEIGHTED LIKELIHOOD NEGATIVE BINOMIAL REGRESSION WEIGHTED LIKELIHOOD NEGATIVE BINOMIAL REGRESSION Michael Amiguet 1, Alfio Marazzi 1, Victor Yohai 2 1 - University of Lausanne, Institute for Social and Preventive Medicine, Lausanne, Switzerland 2 - University

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Leverage effects on Robust Regression Estimators

Leverage effects on Robust Regression Estimators Leverage effects on Robust Regression Estimators David Adedia 1 Atinuke Adebanji 2 Simon Kojo Appiah 2 1. Department of Basic Sciences, School of Basic and Biomedical Sciences, University of Health and

More information

A general linear model OPTIMAL BIAS BOUNDS FOR ROBUST ESTIMATION IN LINEAR MODELS

A general linear model OPTIMAL BIAS BOUNDS FOR ROBUST ESTIMATION IN LINEAR MODELS OPTIMAL BIAS BOUNDS FOR ROBUST ESTIMATION IN LINEAR MODELS CHRISTINE H. MOLLER Freie Universitbt Berlin 1. Mathematisches Instilut Arnimallee 2-6 0-14195 Berlin Germany Abstract. A conditionally contaminated

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR ph CONTROL

DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR ph CONTROL DESIGN OF AN ON-LINE TITRATOR FOR NONLINEAR CONTROL Alex D. Kalafatis Liuping Wang William R. Cluett AspenTech, Toronto, Canada School of Electrical & Computer Engineering, RMIT University, Melbourne,

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

Robust estimation of the Pareto index: A Monte Carlo Analysis

Robust estimation of the Pareto index: A Monte Carlo Analysis Working Papers No. 32/2013 (117) MICHAŁ BRZEZIŃSKI Robust estimation of the Pareto index: A Monte Carlo Analysis Warsaw 2013 Robust estimation of the Pareto index: A Monte Carlo Analysis MICHAL BRZEZINSKI

More information

Multiview Geometry and Bundle Adjustment. CSE P576 David M. Rosen

Multiview Geometry and Bundle Adjustment. CSE P576 David M. Rosen Multiview Geometry and Bundle Adjustment CSE P576 David M. Rosen 1 Recap Previously: Image formation Feature extraction + matching Two-view (epipolar geometry) Today: Add some geometry, statistics, optimization

More information

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract Far East J. Theo. Stat. 0() (006), 179-196 COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS Department of Statistics University of Manitoba Winnipeg, Manitoba, Canada R3T

More information

A Bayesian perspective on GMM and IV

A Bayesian perspective on GMM and IV A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all

More information

Fast approximations for the Expected Value of Partial Perfect Information using R-INLA

Fast approximations for the Expected Value of Partial Perfect Information using R-INLA Fast approximations for the Expected Value of Partial Perfect Information using R-INLA Anna Heath 1 1 Department of Statistical Science, University College London 22 May 2015 Outline 1 Health Economic

More information

In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence function. In the present chapter, the

In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence function. In the present chapter, the Chapter 3 Robustness Properties of the Student t Based Pseudo Maximum Likelihood Estimator In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence

More information

Bayesian Models for Regularization in Optimization

Bayesian Models for Regularization in Optimization Bayesian Models for Regularization in Optimization Aleksandr Aravkin, UBC Bradley Bell, UW Alessandro Chiuso, Padova Michael Friedlander, UBC Gianluigi Pilloneto, Padova Jim Burke, UW MOPTA, Lehigh University,

More information

Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems

Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems Modern Applied Science; Vol. 9, No. ; 05 ISSN 9-844 E-ISSN 9-85 Published by Canadian Center of Science and Education Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity

More information

Nonlinear ph Control Using a Three Parameter Model

Nonlinear ph Control Using a Three Parameter Model 130 ICASE: The Institute of Control, Automation and Systems Engineers, KOREA Vol. 2, No. 2, June, 2000 Nonlinear ph Control Using a Three Parameter Model Jietae Lee and Ho-Cheol Park Abstract: A two parameter

More information

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions

A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions A Shape Constrained Estimator of Bidding Function of First-Price Sealed-Bid Auctions Yu Yvette Zhang Abstract This paper is concerned with economic analysis of first-price sealed-bid auctions with risk

More information

Robust estimation, efficiency, and Lasso debiasing

Robust estimation, efficiency, and Lasso debiasing Robust estimation, efficiency, and Lasso debiasing Po-Ling Loh University of Wisconsin - Madison Departments of ECE & Statistics WHOA-PSI workshop Washington University in St. Louis Aug 12, 2017 Po-Ling

More information

Relevance Vector Machines for Earthquake Response Spectra

Relevance Vector Machines for Earthquake Response Spectra 2012 2011 American American Transactions Transactions on on Engineering Engineering & Applied Applied Sciences Sciences. American Transactions on Engineering & Applied Sciences http://tuengr.com/ateas

More information

Midwest Big Data Summer School: Introduction to Statistics. Kris De Brabanter

Midwest Big Data Summer School: Introduction to Statistics. Kris De Brabanter Midwest Big Data Summer School: Introduction to Statistics Kris De Brabanter kbrabant@iastate.edu Iowa State University Department of Statistics Department of Computer Science June 20, 2016 1/27 Outline

More information

Linear model selection and regularization

Linear model selection and regularization Linear model selection and regularization Problems with linear regression with least square 1. Prediction Accuracy: linear regression has low bias but suffer from high variance, especially when n p. It

More information

ROBUST TESTS ON FRACTIONAL COINTEGRATION 1

ROBUST TESTS ON FRACTIONAL COINTEGRATION 1 ROBUST TESTS ON FRACTIONAL COINTEGRATION 1 by Andrea Peters and Philipp Sibbertsen Institut für Medizininformatik, Biometrie und Epidemiologie, Universität Erlangen, D-91054 Erlangen, Germany Fachbereich

More information

Data reconciliation: a robust approach using contaminated distribution.

Data reconciliation: a robust approach using contaminated distribution. Data reconciliation: a robust approach using contaminated distribution. Moustapha Alhaj-Dibo, Didier Maquin, José Ragot To cite this version: Moustapha Alhaj-Dibo, Didier Maquin, José Ragot. Data reconciliation:

More information

Adaptive Filter Theory

Adaptive Filter Theory 0 Adaptive Filter heory Sung Ho Cho Hanyang University Seoul, Korea (Office) +8--0-0390 (Mobile) +8-10-541-5178 dragon@hanyang.ac.kr able of Contents 1 Wiener Filters Gradient Search by Steepest Descent

More information

COMP90051 Statistical Machine Learning

COMP90051 Statistical Machine Learning COMP90051 Statistical Machine Learning Semester 2, 2017 Lecturer: Trevor Cohn 2. Statistical Schools Adapted from slides by Ben Rubinstein Statistical Schools of Thought Remainder of lecture is to provide

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Institute of Statistics and Econometrics Georg-August-University Göttingen Department of Statistics

More information

Bayesian Networks. Motivation

Bayesian Networks. Motivation Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations

More information

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Applied Econometrics (MSc.) Lecture 3 Instrumental Variables Estimation - Theory Department of Economics University of Gothenburg December 4, 2014 1/28 Why IV estimation? So far, in OLS, we assumed independence.

More information

Modeling Multiscale Differential Pixel Statistics

Modeling Multiscale Differential Pixel Statistics Modeling Multiscale Differential Pixel Statistics David Odom a and Peyman Milanfar a a Electrical Engineering Department, University of California, Santa Cruz CA. 95064 USA ABSTRACT The statistics of natural

More information

Measuring robustness

Measuring robustness Measuring robustness 1 Introduction While in the classical approach to statistics one aims at estimates which have desirable properties at an exactly speci ed model, the aim of robust methods is loosely

More information

Bootstrap Goodness-of-fit Testing for Wehrly Johnson Bivariate Circular Models

Bootstrap Goodness-of-fit Testing for Wehrly Johnson Bivariate Circular Models Overview Bootstrap Goodness-of-fit Testing for Wehrly Johnson Bivariate Circular Models Arthur Pewsey apewsey@unex.es Mathematics Department University of Extremadura, Cáceres, Spain ADISTA14 (BRUSSELS,

More information

Robust Variable Selection Through MAVE

Robust Variable Selection Through MAVE Robust Variable Selection Through MAVE Weixin Yao and Qin Wang Abstract Dimension reduction and variable selection play important roles in high dimensional data analysis. Wang and Yin (2008) proposed sparse

More information

Definitions of ψ-functions Available in Robustbase

Definitions of ψ-functions Available in Robustbase Definitions of ψ-functions Available in Robustbase Manuel Koller and Martin Mächler July 18, 2018 Contents 1 Monotone ψ-functions 2 1.1 Huber.......................................... 3 2 Redescenders

More information

7. Forecasting with ARIMA models

7. Forecasting with ARIMA models 7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability

More information

A COLLABORATIVE 20 QUESTIONS MODEL FOR TARGET SEARCH WITH HUMAN-MACHINE INTERACTION

A COLLABORATIVE 20 QUESTIONS MODEL FOR TARGET SEARCH WITH HUMAN-MACHINE INTERACTION A COLLABORATIVE 20 QUESTIONS MODEL FOR TARGET SEARCH WITH HUMAN-MACHINE INTERACTION Theodoros Tsiligkaridis, Brian M Sadler and Alfred O Hero III, University of Michigan, EECS Dept and Dept Statistics,

More information