Multitask Learning of Environmental Spatial Data

Similar documents
Machine Learning of Environmental Spatial Data Mikhail Kanevski 1, Alexei Pozdnoukhov 2, Vasily Demyanov 3

Machine Learning Algorithms for GeoSpatial Data. Applications and Software Tools

Advanced analysis and modelling tools for spatial environmental data. Case study: indoor radon data in Switzerland

Decision-Oriented Environmental Mapping with Radial Basis Function Neural Networks

Analysis of Interest Rate Curves Clustering Using Self-Organising Maps

Environmental Data Mining and Modelling Based on Machine Learning Algorithms and Geostatistics

arxiv: v1 [q-fin.st] 27 Sep 2007

Space-Time Kernels. Dr. Jiaqiu Wang, Dr. Tao Cheng James Haworth University College London

Follow this and additional works at:

Learning Gaussian Process Models from Uncertain Data

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

MTTTS16 Learning from Multiple Sources

RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS

PATTERN CLASSIFICATION

Statistical Machine Learning from Data

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

Jeff Howbert Introduction to Machine Learning Winter

Transactions on Information and Communications Technologies vol 16, 1996 WIT Press, ISSN

Pattern Recognition and Machine Learning

A. Pelliccioni (*), R. Cotroneo (*), F. Pungì (*) (*)ISPESL-DIPIA, Via Fontana Candida 1, 00040, Monteporzio Catone (RM), Italy.

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

18.6 Regression and Classification with Linear Models

Machine Learning 2010

Reading, UK 1 2 Abstract

Neural networks and support vector machines

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Performance Analysis of Some Machine Learning Algorithms for Regression Under Varying Spatial Autocorrelation

Nonlinear Classification

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

Relevance Vector Machines for Earthquake Response Spectra

Nonparametric Bayesian Methods (Gaussian Processes)

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

Influence of parameter estimation uncertainty in Kriging: Part 2 Test and case study applications

Analysis of Fast Input Selection: Application in Time Series Prediction

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Markov chain optimisation for energy systems (MC-ES)

SINGLE-TASK AND MULTITASK SPARSE GAUSSIAN PROCESSES

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Linear & nonlinear classifiers

Mining Classification Knowledge

Leaving The Span Manfred K. Warmuth and Vishy Vishwanathan

Classifier s Complexity Control while Training Multilayer Perceptrons

Introduction to Support Vector Machines

Neural Networks and the Back-propagation Algorithm

Online Estimation of Discrete Densities using Classifier Chains

Lecture 5: Logistic Regression. Neural Networks

Geographical General Regression Neural Network (GGRNN) Tool For Geographically Weighted Regression Analysis

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Using Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Introduction to Machine Learning

In this chapter, we provide an introduction to covariate shift adaptation toward machine learning in a non-stationary environment.

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data

3.4 Linear Least-Squares Filter

Table of Contents. Multivariate methods. Introduction II. Introduction I

Introduction to Machine Learning

Discriminative Direction for Kernel Classifiers

Unsupervised Learning with Permuted Data

HMM and IOHMM Modeling of EEG Rhythms for Asynchronous BCI Systems

Machine Learning Techniques

Deep unsupervised learning

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Linear & nonlinear classifiers

Learning Multiple Tasks with a Sparse Matrix-Normal Penalty

Radial Basis Function Networks. Ravi Kaushik Project 1 CSC Neural Networks and Pattern Recognition

Classification for High Dimensional Problems Using Bayesian Neural Networks and Dirichlet Diffusion Trees

Statistical Pattern Recognition

Deep Feedforward Networks

Spatial Statistics & R

Classification and Pattern Recognition

Supervised Learning Part I

Artificial neural networks

10-701/ Machine Learning, Fall

Active Sonar Target Classification Using Classifier Ensembles

Machine Learning Support Vector Machines. Prof. Matteo Matteucci

Machine Learning. Regression. Manfred Huber

10-701/ Machine Learning - Midterm Exam, Fall 2010

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

Data Mining Part 5. Prediction

In the Name of God. Lectures 15&16: Radial Basis Function Networks

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data

Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Unsupervised Learning Methods

LEARNING & LINEAR CLASSIFIERS

Support Vector Machines: Training with Stochastic Gradient Descent. Machine Learning Fall 2017

18.9 SUPPORT VECTOR MACHINES

MLPR: Logistic Regression and Neural Networks

Content. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition

Statistical Machine Learning from Data

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Outline. MLPR: Logistic Regression and Neural Networks Machine Learning and Pattern Recognition. Which is the correct model? Recap.

The exam is closed book, closed notes except your one-page cheat sheet.

INTRODUCTION TO DATA SCIENCE

Multi-Layer Boosting for Pattern Recognition

Radial-Basis Function Networks

Transcription:

9th International Congress on Environmental Modelling and Software Brigham Young University BYU ScholarsArchive 6th International Congress on Environmental Modelling and Software - Leipzig, Germany - July 2012 Jul 1st, 12:00 AM Multitask Learning of Environmental Spatial Data Mikhail Kanevski Follow this and additional works at: https://scholarsarchive.byu.edu/iemssconference Kanevski, Mikhail, "Multitask Learning of Environmental Spatial Data" (2012). 9th International Congress on Environmental Modelling and Software. 89. https://scholarsarchive.byu.edu/iemssconference/2012/stream-b/89 This Event is brought to you for free and open access by the Civil and Environmental Engineering at BYU ScholarsArchive. It has been accepted for inclusion in 9th International Congress on Environmental Modelling and Software by an authorized administrator of BYU ScholarsArchive. For more information, please contact scholarsarchive@byu.edu.

International Environmental Modelling and Software Society (iemss) 2012 International Congress on Environmental Modelling and Software Managing Resources of a Limited Planet, Sixth Biennial Meeting, Leipzig, Germany R. Seppelt, A.A. Voinov, S. Lange, D. Bankamp (Eds.) http://www.iemss.org/society/index.php/iemss-2012-proceedings Multitask Learning of Environmental Spatial Data Mikhail Kanevski Institute of Geomatics and Analysis of Risk, University of Lausanne (Mikhail.Kanevski@unil.ch) Abstract: The present research deals with an application of artificial neural networks for multitask learning from spatial environmental data. The real case study (sediments contamination of Geneva Lake) consists of 8 pollutants. There are different relationships between these variables, from linear correlations to strong nonlinear dependencies. The main idea is to construct a subsets of pollutants which can be efficiently modeled together within the multitask framework. The proposed two-step approach is based on: 1) the criterion of nonlinear predictability of each variable k by analyzing all possible models composed from the rest of the variables by using a General Regression Neural Network (GRNN) as a model; 2) a multitask learning of the best model using multilayer perceptron and spatial predictions. The results of the study are analyzed using both machine learning and geostatistical tools. Keywords: Machine learning algorithms; multitask learning; environmental multivariate data; geostatistics 1 INTRODUCTION Multitask learning (MTL) is an important recent trend in machine learning research. Loosely speaking the idea is that learning together several related (to be defined) tasks is more efficient, easier and gives rise to more parsimonious models. In a nonlinear modelling framework, e.g. multilayer perceptron, it means that several tasks (outputs) are learned together by sharing neurons in a hidden layer(s). One of the first detailed researches on this problem was published by Caruana (1997). If the tasks, are not related, for example, output variables are independent, the resulting network is equivalent to two non-interacting ones. There are many variants of MTL and recently the concept was generalized to kernel machines, Evgeniu et al. (2005), and to Gaussian Processes, Bonilla and Williams (2009). The important problem of relatedness of the tasks to be learned can be considered in a different way: from quite simple, like the tasks sharing the same covariance structures to more advanced approach by using SVM-like optimization procedures. The simple difference between single task and multitask learning is illustrated in Figure 1. In the present study we apply multitask learning concept to model and to predict the spatial pollution of sediments in Geneva Lake. There are eight output (dependent) variables (pollutants) measured at 200 points in the Lake. The general problem is the following: in order to make a better prediction of the variable k, which other variables should be added/used to improve the quality of the analysis (better generalization and smaller uncertainty)?

We accept the hypothesis that the input space (independent variables) is fixed for all tasks, but not necessarily known a priori. Input space can include not only the geographical coordinates, but also bathymetry, distance to pollution sources, currents etc., which are potentially important for the predictions. It can be quite complex and high dimensional, like it was observed for wind speed modelling in Foresti et al. (2011). In a more general setting the input spaces can be different. Let us have a look at the output space of the problem. In order to select a group of related variables there are many approaches proposed in machine learning: linear measures like correlations, nonlinear ones like mutual information, clustering of variables according to some similarity criteria, for example, by using self-organizing maps, see Kanevski et al. (2009). The models like PCA can be used to detect corresponding linear relationships between variables. In general, the relationships between variables are non-linear, see Figure 2, where the scatterplot matrix of real data is given. Therefore, the nonlinear modelling algorithm GRNN, is applied in the present study. GRNN is a nonparametric, nonlinear and efficient tool for multidimensional regression problems. Let us remind, that in case of classification problems an analogue of GRNN for categorical data Probabilistic Neural Networks can be used within the framework of the proposed methodology. The description of PNN and their applications can be found in Kanevski et al. (2009). In the present research in order to select the most related (in the sense of MTL) group of variables a criterion based on the predictability of output variable of interest k by using all combinations of other outputs is considered. For example, in order to select the group for Ni pollutant, GRNN considers as an input space all possible combinations of other output variables in order to predict one GRNN output - Ni. The number of possible models is [2 (n-1) 1], where n is the dimension of the input space of GRNN: the number of GRNN dependent variables constructed by using other pollutants. For n=8 it equals to 127 models. The combination giving the minimum of cross-validation error is then chosen as the best. The procedure does not depend explicitly on the structure of the original input space (independent variables) because at this step only the dependent variables (pollutants) are used. Thus, the relatedness of the tasks is quantified via the criterion of nonlinear predictability. From another point of view, it looks like a problem of feature selection the selection of the group of the most relevant features, see, for example, Kanevski and Timonin (2010), Foresti et al. (2011). Figure 1. Principles of single task (left) and multitask right (5 tasks) learning using multilayer perceptrons

In the following sections the methodology and corresponding case study along with some representative results are given. Figure 2. Scatterplot matrix of eight output variables: Cd, Hg, Zn, Cu, Ti, Cr, V, Ni. Table 1. Left relatedness between dependent variables estimated by the best multiple linear regression model. Variables with significant beta (p=0.05) are marked. Right relatedness between dependent variables estimated by GRNN. 2 METHODOLOGY The procedure of using the predictability criterion for the quantification of the relatedness was performed for all pollutants and the corresponding information is summarized in a matrix form (matrix of relatedness ). The matrix of relatedness based on linear multivariate analysis (linear multivariate regression) is shown in Table 1 (left). The modelled variables are given at the left of the matrix while the most linearly related variables (nonzero betas with a p-value = 0.05) are given in

the corresponding rows. The matrix of nonlinear relatedness estimated by the GRNN is also shown in Table 1 (right). The selected models (group of variables) can be then analysed at the second step by using any multivariate model from machine learning algorithms. Thus, the generic methodology proposed consists of two fundamental phases: 1. pre-processing of multivariate data and preparing of multitask related groups for each environmental variable to be studied, 2. multitask learning, testing/validation of the models and spatial predictions. The first phase is based on GRNN. General Regression Neural Networks were introduced by Specht (1991). Below a short description of GRNN is given following Kanevski and Timonin (2010). Let us consider data measurements Z n (n=1, N) in m-dimensional space (x 1,, x m ). According to GRNN, the prediction at unknown point Z(x) is defined by the following formula N m 2 i in Zn exp 2 n= 1 i= 1 2σ i 1 xm = N m 2 D ( xi, xin ) exp 2 n= 1 i= 1 2σ i Z( x,... ) D ( x, x ) where D 2 ( xi, x in) is a distance between point of prediction and data, σ i is a kernel bandwidth corresponding to the i-th input dimension (i=1, m). It is supposed that interaction between input variables is negligible: the procedure of feature selection and not a feature extraction is applied. Nevertheless, in a more general setting a full covariance matrix (Mahalanobis distance) or some data whitening procedures can be considered. The only unknown parameters in GRNN model are the kernel bandwidths ( σ i ). These parameters are tuned either by splitting of data into training and validation subsets and trying to minimize the validation error by choosing corresponding bandwidths or (if there are not too many data) by using the leave-one-out criterion (called the cross-validation in geostatistics). Gradient search from a good starting point, for example estimated by isotropic kernel model, is another fast and efficient solution to minimize the crossvalidation (or validation) error. Let us remind some useful properties of GRNN for data processing: 1) for an isotopic kernel and the cross-validation error the minimisation solution is unique; 2) when there are no spatial structures, i.e. only spatially not correlated noise, there is no minimum on a cross-validation curve and the kernel bandwidth tend to infinity (i.e. it is much larger than the region of the study or maximal distance between measurements). This is also true when only some coordinates are very noisy and useless for the prediction (if k-input variable is a noise, thenσ k ). In such cases 2 2 D ( xk, x ) kn D ( xk, x ) kn from: 1 it follows that: exp 1 2 2 2σ k 2σ k and the corresponding coordinate k disappears (does not contribute to the solution). This phenomenon was observed in Specht and Romsdahl (1994).

Recently the same idea was applied for the feature selection tasks in Gheyas and Smith (2010). More details on theory, training and applications of GRNN for spatial environmental data, can be found in Kanevski et al. (2009). GRNN can also be used to analyse the residuals after the modelling by GRNN or other algorithms, like multilayer perceptrons, support vector machines etc. taking into account useful properties of GRNN mentioned above. GRNN is a good discriminatory (pattern/noise) tool. A straightforward and direct approach in using isotropic GRNN for the selection of good group of variables/features is to consider all possible combinations of the models = [2 (n-1) -1], and to select the best one, i.e. the one, which minimizes the leave-one-out error. This approach was considered in Kanevski and Timonin (2010) and Gheyas and Smith (2010), where the traditional feature selection task was considered - the selection of the best combination of the most relevant inputs. When the number of variables is large, let us say >20, this approach is practically not possible and different approximate algorithms can be used, e.g. genetic algorithms, simulated annealing, to make the problem feasible, see Gheyas and Smith (2010). For the second phase a multilayer perceptron (MLP) will be used in the present study. Now, an input and output spaces have to be constructed. The output space was prepared during the first phase: the most related groups of dependent variables were selected by GRNN. The simplest input space geographical space composed of only longitude and latitude is considered in this paper. This space is typical in geostatistical studies of pollution mapping. Moreover, working in a geographical space permits also to use the geostatistical tools variography, to control and to quantify the quality of MLP estimates, like it was proposed in Kanevski and Maignan (2004). In fact, MLP can be replaced by any other nonlinear multivariate model, but it has some interesting for this research properties: after training the weights can be analysed to understand the results of MTL, an application of shuffled variables can be considered as a particular regularization procedure (noise injection), as it was demonstrated, for example, in Kanevski et. al. (2009). It is well known that MLP application for environmental tasks is not a trivial problem, see ASCE Tasks (2000). Many of the problems - model selection, training/overfitting, etc. can be solved if the level of noise in data is estimated independently. Then, only the structured information can be extracted from the data without fitting the noise. In the present paper the variography is used. The variography is applied both for the original data and for the residuals (see Figure 5). The level of noise and small scale variability in data were modelled with a nugget effect in the variogram. Finally, the problem of MLP selection is simplified, e.g. by applying an early stopping criterion at the level of noise. 3 RESULTS Let us consider a real case study and some results. In the present paper, for the brevity of presentation, the main results only for Ni are given. In order to test the efficiency of the approach and algorithm two spatially non-structured variables were additionally generated by shuffling the original data (shuffling = randomization, which preserves the global distribution but destroys the spatial patterns). Therefore, 10 output variables were considered (8 real - Cd, Hg, Zn, Cu, Ti, Cr, V, Ni; + 2 shuffled Zn and Ni) which gives rise to the total number [(2 (n-1) -1) = 511] of possible models. It is expected that the models with shuffled data will be neglected according to the cross-validation error criterion.

In order to compare the results of modelling at the second step of the analysis, the original 200 measurements were split into training (150) and validation (50) subsets. The splitting of data was carried out using spatial declustering procedures, which guaranties that all geographical regions are represented and the variability of data are preserved in both subsets. The same procedure was performed also for MTL using the non-optimal models as well. Because of the different variability of pollutants all variables before modelling were projected from original min-max values to the interval [0,1]. It helps during the training procedures and in the interpretation of the results. In Figure 3 the results for Ni (models of GRNN MTL pre-processing), are ranked according to the cross-validation error (the best model is left). Only the top ten models are shown. The dark colour codes the selected variables for a particular model/group, organised by columns. The models including the last two variables, shuffled Ni and Zn, were never selected. The same analysis was carried out for all eight pollutants and the results are summarized in Table 1 (right). Figure 3. The structure of best (from left to right) models for Ni prediction. Important co-variables are the following: Cd, Zn, Cr, and V. Two variables at the bottom are shuffled data on Zn and Ni. Figure 4. The structure of the best (from left to right) models for Zn prediction. For the best (left) model all co-variables with the exception of shuffled are important.

The ranking of models according to the cross-validation error for Zn pollutant is shown in Figure 4. It follows, that for Zn multitask learning all other pollutants should be taken into account. After the detailed pre-analysis described above, a nonlinear multitask MLP model was developed and applied. The input layer of MLP consists of geographical coordinates, one or two hidden layers with a varying number of neurons selected by minimizing the validation error. Typical number of hidden neurons was between 7 and 15. The output layer includes the best group of variables selected by GRNN at the MTL pre-processing step. The best model developed was compared also with other models when the variables in MTL group were replaced either by shuffled ones or by non-optimal variables, for example for Ni instead of Zn, Hg was used. For the best MTL model the mean square error was always lower (10% - 30%). Finally, the quality of the modelling procedure was also controlled by a geostatistical tool variogram, see Kanevski and Maignan (2004). This test can be easily done in a low dimensional space. In Figure 5 the experimental variogram of raw Ni data and the theoretical variogram model are compared with the variogram of the training residuals of the best MTL model. For the clarity of interpretation only the isotropic variograms are shown. The variogram of the residuals demonstrates pure nugget effect no spatial structures in the residuals. Moreover, the level of the nugget exactly corresponds to the nugget value in the theoretical model, which means that after modelling all structured information was correctly extracted and the residuals correspond to the noise level in raw data. Figure 5. The results of Ni mapping (top) and the variography of raw data and training residuals. 4 CONCLUSIONS The paper presents a new two stage approach to multitask learning of spatial environmental multivariate data: 1) pre-processing of data for multitask learning and preparing of related groups of the variables; 2) analysis, testing and nonlinear multivariate modelling by using the best MTL group.

The approach is quite generic, does not depend on the structure of the input space and can be applied for a variety of environmental data mining tasks: spatial pollution data, environmental monitoring time series data, renewable resources assessments, natural hazards, etc. The relatedness of the variables is quantified using the nonlinear predictability criterion. For a reasonable number of dependent variables ( 20) the complete space of solutions can be explored in order to select the best group. When the number of variables is higher some empirical optimization procedures, like simulated annealing, genetic algorithms etc. can be applied. The obtained results demonstrate the efficiency of the approach for the multitask learning of environmental phenomena. Moreover, the criterion of nonlinear predictability can be useful itself for better understanding of the relationships between the variables. The future research will deal with the generalization of the approach to spatio-temporal data and the application for other challenging case studies. ACKNOWLEDGMENTS The research was carried out within the framework of Swiss National Science Foundation projects 200020-121835 and 200021-126505. The author thanks to Dr. V. Timonin for many useful discussions on GRNN modelling. The constructive comments of the reviewers enabled to improve the paper. REFERENCES ASCE Task Committee on Application of Artificial Neural Networks in Hydrology. Journal of Hydrologic Engineering, vol. 5, 2000. Bonilla E., K., and Ch. Williams. Multi-task Gaussian Process Prediction, Advances in Neural Information Processing Systems, 20, MIT Press, 2009. Caruana R., Multi-Task Learning. Machine Learning, 28, 41-75, 1997. Evgeniu T., Ch. Micchelli, and M. Pontil. Learning Multiple Tasks with Kernel Methods. Journal of Machine Learning Research, 6, 615-636, 2005. Foresti L., D. Tuia, M. Kanevski, and A. Pozdnoukhov. Learning wind fields with multiple kernels. Stochastic Environmental Research and Risk Assessment, vol. 25, 51-66, 2011. Gheyas I, and L. Smith. Feature subset selection in large dimensionality domains. Pattern recognition, 43, 5-13, 2010. Kanevski M. and M. Maignan. Analysis and Modelling of Spatial Environmental Data. EPFL Press, Lausanne Switzerland, 2004. Kanevski M. and V. Timonin. On Automatic Mapping of Environmental Data Using Adaptive General Regression Neural Network, paper presented at GIS Research UK conference, London 2010. Kanevski, M., A. Pozdnoukhov, and V. Timonin. Machine Learning for Spatial Environmental Data: Theory, Applications and Software. EPFL Press, Lausanne, 2009. Specht D. and H. Romsdahl. Experience with Adaptive Probabilistic Neural Networks and Adaptive General Regression Neural Networks. Proceedings of the IEEE World Congress on Computational Intelligence, 2, 203-1208, 1994. Specht D. A General Regression Neural Network. IEEE Transaction on Neural Networks., 2 pp. 568-576, 1991.