Deep Neural Network Based Demand Side Short Term Load Forecasting
|
|
- Christiana Pope
- 5 years ago
- Views:
Transcription
1 Deep Neural Network Based Demand Side Short Term Load Forecasting Seunghyoung Ryu Department of Electronic Engineering Sogang University Seoul, Korea Jaekoo Noh KEPCO Research Institute Korea Electric Power Corporation Daejeon, Korea Hongseok Kim Department of Electronic Engineering Sogang University Seoul, Korea Abstract In smart grid, one of the most important research areas is load forecasting; it spans from traditional time series analysis to recent machine learning approach and mostly focuses on forecasting aggregated electricity consumption. However, the importance of demand side energy management including individual load forecasting is becoming critical. In this paper, we propose deep neural network (DNN) based load forecasting models, and apply them to demand side empirical load database. DNNs are trained by two different ways: pre-training restricted Boltzmann machine and using rectified linear unit without pre-training. DNN forecasting models are trained by individual customer s electricity consumption data and regional meteorological elements. To verify the performance of DNNs, forecasting results are compared with shallow neural network (SNN), and double seasonal Holt- Winters (DSHW) model. The mean absolute percentage error (MAPE) and relative root mean square error (RRMSE) are used for verification. The results show that DNNs exhibit accurate and robust forecasts compared to other forecasting models, e.g., MAPE and RRMSE are reduced by up to 17% and 22% compared to SNN, and 9% and 29% compared to DSHW. I. INTRODUCTION The electricity load forecasting is one of the main research areas in smart grid and can be classified into several categories according to its target forecasting ranges. Among these, short term load forecasting (STLF), which focuses on forecasting load of hours or days ahead, secures attentions in the smart grid because of its wide applicability to demand side management, energy storage operation, peak load reduction, etc. One of the main characteristics of the smart grid is bidirectional communications, which breaks down the border between electricity generation and consumption. Bidirectional communications is realized with the deployment of advanced metering infrastructure (AMI), which is a prerequisite of many smart grid services such as 15 minute interval load monitoring, real time pricing, demand response, and trouble detection. However, the collapse of the border between generation and consumption increases the grid complexity. Now consumers can produce electricity from renewable sources such as photovoltaic cells and wind turbines. Furthermore, by using energy storage system (ESS), consumers can actively shape their power demand. In addition to these changes, deregulation and privatization of electricity markets make utility companies offer smart end-user services to their customers, and thus customer load forecasting becomes one of the basic elements in user services. These new characteristics imply that the roll of consumers would be more significant than before because the scope of forecasting is shifted to microscopic view from macroscopic view for stable electricity operation. There are many research activities on forecasting models from classical time series analysis to recent machine learning approach. Time series analysis [1], one of the most popular load forecasting methods, includes autoregressive integrated moving average (ARIMA), exponential smoothing, and Kalman filter. In [2], the authors developed a double seasonal Holt-Winters model (exponential smoothing) that showed better results than other models like ARIMA or neural networks. Another branch of load forecasting is to use an artificial neural network (ANN) model [3], [4]. Since ANN can model nonlinearity, it is widely used for various forecasting applications. The most common type of ANNs is a multilayer perceptron (MLP) that forecasts a load profile using previous load data (e.g., [5]). ANN based STLF models are well summarized in [6]. A deep neural network (DNN) is an ANN with more layers than typical three layered MLP. Deep structure increases the feature abstraction capability of neural networks. Recently, the progress of Internet of things (IoT) and big data enables the adoption of DNNs in diverse research fields. However, to the best of our knowledges, we could barely find DNN based electricity load forecasting models. For estimating building energy consumption, conditional restricted Boltzmann machine and factored conditional restricted Boltzmann machine are used in [7], and deep belief network (DBN) based reinforcement learning forecasting model is developed in [8]. In [9], DNN is used for time series wind forecasting, and DBN was used as a part of ensemble learning in [1]. In [11], the author trained DNN with discriminative pre-training method for electricity load forecasting. In this paper, we adopt DNN as a demand side day-ahead STLF model and analyze its forecasting results of 4 dominant industrial customers in Korea. We train DNNs in two different ways: pre-training restricted Boltzmann machine (RBM) and using rectified linear unit (ReLU) without pre-training. We compare DNN forecasting results with typical three layered shallow neural network (SNN), ARIMA, and double seasonal Holt-Winters (DSHW) model /16/$ IEEE
2 The rest of this paper is organized as follows. In Section II, basic ANNs and DNNs will be introduced. In Section III, we propose a DNN-based framework, and implementation issues of forecasting model will be followed in Section IV. Experimental results and conclusion will be given in Section V and VI, respectively. II. METHODOLOGY Recently, DNNs show tremendous progresses in many research fields such as acoustic modeling, natural language processing, and image recognition. A layout of ANNs had remained shallow due to the problems while training DNNs such as vanishing gradient, overfitting, and lack of computational power. In 26, Hinton et al., published a seminal paper about deep learning [12], and after that deep learning has been developing rapidly. With deeper layers, DNNs have an ability to capture highly abstracted features from data. Since load profiles have nonlinear characteristics among various factors that affect the shapes of load patterns, it is reasonable to use DNN as a forecast model. In this paper, we adopt two different DNN models to learn complicated relations between weather variables, date, and previous consumptions for individual customers. Then, the DNN produces a dayahead forecast of 24 hours load profile according to past observations. A. Artificial neural networks ANN is a system of connected artificial neurons, and it has an ability to model any arbitrary nonlinear function [13]. A typical structure of ANN forecasting model is based on MLP. MLP has three-layered structure; input, hidden, and output layers. Each layer consists of neurons without intralayer connections. However, between the layers, there exist full weighted connections among neurons. The elements of MLP are as follows. By using a superscript l to denote a layer, the output vector y of the layer l is y l = h(z l )=h((w l ) T y l 1 + b l ) (1) where h( ) represents a nonlinear vector activation function, W l =[wij l ] and bl represent a weight matrix and a bias vector, respectively. The back propagation (BP) algorithm is used to update weight matrices to minimize the error between target value and the value produced by MLP. Then, wij l is updated by w l l ij where represents learning rate and E is the error. wij l can be derived by the chain rule [14]. In the BP algorithm, error signal propagates backwards to calculate wij l. In SNN, propagates well to the input layer. However in DNN, due to the gradient of typical activation functions is zero in the saturation area, also goes to zero as it propagates toward input layer. This problem of training DNNs is called vanishing gradient and made it hard to train DNNs. (2) B. Deep neural networks We adopt two techniques to overcome the difficulties of training deep structure. The first method is to pre-train the DNN by stacking RBMs, which is known as DBN. It is known that neural networks with multiple hidden layers can be well trained by the BP algorithm when weights are initialized by stacked RBMs [12]. Also, pre-training leads the network to better local minima that support better generalization [15]. RBM is a neural network that has only two layers called visible and hidden layer. Typical RBM has binary nodes in each layer denoted by v = {v i } and h = {h j }, respectively. For any (v i,h j ) pair, there exists a bidirectional connection, but no intra-layer connection. The probability to have specific (v, h) configuration is determined by its energy E(v, h) in the form of p(v, h) = 1 Z e E(v,h) (3) where Z = P v,h e E(v,h) is a normalization factor. With bias vector b v for visible, b h for hidden and weight matrix W between v and h, the energy of RBM is defined as [16] E(v, h) = b T v v b T h h v T Wh. (4) Since there is no intra-layer connection, the states of the visible and hidden layer are conditionally independent to each other. Then the probability of a specific node to turn on is p(h j =1 v) = (b h,j + v T W j ), (5) p(v i =1 h) = (b v,i + h T (W T ) i ) (6) where (x) = 1/(1 + e x ) and W j represents the j-th column vector of the matrix W. RBM is then trained to maximize the log probability of v over given training data. The contrastive divergence algorithm is widely known to update W. It performs k-step Gibbs sampling to generate a reconstructed image of given training data and updates W with w ij = (hv, hi data hv, hi recon ) (7) where is a learning rate and the angle brackets are expectations under the distribution of subscript [16]. Geometrically, W is the direction of lowering/increasing the energy of training/reconstructed data. The RBM pre-training method initializes DNN with weights of stacked RBMs. RBM stacking process is as follows. First, train the RBM with training data. Then its upper RBM is trained from the hidden layer of the previous RBM. The hidden layer of the lower RBM becomes the visible layer of its new upper RBM. In this manner, learning and stacking procedure iterates until having a desired network structure. After unsupervised RBM pre-training, supervised fine tuning is performed with the BP algorithm. Because the network already knows the feature of the given training data, pretraining method helps to train DNN. Another method to train DNN is much simpler than pretraining: using ReLUs instead of typical sigmoid units [17]. The ReLU has a rectifier h(z) = max(,z) as an activation
3 Fig. 1. The proposed DNN short term load forecasting framework. function. Since the gradient of the rectifier is 1 if z, and otherwise, error signal is still alive when it propagates backward to the lowest layer, and DNN can be trained well with the BP algorithm without the vanishing gradient problem. Also using ReLU has advantage in training time due to the characteristic of its gradient. III. PROPOSED FRAMEWORK In this section, we describe the proposed load forecasting framework using DNNs as depicted in Fig. 1. The framework is based on the knowledge discovery in databases (KDD) process and consists of three modules: data processing module, DNN training module, and DNN forecasting module. A. Data processing The upper part of Fig. 1 shows data processing module. First, we acquire hourly measured customer load data provided by Korea Electric Power Corporation (KEPCO). Each customer has information about time, power consumptions, province, and industrial category in accordance with the Korea Standard Industrial Classification (KSIC). Customers consist of dominant power contractors in their province and industrial category. There are eight representative industrial categories where we randomly select five customers per category for this study. We also use weather data to forecast load profiles. From the regional weather database of the National Climate Data Service System (NCDSS), we select five parameters after examining correlation analysis result: daily average temperature, humidity, solar radiations, cloud cover, and wind speed. Next, accumulated time series load and weather data go through data preprocessing step. Data preprocessing contains data cleansing, normalization, and structure change. We find that load measurements have some defective data such as missing values and zero values. In data cleansing step, these defects are replaced with the average of previous loads. After Parameter y 1,y 2,...,y 24 x 1,x 2,...,x 24n x 24 n+1,...,x 26 n x 26 n+1,...,x 26 n+5 x 26 n+6,...,x 26 n+9 TABLE I TRAINING DATA COMPOSITION Configuration label, 24 hours normalized load profile past n normalized load profile date information of past n days (day of the week, weekday) weather parameters correspond to label date information of label (season, month, day of the week, weekday) data cleansing, all load and weather data are normalized with its maximum value. If training data has large values, weight matrix need be extremely small to make the weighted sum within appropriate input range of sigmoid function. However, small weights make it hard to use the BP algorithm, so we normalize the database first, and denormalize forecasting result to obtain valid load forecasts. The third part of preprocessing is restructuring the given database into training and test data. The training data contains labels (desired output) and input variables. For example, each label is a normalized daily load profile with 24 dimensions to obtain daily load profile from corresponding input vector. Input variables could be any parameters that affect electricity consumption. The input data consists of past electricity consumptions, weather parameters for target date, season, month, and date indicators. With a time window index n, the composition of training vector is described in Table I. Test data is used for evaluating accuracy of the proposed forecasting model and they are not used in training step. We construct validation data and exploit validation error as an indicator for selecting proper parameters. In the experiment, we apply 1-fold cross validation to determine the number of epochs. B. Training & forecasting The DNN training module is the main part of this framework, and the implementation details are given in Section IV. We first determine the structure of DNN, e.g., the numbers of hidden layers and neurons. The number of neurons in input and output layers are fixed to the dimension of training data and the period of forecast load profile. In the case of using 15 minute load data, 15 minute interval load forecasts can be obtained by adding more input and output neurons that correspond to each time slot. Once DNN is created, we select a way to train DNN (pre-training or ReLU), and DNN learns nonlinear relations between load profiles and past observations. The last part of the framework is the forecasting module. By using the trained DNN, we obtain forecasts of target dates. Then model accuracy is evaluated by test data. The mean absolute percentage error (MAPE) and relative root mean square error (RRMSE) are used for error measure. IV. IMPLEMENTATION OF DNNS We construct DNNs for each customer using R with darch package [18], and train with various customer s electricity consumption data.
4 12 1 Training Data Test Data Fig. 2. The structure of DNN in proposed STLF framework. The number of neurons is shown above of each layer. A. Size of data There could be some issues of using DNNs as a forecasting model if the size of load data is limited. DNNs shows remarkable performance in other fields with huge training database. For example, the ImageNet large scale visual recognition challenge in 214 has more than.4m training images. Using pre-training is not necessary with large training data. But load data is currently limited in numbers because gathering demand side load data just started recently. We could obtain roughly 75 daily load profiles per customer for three years except holiday. In contrast, the number of parameters in DNN may outnumber training data. This size-related problem could lead overfitting that memorizes the training data rather than learning. Therefore, we need to consider proper DNN structure and overfitting. B. Structure of DNNs The structure of DNNs is mainly determined by the numbers of layers and neurons. The numbers of input neurons and output neurons are automatically fixed by the dimension of training data and forecasting period. By adjusting the number of input and output neurons, DNN could deal with different measuring interval and forecasting period. In [19], the authors used heuristic method to choose the number of hidden neurons. The heuristics may work with SNN because there are only two possible actions: increase or decrease. However, in the case of DNNs, it is not feasible to test all possible combinations of layers and neurons since we need to consider both width and depth of network. Therefore we determine the structure by trial and error (adding or subtracting 5 neurons for each layer), and finally select four hidden layers and 15 neurons per layer. Hidden layers have either sigmoid unit for pre-training or ReLU, and the output layer has linear units to construct load profile. The proposed DNN structure is described in Fig. 2. In the experiment, we train DNNs with mini batch and resilient backpropagation [2]. In addition we train DNNs without nonbusiness day to focus on load profile of business day, and time window of five is used. C. Resolving overfitting Overfitting can be recognized by observing a learning curve that describes error rate according to the number of epochs. MAPE (%) Epochs Fig. 3. An example of customer learning curve. Errors of both training and test data gradually decrease. When overfitting occurs, there exists a certain point where test error starts to increase while training error still decreases. This means that model memorizes the given training data, but it cannot forecast well in real situation. Thus too much training with complex forecasting model leads to bad generalization. There are several methods to prevent overfitting such as dropout [21] or early stopping. However, when we observe the learning curve of DNNs, we did not observe overfitting occurs; test error also gradually decreases as training error does in both pre-training and ReLU models. An example of customer s learning curve with 1 days of test data is depicted in Fig. 3. Since customers have different learning curves, we apply k- fold cross validation to determine stopping criterion. From this study, the size of electricity consumption data seems less critical, but it may require further verification with different network structures and/or when more data are available later. V. EXPERIMENTAL RESULTS In this section, experimental results of STLF by DNNs are given. MAPE and RRMSE are used to measure the accuracy of daily forecasts. Experiments are focused on forecasting load profiles of business days. Forecasting results of DNNs are compared with SNN, ARIMA, and DSHW model in order to verify DNN as a forecasting model of individual customers. ARIMA and DSHW models are created for each daily forecast with past 8 weeks. A. Case 1: single load type We first present the result about one selected customer and analyze the forecasting results of four seasons. The customer belongs to networking business category and has typical daily load patterns with seasonal variation. Fig. 4 shows the customer s electricity consumptions of 212 to 214. There exists significant load increase in summer, and the average hourly electricity consumption is 1,642 kwh. We forecast load profiles of three weeks in the middle of each season. As an example, forecasts in summer are described in Fig. 5. An example of daily load profile is zoomed in the box. ARIMA with daily seasonality shows the worst forecasts; load shapes are similar to real loads, but ARIMA could not reflect the recent trend. The DSHW model considers both daily and weekly seasonality. It shows better accuracy than the ARIMA
5 Season Date TABLE II SEASONAL ERROR TABLE (%) OF THE SELECTED CUSTOMER. MAPE (%) RRMSE (%) RBM ReLU SNN DSHW ARIMA RBM ReLU SNN DSHW ARIMA Spring Summer Fall Winter Average than ARIMA, i.e., 69%. Furthermore, DNN models forecast well without being affected by seasonal variation. MWh Date index Time (hour) Fig. 4. An example of customer s load pattern. Daily load profiles can be recognized with load and time axes while seasonal load variations are expressed with load and date index axes. Load (kw) Time (hour) 2 Real DNN(RBM) DNN(ReLU) SNN DSHW ARIMA Fig. 5. Seasonal forecasting results of selected customer in summer. This figure includes hourly load forecasts of 15 days described in Table II. model. However, DSHW sometimes shows unusual patterns depicted in the box of Fig. 5 which seem to be the imprint of past abnormal consumptions. In contrast, the forecasts by two proposed DNN models are more stable than the ARIMA and DSHW models. When there exists a gap between real and DNN forecasts, other models exhibit wider gaps. However it is hard to find opposite cases. In all four seasons, forecasts from DNNs follow the recent trend and real load patterns well. The average errors of all seasons are shown in Table II. As can be seen, two DNN forecasting models show better accuracy than the SNN, DSHW, and ARIMA models in overall. DNN with pre-training shows the best accuracy in average by 3.2% and 4.1% in terms of MAPE and RRMSE. MAPE decreases by 27% than SNN, and 12% than DSHW. Due to the average gap of forecasts by ARIMA, MAPE of DNN is much lower B. Case 2: various load types To identify generalized performance of the proposed DNNs, we apply the DNNs to a large set of load database. We select eight industrial categories, and five customers are randomly selected in each category. We forecast daily load profiles of 1 weeks from Sep. 13 to Dec. 19 of 214. Considering the inaccurate result of Case 1 and burden of model selection, we exclude ARIMA in Case 2. As an example, error results of customers in the public administration category and average errors of total eight industries are shown in Table III. Due to space limitation, we can show only one industrial category in detail and present the average of eight industries with 4 customers in the bottom line. MAPE over 1% are not counted in the result. As described in the table, both DNN models show accurate forecasting result even though SNN or DSHW sometimes exhibits better result than the DNNs. In overall the average MAPE of DNN with ReLU decreases by 17% and 9% compared to the SNN and DSHW, respectively. Decrease in the RRMSE is more drastic; 22% & 29% lower than that of the SNN and DSHW. For further analysis, we compare the daily errors of customers. Fig. 6 shows the empirical cumulative distribution function (CDF) of daily RRMSEs. In terms of RRMSE CDF, two DNN models are always better than the SNN model. Since DNNs has more computational power, this result may not be surprising. Compared with DSHW, two DNN models and DSHW exhibits similar performance in small error region. Roughly when errors are under 3%, performance of DSHW is better than DNNs, but after that, DNNs begin to be better than DSHW; DNNs forecast more reliably even if loads are uncertain with high forecasting errors. These results imply the robustness of DNN models. Between two DNN models, there is no significant difference in accuracy. Since the RBM pre-training needs more computational power and time, the DNN with ReLU could be a better choice for demand side forecasting. VI. CONCLUSION In this paper, we proposed DNN-based demand side STLF framework. The proposed framework contains data processing, DNN training and DNN forecasting step, and forecast daily load profile day ahead based on weather, date and past electricity consumptions. We found that DNNs are effective to learn characteristics of load patterns and show better performance compared to other forecasting models. Our extensive
6 Industry TABLE III ERROR TABLE OF PUBLIC ADMINISTRATION CATEGORY AND AVERAGE ERRORS OF 8 INDUSTRIES (4 CUSTOMERS) Customer Peak MAPE (%) RRMSE (%) ID Load (kw) DNN(RBM) DNN(ReLU) SNN DSHW DNN(RBM) DNN(ReLU) SNN DSHW Public administration Average Total 4 customers Average CDF DNN(RBM) DNN(ReLU) SNN DSHW RRMSE (%) Fig. 6. Empirical CDF of daily RRMSE. experiments show that two proposed approaches, DNN with pre-training and DNN with ReLU, can be trained well with up to three years of customer load data without overfitting. The performance of forecasting model is evaluated based on the MAPE and RRMSE. Due to the space limitation, we cannot provide all the test results performed for 4 customers in eight industrial categories but we verify that in overall DNNs perform better than existing methods, e.g., MAPE and RRMSE are reduced by up to 17% and 22% than SNN, and 9% and 29% by DSHW. The results of this paper can be further extended in several directions. We are currently gathering nation-wide data and thus expect to train the DNN better when big data is available. Then, it would be possible to explore a better DNN structure, e.g., consider deep recurrent neural network as a time series forecasting model. In addition, analyzing the conditions (e.g., load type, load pattern, date, etc) when DNN is accurate, would be helpful to select a proper forecasting model or to use different data processing scheme for individual customers. ACKNOWLEDGMENT This research was partially supported by Korea Electric Power Corporation and Korea Electrical Engineering & Science Research Institute (grant number : R14XA2-5) and by Korea Institute of Energy Technology Evaluation and Planning (grand number : ). REFERENCES [1] M. T. Hagan and S. M. Behr, The time series approach to short term load forecasting, Power Systems, IEEE Transactions on, vol. 2, no. 3, pp , [2] J. W. Taylor, Short-term electricity demand forecasting using double seasonal exponential smoothing, Journal of the Operational Research Society, vol. 54, no. 8, pp , 23. [3] D. C. Park, M. El-Sharkawi, R. Marks, L. Atlas, M. Damborg et al., Electric load forecasting using an artificial neural network, Power Systems, IEEE Transactions on, vol. 6, no. 2, pp , [4] L. Hernandez, C. Baladrón, J. M. Aguiar, B. Carro, A. J. Sanchez- Esguevillas, and J. Lloret, Short-term load forecasting for microgrids based on artificial neural networks, Energies, vol. 6, no. 3, pp , 213. [5] A. G. Bakirtzis, V. Petridis, S. Kiartzis, and M. C. Alexiadis, A neural network short term load forecasting model for the Greek power system, Power Systems, IEEE Transactions on, vol. 11, no. 2, pp , [6] H. S. Hippert, C. E. Pedreira, and R. C. Souza, Neural networks for short-term load forecasting: A review and evaluation, Power Systems, IEEE Transactions on, vol. 16, no. 1, pp , 21. [7] E. Mocanu, P. H. Nguyen, M. Gibescu, and W. L. Kling, Deep learning for estimating building energy consumption, Sustainable Energy, Grids and Networks, vol. 6, pp , 216. [8] E. Mocanu, P. H. Nguyen, W. L. Kling, and M. Gibescu, Unsupervised energy prediction in a Smart Grid context using reinforcement crossbuilding transfer learning, Energy and Buildings, vol. 116, pp , 216. [9] M. Dalto, J. Matusko, and M. Vasak, Deep neural networks for ultrashort-term wind forecasting, in Industrial Technology (ICIT), 215 IEEE International Conference on. IEEE, 215, pp [1] X. Qiu, L. Zhang, Y. Ren, P. N. Suganthan, and G. Amaratunga, Ensemble deep learning for regression and time series forecasting. in CIEL, 214, pp [11] W. He, Deep neural network based load forecast, Computer Modeling and New Technologies, vol. 18, no. 3, pp , 214. [12] G. E. Hinton and R. R. Salakhutdinov, Reducing the dimensionality of data with neural networks, Science, vol. 313, no. 5786, pp , 26. [13] M. Leshno, V. Y. Lin, A. Pinkus, and S. Schocken, Multilayer feedforward networks with a nonpolynomial activation function can approximate any function, Neural networks, vol. 6, no. 6, pp , [14] Y. S. Abu-Mostafa, M. Magdon-Ismail, and H.-T. Lin, Learning from data. AMLBook Berlin, Germany, 212. [15] D. Erhan, Y. Bengio, A. Courville, P.-A. Manzagol, P. Vincent, and S. Bengio, Why does unsupervised pre-training help deep learning? The Journal of Machine Learning Research, vol. 11, pp , 21. [16] G. Hinton, A practical guide to training restricted Boltzmann machines, Momentum, vol. 9, no. 1, p. 926, 21. [17] A. L. Maas, A. Y. Hannun, and A. Y. Ng, Rectifier nonlinearities improve neural network acoustic models, in Proc. ICML, vol. 3, 213, p. 1. [18] M. Drees, Implementierung und Analyse von tiefen Architekturen in R, Master s thesis, Fachhochschule Dortmund, 213. [19] H. Shayeghi, H. Shayanfar, and G. Azimi, Intelligent neural network based STLF, International Journal of Computer Systems Science and Engineering, vol. 4, no. 1, 29. [2] M. Riedmiller and H. Braun, A direct adaptive method for faster backpropagation learning: The RPROP algorithm, in Neural Networks, 1993., IEEE International Conference On. IEEE, 1993, pp [21] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, Dropout: A simple way to prevent neural networks from overfitting, The Journal of Machine Learning Research, vol. 15, no. 1, pp , 214.
Deep Learning Architecture for Univariate Time Series Forecasting
CS229,Technical Report, 2014 Deep Learning Architecture for Univariate Time Series Forecasting Dmitry Vengertsev 1 Abstract This paper studies the problem of applying machine learning with deep architecture
More informationNeed for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels
Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)
More informationDeep Neural Networks
Deep Neural Networks DT2118 Speech and Speaker Recognition Giampiero Salvi KTH/CSC/TMH giampi@kth.se VT 2015 1 / 45 Outline State-to-Output Probability Model Artificial Neural Networks Perceptron Multi
More informationElectric Load Forecasting Using Wavelet Transform and Extreme Learning Machine
Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Song Li 1, Peng Wang 1 and Lalit Goel 1 1 School of Electrical and Electronic Engineering Nanyang Technological University
More informationNeed for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels
Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)
More informationDeep unsupervised learning
Deep unsupervised learning Advanced data-mining Yongdai Kim Department of Statistics, Seoul National University, South Korea Unsupervised learning In machine learning, there are 3 kinds of learning paradigm.
More informationSpeaker Representation and Verification Part II. by Vasileios Vasilakakis
Speaker Representation and Verification Part II by Vasileios Vasilakakis Outline -Approaches of Neural Networks in Speaker/Speech Recognition -Feed-Forward Neural Networks -Training with Back-propagation
More informationClassification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton Motivation Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses
More informationJakub Hajic Artificial Intelligence Seminar I
Jakub Hajic Artificial Intelligence Seminar I. 11. 11. 2014 Outline Key concepts Deep Belief Networks Convolutional Neural Networks A couple of questions Convolution Perceptron Feedforward Neural Network
More informationUNSUPERVISED LEARNING
UNSUPERVISED LEARNING Topics Layer-wise (unsupervised) pre-training Restricted Boltzmann Machines Auto-encoders LAYER-WISE (UNSUPERVISED) PRE-TRAINING Breakthrough in 2006 Layer-wise (unsupervised) pre-training
More informationWEATHER DEPENENT ELECTRICITY MARKET FORECASTING WITH NEURAL NETWORKS, WAVELET AND DATA MINING TECHNIQUES. Z.Y. Dong X. Li Z. Xu K. L.
WEATHER DEPENENT ELECTRICITY MARKET FORECASTING WITH NEURAL NETWORKS, WAVELET AND DATA MINING TECHNIQUES Abstract Z.Y. Dong X. Li Z. Xu K. L. Teo School of Information Technology and Electrical Engineering
More informationIntroduction to Convolutional Neural Networks (CNNs)
Introduction to Convolutional Neural Networks (CNNs) nojunk@snu.ac.kr http://mipal.snu.ac.kr Department of Transdisciplinary Studies Seoul National University, Korea Jan. 2016 Many slides are from Fei-Fei
More informationBased on the original slides of Hung-yi Lee
Based on the original slides of Hung-yi Lee Google Trends Deep learning obtains many exciting results. Can contribute to new Smart Services in the Context of the Internet of Things (IoT). IoT Services
More informationIntroduction to Neural Networks
CUONG TUAN NGUYEN SEIJI HOTTA MASAKI NAKAGAWA Tokyo University of Agriculture and Technology Copyright by Nguyen, Hotta and Nakagawa 1 Pattern classification Which category of an input? Example: Character
More informationShort Term Load Forecasting Based Artificial Neural Network
Short Term Load Forecasting Based Artificial Neural Network Dr. Adel M. Dakhil Department of Electrical Engineering Misan University Iraq- Misan Dr.adelmanaa@gmail.com Abstract Present study develops short
More informationARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD
ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided
More informationNeural Networks. Nicholas Ruozzi University of Texas at Dallas
Neural Networks Nicholas Ruozzi University of Texas at Dallas Handwritten Digit Recognition Given a collection of handwritten digits and their corresponding labels, we d like to be able to correctly classify
More informationDeep Recurrent Neural Networks
Deep Recurrent Neural Networks Artem Chernodub e-mail: a.chernodub@gmail.com web: http://zzphoto.me ZZ Photo IMMSP NASU 2 / 28 Neuroscience Biological-inspired models Machine Learning p x y = p y x p(x)/p(y)
More informationMODELLING ENERGY DEMAND FORECASTING USING NEURAL NETWORKS WITH UNIVARIATE TIME SERIES
MODELLING ENERGY DEMAND FORECASTING USING NEURAL NETWORKS WITH UNIVARIATE TIME SERIES S. Cankurt 1, M. Yasin 2 1&2 Ishik University Erbil, Iraq 1 s.cankurt@ishik.edu.iq, 2 m.yasin@ishik.edu.iq doi:10.23918/iec2018.26
More informationIntroduction to Deep Neural Networks
Introduction to Deep Neural Networks Presenter: Chunyuan Li Pattern Classification and Recognition (ECE 681.01) Duke University April, 2016 Outline 1 Background and Preliminaries Why DNNs? Model: Logistic
More informationA Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems
A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems Hiroyuki Mori Dept. of Electrical & Electronics Engineering Meiji University Tama-ku, Kawasaki
More informationA Hybrid Deep Learning Approach For Chaotic Time Series Prediction Based On Unsupervised Feature Learning
A Hybrid Deep Learning Approach For Chaotic Time Series Prediction Based On Unsupervised Feature Learning Norbert Ayine Agana Advisor: Abdollah Homaifar Autonomous Control & Information Technology Institute
More informationANN and Statistical Theory Based Forecasting and Analysis of Power System Variables
ANN and Statistical Theory Based Forecasting and Analysis of Power System Variables Sruthi V. Nair 1, Poonam Kothari 2, Kushal Lodha 3 1,2,3 Lecturer, G. H. Raisoni Institute of Engineering & Technology,
More informationCSCI567 Machine Learning (Fall 2018)
CSCI567 Machine Learning (Fall 2018) Prof. Haipeng Luo U of Southern California Sep 12, 2018 September 12, 2018 1 / 49 Administration GitHub repos are setup (ask TA Chi Zhang for any issues) HW 1 is due
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationLearning Deep Architectures for AI. Part II - Vijay Chakilam
Learning Deep Architectures for AI - Yoshua Bengio Part II - Vijay Chakilam Limitations of Perceptron x1 W, b 0,1 1,1 y x2 weight plane output =1 output =0 There is no value for W and b such that the model
More informationCSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer
CSE446: Neural Networks Spring 2017 Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer Human Neurons Switching time ~ 0.001 second Number of neurons 10 10 Connections per neuron 10 4-5 Scene
More informationFeature Design. Feature Design. Feature Design. & Deep Learning
Artificial Intelligence and its applications Lecture 9 & Deep Learning Professor Daniel Yeung danyeung@ieee.org Dr. Patrick Chan patrickchan@ieee.org South China University of Technology, China Appropriately
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationLarge-Scale Feature Learning with Spike-and-Slab Sparse Coding
Large-Scale Feature Learning with Spike-and-Slab Sparse Coding Ian J. Goodfellow, Aaron Courville, Yoshua Bengio ICML 2012 Presented by Xin Yuan January 17, 2013 1 Outline Contributions Spike-and-Slab
More informationApplication of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption
Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/59 Lecture 4a Feedforward neural network October 30, 2015 2/59 Table of contents 1 1. Objectives of Lecture
More informationShort-term water demand forecast based on deep neural network ABSTRACT
Short-term water demand forecast based on deep neural network Guancheng Guo 1, Shuming Liu 2 1,2 School of Environment, Tsinghua University, 100084, Beijing, China 2 shumingliu@tsinghua.edu.cn ABSTRACT
More informationNonlinear system modeling with deep neural networks and autoencoders algorithm
Nonlinear system modeling with deep neural networks and autoencoders algorithm Erick De la Rosa, Wen Yu Departamento de Control Automatico CINVESTAV-IPN Mexico City, Mexico yuw@ctrl.cinvestav.mx Xiaoou
More informationComputational Intelligence Winter Term 2017/18
Computational Intelligence Winter Term 207/8 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS ) Fakultät für Informatik TU Dortmund Plan for Today Single-Layer Perceptron Accelerated Learning
More informationGreedy Layer-Wise Training of Deep Networks
Greedy Layer-Wise Training of Deep Networks Yoshua Bengio, Pascal Lamblin, Dan Popovici, Hugo Larochelle NIPS 2007 Presented by Ahmed Hefny Story so far Deep neural nets are more expressive: Can learn
More informationSynaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics
Synaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics Byung-Gook Park Inter-university Semiconductor Research Center & Department of Electrical and Computer Engineering Seoul National
More informationNeural Networks and Deep Learning
Neural Networks and Deep Learning Professor Ameet Talwalkar November 12, 2015 Professor Ameet Talwalkar Neural Networks and Deep Learning November 12, 2015 1 / 16 Outline 1 Review of last lecture AdaBoost
More informationDEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY
DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY 1 On-line Resources http://neuralnetworksanddeeplearning.com/index.html Online book by Michael Nielsen http://matlabtricks.com/post-5/3x3-convolution-kernelswith-online-demo
More informationReading Group on Deep Learning Session 4 Unsupervised Neural Networks
Reading Group on Deep Learning Session 4 Unsupervised Neural Networks Jakob Verbeek & Daan Wynen 206-09-22 Jakob Verbeek & Daan Wynen Unsupervised Neural Networks Outline Autoencoders Restricted) Boltzmann
More informationShort Term Load Forecasting Using Multi Layer Perceptron
International OPEN ACCESS Journal Of Modern Engineering Research (IJMER) Short Term Load Forecasting Using Multi Layer Perceptron S.Hema Chandra 1, B.Tejaswini 2, B.suneetha 3, N.chandi Priya 4, P.Prathima
More informationDeep Learning Architectures and Algorithms
Deep Learning Architectures and Algorithms In-Jung Kim 2016. 12. 2. Agenda Introduction to Deep Learning RBM and Auto-Encoders Convolutional Neural Networks Recurrent Neural Networks Reinforcement Learning
More informationDeep Generative Models. (Unsupervised Learning)
Deep Generative Models (Unsupervised Learning) CEng 783 Deep Learning Fall 2017 Emre Akbaş Reminders Next week: project progress demos in class Describe your problem/goal What you have done so far What
More informationComputational Intelligence
Plan for Today Single-Layer Perceptron Computational Intelligence Winter Term 00/ Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS ) Fakultät für Informatik TU Dortmund Accelerated Learning
More informationCourse 395: Machine Learning - Lectures
Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture
More informationThis paper presents the
ISESCO JOURNAL of Science and Technology Volume 8 - Number 14 - November 2012 (2-8) A Novel Ensemble Neural Network based Short-term Wind Power Generation Forecasting in a Microgrid Aymen Chaouachi and
More informationShort-term wind forecasting using artificial neural networks (ANNs)
Energy and Sustainability II 197 Short-term wind forecasting using artificial neural networks (ANNs) M. G. De Giorgi, A. Ficarella & M. G. Russo Department of Engineering Innovation, Centro Ricerche Energia
More informationSTA 414/2104: Lecture 8
STA 414/2104: Lecture 8 6-7 March 2017: Continuous Latent Variable Models, Neural networks With thanks to Russ Salakhutdinov, Jimmy Ba and others Outline Continuous latent variable models Background PCA
More informationMachine Learning
Machine Learning 10-315 Maria Florina Balcan Machine Learning Department Carnegie Mellon University 03/29/2019 Today: Artificial neural networks Backpropagation Reading: Mitchell: Chapter 4 Bishop: Chapter
More informationCS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes
CS 6501: Deep Learning for Computer Graphics Basics of Neural Networks Connelly Barnes Overview Simple neural networks Perceptron Feedforward neural networks Multilayer perceptron and properties Autoencoders
More informationA Hybrid Model of Wavelet and Neural Network for Short Term Load Forecasting
International Journal of Electronic and Electrical Engineering. ISSN 0974-2174, Volume 7, Number 4 (2014), pp. 387-394 International Research Publication House http://www.irphouse.com A Hybrid Model of
More informationLecture 17: Neural Networks and Deep Learning
UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions
More informationEnhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines
Enhanced Gradient and Adaptive Learning Rate for Training Restricted Boltzmann Machines KyungHyun Cho KYUNGHYUN.CHO@AALTO.FI Tapani Raiko TAPANI.RAIKO@AALTO.FI Alexander Ilin ALEXANDER.ILIN@AALTO.FI Department
More informationImplementation of Artificial Neural Network for Short Term Load Forecasting
Implementation of Artificial Neural Network for Short Term Load Forecasting Anju Bala * * Research Scholar, E.E. Department, D.C.R. University of Sci. & Technology, Murthal, anju8305@gmail.com N. K. Yadav
More informationNeural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /9/17
3/9/7 Neural Networks Emily Fox University of Washington March 0, 207 Slides adapted from Ali Farhadi (via Carlos Guestrin and Luke Zettlemoyer) Single-layer neural network 3/9/7 Perceptron as a neural
More informationUnit III. A Survey of Neural Network Model
Unit III A Survey of Neural Network Model 1 Single Layer Perceptron Perceptron the first adaptive network architecture was invented by Frank Rosenblatt in 1957. It can be used for the classification of
More informationHYBRID ARTIFICIAL NEURAL NETWORK SYSTEM FOR SHORT-TERM LOAD FORECASTING
Ilić, S. A. et al.: Hybrid Artificial Neural Network System for Short-Term... S215 HYBRID ARTIFICIAL NEURAL NETWORK SYSTEM FOR SHORT-TERM LOAD FORECASTING by Slobodan A. ILIĆ*, Srdjan M. VUKMIROVIĆ, Aleksandar
More informationA Novel 2-D Model Approach for the Prediction of Hourly Solar Radiation
A Novel 2-D Model Approach for the Prediction of Hourly Solar Radiation F Onur Hocao glu, Ö Nezih Gerek, and Mehmet Kurban Anadolu University, Dept of Electrical and Electronics Eng, Eskisehir, Turkey
More informationKnowledge Extraction from DBNs for Images
Knowledge Extraction from DBNs for Images Son N. Tran and Artur d Avila Garcez Department of Computer Science City University London Contents 1 Introduction 2 Knowledge Extraction from DBNs 3 Experimental
More informationA new method for short-term load forecasting based on chaotic time series and neural network
A new method for short-term load forecasting based on chaotic time series and neural network Sajjad Kouhi*, Navid Taghizadegan Electrical Engineering Department, Azarbaijan Shahid Madani University, Tabriz,
More informationIntroduction to Machine Learning
Introduction to Machine Learning Neural Networks Varun Chandola x x 5 Input Outline Contents February 2, 207 Extending Perceptrons 2 Multi Layered Perceptrons 2 2. Generalizing to Multiple Labels.................
More informationHow to do backpropagation in a brain
How to do backpropagation in a brain Geoffrey Hinton Canadian Institute for Advanced Research & University of Toronto & Google Inc. Prelude I will start with three slides explaining a popular type of deep
More informationMachine Learning. Neural Networks
Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE
More informationMachine Learning Approach to Forecast Global Solar Radiation Time Series
University of New Mexico UNM Digital Repository Electrical and Computer Engineering ETDs Engineering ETDs 8-25-2016 Machine Learning Approach to Forecast Global Solar Radiation Time Series Guillermo Terren-Serrano
More informationDeep Learning: a gentle introduction
Deep Learning: a gentle introduction Jamal Atif jamal.atif@dauphine.fr PSL, Université Paris-Dauphine, LAMSADE February 8, 206 Jamal Atif (Université Paris-Dauphine) Deep Learning February 8, 206 / Why
More informationHybrid PSO-ANN Application for Improved Accuracy of Short Term Load Forecasting
Hybrid PSO-ANN Application for Improved Accuracy of Short Term Load Forecasting A. G. ABDULLAH, G. M. SURANEGARA, D.L. HAKIM Electrical Engineering Education Department Indonesia University of Education
More informationDeep Feedforward Networks
Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3
More informationData and prognosis for renewable energy
The Hong Kong Polytechnic University Department of Electrical Engineering Project code: FYP_27 Data and prognosis for renewable energy by Choi Man Hin 14072258D Final Report Bachelor of Engineering (Honours)
More informationApprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning
Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Nicolas Thome Prenom.Nom@cnam.fr http://cedric.cnam.fr/vertigo/cours/ml2/ Département Informatique Conservatoire
More informationCSC 578 Neural Networks and Deep Learning
CSC 578 Neural Networks and Deep Learning Fall 2018/19 3. Improving Neural Networks (Some figures adapted from NNDL book) 1 Various Approaches to Improve Neural Networks 1. Cost functions Quadratic Cross
More informationESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS 1. INTRODUCTION
Mathematical and Computational Applications, Vol. 11, No. 3, pp. 215-224, 2006. Association for Scientific Research ESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS Ömer Altan
More informationExplanatory Information Analysis for Day-Ahead Price Forecasting in the Iberian Electricity Market
Energies 2015, 8, 10464-10486; doi:10.3390/en80910464 Article OPEN ACCESS energies ISSN 1996-1073 www.mdpi.com/journal/energies Explanatory Information Analysis for Day-Ahead Price Forecasting in the Iberian
More informationMeasuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information
Measuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information Mathias Berglund, Tapani Raiko, and KyungHyun Cho Department of Information and Computer Science Aalto University
More informationNotes on Back Propagation in 4 Lines
Notes on Back Propagation in 4 Lines Lili Mou moull12@sei.pku.edu.cn March, 2015 Congratulations! You are reading the clearest explanation of forward and backward propagation I have ever seen. In this
More informationAbout Nnergix +2, More than 2,5 GW forecasted. Forecasting in 5 countries. 4 predictive technologies. More than power facilities
About Nnergix +2,5 5 4 +20.000 More than 2,5 GW forecasted Forecasting in 5 countries 4 predictive technologies More than 20.000 power facilities Nnergix s Timeline 2012 First Solar Photovoltaic energy
More informationData Mining Part 5. Prediction
Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,
More informationModelling and Prediction of 150KW PV Array System in Northern India using Artificial Neural Network
International Journal of Engineering Science Invention ISSN (Online): 2319 6734, ISSN (Print): 2319 6726 Volume 5 Issue 5 May 2016 PP.18-25 Modelling and Prediction of 150KW PV Array System in Northern
More information10. Artificial Neural Networks
Foundations of Machine Learning CentraleSupélec Fall 217 1. Artificial Neural Networks Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe-agathe.azencott@mines-paristech.fr Learning
More informationFrom perceptrons to word embeddings. Simon Šuster University of Groningen
From perceptrons to word embeddings Simon Šuster University of Groningen Outline A basic computational unit Weighting some input to produce an output: classification Perceptron Classify tweets Written
More informationArtificial Neural Networks. Historical description
Artificial Neural Networks Historical description Victor G. Lopez 1 / 23 Artificial Neural Networks (ANN) An artificial neural network is a computational model that attempts to emulate the functions of
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward
More informationA summary of Deep Learning without Poor Local Minima
A summary of Deep Learning without Poor Local Minima by Kenji Kawaguchi MIT oral presentation at NIPS 2016 Learning Supervised (or Predictive) learning Learn a mapping from inputs x to outputs y, given
More informationFrom CDF to PDF A Density Estimation Method for High Dimensional Data
From CDF to PDF A Density Estimation Method for High Dimensional Data Shengdong Zhang Simon Fraser University sza75@sfu.ca arxiv:1804.05316v1 [stat.ml] 15 Apr 2018 April 17, 2018 1 Introduction Probability
More informationDeep Belief Networks are compact universal approximators
1 Deep Belief Networks are compact universal approximators Nicolas Le Roux 1, Yoshua Bengio 2 1 Microsoft Research Cambridge 2 University of Montreal Keywords: Deep Belief Networks, Universal Approximation
More informationDeep Neural Networks As Time Series Forecasters of Energy Demand
Marquette University e-publications@marquette Master's Theses (2009 -) Dissertations, Theses, and Professional Projects Deep Neural Networks As Time Series Forecasters of Energy Demand Gregory Merkel Marquette
More informationCourse Structure. Psychology 452 Week 12: Deep Learning. Chapter 8 Discussion. Part I: Deep Learning: What and Why? Rufus. Rufus Processed By Fetch
Psychology 452 Week 12: Deep Learning What Is Deep Learning? Preliminary Ideas (that we already know!) The Restricted Boltzmann Machine (RBM) Many Layers of RBMs Pros and Cons of Deep Learning Course Structure
More informationTopic 3: Neural Networks
CS 4850/6850: Introduction to Machine Learning Fall 2018 Topic 3: Neural Networks Instructor: Daniel L. Pimentel-Alarcón c Copyright 2018 3.1 Introduction Neural networks are arguably the main reason why
More informationUnivariate versus Multivariate Models for Short-term Electricity Load Forecasting
Univariate versus Multivariate Models for Short-term Electricity Load Forecasting Guilherme Guilhermino Neto 1, Samuel Belini Defilippo 2, Henrique S. Hippert 3 1 IFES Campus Linhares. guilherme.neto@ifes.edu.br
More informationDeep Learning Srihari. Deep Belief Nets. Sargur N. Srihari
Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous
More informationRecurrent Neural Network Based Gating for Natural Gas Load Prediction System
Recurrent Neural Network Based Gating for Natural Gas Load Prediction System Petr Musilek, Member, IEEE, Emil Pelikán, Tomáš Brabec and Milan Šimůnek Abstract Prediction of natural gas consumption is an
More informationSHORT TERM LOAD FORECASTING
Indian Institute of Technology Kanpur (IITK) and Indian Energy Exchange (IEX) are delighted to announce Training Program on "Power Procurement Strategy and Power Exchanges" 28-30 July, 2014 SHORT TERM
More informationDemand Forecasting in Deregulated Electricity Markets
International Journal of Computer Applications (975 8887) Demand Forecasting in Deregulated Electricity Marets Anamia Electrical & Electronics Engineering Department National Institute of Technology Jamshedpur
More information4. Multilayer Perceptrons
4. Multilayer Perceptrons This is a supervised error-correction learning algorithm. 1 4.1 Introduction A multilayer feedforward network consists of an input layer, one or more hidden layers, and an output
More informationRetrieval of Cloud Top Pressure
Master Thesis in Statistics and Data Mining Retrieval of Cloud Top Pressure Claudia Adok Division of Statistics and Machine Learning Department of Computer and Information Science Linköping University
More informationMulti-Plant Photovoltaic Energy Forecasting Challenge with Regression Tree Ensembles and Hourly Average Forecasts
Multi-Plant Photovoltaic Energy Forecasting Challenge with Regression Tree Ensembles and Hourly Average Forecasts Kathrin Bujna 1 and Martin Wistuba 2 1 Paderborn University 2 IBM Research Ireland Abstract.
More informationShort Term Power Load Forecasting Using Deep Neural Networks
Short Term Power Load Forecasting Using Deep Neural Networks Ghulam Mohi Ud Din Department of Computer Science Liverpool John Moores University United Kingdom g.mohiuddin@215.ljmu.ac.uk Angelos K. Marnerides
More informationNumerical Interpolation of Aircraft Aerodynamic Data through Regression Approach on MLP Network
, pp.12-17 http://dx.doi.org/10.14257/astl.2018.151.03 Numerical Interpolation of Aircraft Aerodynamic Data through Regression Approach on MLP Network Myeong-Jae Jo 1, In-Kyum Kim 1, Won-Hyuck Choi 2 and
More informationAdministration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6
Administration Registration Hw3 is out Due on Thursday 10/6 Questions Lecture Captioning (Extra-Credit) Look at Piazza for details Scribing lectures With pay; come talk to me/send email. 1 Projects Projects
More informationStochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence
ESANN 0 proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning. Bruges (Belgium), 7-9 April 0, idoc.com publ., ISBN 97-7707-. Stochastic Gradient
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) Human Brain Neurons Input-Output Transformation Input Spikes Output Spike Spike (= a brief pulse) (Excitatory Post-Synaptic Potential)
More information