Multiple Wavelet Coefficients Fusion in Deep Residual Networks for Fault Diagnosis

Similar documents
Bearing fault diagnosis based on EMD-KPCA and ELM

Bearing fault diagnosis based on Shannon entropy and wavelet package decomposition

Lecture 7 Convolutional Neural Networks

Introduction to Convolutional Neural Networks 2018 / 02 / 23

Sajid Anwar, Kyuyeon Hwang and Wonyong Sung

Deep Residual. Variations

WHY ARE DEEP NETS REVERSIBLE: A SIMPLE THEORY,

Convolutional Neural Network Architecture

COMPLEX INPUT CONVOLUTIONAL NEURAL NETWORKS FOR WIDE ANGLE SAR ATR

An overview of deep learning methods for genomics

Machine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Benchmarking Functional Link Expansions for Audio Classification Tasks

Nonlinear Models. Numerical Methods for Deep Learning. Lars Ruthotto. Departments of Mathematics and Computer Science, Emory University.

Single-Image-Based Rain and Snow Removal Using Multi-guided Filter

<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)

Two at Once: Enhancing Learning and Generalization Capacities via IBN-Net

Neural Networks 2. 2 Receptive fields and dealing with image inputs

Cheng Soon Ong & Christian Walder. Canberra February June 2018

The Fault extent recognition method of rolling bearing based on orthogonal matching pursuit and Lempel-Ziv complexity

Towards understanding feedback from supermassive black holes using convolutional neural networks

WHEELSET BEARING VIBRATION ANALYSIS BASED ON NONLINEAR DYNAMICAL METHOD

Agenda. Digit Classification using CNN Digit Classification using SAE Visualization: Class models, filters and saliency 2 DCT

CSE 591: Introduction to Deep Learning in Visual Computing. - Parag S. Chandakkar - Instructors: Dr. Baoxin Li and Ragav Venkatesan

Statistical Machine Learning

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Binary Convolutional Neural Network on RRAM

Convolutional Neural Networks II. Slides from Dr. Vlad Morariu

Lecture 8: Introduction to Deep Learning: Part 2 (More on backpropagation, and ConvNets)

Novelty Detection based on Extensions of GMMs for Industrial Gas Turbines

Very Deep Residual Networks with Maxout for Plant Identification in the Wild Milan Šulc, Dmytro Mishkin, Jiří Matas

Introduction to Deep Learning

Deep Learning (CNNs)

Introduction to Convolutional Neural Networks (CNNs)

Determination of Linear Force- Free Magnetic Field Constant αα Using Deep Learning

Invariant Scattering Convolution Networks

Bearing fault diagnosis based on TEO and SVM

Introduction to Machine Learning (67577)

A novel intelligent predictive maintenance procedure for electrical machines

DYNAMIC TEXTURE RECOGNITION USING ENHANCED LBP FEATURES

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

arxiv: v1 [cs.lg] 25 Sep 2018

arxiv: v2 [cs.sd] 7 Feb 2018

Normalization Techniques in Training of Deep Neural Networks

Jakub Hajic Artificial Intelligence Seminar I

Explaining Predictions of Non-Linear Classifiers in NLP

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

A practical theory for designing very deep convolutional neural networks. Xudong Cao.

Introduction to Deep Learning

1251. An approach for tool health assessment using the Mahalanobis-Taguchi system based on WPT-AR

Lecture 17: Neural Networks and Deep Learning

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

Action-Decision Networks for Visual Tracking with Deep Reinforcement Learning

Explaining Predictions of Non-Linear Classifiers in NLP

Convolution and Pooling as an Infinitely Strong Prior

CS 1674: Intro to Computer Vision. Final Review. Prof. Adriana Kovashka University of Pittsburgh December 7, 2016

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Continuous NN. Numerical Methods for Deep Learning. Lars Ruthotto. Departments of Mathematics and Computer Science, Emory University.

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

RESEARCH ON COMPLEX THREE ORDER CUMULANTS COUPLING FEATURES IN FAULT DIAGNOSIS

Neural Networks. David Rosenberg. July 26, New York University. David Rosenberg (New York University) DS-GA 1003 July 26, / 35

Google s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

Tasks ADAS. Self Driving. Non-machine Learning. Traditional MLP. Machine-Learning based method. Supervised CNN. Methods. Deep-Learning based

P-TELU : Parametric Tan Hyperbolic Linear Unit Activation for Deep Neural Networks

2256. Application of empirical mode decomposition and Euclidean distance technique for feature selection and fault diagnosis of planetary gearbox

SHAKE-SHAKE REGULARIZATION OF 3-BRANCH

Semi-Orthogonal Low-Rank Matrix Factorization for Deep Neural Networks

Topics in AI (CPSC 532L): Multimodal Learning with Vision, Language and Sound. Lecture 3: Introduction to Deep Learning (continued)

Convolutional Dictionary Learning and Feature Design

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Simon Trebst. Quantitative Modeling of Complex Systems

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

Deep Learning Autoencoder Models

arxiv: v2 [cs.cv] 12 Apr 2016

Very Deep Convolutional Neural Networks for LVCSR

ECE521 W17 Tutorial 1. Renjie Liao & Min Bai

Myoelectrical signal classification based on S transform and two-directional 2DPCA

CSC411: Final Review. James Lucas & David Madras. December 3, 2018

2262. Remaining life prediction of rolling bearing based on PCA and improved logistic regression model

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Machine Learning Lecture 14

Probabilistic Class-Specific Discriminant Analysis

Orientation Map Based Palmprint Recognition

Introduction to Neural Networks

Based on the original slides of Hung-yi Lee

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning

Machine Learning Techniques

Misalignment Fault Detection in Dual-rotor System Based on Time Frequency Techniques

TENSOR LAYERS FOR COMPRESSION OF DEEP LEARNING NETWORKS. Cris Cecka Senior Research Scientist, NVIDIA GTC 2018

arxiv: v1 [astro-ph.im] 20 Jan 2017

Introduction to Deep Neural Networks

Introduction to (Convolutional) Neural Networks

Deep Feedforward Networks. Han Shao, Hou Pong Chan, and Hongyi Zhang

Open Access Recognition of Pole Piece Defects of Lithium Battery Based on Sparse Decomposition

Dense Fusion Classmate Network for Land Cover Classification

Segmentation of Cell Membrane and Nucleus using Branches with Different Roles in Deep Neural Network

Method for Recognizing Mechanical Status of Container Crane Motor Based on SOM Neural Network

Quantum Convolutional Neural Networks

Transcription:

Multiple Wavelet Coefficients Fusion in Deep Residual Networks for Fault Diagnosis Minghang Zhao, Myeongsu Kang, Baoping Tang, Michael Pecht 1

Backgrounds Accurate fault diagnosis is important to ensure the safety of automobiles and helicopters, long-term generation of electric power, and reliable operating of other electrical and mechanical systems. Discrete wavelet packet transform (DWPT), an effective tool to decompose non-stationary vibration signals into various frequency bands, has been widely applied for machine fault diagnosis [1]. Besides, the usage of deep learning methods is becoming more and more popular to automatically learn discriminative features from vibration signals for improving diagnostic accuracies [2]. 2

Motivations However, there is still no consensus as to which wavelet (e.g., DB1, DB2, and DB3) can achieve an optimal performance in fault diagnosis. Besides, different wavelets may be optimal for recognizing different kinds of faults under different working conditions. It is very unlikely for one certain wavelet to be the most effective in recognizing all kinds of faults (such as bearing inner raceway faults, outer raceway faults, and rolling element faults). Therefore, the fusion of multiple wavelets into deep neural networks has an potential to improve the accuracy of a fault diagnostic task which involves the recognition of various fault types. 3

Input Data Configuration The wavelet coefficients at various frequency bands obtained using a certain wavelet can be stacked to be a 2D matrix; then, the 2D matrices derived from multiple wavelets can be formed to be a 3D matrix. Wavelet coefficients at the 1 st decomposition level 2D matrices of wavelet coefficients at the ii th decomposition level Decomposition using various wavelets Signal 2 nd wavelet 1 st wavelet WW 1,0 WW 1,1 WW 1,0 WW 1,1 NN ww th wavelet WW 1,0 WW 1,1 WW ii,0 WW ii,1 WW ii,2 WW ii,0 WW ii,1 WW ii,2 WW ii,3 WW ii,3 WW ii,2 ii 2 WW ii,2 ii 2 WW ii,2 ii 1 WW ii,0 WW ii,1 WW ii,2 WW ii,3 WW ii,2 ii 2 WW ii,2 ii 1 Frequency band Time WW ii,2 ii 1 4

An Overview of Deep Residual Networks The deep residual network (DRN) is an improved variant of convolutional neural networks (CNNs), which uses identity shortcuts to ease the difficulty of training [3]-[4]. Input BN ReLU Conv 3 3 BN ReLU Conv 3 3 Conv 3 3 BN, ReLU, Conv 3 3 BN, ReLU, Conv 3 3 BN, ReLU, Conv 3 3 BN, ReLU, Conv 3 3 A number of RBUs A residual building unit (RBU) BN, ReLU, GAP Fully connected output layer A deep residual network BN: Batch normalization ReLU: Rectifier linear unit Conv 3 3: Convolution with kernels in the size of 3 3 GAP: Global average pooling 5

The First Developed Method To achieve multiple wavelet coefficients fusion, a simple method is to concatenate these 2D matrices of wavelet coefficients and feed them into a DRN. The method was named as Multiple Wavelet Coefficients Fusion in a Deep Residual Network by Concatenation (MWCF-DRN-C). 2D matrix 1 A vibration signal + DWPTs 2D matrix 2 2D matrix 3 2D matrix N A concatenation layer BN, ReLU, BN, ReLU, Conv, m BN, ReLU, GAP (Dropout) Fully connected output layer m: an indicator of the number of convolutional kernels 6

The Second Developed Method An individual convolutional layer with trainable parameters is applied to each 2D matrix of wavelet coefficients with the goal of converting the important wavelet coefficients to be large features. Then, the element-wise maximum features are chosen to be the output in the maximization layer [5]. The method was named as Multiple Wavelet Coefficients Fusion in a Deep Residual Network by Maximization (MWCF-DRN-M). 2D matrix 1 A vibration signal + DWPTs 2D matrix 2 2D matrix 3 A maximization layer BN, ReLU, BN, ReLU, Conv, m BN, ReLU, GAP (Dropout) Fully connected output layer 2D matrix N 7

Explanations on the Second Developed Method The 2D matrices of wavelet coefficients are different representations of the same vibration signal. It is unavoidable that these 2D matrices of wavelet coefficients contain much redundant/repetitive information. Much redundancy 2D matrix 1 A vibration signal + DWPTs 2D matrix 2 2D matrix 3 A maximization layer BN, ReLU, BN, ReLU, Conv, m BN, ReLU, GAP (Dropout) Fully connected output layer 2D matrix N 8

Explanations on the Second Developed Method The maximization layer and the convolutional layers before it can be interpreted as a trainable feature selection process, which allows the important features to be passed to the subsequent layers while the relatively unimportant features being abandoned. Much redundancy Trainable feature selection 2D matrix 1 A vibration signal + DWPTs 2D matrix 2 2D matrix 3 A maximization layer BN, ReLU, BN, ReLU, Conv, m BN, ReLU, GAP (Dropout) Fully connected output layer 2D matrix N 9

Experimental Setup A drivetrain dynamics simulator [6] was used to simulate the faults. Experiments were conducted under the 10-fold cross-validation scheme. Comparisons were made with the conventional CNN and DRN to demonstrate the efficacy of the developed MWCF-DRN-C and MWCF-DRN-M. 10

Results 11

Conclusions The fusion of multiple wavelet coefficients in deep neural networks can be able to improve the fault diagnostic performance. In the experimental result, the MWCF-DRN-M method was slightly better than the MWCF-DRN-C method by yielding a 0.80% improvement in terms of overall average testing accuracy. 12

References 1. R. Yan, R. X. Gao, and X. Chen, Wavelets for fault diagnosis of rotary machines: A review with applications, Signal Process., vol. 96, pp. 1 15, 2014. 2. M. Zhao, M. Kang, B. Tang, and M. Pecht, Deep Residual Networks With Dynamically Weighted Wavelet Coefficients for Fault Diagnosis of Planetary Gearboxes, IEEE Transactions on Industrial Electronics, vol. 65, no. 5, pp. 4290 4300, 2018. 3. K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in Proc. IEEE Conf. Comput. Vision Pattern Recognit., Seattle, WA, USA, Jun. 27 30, 2016, pp. 770 778. 4. K. He, X. Zhang, S. Ren, and J. Sun, Identity mappings in deep residual networks, in Computer Vision ECCV 2016 (Lecture Notes in Computer Science 9908), B. Leibe, J. Matas, N. Sebe, and M. Welling, Eds., Cham, Switzerland: Springer, 2016, pp. 630 645. 5. Z. Liao and C. Gustavo, A deep convolutional neural network module that promotes competition of multiple-size filters, Pattern Recognit., vol. 71, pp. 94 105, 2017. 6. Drivetrain Diagnostics Simulator. SpectraQuest, Richmond, VA, USA, [Online]. Available: http://spectraquest.com/drivetrains/details/dds/ 13