Group Equivariant Convolutional Networks. Taco Cohen
|
|
- Elfreda Hall
- 5 years ago
- Views:
Transcription
1 Group Equivariant Convolutional Networks Taco Cohen
2 Outline 1. Symmetry & Deep Learning: Statistical Power from Symmetry Invariance vs Equivariance Equivariance in Deep Learning 2.Group Theory Symmetry, Groups Subgroups, Cosets, Quotients Wallpaper groups 3.Group Equivariant Networks CNNs and translation equivariance G-Convolutions Equivariance of non-linearities Equivariance of G-pooling operator Backpropagation 4. Algorithms Spatial & Spectral G-convs 5. Results
3 Background: Invariance, Equivariance & Symmetry
4 Symmetry in ML A symmetry of a function is a transformation that leaves that function invariant In ML: look for symmetries of densities, factors, label functions, Shariff, R. (2015). Exploiting Symmetries To Construct Efficient Mcmc Algorithms.
5 Invariance Face Eyes Nose Mouth The Picasso Problem
6 Equivariance T 2 g Z 1 Z 2 T 1 g (T 1 g x) =T 2 g (x) Hinton, G., Krizhevsky, A., & Wang, S. (2011). Transforming auto-encoders. ICANN-11 Lenc, K., & Vedaldi, A. (2015). Understanding image representations by measuring their equivariance and equivalence (CVPR) Cohen, T., & Welling, M. (2014). Learning the Irreducible Representations of Commutative Lie Groups. (ICML)
7 Symmetry in DL Why do CNNs work so well? They exploit translational symmetry In deep nets, each layer should preserve the symmetry The representation should be an equivariant map for the symmetry group. Szegedy et al. (2015) Deeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeep!
8 ConvNets are Translation Equivariant Source:
9 Are ConvNets Rotation-Equivariant? Source:
10 CNNs want to be Equivariant
11 Visual Group Theory With figures from Visual Group Theory by Nathan Carter (2009)
12 Symmetries
13 Groups A group is: a set, G, together with a binary operation that combines two elements a, b in G and produces another element ab, that satisfies the group axioms: 1. Identity: there exists an element e in G, such that for every element a in G, ea = ae = a 2. Associativity: For all a, b and c in G, (a b) c = a (b c). 3. Closure: for all a, b in G, the composition ab is also in G 4. Inverse: for each a in G, there exists an element b in G such that ab = ba = e
14 The Symmetries of an Object form a Group hh -1 =e 1 h h hv=r 1 h 1 v
15 Examples Lie Groups Discrete Groups Topological Groups Galois Groups Wallpaper Groups Space Groups
16 Cayley Graphs
17 Subgroups A subgroup of a group is a subset of said group, that is itself a group
18 Cosets The left coset of subgroup H in G with respect to g is the set: gh = {gh h 2 H} Question: when is a coset a subgroup? Question: do the cosets always partition the group?
19 Quotients The quotient of G by subgroup H is the set of cosets of H in G G/H = {gh g 2 G}
20 Wallpaper Groups
21 The Groups p4 & p4m 0 1 ya 1 g(m, r, u, v) = = 2 ( 1) m cos(r /2) sin(r /2) sin(r /2) cos(r /2) ( 1) m cos(r /2) ( 1) m sin(r /2) u 4 sin(r /2) cos(r /2) v u 40 1 v
22 The Goups p6 & p6m g(m, r, u, v) = = 2 ( 1) m cos(r /3) sin(r /3) sin(r /3) cos(r /3) ( 1) m cos(r /3) ( 1) m sin(r /3) u 4 sin(r /3) cos(r /3) v u 40 1 v
23 Group Equivariant CNNs
24 How to think about CNNs A stack of feature maps is a 3D array A stack of feature maps is a vector-valued function f l : Z 2! R K l Mmm Donuts Genus one topological space
25 G-Equivariant Correlation on Z 2 Standard correlation: translate canonical filter and compute inner product G-Correlation: transform canonical filter and compute inner product
26 Translational Correlation Translation [T s f](x) =f(x s) Correlation X X K [f? ](s) = f k (x)[t s ] k (x) x2z 2 Equivariance k=1 [T s f]? = T s [f? ]
27 Group Correlation on Z 2 Transformation [T g f](x) =f(g 1 x) G-Correlation [f? ](g) = X x2z 2 K X k=1 f k (x)[t g ] k (x) Equivariance [T g f]? = T g [f? ]
28 ( Feature Transformation Law (p4) [T g f]? = T g [f? ] Transformation of 2D image??? T r ( = apple R( 0 ) apple R( ) t 0 1 = apple R( 0 ) R( 0 )t 0 1 1
29 ( ( Feature Transformation Law (p4m) [T g f]? = T g [f? ] T r ( = T m ( =
30 G-Conv is Non-Commutative G = p4 f = psi = f??f Same information content: f? (g 1 )=?f(g)
31 Group Correlation on G Transformation [T g f](h) =f(g 1 h) G-Correlation [f? ](g) = X h2g KX k=1 f k (h)[t g ] k (h) Equivariance [T g f]? = T g [f? ]
32 Non-linearities: Equivariant Function Composition Commutes with Domain Transformations [T h f]= (f h 1 )=( f) h 1 = T h [ f] f T h T h
33 Strideless G-Pooling: Equivariant Max-pool over neighborhood gu of g: Pf(g) = max k2g U f(k) Pooling Operator Commutes with G-Action PT h = T h P f T h P P T h
34 Z 2 pooling of a p6 feature map f T r f U P f P T r f T r P f
35 Subsampling Subsample on a subgroup H of G 2Z 2 Z 2 C 4 n 2Z 2 C 4 n Z 2 C 4 C 4 n Z 2 Z 2 C 4 n Z 2
36 Coset Pooling Choose pooling domain U to be a subgroup H gh = {gh h 2 H} Pooled feature map is constant on cosets Pf(gh) = max k2ghh f(k) = max k2gh f(k) That is, the pooled feature map is a function on the quotient G / H
37 Example: p4 Coset Pooling r p4 feature map array x y Pool over U = C 4 p4/c 4 = Z 2 feature map array Pool over U = Z 2 p4/z 2 = C 4 feature map array Pool over U = C 4 n 2Z 2 Z 2 /2Z 2 feature map array
38 Backprop Input grad: involuted correlation L f l 1 j (k) = L f l? l j (k) Filter grad: involuted convolution L li j (k) = L f l i f l j 1 (k)
39 Algorithms: Spatial & Spectral G-Convs
40 Naive Spatial Implementation Compute: r(g) =f (g) For each output channel j: g For each g in G-grid: Warp filter: j g = T g j Compute inner product: t x t y r j [g] =hf, g i = X i X t x X t y X f i [t x,t y, ] ji g [t x,t y, ]
41 Efficient Spatial Implementation G-Correlation [f? ](g) = X h2g KX k=1 f k (h)[t g ] k (h) Decompose into translation and rotation: T g = T tr = T t T r To get output feature plane theta, do a planar convolution with rotated filter: [f? ](t, r) = X h X f(h)l t r (h) k
42 Spectral G-Convolution f t y t y t x t x F F ˆf 0 h i ˆf h i5 ˆf K ˆ0 h ˆ1i h i5 ˆK ˆf 0 ˆ0 h ˆf 1 ˆ1i h i5 ˆf K ˆK F 1 f Folland, G. B. (1995). A Course in Abstract Harmonic Analysis.
43 Results: Rotated MNIST SVM [1] NNet [1] DBN [1] RC-RBM [2] Z2CNN P4CNNRP P4CNN 10,38 17,62 12,11 3,98 5,68 3, [1] Larochelle, H., Erhan, D., Courville, A., Bergstra, J., and Bengio, Y. (2007) An empirical evaluation of deep architectures on problems with many factors of variation. (ICML) [2] Schmidt, U., & Roth, S. (2012). Learning rotation-aware features: From invariant priors to equivariant descriptors. (CVPR)
44 CIFAR-10 All-CNN 3 x 3 conv. 96 ReLU 3 x 3 conv. 96 ReLU 3 x 3 conv. 96 ReLU (stride 2) 3 x 3 conv. 192 ReLU 3 x 3 conv. 192 ReLU 3 x 3 conv. 192 ReLU (stride 2) 3 x 3 conv. 192 ReLU 3 x 3 conv. 96 ReLU 1 x 1 conv. 192 ReLU 1 x 1 conv. 10 ReLU 3 x 3 conv. 96 ReLU Global averaging softmax P4-All-CNN Replace conv by p4-conv, halve number of filters Results
45 Work in Progress..
46 Projects / Theses / Internships taco.cohen@gmail.com
47 Questions?
Group Equivariant Convolutional Networks
Taco S. Cohen University of Amsterdam Max Welling University of Amsterdam University of California Irvine Canadian Institute for Advanced Research T.S.COHEN@UVA.NL M.WELLING@UVA.NL Abstract We introduce
More informationIntroduction to Convolutional Neural Networks (CNNs)
Introduction to Convolutional Neural Networks (CNNs) nojunk@snu.ac.kr http://mipal.snu.ac.kr Department of Transdisciplinary Studies Seoul National University, Korea Jan. 2016 Many slides are from Fei-Fei
More informationMachine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group
Machine Learning for Computer Vision 8. Neural Networks and Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group INTRODUCTION Nonlinear Coordinate Transformation http://cs.stanford.edu/people/karpathy/convnetjs/
More informationMachine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016
Machine Learning for Signal Processing Neural Networks Continue Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016 1 So what are neural networks?? Voice signal N.Net Transcription Image N.Net Text
More informationLecture 7 Convolutional Neural Networks
Lecture 7 Convolutional Neural Networks CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago April 17, 2017 We saw before: ŷ x 1 x 2 x 3 x 4 A series of matrix multiplications:
More informationGeneral Theory of Equivariant CNNs
General Theory of Equivariant CNNs Taco Cohen July 13, 2018 Collaborators Mario Geiger Maurice Weiler Jonas Koehler Marysia Winkels Max Welling Emiel Hoogeboom Jorn Peters Bas Veeling Jim Winkens Jasper
More informationSpatial Transformer Networks
BIL722 - Deep Learning for Computer Vision Spatial Transformer Networks Max Jaderberg Andrew Zisserman Karen Simonyan Koray Kavukcuoglu Contents Introduction to Spatial Transformers Related Works Spatial
More informationDeep Learning. Convolutional Neural Network (CNNs) Ali Ghodsi. October 30, Slides are partially based on Book in preparation, Deep Learning
Convolutional Neural Network (CNNs) University of Waterloo October 30, 2015 Slides are partially based on Book in preparation, by Bengio, Goodfellow, and Aaron Courville, 2015 Convolutional Networks Convolutional
More informationJakub Hajic Artificial Intelligence Seminar I
Jakub Hajic Artificial Intelligence Seminar I. 11. 11. 2014 Outline Key concepts Deep Belief Networks Convolutional Neural Networks A couple of questions Convolution Perceptron Feedforward Neural Network
More informationDeep Learning (CNNs)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deep Learning (CNNs) Deep Learning Readings: Murphy 28 Bishop - - HTF - - Mitchell
More informationCENG 783. Special topics in. Deep Learning. AlchemyAPI. Week 8. Sinan Kalkan
CENG 783 Special topics in Deep Learning AlchemyAPI Week 8 Sinan Kalkan Loss functions Many correct labels case: Binary prediction for each label, independently: L i = σ j max 0, 1 y ij f j y ij = +1 if
More informationConvolutional Neural Networks
Convolutional Neural Networks Books» http://www.deeplearningbook.org/ Books http://neuralnetworksanddeeplearning.com/.org/ reviews» http://www.deeplearningbook.org/contents/linear_algebra.html» http://www.deeplearningbook.org/contents/prob.html»
More informationNeural Architectures for Image, Language, and Speech Processing
Neural Architectures for Image, Language, and Speech Processing Karl Stratos June 26, 2018 1 / 31 Overview Feedforward Networks Need for Specialized Architectures Convolutional Neural Networks (CNNs) Recurrent
More informationCSC321 Lecture 16: ResNets and Attention
CSC321 Lecture 16: ResNets and Attention Roger Grosse Roger Grosse CSC321 Lecture 16: ResNets and Attention 1 / 24 Overview Two topics for today: Topic 1: Deep Residual Networks (ResNets) This is the state-of-the
More informationConvolutional Neural Network Architecture
Convolutional Neural Network Architecture Zhisheng Zhong Feburary 2nd, 2018 Zhisheng Zhong Convolutional Neural Network Architecture Feburary 2nd, 2018 1 / 55 Outline 1 Introduction of Convolution Motivation
More informationStructured deep models: Deep learning on graphs and beyond
Structured deep models: Deep learning on graphs and beyond Hidden layer Hidden layer Input Output ReLU ReLU, 25 May 2018 CompBio Seminar, University of Cambridge In collaboration with Ethan Fetaya, Rianne
More informationOutline. Convolution. Filtering
Filtering Outline Convolution Filtering Logistics HW1 HW2 - out tomorrow Recall: what is a digital (grayscale) image? Matrix of integer values Images as height fields Let s think of image as zero-padded
More informationLecture 17: Neural Networks and Deep Learning
UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions
More informationCS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS
CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS LAST TIME Intro to cudnn Deep neural nets using cublas and cudnn TODAY Building a better model for image classification Overfitting
More informationConvolution and Pooling as an Infinitely Strong Prior
Convolution and Pooling as an Infinitely Strong Prior Sargur Srihari srihari@buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Convolutional
More informationDeep Generative Models. (Unsupervised Learning)
Deep Generative Models (Unsupervised Learning) CEng 783 Deep Learning Fall 2017 Emre Akbaş Reminders Next week: project progress demos in class Describe your problem/goal What you have done so far What
More informationRAGAV VENKATESAN VIJETHA GATUPALLI BAOXIN LI NEURAL DATASET GENERALITY
RAGAV VENKATESAN VIJETHA GATUPALLI BAOXIN LI NEURAL DATASET GENERALITY SIFT HOG ALL ABOUT THE FEATURES DAISY GABOR AlexNet GoogleNet CONVOLUTIONAL NEURAL NETWORKS VGG-19 ResNet FEATURES COMES FROM DATA
More informationIntroduction to Deep Neural Networks
Introduction to Deep Neural Networks Presenter: Chunyuan Li Pattern Classification and Recognition (ECE 681.01) Duke University April, 2016 Outline 1 Background and Preliminaries Why DNNs? Model: Logistic
More informationBayesian Networks (Part I)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks (Part I) Graphical Model Readings: Murphy 10 10.2.1 Bishop 8.1,
More informationUnderstanding CNNs using visualisation and transformation analysis
Understanding CNNs using visualisation and transformation analysis Andrea Vedaldi Medical Imaging Summer School August 2016 Image representations 2 encoder Φ representation An encoder maps the data into
More informationNeural Networks 2. 2 Receptive fields and dealing with image inputs
CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There
More informationAgenda. Digit Classification using CNN Digit Classification using SAE Visualization: Class models, filters and saliency 2 DCT
versus 1 Agenda Deep Learning: Motivation Learning: Backpropagation Deep architectures I: Convolutional Neural Networks (CNN) Deep architectures II: Stacked Auto Encoders (SAE) Caffe Deep Learning Toolbox:
More informationFinal Examination CS 540-2: Introduction to Artificial Intelligence
Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11
More informationReliable and Interpretable Artificial Intelligence
(Sept 15, 2017) Reliable and Interpretable Artificial Intelligence How good (robust) is your neural net? [1] Pei et. al., DeepXplore: Automated Whitebox Testing of Deep Learning Systems, SOSP 2017 Attacks
More informationConvolutional neural networks
11-1: Convolutional neural networks Prof. J.C. Kao, UCLA Convolutional neural networks Motivation Biological inspiration Convolution operation Convolutional layer Padding and stride CNN architecture 11-2:
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationTeddy Einstein Math 4320
Teddy Einstein Math 4320 HW4 Solutions Problem 1: 2.92 An automorphism of a group G is an isomorphism G G. i. Prove that Aut G is a group under composition. Proof. Let f, g Aut G. Then f g is a bijective
More informationMaxout Networks. Hien Quoc Dang
Maxout Networks Hien Quoc Dang Outline Introduction Maxout Networks Description A Universal Approximator & Proof Experiments with Maxout Why does Maxout work? Conclusion 10/12/13 Hien Quoc Dang Machine
More informationLearning Deep Architectures
Learning Deep Architectures Yoshua Bengio, U. Montreal Microsoft Cambridge, U.K. July 7th, 2009, Montreal Thanks to: Aaron Courville, Pascal Vincent, Dumitru Erhan, Olivier Delalleau, Olivier Breuleux,
More informationWHY ARE DEEP NETS REVERSIBLE: A SIMPLE THEORY,
WHY ARE DEEP NETS REVERSIBLE: A SIMPLE THEORY, WITH IMPLICATIONS FOR TRAINING Sanjeev Arora, Yingyu Liang & Tengyu Ma Department of Computer Science Princeton University Princeton, NJ 08540, USA {arora,yingyul,tengyu}@cs.princeton.edu
More informationLocal Affine Approximators for Improving Knowledge Transfer
Local Affine Approximators for Improving Knowledge Transfer Suraj Srinivas & François Fleuret Idiap Research Institute and EPFL {suraj.srinivas, francois.fleuret}@idiap.ch Abstract The Jacobian of a neural
More informationDeep Learning: Self-Taught Learning and Deep vs. Shallow Architectures. Lecture 04
Deep Learning: Self-Taught Learning and Deep vs. Shallow Architectures Lecture 04 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Self-Taught Learning 1. Learn
More informationCSCI 315: Artificial Intelligence through Deep Learning
CSCI 35: Artificial Intelligence through Deep Learning W&L Fall Term 27 Prof. Levy Convolutional Networks http://wernerstudio.typepad.com/.a/6ad83549adb53ef53629ccf97c-5wi Convolution: Convolution is
More informationGlobal Optimality in Matrix and Tensor Factorization, Deep Learning & Beyond
Global Optimality in Matrix and Tensor Factorization, Deep Learning & Beyond Ben Haeffele and René Vidal Center for Imaging Science Mathematical Institute for Data Science Johns Hopkins University This
More informationUNSUPERVISED LEARNING
UNSUPERVISED LEARNING Topics Layer-wise (unsupervised) pre-training Restricted Boltzmann Machines Auto-encoders LAYER-WISE (UNSUPERVISED) PRE-TRAINING Breakthrough in 2006 Layer-wise (unsupervised) pre-training
More informationUnderstanding How ConvNets See
Understanding How ConvNets See Slides from Andrej Karpathy Springerberg et al, Striving for Simplicity: The All Convolutional Net (ICLR 2015 workshops) CSC321: Intro to Machine Learning and Neural Networks,
More informationNeural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications
Neural Networks Bishop PRML Ch. 5 Alireza Ghane Neural Networks Alireza Ghane / Greg Mori 1 Neural Networks Neural networks arise from attempts to model human/animal brains Many models, many claims of
More informationSpatial Transformation
Spatial Transformation Presented by Liqun Chen June 30, 2017 1 Overview 2 Spatial Transformer Networks 3 STN experiments 4 Recurrent Models of Visual Attention (RAM) 5 Recurrent Models of Visual Attention
More informationConvolutional Neural Networks. Srikumar Ramalingam
Convolutional Neural Networks Srikumar Ramalingam Reference Many of the slides are prepared using the following resources: neuralnetworksanddeeplearning.com (mainly Chapter 6) http://cs231n.github.io/convolutional-networks/
More informationGroup Invariance, Stability to Deformations, and Complexity of Deep Convolutional Representations
Group Invariance, Stability to Deformations, and Complexity of Deep Convolutional Representations Alberto Bietti Julien Mairal Inria, Grenoble Journées SMAI-MODE 2018, Autrans Alberto Bietti Deep Convolutional
More informationTwo at Once: Enhancing Learning and Generalization Capacities via IBN-Net
Two at Once: Enhancing Learning and Generalization Capacities via IBN-Net Supplementary Material Xingang Pan 1, Ping Luo 1, Jianping Shi 2, and Xiaoou Tang 1 1 CUHK-SenseTime Joint Lab, The Chinese University
More informationClassification of Hand-Written Digits Using Scattering Convolutional Network
Mid-year Progress Report Classification of Hand-Written Digits Using Scattering Convolutional Network Dongmian Zou Advisor: Professor Radu Balan Co-Advisor: Dr. Maneesh Singh (SRI) Background Overview
More informationThe XOR problem. Machine learning for vision. The XOR problem. The XOR problem. x 1 x 2. x 2. x 1. Fall Roland Memisevic
The XOR problem Fall 2013 x 2 Lecture 9, February 25, 2015 x 1 The XOR problem The XOR problem x 1 x 2 x 2 x 1 (picture adapted from Bishop 2006) It s the features, stupid It s the features, stupid The
More informationGroups and Symmetries
Groups and Symmetries Definition: Symmetry A symmetry of a shape is a rigid motion that takes vertices to vertices, edges to edges. Note: A rigid motion preserves angles and distances. Definition: Group
More informationLearning Deep Architectures
Learning Deep Architectures Yoshua Bengio, U. Montreal CIFAR NCAP Summer School 2009 August 6th, 2009, Montreal Main reference: Learning Deep Architectures for AI, Y. Bengio, to appear in Foundations and
More informationDeep Learning Srihari. Deep Belief Nets. Sargur N. Srihari
Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous
More informationLarge-Scale Feature Learning with Spike-and-Slab Sparse Coding
Large-Scale Feature Learning with Spike-and-Slab Sparse Coding Ian J. Goodfellow, Aaron Courville, Yoshua Bengio ICML 2012 Presented by Xin Yuan January 17, 2013 1 Outline Contributions Spike-and-Slab
More informationDeep Learning Autoencoder Models
Deep Learning Autoencoder Models Davide Bacciu Dipartimento di Informatica Università di Pisa Intelligent Systems for Pattern Recognition (ISPR) Generative Models Wrap-up Deep Learning Module Lecture Generative
More informationALGEBRAIC GROUPS J. WARNER
ALGEBRAIC GROUPS J. WARNER Let k be an algebraically closed field. varieties unless otherwise stated. 1. Definitions and Examples For simplicity we will work strictly with affine Definition 1.1. An algebraic
More information2) e = e G G such that if a G 0 =0 G G such that if a G e a = a e = a. 0 +a = a+0 = a.
Chapter 2 Groups Groups are the central objects of algebra. In later chapters we will define rings and modules and see that they are special cases of groups. Also ring homomorphisms and module homomorphisms
More informationFeature Design. Feature Design. Feature Design. & Deep Learning
Artificial Intelligence and its applications Lecture 9 & Deep Learning Professor Daniel Yeung danyeung@ieee.org Dr. Patrick Chan patrickchan@ieee.org South China University of Technology, China Appropriately
More informationIntroduction to Machine Learning (67577)
Introduction to Machine Learning (67577) Shai Shalev-Shwartz School of CS and Engineering, The Hebrew University of Jerusalem Deep Learning Shai Shalev-Shwartz (Hebrew U) IML Deep Learning Neural Networks
More informationNeed for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels
Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)
More informationLearning Deep Architectures for AI. Part II - Vijay Chakilam
Learning Deep Architectures for AI - Yoshua Bengio Part II - Vijay Chakilam Limitations of Perceptron x1 W, b 0,1 1,1 y x2 weight plane output =1 output =0 There is no value for W and b such that the model
More informationUnsupervised Learning of Hierarchical Models. in collaboration with Josh Susskind and Vlad Mnih
Unsupervised Learning of Hierarchical Models Marc'Aurelio Ranzato Geoff Hinton in collaboration with Josh Susskind and Vlad Mnih Advanced Machine Learning, 9 March 2011 Example: facial expression recognition
More informationRegML 2018 Class 8 Deep learning
RegML 2018 Class 8 Deep learning Lorenzo Rosasco UNIGE-MIT-IIT June 18, 2018 Supervised vs unsupervised learning? So far we have been thinking of learning schemes made in two steps f(x) = w, Φ(x) F, x
More informationNotas de Aula Grupos Profinitos. Martino Garonzi. Universidade de Brasília. Primeiro semestre 2018
Notas de Aula Grupos Profinitos Martino Garonzi Universidade de Brasília Primeiro semestre 2018 1 Le risposte uccidono le domande. 2 Contents 1 Topology 4 2 Profinite spaces 6 3 Topological groups 10 4
More informationIntroduction to Deep Learning CMPT 733. Steven Bergner
Introduction to Deep Learning CMPT 733 Steven Bergner Overview Renaissance of artificial neural networks Representation learning vs feature engineering Background Linear Algebra, Optimization Regularization
More informationLecture 35: Optimization and Neural Nets
Lecture 35: Optimization and Neural Nets CS 4670/5670 Sean Bell DeepDream [Google, Inceptionism: Going Deeper into Neural Networks, blog 2015] Aside: CNN vs ConvNet Note: There are many papers that use
More informationSGD and Deep Learning
SGD and Deep Learning Subgradients Lets make the gradient cheating more formal. Recall that the gradient is the slope of the tangent. f(w 1 )+rf(w 1 ) (w w 1 ) Non differentiable case? w 1 Subgradients
More informationarxiv: v1 [cs.cv] 14 Sep 2016
Understanding Convolutional Neural Networks with A Mathematical Model arxiv:1609.04112v1 [cs.cv] 14 Sep 2016 Abstract C.-C. Jay Kuo Ming-Hsieh Department of Electrical Engineering University of Southern
More informationAu-delà de la Machine de Boltzmann Restreinte. Hugo Larochelle University of Toronto
Au-delà de la Machine de Boltzmann Restreinte Hugo Larochelle University of Toronto Introduction Restricted Boltzmann Machines (RBMs) are useful feature extractors They are mostly used to initialize deep
More informationConvolutional Neural Networks II. Slides from Dr. Vlad Morariu
Convolutional Neural Networks II Slides from Dr. Vlad Morariu 1 Optimization Example of optimization progress while training a neural network. (Loss over mini-batches goes down over time.) 2 Learning rate
More informationSajid Anwar, Kyuyeon Hwang and Wonyong Sung
Sajid Anwar, Kyuyeon Hwang and Wonyong Sung Department of Electrical and Computer Engineering Seoul National University Seoul, 08826 Korea Email: sajid@dsp.snu.ac.kr, khwang@dsp.snu.ac.kr, wysung@snu.ac.kr
More information120A LECTURE OUTLINES
120A LECTURE OUTLINES RUI WANG CONTENTS 1. Lecture 1. Introduction 1 2 1.1. An algebraic object to study 2 1.2. Group 2 1.3. Isomorphic binary operations 2 2. Lecture 2. Introduction 2 3 2.1. The multiplication
More informationLecture 8: Introduction to Deep Learning: Part 2 (More on backpropagation, and ConvNets)
COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 8: Introduction to Deep Learning: Part 2 (More on backpropagation, and ConvNets) Sanjeev Arora Elad Hazan Recap: Structure of a deep
More informationExpressive Efficiency and Inductive Bias of Convolutional Networks:
Expressive Efficiency and Inductive Bias of Convolutional Networks: Analysis & Design via Hierarchical Tensor Decompositions Nadav Cohen The Hebrew University of Jerusalem AAAI Spring Symposium Series
More informationModularity Matters: Learning Invariant Relational Reasoning Tasks
Summer Review 3 Modularity Matters: Learning Invariant Relational Reasoning Tasks Jason Jo 1, Vikas Verma 2, Yoshua Bengio 1 1: MILA, Universit de Montral 2: Department of Computer Science Aalto University,
More informationArtificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino
Artificial Neural Networks Data Base and Data Mining Group of Politecnico di Torino Elena Baralis Politecnico di Torino Artificial Neural Networks Inspired to the structure of the human brain Neurons as
More informationExpressive Efficiency and Inductive Bias of Convolutional Networks:
Expressive Efficiency and Inductive Bias of Convolutional Networks: Analysis & Design via Hierarchical Tensor Decompositions Nadav Cohen The Hebrew University of Jerusalem AAAI Spring Symposium Series
More informationDeep Residual. Variations
Deep Residual Network and Its Variations Diyu Yang (Originally prepared by Kaiming He from Microsoft Research) Advantages of Depth Degradation Problem Possible Causes? Vanishing/Exploding Gradients. Overfitting
More information<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)
Learning for Deep Neural Networks (Back-propagation) Outline Summary of Previous Standford Lecture Universal Approximation Theorem Inference vs Training Gradient Descent Back-Propagation
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) For this week, Reading Chapter 4: Neural Networks (Mitchell, 1997) See Canvas For subsequent weeks: Scaling Learning Algorithms toward
More informationIntroduction to Deep Learning
Introduction to Deep Learning Some slides and images are taken from: David Wolfe Corne Wikipedia Geoffrey A. Hinton https://www.macs.hw.ac.uk/~dwcorne/teaching/introdl.ppt Feedforward networks for function
More informationImproved Local Coordinate Coding using Local Tangents
Improved Local Coordinate Coding using Local Tangents Kai Yu NEC Laboratories America, 10081 N. Wolfe Road, Cupertino, CA 95129 Tong Zhang Rutgers University, 110 Frelinghuysen Road, Piscataway, NJ 08854
More informationNeed for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels
Need for Deep Networks Perceptron Can only model linear functions Kernel Machines Non-linearity provided by kernels Need to design appropriate kernels (possibly selecting from a set, i.e. kernel learning)
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3
More informationInvariant Scattering Convolution Networks
Invariant Scattering Convolution Networks Joan Bruna and Stephane Mallat Submitted to PAMI, Feb. 2012 Presented by Bo Chen Other important related papers: [1] S. Mallat, A Theory for Multiresolution Signal
More informationComments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms
Neural networks Comments Assignment 3 code released implement classification algorithms use kernels for census dataset Thought questions 3 due this week Mini-project: hopefully you have started 2 Example:
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 6: Neural Networks and Deep Learning January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology
More informationLet your machine do the learning
Let your machine do the learning Neural Networks Daniel Hugo Cámpora Pérez Universidad de Sevilla CERN Inverted CERN School of Computing, 6th - 8th March, 2017 Daniel Hugo Cámpora Pérez Let your machine
More informationDEPARTMENT OF MATHEMATIC EDUCATION MATHEMATIC AND NATURAL SCIENCE FACULTY
HANDOUT ABSTRACT ALGEBRA MUSTHOFA DEPARTMENT OF MATHEMATIC EDUCATION MATHEMATIC AND NATURAL SCIENCE FACULTY 2012 BINARY OPERATION We are all familiar with addition and multiplication of two numbers. Both
More informationTraining Neural Networks Practical Issues
Training Neural Networks Practical Issues M. Soleymani Sharif University of Technology Fall 2017 Most slides have been adapted from Fei Fei Li and colleagues lectures, cs231n, Stanford 2017, and some from
More informationMargin Preservation of Deep Neural Networks
1 Margin Preservation of Deep Neural Networks Jure Sokolić 1, Raja Giryes 2, Guillermo Sapiro 3, Miguel R. D. Rodrigues 1 1 Department of E&EE, University College London, London, UK 2 School of EE, Faculty
More informationWhat Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1
What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single
More informationOutline. CSCI567 Machine Learning (Spring 2019) Outline. Math formulation. Prof. Victor Adamchik. Feb. 12, 2019
Outline CSCI56 Machine Learning (Spring 29) Review of last lecture Prof. Victor Adamchik 2 Convolutional neural networks U of Southern California Feb. 2, 29 Kernel methods February 2, 29 / 48 February
More informationNeural Networks. William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]
Neural Networks William Cohen 10-601 [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ] WHAT ARE NEURAL NETWORKS? William s notation Logis;c regression + 1
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationPart 2. Representation Learning Algorithms
53 Part 2 Representation Learning Algorithms 54 A neural network = running several logistic regressions at the same time If we feed a vector of inputs through a bunch of logis;c regression func;ons, then
More informationAn overview of deep learning methods for genomics
An overview of deep learning methods for genomics Matthew Ploenzke STAT115/215/BIO/BIST282 Harvard University April 19, 218 1 Snapshot 1. Brief introduction to convolutional neural networks What is deep
More informationTUTORIAL PART 1 Unsupervised Learning
TUTORIAL PART 1 Unsupervised Learning Marc'Aurelio Ranzato Department of Computer Science Univ. of Toronto ranzato@cs.toronto.edu Co-organizers: Honglak Lee, Yoshua Bengio, Geoff Hinton, Yann LeCun, Andrew
More informationStatistical Machine Learning
Statistical Machine Learning Lecture 9 Numerical optimization and deep learning Niklas Wahlström Division of Systems and Control Department of Information Technology Uppsala University niklas.wahlstrom@it.uu.se
More informationClassification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton Motivation Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses
More informationDeeply-Supervised Nets
Deeply-Supervised Nets Chen-Yu Lee Saining Xie Patrick W. Gallagher Dept. of ECE, UCSD Dept. of CSE and CogSci, UCSD Dept. of CogSci, UCSD chl60@ucsd.edu s9xie@ucsd.edu patrick.w.gallagher@gmail.com Zhengyou
More informationCSE 591: Introduction to Deep Learning in Visual Computing. - Parag S. Chandakkar - Instructors: Dr. Baoxin Li and Ragav Venkatesan
CSE 591: Introduction to Deep Learning in Visual Computing - Parag S. Chandakkar - Instructors: Dr. Baoxin Li and Ragav Venkatesan Overview Background Why another network structure? Vanishing and exploding
More information