Digital Architecture for Real-Time CNN-based Face Detection for Video Processing

Size: px
Start display at page:

Download "Digital Architecture for Real-Time CNN-based Face Detection for Video Processing"

Transcription

1 ASPC lab Grant # June , / 26 igital Architecture for Real-Time CNN-based Face etection for Video Processing Smrity Bhattarai 1, Arjuna Madanayake 1 Renato J. Cintra 2 Stefan uffner 3, Christophe Garcia 3 epartment of Electrical and Computer Engineering, University of Akron, Akron, OH, USA 1 epartamento de Estatística UFPE, Recife, Brazil 2 LIRIS, INSA Lyon, France 3 June 27, 2017

2 Outline 1 Introduction and Motivation to the Research 2 Proposed Approach 3 Hardware Implementation 4 Conclusions ASPC lab sb229@zips.uakron.edu June 27, / 26

3 Image Processing and Face etection The method of converting an image into a digital form to get an enhanced image or extract useful information by performing some operations is called image processing Processing an image signal using mathematical operations for which an input can be either image, video or graph and the output is either image or set of characteristics of the corresponding image Image processing includes applying different sets of signal processing methods on images by treating them mostly as two dimensional (2) signals ASPC lab sb229@zips.uakron.edu June 27, / 26

4 Applications of Image Processing 1 Face etection and Recognition 2 efense Surveillance 3 Moving Object Tracking 4 Content based Image Retrieval 5 Image and Video Compression 6 Image Sharpening and Restoration Explored face detection technique that uses concept of feed forward neural network which is one of the most fundamental image processing technique ASPC lab sb229@zips.uakron.edu June 27, / 26

5 Challenges in Face detection Illumination Variation Expression Variation Variation in Image Position and Scale Variation in Quality of Image ASPC lab June 27, / 26

6 Proposed Approach Explore a digital architecture for real-time face detection and recognition system Approximate hardware realization scheme learning algorithm of convolutional neural networks (CNN) [1] to improve the efficiency of the hardware implementation of neural networks and to reduce the hardware complexity Evaluate accuracy of the system under different constraints Analyze hardware complexity of the system ASPC lab sb229@zips.uakron.edu June 27, / 26

7 Block iagram of the Proposed Approach[1] First Stage Second Stage Input Image Convolution and Sub sampling Block A M0 Stage 1 output Main Block Main Block Bias and Scale ecision Main Block Stage 1 output Sub Block Part 1 Stage 1 output Stage 1 output Part 2 Block A Block A Sub Block M1 M2 M3 N0 N1 N2 N3 N4 N5 Neuron Block ecision Output ASPC lab sb229@zips.uakron.edu June 27, / 26

8 Proposed Approach - First Stage Images are matrices of order SS1 is sub-sampling operation K1, K2, K3, K4 are four kernels of order 5 5 b 11, b 12, b 13, b 14 are biases α 1,α 2,α 3,α 4 are scale factors is the decision block 1, 2, 3, 4 are four output images after first stage and are of order H k (z x, z y ) SS1 K1 K2 K3 K4 S1 α 1 α 2 α 3 α 4 Figure: First Stage B1 ec ˆ b 11 ˆ b 12 ˆ b 13 ˆ b ASPC lab sb229@zips.uakron.edu June 27, / 26

9 Proposed Approach Second Stage K={1,2,..,8} p={1,2,..,8} A k H k (z x, z y ) P k H k (z x, z y ) M k O k,1 Q k H k (z x, z y ) O k,2 K={1,2,..,6} p={8,9,..,14} S2 B2 ec Sum B3 ec b 2,p ˆ N 1,p b 3,p β p β p N 2,p ˆ b 2,p N 1,p b 3,p N 2,p B4 ec b 4 O Figure: Second Stage Images are of matrices of order M k, O k,1 and O k,2 are kernels of order 3 3 ˆb 2,p, b 3,p and b 4 are biases β p and are scale factors and decision block respectively N 1,p are neurons of order 6 7 N 2,p is a single neuron ASPC lab sb229@zips.uakron.edu June 27, / 26

10 Hardware Implementation Convolution and Sub-sampling M N J(i k, j k ) = K(m, n).i [(i k 1) + m, (j k 1) + n] (1) m=1 n=1 H k (z x, z y ) = (1 + z 1 x )(1 + zy 1 ) 4 (2) where, J is the convolution operation H is the sub-sampling operation K is kernel matrix of order m n I is image matrix ASPC lab sb229@zips.uakron.edu June 27, / 26

11 Convolution and Sub-sampling From Equations 1 and 2 Q1 Q2 Q5 Q6 Q20 Q21 Q24 Q25 Q C = L+1 i=l S+1 x(i, j) j=s k11 k12 k15 k21 k45 k51 k54 Q1 Q2 Q5 Q6 A 1 Q20 Q21 Q24 Q25 l11 l12 l15 l21 l45 l51 l54 l55 k55 where, L = 2(T \A) + (C 1)\R + 1 S = 2[(T 1)%A] + (C 1)%R + 1 T = Time Stamp R = Order of Kernel = 5 C = Column Number Q 1 = x 11 + x 12 + x 21 + x 22 Q1 Q2 Q5 Q6 A 2 Q20 Q21 Q24 Q25 m11 m12 m15 m21 m45 m51 m54 m55 Q1 Q2 Q5 Q6 A 3 Q20 Q21 Q24 Q25 n11 n12 n15 n21 n45 n51 n54 n55 Q 25 = x 55 + x 56 + x 65 + x 66 A 4 Figure: Combined Convolution and Sub-sampling Block ASPC lab sb229@zips.uakron.edu June 27, / 26

12 Scaling and Bias Block S1 = (α 1, α 2, α 3, α 4 ) A 1 α 1 X 1 A 2 α 2 X 2 A 3 α 3 X 3 A 4 α 4 X 4 Figure: Scaling Block B1 = b 1,k ˆ b 1,k ˆ = b 1,k.α k + b 1,2,k b 1,2,k ˆ b 1,k b 1,k α k ˆ b 1,k X k ˆX k k = 1, 2, 3, 4 Figure: Bias Block ASPC lab sb229@zips.uakron.edu June 27, / 26

13 ecision Block Input 2 a b 2 a > b x 1 Sel MSB Sel d 0 â Output 1 d 0 d 1 1 d 1 Figure: ecision Block The decision block is basically linear-2 approximation[2] and given as: 1 if x < 2 σ(x) = â. x/2 if 2 x < 2 1 if x 2 where, â = 7/4 ASPC lab sb229@zips.uakron.edu June 27, / 26

14 Hardware Implementation etailed Block iagram of Second Stage SS C2 B1 S1 B2 ec N1 SumB3ec N2 b2,1 β1 b2,2,1 N1,1 NB1 N2,1 A M1 M2 b2,2 β2 b2,2,2 N1,2 NB2 N2,2 b2,3 β3 b2,2,3 N1,3 NB3 N2,3 B M3 M4 b2,4 β4 b2,2,4 N1,4 NB4 N2,4 b2,5 β5 b2,2,5 N1,5 NB5 N2,5 C M5 M6 b2,6 β6 b2,2,6 N1,6 NB6 N2,6 b2,7 β7 b2,2,7 N1,7 NB7 N2,7 M7 b2,8 β8 b2,2,8 N1,8 NB8 N2,8 M8 A B R11 R12 b2,9 β9 b2,2,9 N1,9 NB9 N2,9 B4 ec Output A R21 b2,10 β10 b2,2,10 N1,10 NB10 N2,10 C R22 Sum All A R31 R32 b2,11 β11 b2,2,11 N1,11 NB11 N2,11 B C R41 R42 b2,12 β12 b2,2,12 N1,12 NB12 N2,12 B R51 R52 b2,13 β13 b2,2,13 N1,13 NB13 N2,13 C R61 R62 b2,14 β14 b2,2,14 N1,14 NB14 N2,14 ASPC lab sb229@zips.uakron.edu June 27, / 26

15 Convolution and Sub-sampling Q1 Q2 Q3 Q4 Q5 Q6 Q7 Q8 Q9 a11 a12 a13 a21 a22 a23 a31 a32 a33 F 1 Figure: 1st part of 2nd Stage of Combined Convolution Sub-sampling Block Q1 Q2 Q3 Q4 Q5 Q6 Q7 Q8 Q9 i11 i12 i13 i21 i22 i23 i31 i32 i33 F 9 j11 j12 j13 j21 j22 j23 j31 j32 j33 Q1 Q2 Q3 Q4 Q5 Q6 Q7 Q8 Q9 Figure: 2nd part of 2nd Stage of Combined Convolution Sub-sampling Block Followed by a scale, bias and decision block like in first stage ASPC lab sb229@zips.uakron.edu June 27, / 26

16 Neuron Block including Bias and ecision Block N 1,p NB p Input 42 ec Output z 1 Figure: Neuron Block Total 14 matrices of size 6 7 and is represented by N p where p = 1, 2,..14 Contains bias for each neuron block represented by the set NB p NB p = ˆ b 3,p Followed by an activation function represented by decision block Scalar multiplied with the matrix generated after second stage of decision block element by element and passes through the added decision block ASPC lab sb229@zips.uakron.edu June 27, / 26

17 Neuron Multiplication Block with Final bias and ecision Block Ĵ p is the output of Neuron Block The set Ĵ p is multiplied with corresponding neurons set called N 2,p Ĵ1 Ĵ2 N2,1 N2,2 All the elements of the resulting set ˆV p is added together to get a value W The value W obtained after Neuron Multiplication is biased with bias B4 The output is then passed through the decision block giving rise to obtain final output Final Output If the final output is positive, the image taken at the beginning is a face else the image taken is not a face ˆ J13 ˆ J14 N2,13 N2,14 W Figure: Neuron Multiplication Block B4 W ec Output Figure: Final Bias and ecision Block ASPC lab sb229@zips.uakron.edu June 27, / 26

18 Analysis and Results Tests were carried out using different set of images All the input images were standardized to size and in.pgm (Portable Gray Map) format The algorithm was first modelled and tested in MATLAB Simulink using Xilinx block set Tested on Xilinx Virtex-6 ML605 FPGA Evaluation Kit 1164 images and were processed Figure: Sample test images ASPC lab June 27, / 26

19 Accuracy etails MATLAB Accuracy etails Tested Images 1164 MATLAB Correctly etected 1124 Accuracy of MATLAB 96.56% Change in Accuracy of esign After Changing the Word-Length of Scale and Biases Word Length Part Scale Bias Integer 8 8 Fractional 5 5 Integer 7 7 Fractional 4 4 Integer 6 6 Fractional 3 3 Integer 5 5 Fractional 2 2 Integer 4 4 Fractional 1 1 Correctly etected Xilinx Accuracy % % % % % ASPC lab sb229@zips.uakron.edu June 27, / 26

20 Hardware Complexity Hardware Consumption Along with Timing etails Scale and Bias Hardware Consumption Stage T Integer Fractional LUTs FFs cpd F max (ns) (MHz) ASPC lab sb229@zips.uakron.edu June 27, / 26

21 Hardware Complexity and Accuracy Change in Accuracy After Changing the Word-Length of Kernel from the First stage Word - Length Correctly etected Xilinx Integer Fractional Accuracy % % % % % Hardware Consumption Along with Timing etails Kernel from first stage Hardware Consumption Integer Fractional LUTs FFs T cpd F max (ns) (MHz) ASPC lab sb229@zips.uakron.edu June 27, / 26

22 Hardware Complexity and Accuracy Change in Accuracy After Changing the Word-Length of Kernel from the Second Stage Word - Length Correctly etected Xilinx Integer Fractional Accuracy % % % % % Hardware Consumption Along with Timing etails Kernel from second stage Hardware Consumption Integer Fractional LUTs FFs T cpd F max (ns) (MHz) ASPC lab sb229@zips.uakron.edu June 27, / 26

23 Hardware Complexity and Accuracy Change in Accuracy After Changing the Word-Length of Neuron Biases Word - Length Correctly etected Integer Fractional Xilinx Accuracy % % % % % Hardware Consumption Along with Timing etails Neuron Hardware Consumption Integer Fractional LUTs FFs T cpd F max (ns) (MHz) ASPC lab sb229@zips.uakron.edu June 27, / 26

24 Conclusions A hardware system capable of face detection and recognition was designed The implementation paves a path for an efficient architecture designed to detect highly variable face patterns The system is a fully trained architecture comprising of a pipeline of convolution, sub-sampling and neural networks, such that no pre-processing of the image is required The experimental results show high accuracy ( 96%) The proposed model can be a good solution in various facial recognition and detection application ASPC lab sb229@zips.uakron.edu June 27, / 26

25 References I [1] C. Garcia and M. elakis, Convolutional face finder: A neural architecture for fast and robust face detection, IEEE Transactions on pattern analysis and machine intelligence, vol. 26, no. 11, pp , [2] J. Schmidhuber, eep learning in neural networks: An overview, Neural Networks, vol. 61, pp , ASPC lab sb229@zips.uakron.edu June 27, / 26

26 Thank You

Intelligent Systems Discriminative Learning, Neural Networks

Intelligent Systems Discriminative Learning, Neural Networks Intelligent Systems Discriminative Learning, Neural Networks Carsten Rother, Dmitrij Schlesinger WS2014/2015, Outline 1. Discriminative learning 2. Neurons and linear classifiers: 1) Perceptron-Algorithm

More information

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen Neural Networks - I Henrik I Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I Christensen (RIM@GT) Neural Networks 1 /

More information

A Parallel Method for the Computation of Matrix Exponential based on Truncated Neumann Series

A Parallel Method for the Computation of Matrix Exponential based on Truncated Neumann Series A Parallel Method for the Computation of Matrix Exponential based on Truncated Neumann Series V. S. Dimitrov 12, V. Ariyarathna 3, D. F. G. Coelho 1, L. Rakai 1, A. Madanayake 3, R. J. Cintra 4 1 ECE Department,

More information

A Deep Convolutional Neural Network Based on Nested Residue Number System

A Deep Convolutional Neural Network Based on Nested Residue Number System A Deep Convolutional Neural Network Based on Nested Residue Number System Hiroki Nakahara Tsutomu Sasao Ehime University, Japan Meiji University, Japan Outline Background Deep convolutional neural network

More information

The z-vertex Trigger for Belle II

The z-vertex Trigger for Belle II The z-vertex Trigger for Belle II Sebastian Skambraks Technische Universität München IMPRS Young Scientist Workshop at Ringberg Castle July 10, 2015 1 / 13 Outline Introduction Motivation Signal Flow Multi

More information

CSC321 Lecture 5: Multilayer Perceptrons

CSC321 Lecture 5: Multilayer Perceptrons CSC321 Lecture 5: Multilayer Perceptrons Roger Grosse Roger Grosse CSC321 Lecture 5: Multilayer Perceptrons 1 / 21 Overview Recall the simple neuron-like unit: y output output bias i'th weight w 1 w2 w3

More information

Jakub Hajic Artificial Intelligence Seminar I

Jakub Hajic Artificial Intelligence Seminar I Jakub Hajic Artificial Intelligence Seminar I. 11. 11. 2014 Outline Key concepts Deep Belief Networks Convolutional Neural Networks A couple of questions Convolution Perceptron Feedforward Neural Network

More information

A Gray Code Based Time-to-Digital Converter Architecture and its FPGA Implementation

A Gray Code Based Time-to-Digital Converter Architecture and its FPGA Implementation A Gray Code Based Time-to-Digital Converter Architecture and its FPGA Implementation Congbing Li Haruo Kobayashi Gunma University Gunma University Kobayashi Lab Outline Research Objective & Background

More information

Reduced-Area Constant-Coefficient and Multiple-Constant Multipliers for Xilinx FPGAs with 6-Input LUTs

Reduced-Area Constant-Coefficient and Multiple-Constant Multipliers for Xilinx FPGAs with 6-Input LUTs Article Reduced-Area Constant-Coefficient and Multiple-Constant Multipliers for Xilinx FPGAs with 6-Input LUTs E. George Walters III Department of Electrical and Computer Engineering, Penn State Erie,

More information

ECE521 Lectures 9 Fully Connected Neural Networks

ECE521 Lectures 9 Fully Connected Neural Networks ECE521 Lectures 9 Fully Connected Neural Networks Outline Multi-class classification Learning multi-layer neural networks 2 Measuring distance in probability space We learnt that the squared L2 distance

More information

Sajid Anwar, Kyuyeon Hwang and Wonyong Sung

Sajid Anwar, Kyuyeon Hwang and Wonyong Sung Sajid Anwar, Kyuyeon Hwang and Wonyong Sung Department of Electrical and Computer Engineering Seoul National University Seoul, 08826 Korea Email: sajid@dsp.snu.ac.kr, khwang@dsp.snu.ac.kr, wysung@snu.ac.kr

More information

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann ECLT 5810 Classification Neural Networks Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann Neural Networks A neural network is a set of connected input/output

More information

Spatial Transformation

Spatial Transformation Spatial Transformation Presented by Liqun Chen June 30, 2017 1 Overview 2 Spatial Transformer Networks 3 STN experiments 4 Recurrent Models of Visual Attention (RAM) 5 Recurrent Models of Visual Attention

More information

Artificial Neural Networks

Artificial Neural Networks Introduction ANN in Action Final Observations Application: Poverty Detection Artificial Neural Networks Alvaro J. Riascos Villegas University of los Andes and Quantil July 6 2018 Artificial Neural Networks

More information

Deep Feedforward Networks. Sargur N. Srihari

Deep Feedforward Networks. Sargur N. Srihari Deep Feedforward Networks Sargur N. srihari@cedar.buffalo.edu 1 Topics Overview 1. Example: Learning XOR 2. Gradient-Based Learning 3. Hidden Units 4. Architecture Design 5. Backpropagation and Other Differentiation

More information

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES

More information

Online Videos FERPA. Sign waiver or sit on the sides or in the back. Off camera question time before and after lecture. Questions?

Online Videos FERPA. Sign waiver or sit on the sides or in the back. Off camera question time before and after lecture. Questions? Online Videos FERPA Sign waiver or sit on the sides or in the back Off camera question time before and after lecture Questions? Lecture 1, Slide 1 CS224d Deep NLP Lecture 4: Word Window Classification

More information

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture Neural Networks for Machine Learning Lecture 2a An overview of the main types of neural network architecture Geoffrey Hinton with Nitish Srivastava Kevin Swersky Feed-forward neural networks These are

More information

Simple neuron model Components of simple neuron

Simple neuron model Components of simple neuron Outline 1. Simple neuron model 2. Components of artificial neural networks 3. Common activation functions 4. MATLAB representation of neural network. Single neuron model Simple neuron model Components

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Yongjin Park 1 Goal of Feedforward Networks Deep Feedforward Networks are also called as Feedforward neural networks or Multilayer Perceptrons Their Goal: approximate some function

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Neural Networks Varun Chandola x x 5 Input Outline Contents February 2, 207 Extending Perceptrons 2 Multi Layered Perceptrons 2 2. Generalizing to Multiple Labels.................

More information

Neural Networks 2. 2 Receptive fields and dealing with image inputs

Neural Networks 2. 2 Receptive fields and dealing with image inputs CS 446 Machine Learning Fall 2016 Oct 04, 2016 Neural Networks 2 Professor: Dan Roth Scribe: C. Cheng, C. Cervantes Overview Convolutional Neural Networks Recurrent Neural Networks 1 Introduction There

More information

Artificial neural networks

Artificial neural networks Artificial neural networks Chapter 8, Section 7 Artificial Intelligence, spring 203, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 8, Section 7 Outline Brains Neural

More information

Design of Basic Logic Gates using CMOS and Artificial Neural Networks (ANN)

Design of Basic Logic Gates using CMOS and Artificial Neural Networks (ANN) AMSE JOURNALS-2016-Series: Advances D; Vol. 21; N 1; pp 54-65 Submitted July 2016; Revised Oct. 11, 2016, Accepted Nov. 15, 2016 Design of Basic Logic Gates using CMOS and Artificial Neural Networks (ANN)

More information

<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)

<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation) Learning for Deep Neural Networks (Back-propagation) Outline Summary of Previous Standford Lecture Universal Approximation Theorem Inference vs Training Gradient Descent Back-Propagation

More information

Deep Learning: Approximation of Functions by Composition

Deep Learning: Approximation of Functions by Composition Deep Learning: Approximation of Functions by Composition Zuowei Shen Department of Mathematics National University of Singapore Outline 1 A brief introduction of approximation theory 2 Deep learning: approximation

More information

Higher-order Statistical Modeling based Deep CNNs (Introduction) Peihua Li Dalian University of Technology

Higher-order Statistical Modeling based Deep CNNs (Introduction) Peihua Li Dalian University of Technology Higher-order Statistical Modeling based Deep CNNs (Introduction) Peihua Li Dalian University of Technology http://peihuali.org Outline What is Higher-order? Why We Study High-order Overview of Speaker

More information

Classification with Perceptrons. Reading:

Classification with Perceptrons. Reading: Classification with Perceptrons Reading: Chapters 1-3 of Michael Nielsen's online book on neural networks covers the basics of perceptrons and multilayer neural networks We will cover material in Chapters

More information

Lecture: Face Recognition

Lecture: Face Recognition Lecture: Face Recognition Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 12-1 What we will learn today Introduction to face recognition The Eigenfaces Algorithm Linear

More information

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight) CSCE 636 Neural Networks Instructor: Yoonsuck Choe Deep Learning What Is Deep Learning? Learning higher level abstractions/representations from data. Motivation: how the brain represents sensory information

More information

Convolutional Neural Network Architecture

Convolutional Neural Network Architecture Convolutional Neural Network Architecture Zhisheng Zhong Feburary 2nd, 2018 Zhisheng Zhong Convolutional Neural Network Architecture Feburary 2nd, 2018 1 / 55 Outline 1 Introduction of Convolution Motivation

More information

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA   1/ 21 Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural

More information

Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

More information

Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks

Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks Yufei Ma, Yu Cao, Sarma Vrudhula,

More information

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari Deep Belief Nets Sargur N. Srihari srihari@cedar.buffalo.edu Topics 1. Boltzmann machines 2. Restricted Boltzmann machines 3. Deep Belief Networks 4. Deep Boltzmann machines 5. Boltzmann machines for continuous

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Machine Learning with Quantum-Inspired Tensor Networks

Machine Learning with Quantum-Inspired Tensor Networks Machine Learning with Quantum-Inspired Tensor Networks E.M. Stoudenmire and David J. Schwab Advances in Neural Information Processing 29 arxiv:1605.05775 RIKEN AICS - Mar 2017 Collaboration with David

More information

Convolutional Associative Memory: FIR Filter Model of Synapse

Convolutional Associative Memory: FIR Filter Model of Synapse Convolutional Associative Memory: FIR Filter Model of Synapse Rama Murthy Garimella 1, Sai Dileep Munugoti 2, Anil Rayala 1 1 International Institute of Information technology, Hyderabad, India. rammurthy@iiit.ac.in,

More information

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Speaker Representation and Verification Part II. by Vasileios Vasilakakis Speaker Representation and Verification Part II by Vasileios Vasilakakis Outline -Approaches of Neural Networks in Speaker/Speech Recognition -Feed-Forward Neural Networks -Training with Back-propagation

More information

Final Examination CS540-2: Introduction to Artificial Intelligence

Final Examination CS540-2: Introduction to Artificial Intelligence Final Examination CS540-2: Introduction to Artificial Intelligence May 9, 2018 LAST NAME: SOLUTIONS FIRST NAME: Directions 1. This exam contains 33 questions worth a total of 100 points 2. Fill in your

More information

Pipelined Viterbi Decoder Using FPGA

Pipelined Viterbi Decoder Using FPGA Research Journal of Applied Sciences, Engineering and Technology 5(4): 1362-1372, 2013 ISSN: 2040-7459; e-issn: 2040-7467 Maxwell Scientific Organization, 2013 Submitted: July 05, 2012 Accepted: August

More information

Research Article Architecture Analysis of an FPGA-Based Hopfield Neural Network

Research Article Architecture Analysis of an FPGA-Based Hopfield Neural Network Artificial Neural Systems, Article ID 602325, 10 pages http://dx.doi.org/10.1155/2014/602325 Research Article Architecture Analysis of an FPGA-Based Hopfield Neural Network Miguel Angelo de Abreu de Sousa,

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Multi-scale Geometric Summaries for Similarity-based Upstream S

Multi-scale Geometric Summaries for Similarity-based Upstream S Multi-scale Geometric Summaries for Similarity-based Upstream Sensor Fusion Duke University, ECE / Math 3/6/2019 Overall Goals / Design Choices Leverage multiple, heterogeneous modalities in identification

More information

Binary Convolutional Neural Network on RRAM

Binary Convolutional Neural Network on RRAM Binary Convolutional Neural Network on RRAM Tianqi Tang, Lixue Xia, Boxun Li, Yu Wang, Huazhong Yang Dept. of E.E, Tsinghua National Laboratory for Information Science and Technology (TNList) Tsinghua

More information

Neural networks. Chapter 19, Sections 1 5 1

Neural networks. Chapter 19, Sections 1 5 1 Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10

More information

A hardware architecture for fast video object recognition using SVM and Zernike Moments

A hardware architecture for fast video object recognition using SVM and Zernike Moments A hardware architecture for fast video object recognition using SVM and Zernike Moments Cedric LEMAITRE a, Johel MITERAN a, Olivier AUBRETON b and Romuald MOSQUERON a a Laboratoire Lei (UMR CNRS 5158),

More information

EECS Components and Design Techniques for Digital Systems. Lec 26 CRCs, LFSRs (and a little power)

EECS Components and Design Techniques for Digital Systems. Lec 26 CRCs, LFSRs (and a little power) EECS 150 - Components and esign Techniques for igital Systems Lec 26 CRCs, LFSRs (and a little power) avid Culler Electrical Engineering and Computer Sciences University of California, Berkeley http://www.eecs.berkeley.edu/~culler

More information

Advanced Hardware Architecture for Soft Decoding Reed-Solomon Codes

Advanced Hardware Architecture for Soft Decoding Reed-Solomon Codes Advanced Hardware Architecture for Soft Decoding Reed-Solomon Codes Stefan Scholl, Norbert Wehn Microelectronic Systems Design Research Group TU Kaiserslautern, Germany Overview Soft decoding decoding

More information

Lecture 6. Notes on Linear Algebra. Perceptron

Lecture 6. Notes on Linear Algebra. Perceptron Lecture 6. Notes on Linear Algebra. Perceptron COMP90051 Statistical Machine Learning Semester 2, 2017 Lecturer: Andrey Kan Copyright: University of Melbourne This lecture Notes on linear algebra Vectors

More information

HARDWARE IMPLEMENTATION OF FIR/IIR DIGITAL FILTERS USING INTEGRAL STOCHASTIC COMPUTATION. Arash Ardakani, François Leduc-Primeau and Warren J.

HARDWARE IMPLEMENTATION OF FIR/IIR DIGITAL FILTERS USING INTEGRAL STOCHASTIC COMPUTATION. Arash Ardakani, François Leduc-Primeau and Warren J. HARWARE IMPLEMENTATION OF FIR/IIR IGITAL FILTERS USING INTEGRAL STOCHASTIC COMPUTATION Arash Ardakani, François Leduc-Primeau and Warren J. Gross epartment of Electrical and Computer Engineering McGill

More information

Implementation of Nonlinear Template Runner Emulated Digital CNN-UM on FPGA

Implementation of Nonlinear Template Runner Emulated Digital CNN-UM on FPGA Implementation of Nonlinear Template Runner Emulated Digital CNN-UM on FPGA Z. Kincses * Z. Nagy P. Szolgay * Department of Image Processing and Neurocomputing, University of Pannonia, Hungary * e-mail:

More information

Elliptic Curve Group Core Specification. Author: Homer Hsing

Elliptic Curve Group Core Specification. Author: Homer Hsing Elliptic Curve Group Core Specification Author: Homer Hsing homer.hsing@gmail.com Rev. 0.1 March 4, 2012 This page has been intentionally left blank. www.opencores.org Rev 0.1 ii Revision History Rev.

More information

Intelligent Handwritten Digit Recognition using Artificial Neural Network

Intelligent Handwritten Digit Recognition using Artificial Neural Network RESEARCH ARTICLE OPEN ACCESS Intelligent Handwritten Digit Recognition using Artificial Neural Networ Saeed AL-Mansoori Applications Development and Analysis Center (ADAC), Mohammed Bin Rashid Space Center

More information

SP-CNN: A Scalable and Programmable CNN-based Accelerator. Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay

SP-CNN: A Scalable and Programmable CNN-based Accelerator. Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay SP-CNN: A Scalable and Programmable CNN-based Accelerator Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay Motivation Power is a first-order design constraint, especially for embedded devices. Certain

More information

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single

More information

Simple Neural Nets For Pattern Classification

Simple Neural Nets For Pattern Classification CHAPTER 2 Simple Neural Nets For Pattern Classification Neural Networks General Discussion One of the simplest tasks that neural nets can be trained to perform is pattern classification. In pattern classification

More information

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight) CSCE 636 Neural Networks Instructor: Yoonsuck Choe Deep Learning What Is Deep Learning? Learning higher level abstractions/representations from data. Motivation: how the brain represents sensory information

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Matrix Notation Mark Schmidt University of British Columbia Winter 2017 Admin Auditting/registration forms: Submit them at end of class, pick them up end of next class. I need

More information

An Approximate Parallel Multiplier with Deterministic Errors for Ultra-High Speed Integrated Optical Circuits

An Approximate Parallel Multiplier with Deterministic Errors for Ultra-High Speed Integrated Optical Circuits An Approximate Parallel Multiplier with Deterministic Errors for Ultra-High Speed Integrated Optical Circuits Jun Shiomi 1, Tohru Ishihara 1, Hidetoshi Onodera 1, Akihiko Shinya 2, Masaya Notomi 2 1 Graduate

More information

A Novel Activity Detection Method

A Novel Activity Detection Method A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE 4: Linear Systems Summary # 3: Introduction to artificial neural networks DISTRIBUTED REPRESENTATION An ANN consists of simple processing units communicating with each other. The basic elements of

More information

FPGA-BASED HOUSEHOLDER IMPLEMENTATION FOR EFFICIENT EIGENVALUES CALCULATION. Imtiaz Ahmed Sajid, Muhammad Mansoor Ahmed and Muhammad Sagheer

FPGA-BASED HOUSEHOLDER IMPLEMENTATION FOR EFFICIENT EIGENVALUES CALCULATION. Imtiaz Ahmed Sajid, Muhammad Mansoor Ahmed and Muhammad Sagheer International Journal of Innovative Computing, Information and Control ICIC International c 2011 ISSN 1349-4198 Volume 7, Number 10, October 2011 pp. 5939 5946 FPGA-BASED HOUSEHOLDER IMPLEMENTATION FOR

More information

Lab 5: 16 th April Exercises on Neural Networks

Lab 5: 16 th April Exercises on Neural Networks Lab 5: 16 th April 01 Exercises on Neural Networks 1. What are the values of weights w 0, w 1, and w for the perceptron whose decision surface is illustrated in the figure? Assume the surface crosses the

More information

Robust Controller Design for Speed Control of an Indirect Field Oriented Induction Machine Drive

Robust Controller Design for Speed Control of an Indirect Field Oriented Induction Machine Drive Leonardo Electronic Journal of Practices and Technologies ISSN 1583-1078 Issue 6, January-June 2005 p. 1-16 Robust Controller Design for Speed Control of an Indirect Field Oriented Induction Machine Drive

More information

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation 1 Introduction A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation J Wesley Hines Nuclear Engineering Department The University of Tennessee Knoxville, Tennessee,

More information

Neural networks. Chapter 20, Section 5 1

Neural networks. Chapter 20, Section 5 1 Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

More information

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS Learning Neural Networks Classifier Short Presentation INPUT: classification data, i.e. it contains an classification (class) attribute.

More information

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS LAST TIME Intro to cudnn Deep neural nets using cublas and cudnn TODAY Building a better model for image classification Overfitting

More information

Lecture 4: Perceptrons and Multilayer Perceptrons

Lecture 4: Perceptrons and Multilayer Perceptrons Lecture 4: Perceptrons and Multilayer Perceptrons Cognitive Systems II - Machine Learning SS 2005 Part I: Basic Approaches of Concept Learning Perceptrons, Artificial Neuronal Networks Lecture 4: Perceptrons

More information

Learning Methods for Linear Detectors

Learning Methods for Linear Detectors Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2011/2012 Lesson 20 27 April 2012 Contents Learning Methods for Linear Detectors Learning Linear Detectors...2

More information

THE MOST IMPORTANT BIT

THE MOST IMPORTANT BIT NEURAL NETWORKS THE MOST IMPORTANT BIT A neural network represents a function f : R d R d 2. Peter Orbanz Applied Data Mining 262 BUILDING BLOCKS Units The basic building block is a node or unit: φ The

More information

Structure Design of Neural Networks Using Genetic Algorithms

Structure Design of Neural Networks Using Genetic Algorithms Structure Design of Neural Networks Using Genetic Algorithms Satoshi Mizuta Takashi Sato Demelo Lao Masami Ikeda Toshio Shimizu Department of Electronic and Information System Engineering, Faculty of Science

More information

Deep Learning Lab Course 2017 (Deep Learning Practical)

Deep Learning Lab Course 2017 (Deep Learning Practical) Deep Learning Lab Course 207 (Deep Learning Practical) Labs: (Computer Vision) Thomas Brox, (Robotics) Wolfram Burgard, (Machine Learning) Frank Hutter, (Neurorobotics) Joschka Boedecker University of

More information

Bidirectional Representation and Backpropagation Learning

Bidirectional Representation and Backpropagation Learning Int'l Conf on Advances in Big Data Analytics ABDA'6 3 Bidirectional Representation and Bacpropagation Learning Olaoluwa Adigun and Bart Koso Department of Electrical Engineering Signal and Image Processing

More information

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann (Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for

More information

Deep Belief Networks are Compact Universal Approximators

Deep Belief Networks are Compact Universal Approximators Deep Belief Networks are Compact Universal Approximators Franck Olivier Ndjakou Njeunje Applied Mathematics and Scientific Computation May 16, 2016 1 / 29 Outline 1 Introduction 2 Preliminaries Universal

More information

Neural Networks: Introduction

Neural Networks: Introduction Neural Networks: Introduction Machine Learning Fall 2017 Based on slides and material from Geoffrey Hinton, Richard Socher, Dan Roth, Yoav Goldberg, Shai Shalev-Shwartz and Shai Ben-David, and others 1

More information

CITS 4402 Computer Vision

CITS 4402 Computer Vision CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh Lecture 06 Object Recognition Objectives To understand the concept of image based object recognition To learn how to match images

More information

Conditional Language modeling with attention

Conditional Language modeling with attention Conditional Language modeling with attention 2017.08.25 Oxford Deep NLP 조수현 Review Conditional language model: assign probabilities to sequence of words given some conditioning context x What is the probability

More information

Vector Quantization for Machine Vision

Vector Quantization for Machine Vision 1 Vector Quantization for Machine Vision Vincenzo Liguori Ocean Logic Pty Ltd, Australia Email: enzo@ocean-logic.com Abstract This paper shows how to reduce the computational cost for a variety of common

More information

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY [Gaurav, 2(1): Jan., 2013] ISSN: 2277-9655 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY Face Identification & Detection Using Eigenfaces Sachin.S.Gurav *1, K.R.Desai 2 *1

More information

CSC411: Final Review. James Lucas & David Madras. December 3, 2018

CSC411: Final Review. James Lucas & David Madras. December 3, 2018 CSC411: Final Review James Lucas & David Madras December 3, 2018 Agenda 1. A brief overview 2. Some sample questions Basic ML Terminology The final exam will be on the entire course; however, it will be

More information

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller 2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks Todd W. Neller Machine Learning Learning is such an important part of what we consider "intelligence" that

More information

An artificial neural networks (ANNs) model is a functional abstraction of the

An artificial neural networks (ANNs) model is a functional abstraction of the CHAPER 3 3. Introduction An artificial neural networs (ANNs) model is a functional abstraction of the biological neural structures of the central nervous system. hey are composed of many simple and highly

More information

Convolutional Neural Networks. Srikumar Ramalingam

Convolutional Neural Networks. Srikumar Ramalingam Convolutional Neural Networks Srikumar Ramalingam Reference Many of the slides are prepared using the following resources: neuralnetworksanddeeplearning.com (mainly Chapter 6) http://cs231n.github.io/convolutional-networks/

More information

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton Motivation Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses

More information

Notes on Back Propagation in 4 Lines

Notes on Back Propagation in 4 Lines Notes on Back Propagation in 4 Lines Lili Mou moull12@sei.pku.edu.cn March, 2015 Congratulations! You are reading the clearest explanation of forward and backward propagation I have ever seen. In this

More information

arxiv: v1 [astro-ph.im] 20 Jan 2017

arxiv: v1 [astro-ph.im] 20 Jan 2017 IAU Symposium 325 on Astroinformatics Proceedings IAU Symposium No. xxx, xxx A.C. Editor, B.D. Editor & C.E. Editor, eds. c xxx International Astronomical Union DOI: 00.0000/X000000000000000X Deep learning

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

@SoyGema GEMA PARREÑO PIQUERAS

@SoyGema GEMA PARREÑO PIQUERAS @SoyGema GEMA PARREÑO PIQUERAS WHAT IS AN ARTIFICIAL NEURON? WHAT IS AN ARTIFICIAL NEURON? Image Recognition Classification using Softmax Regressions and Convolutional Neural Networks Languaje Understanding

More information

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

COMP 551 Applied Machine Learning Lecture 14: Neural Networks COMP 551 Applied Machine Learning Lecture 14: Neural Networks Instructor: Ryan Lowe (ryan.lowe@mail.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted,

More information

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY Volume 6, Issue 3, March 2016 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Eigenface and

More information

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY 1 On-line Resources http://neuralnetworksanddeeplearning.com/index.html Online book by Michael Nielsen http://matlabtricks.com/post-5/3x3-convolution-kernelswith-online-demo

More information

Deep Learning (CNNs)

Deep Learning (CNNs) 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deep Learning (CNNs) Deep Learning Readings: Murphy 28 Bishop - - HTF - - Mitchell

More information

Arithmetic operators for pairing-based cryptography

Arithmetic operators for pairing-based cryptography 7. Kryptotag November 9 th, 2007 Arithmetic operators for pairing-based cryptography Jérémie Detrey Cosec, B-IT, Bonn, Germany jdetrey@bit.uni-bonn.de Joint work with: Jean-Luc Beuchat Nicolas Brisebarre

More information

Sparse Kernel Machines - SVM

Sparse Kernel Machines - SVM Sparse Kernel Machines - SVM Henrik I. Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I. Christensen (RIM@GT) Support

More information

Deep Learning Autoencoder Models

Deep Learning Autoencoder Models Deep Learning Autoencoder Models Davide Bacciu Dipartimento di Informatica Università di Pisa Intelligent Systems for Pattern Recognition (ISPR) Generative Models Wrap-up Deep Learning Module Lecture Generative

More information

Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch.

Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch. Neural Networks Oliver Schulte - CMPT 726 Bishop PRML Ch. 5 Neural Networks Neural networks arise from attempts to model human/animal brains Many models, many claims of biological plausibility We will

More information

Deep Neural Networks (1) Hidden layers; Back-propagation

Deep Neural Networks (1) Hidden layers; Back-propagation Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 2 October 2018 http://www.inf.ed.ac.u/teaching/courses/mlp/ MLP Lecture 3 / 2 October 2018 Deep

More information