RE-ENGINEERING COMPUTING WITH NEURO- MIMETIC DEVICES, CIRCUITS, AND ALGORITHMS

Similar documents
NEUROMORPHIC COMPUTING WITH MAGNETO-METALLIC NEURONS & SYNAPSES: PROSPECTS AND PERSPECTIVES

Center for Spintronic Materials, Interfaces, and Novel Architectures. Spintronics Enabled Efficient Neuromorphic Computing: Prospects and Perspectives

Addressing Challenges in Neuromorphic Computing with Memristive Synapses

Neuromorphic computing with Memristive devices. NCM group

Synaptic Devices and Neuron Circuits for Neuron-Inspired NanoElectronics

Magnetic tunnel junction beyond memory from logic to neuromorphic computing WANJUN PARK DEPT. OF ELECTRONIC ENGINEERING, HANYANG UNIVERSITY

Artificial Neural Network

Introduction. Previous work has shown that AER can also be used to construct largescale networks with arbitrary, configurable synaptic connectivity.

CMSC 421: Neural Computation. Applications of Neural Networks

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Neural Networks Introduction

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Network and Fuzzy Logic

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Data Mining Part 5. Prediction

STDP Learning of Image Patches with Convolutional Spiking Neural Networks

Post Von Neumann Computing

CISC 3250 Systems Neuroscience

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

An Overview of Spin-based Integrated Circuits

An Introductory Course in Computational Neuroscience

A Technology-Agnostic MTJ SPICE Model with User-Defined Dimensions for STT-MRAM Scalability Studies

The N3XT Technology for. Brain-Inspired Computing

Lecture 7 Artificial neural networks: Supervised learning

A 68 Parallel Row Access Neuromorphic Core with 22K Multi-Level Synapses Based on Logic- Compatible Embedded Flash Memory Technology

Using a Hopfield Network: A Nuts and Bolts Approach

COMP-4360 Machine Learning Neural Networks

Lecture 4: Feed Forward Neural Networks

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Memory and computing beyond CMOS

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Artificial Intelligence

Part 8: Neural Networks

EEE 241: Linear Systems

Consider the following spike trains from two different neurons N1 and N2:

Proposal for an All-Spin Artificial Neural Network: Emulating Neural and Synaptic Functionalities Through Domain Wall Motion in Ferromagnets

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Event-Driven Random Backpropagation: Enabling Neuromorphic Deep Learning Machines

Implementing Synaptic Plasticity in a VLSI Spiking Neural Network Model

PV021: Neural networks. Tomáš Brázdil

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Introduction and Perceptron Learning

Artifical Neural Networks

Deep Learning with Coherent Nanophotonic Circuits

3 Detector vs. Computer

Neuromorphic Network Based on Carbon Nanotube/Polymer Composites

Computational Models of Human Cognition

Quantum Artificial Intelligence and Machine Learning: The Path to Enterprise Deployments. Randall Correll. +1 (703) Palo Alto, CA

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

Causality and communities in neural networks

How to do backpropagation in a brain

Learning Spatio-Temporally Encoded Pattern Transformations in Structured Spiking Neural Networks 12

Novel VLSI Implementation for Triplet-based Spike-Timing Dependent Plasticity

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Linear Regression, Neural Networks, etc.

Artificial Neural Networks

arxiv: v1 [cs.et] 20 Nov 2014

Instituto Tecnológico y de Estudios Superiores de Occidente Departamento de Electrónica, Sistemas e Informática. Introductory Notes on Neural Networks

New Delhi & Affiliated to VTU, Belgavi ) Oorgaum, Kolar Dist ,Karnataka

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Featured Articles Advanced Research into AI Ising Computer

Neural Network Based Density Measurement

A brain-inspired neuromorphic architecture for robust neural computation

Neuron. Detector Model. Understanding Neural Components in Detector Model. Detector vs. Computer. Detector. Neuron. output. axon

INTRODUCTION TO NEURAL NETWORKS

CS 229 Project Final Report: Reinforcement Learning for Neural Network Architecture Category : Theory & Reinforcement Learning

Spike-based Long Short-Term Memory networks

A gradient descent rule for spiking neurons emitting multiple spikes

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

arxiv: v1 [physics.app-ph] 1 May 2017

Lecture 4: Perceptrons and Multilayer Perceptrons

Deep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017

CN2 1: Introduction. Paul Gribble. Sep 10,

Introduction Biologically Motivated Crude Model Backpropagation

How can ideas from quantum computing improve or speed up neuromorphic models of computation?

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Introduction To Artificial Neural Networks

Probabilistic and Approximate Computing. Dr. Laura Monroe

DEVS Simulation of Spiking Neural Networks

Nature-inspired Analog Computing on Silicon

The Perceptron. Volker Tresp Summer 2014

arxiv: v2 [cs.et] 26 Jan 2018

In the Name of God. Lecture 9: ANN Architectures

Unit 8: Introduction to neural networks. Perceptrons

Processor & SOC Architecture

COMP304 Introduction to Neural Networks based on slides by:

Applications of Memristors in ANNs

Liquid Computing in a Simplified Model of Cortical Layer IV: Learning to Balance a Ball

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Artificial neural networks

Artificial Neural Networks. Part 2

Algorithm/Architecture Co-Design for Low-Power Neuromorphic Computing

arxiv: v3 [cs.lg] 14 Jan 2018

LEADING THE EVOLUTION OF COMPUTE MARK KACHMAREK HPC STRATEGIC PLANNING MANAGER APRIL 17, 2018

This document is an author-formatted work. The definitive version for citation appears as:

Multilayer Perceptron

Lecture 11 : Simple Neuron Models. Dr Eileen Nugent

Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network

Microsystems for Neuroscience and Medicine. Lecture 9

GPGPU ACCELERATED SIMULATION AND PARAMETER TUNING FOR NEUROMORPHIC APPLICATIONS

Transcription:

RE-ENGINEERING COMPUTING WITH NEURO- MIMETIC DEVICES, CIRCUITS, AND ALGORITHMS Kaushik Roy Abhronil Sengupta, Gopal Srinivasan, Aayush Ankit, Priya Panda, Xuanyao Fong, Deliang Fan, Jason Allred School of Electrical & Computer Engineering Purdue University Center for Spintronics, National Science Foundation, DARPA, Vannevar Bush Fellowship, Office of Naval Research, Semiconductor Research Corporation, Intel

HUMAN VS. MACHINE CHRONICLES 2016 2011 1997 IBM Deep Blue vs. Kasparov IBM RS/6000 32-node server (Power2 + 8 dedicated chips) ~15000 W IBM Watson vs. Brad Ritter & Ken Jennings 90 Power 750 Express servers (4 8-core CPUs) ~200000W Google AlphaGo vs. Lee Sedol (1920 CPUs, 280 GPUs) ~300000W 20W

WHAT ARE THE ISSUES WITH DLN? Requires massive amount of training data Learn with much less data Supervised learning Need for higher bio-fidelity Incremental/adaptive learning is difficult catastrophic forgetting Life-long learning Huge power consumption Event driven evaluation can help; Approximate hardware Well-suited for image, speech, text recognition.. Need for cognitive systems to perform larger range of functions not just sensory processing, but also reasoning and decision making Can neuro-mimetic devices help? von-neumann architecture not suitable Need for in-memory-computing, efficient neurons and synapses

Neuromorphic Computing: An Overview Big Data Analytics Optical Character Recognition Medical diagnostics Facial / Emotion recognition Surveillance systems Video analysis Bio-Inspired Computing and Learning Models Deep Learning Spiking Networks Synaptic Plasticity Coupled Oscillators Recurrent Networks Visual-Attention Inspired Low-Power Algorithms/Hardware Leverage from latest development in neuroscience on visual attention to design energy efficient hardware for deep learning neural networks Input Layer Test Instance CNN Layer 1 C1 CNN Layer 2 C2 CNN Layer 3 C3 Output Layer 2 Class label Test Instance Input Layer CNN Layer 2 C2 CNN Layer 1 C1 CNN Layer 3 C3 Output Layer Input (CNN feature vector) Output Layer Linear Classifie r Linear Classifie r Class label Linear Classifier Spatial Attention Networks in the brain Neuroscience studies Classifier X (Complex) Class Label (a) Traditional Approach Model Y identifies all instances that have common semantic information Common semantic between class 1 & 2 2 Ac va on Instance Apply classifier X on all instances Class 2 Stage 2 Stage 1 1 Test Class 1 (a) Baseline DLNN 1 Test Instance Classifier Y (Simple) Conditional Activation Classifier Y performs semantic based elimination Classifier X (Complex) Class Label Activate X for fraction of input instances passed from Y (b) Proposed Approach with Semantically Decomposed Framework Feature selective Semantic based classification elimination Conditional Learning Mapping to design and circuit optimizations Easy Inputs classified..... Hard Inputs classified in early stages in later stages (b) Condi onal Deep Learning Network (CDLN) Neuromorphic System Architecture Exploiting Error-Resiliency Design programmable and scalable hardware fabrics and explore circuit optimizations for achieving high connectivity Multiplier-less Neuron Efficient Synaptic storage Hybrid 8T-6T SRAM Scalable architecture design addressing the issue of high interconnectivity between neural units Neuro-mimetic CMOS & Post-CMOS Device Models CMOS Spiking Neuron Investigate device physics to mimic neuron/ synapse functionalities Spintronic Neurons/Synapses Crossbar Architecture Experiments & Analysis from Neuroscience Explore neuromorphic computing models inspired from hierarchical layer arrangement and spiking nature of brain networks Recommendation Advertising and Self driving business intelligence vehicles Engines STUDENTS/Post-doc: Abhronil Sengupta, Priyadarshini Panda, Syed Shakib, Gopal Srinivasan, Chamika Liyanagendra, Bing Han, Jason Allred, Parami Wijesinghe, Ayush Ankit, Maryam Parsa, Shuhan Chen and others 7

Cognitive Computing Recognition, Inference, Decision Making to Enable Intelligent Autonomous Systems, Life-long Learning, Neuroinspired learning Hardware Accelerators: Image, Speech and Text Recognition Emerging Devices: STT-devices, Resistive RAMs,.. SW (Multicores/GPUs) Datapath Memory

Device/Circuit/Algorithm Co-Design: Spin/ANN Top-Down Perspective Investigate brain-inspired computing models to provide algorithm-level matching to underlying device physics Device-Circuit-Algorithm co-simulation framework used to generate behavioral models for system-level simulations of neuromorphic systems System Level Solution Bottom-Up Perspective Investigate device physics to mimic neuron/ synapse functionalities Calibration of device models with experiments

SPIKING NEURAL NETWORKS

Spiking Neuron Dynamics Postsynaptic neuron membrane potential V given by Leaky-Integrate- Fire model as follows: Ƭ denotes membrane time constant Postsynaptic neuron spikes when membrane potential crosses a certain threshold and gets reset

Spike Timing Dependent Plasticity: Example I 1 I 2 I 3 O 1 Weight Update Equations time Bi and Poo, The Journal of Neuroscience, December 15, 1998. Strength of the synapse should increase (decrease) as post and pre neurons appear to be temporally correlated (uncorrelated)

Artificial Neural Networks: Simple Model Artifical NN synapses w 1 neuron axon w 2 transmitting neuron Signal transmission w n Weighted Summation Thresholding function Cross-bar array of programmable synapses Spin neurons Spintronic neurons operating at ultra-low terminal voltages interfaced with resistive synapses lead to energy-efficient ANNs

DEVICES: NEURONS, SYNAPSES, IN- MEMORY COMPUTING

Building Primitives: Memory, Neurons, Synapses Lateral Spin Valve (Local & Non-local) I Spin current I REF ± I in I REF Integrator V OUT I + = + v i g i Excitory Input Preset I = v i g i Inhibitory Input FM NM L CH m L CH

Spin Transfer Torque Induced Domain Wall Motion Multi-domain magnets consists of a domain wall (DW) separating regions with opposite magnetic polarizations Domain wall can be moved in the direction of electron flow MTJ resistance varies with domain wall position Decoupled write and read current paths Low current induced Ne el domain wall motion can be achieved by spin-orbit torque generated by spin-hall effect from a heavy metal underlayer in presence of DMI Universal device: Suitable for memory, neuron, synapse, interconnects

Core Computing Blocks

SHE Induced DW Motion: Neuron Non-spiking Neuron IF Spiking Neuron Three terminal spintronic device acting as a neuron (with different degrees of bio-fidelity) and synapse The neuron is interfaced with the axon circuit to generate a corresponding analog output current with variation in the input current / Integrate-Fire spiking neuron can be implemented using a similar device structure where the MTJ is located at the extreme edge of the FM. Synapse, acting as the memory element, encodes the weight in its conductance value which is determined by the domain wall position Sengupta, Roy, TBioCAS 16; Sengupta, Roy, TCAS 16

Core Computing Blocks

Self-learning in Spiking Neural Networks Spike-Timing Dependent Plasticity Spintronic synapse in spiking neural networks exhibits spike timing dependent plasticity observed in biological synapses Programming current flowing through heavy metal varies in a similar nature as STDP curve Decoupled spike transmission and programming current paths assist online learning 48fJ energy consumption per synaptic event which is ~10-100x lower in comparison to emerging devices like PCM

Network for Pattern Recognition Pattern recognition performed in a network of excitatory spiking neurons in presence of lateral inhibition and homeostasis

Self-Learning in Spiking Neural Network

Energy Comparison Spintronic Engine & CMOS Baseline Spintronic architecture achieves energy improvements of 204X 2759X with respect to SNeuE (CMOS baseline) across all benchmarks. Spintronic crossbar enabled in-memory processing enables to overcome the memory domination/bottlenecks in the CMOS engine. Spintronic neurons interfaced with spintronic crossbar (SCA) allows energy-efficient innerproduct computations.

STOCHASTIC NEURAL NETWORKS: SHE BASED MTJ

Spiking Neuron Membrane Potential Biological Spiking Neuron MTJ Spiking Neuron LIF Equation: LLGS Equation: The leaky fire and integrate can be approximated by an MTJ the magnetization dynamics mimics the leaky fire and integrate operation Sengupta, Roy, Scientific Reports 16

Interconnected Crossbars for NNs

SHE Based MTJ: Stochastic Neuron & Synapse Exploit the stochastic switching behavior of a mono-domain MTJ in the presence of thermal noise sigmoidal function Stochastic sigmoidal neuron Replaces multi-bit synapses with a stochastic single bit (binary) synapse.

SHE-Based Switching: Experiments V+ y x I+ I- V- H L Field 100 Oe in-plane field Anomalous Hall effect (AHE) in FM layer: Hall resistances change abruptly when the magnetization switches in FM. Ta(5nm)/CoFeB(1.3nm)/MgO(1.5nm)/Ta(5 nm) (bottom to top) Hall bar structure; CoFeB shows PMA. In-plane field H L applied A current pulse group with changeable magnitude is sent through I+/-. Then resistance change is measured across V+/-

Stochastic Binary Synapse Synaptic strength is updated based on the temporal correlation between pre- and post-spike trains Synaptic learning is embedded in the switching probability of binary synapses Switch the MTJ based on spike timing by passing the required write current

Spiking Neuromorphic Architecture with LT- ST Synapses Crossbar arrangement of the LT-ST synapses with CMOS neurons The significant LT synapses are driven by a higher read voltage Prog. Energy = 10uJ Prog. Energy = 23uJ Network of LT-ST synapses provides 5% improvement in the classification accuracy over one-bit synapses Under iso-accuracy conditions, the LT-ST synapses offer a 2X reduction in the synaptic write energy

From Devices to Circuits and Systems C-SPIN Annual Review, September 27-28, 2017 42

Simulation Framework

Conclusions Other than memory, STT devices also show promise for a class of computing models such as brain-inspired computing, stochastic computing STT-devices as neurons and synapses for both ANN, SNN show the possibility of large improvement in energy compared to CMOS implementation