X/94 $ IEEE 1894

Similar documents
Radial Basis Function Networks. Ravi Kaushik Project 1 CSC Neural Networks and Pattern Recognition

Notes on Regularization and Robust Estimation Psych 267/CS 348D/EE 365 Prof. David J. Heeger September 15, 1998

J. Lazzaro, S. Ryckebusch, M.A. Mahowald, and C. A. Mead California Institute of Technology Pasadena, CA 91125

CHAPTER IX Radial Basis Function Networks

In the Name of God. Lectures 15&16: Radial Basis Function Networks

A Parallel Gradient Descent Method for Learning in Analog VLSI Neural Networks

Plan of Class 4. Radial Basis Functions with moving centers. Projection Pursuit Regression and ridge. Principal Component Analysis: basic ideas

Effects of Moving the Centers in an RBF Network

Application of a radial basis function neural network for diagnosis of diabetes mellitus

IN neural-network training, the most well-known online

E techniques for identification of linear time-invariant

ARTIFICIAL INTELLIGENCE LABORATORY. and CENTER FOR BIOLOGICAL INFORMATION PROCESSING. A.I. Memo No August Federico Girosi.

A Robust PCA by LMSER Learning with Iterative Error. Bai-ling Zhang Irwin King Lei Xu.

Introduction to Neural Networks: Structure and Training

Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems

The Power of Approximating: a Comparison of Activation Functions

Chapter 8: Generalization and Function Approximation

Frequency Selective Surface Design Based on Iterative Inversion of Neural Networks

Introduction to Machine Learning Spring 2018 Note Neural Networks

CSE 190: Reinforcement Learning: An Introduction. Chapter 8: Generalization and Function Approximation. Pop Quiz: What Function Are We Approximating?

Generalization and Function Approximation

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

Computation of Smooth Optical Flow in a Feedback Connected Analog Network

COMS 4771 Introduction to Machine Learning. Nakul Verma

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation

Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso

Neural networks. Chapter 20, Section 5 1

Implementation of Carry Look-Ahead in Domino Logic

When can Deep Networks avoid the curse of dimensionality and other theoretical puzzles

CHAPTER 2 NEW RADIAL BASIS NEURAL NETWORKS AND THEIR APPLICATION IN A LARGE-SCALE HANDWRITTEN DIGIT RECOGNITION PROBLEM

Old painting digital color restoration

An Analog-digital CMOS circuit for motion detection based on direction-selective neural networks

Neural networks. Chapter 19, Sections 1 5 1

Second-order Learning Algorithm with Squared Penalty Term

Adaptive Control of a Class of Nonlinear Systems with Nonlinearly Parameterized Fuzzy Approximators

An Error-Entropy Minimization Algorithm for Supervised Training of Nonlinear Adaptive Systems

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

CS599 Lecture 2 Function Approximation in RL

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

Gaussian Processes for Regression. Carl Edward Rasmussen. Department of Computer Science. Toronto, ONT, M5S 1A4, Canada.

From CDF to PDF A Density Estimation Method for High Dimensional Data

Σ N (d i,p z i,p ) 2 (1)

NEAR-OPTIMAL FLIGHT LOAD SYNTHESIS USING NEURAL NETS

Constructing Transportable Behavioural Models for Nonlinear Electronic Devices

NN V: The generalized delta learning rule

Advanced statistical methods for data analysis Lecture 2

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

18.6 Regression and Classification with Linear Models

PATTERN CLASSIFICATION

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

A Bayesian Local Linear Wavelet Neural Network

Automatic Noise Recognition Based on Neural Network Using LPC and MFCC Feature Parameters

Artificial Neural Network Simulation of Battery Performance

Neural Control for Rolling Mills: Incorporating Domain Theories to Overcome Data Deficiency

On Learning µ-perceptron Networks with Binary Weights

CSE 190: Reinforcement Learning: An Introduction. Chapter 8: Generalization and Function Approximation. Pop Quiz: What Function Are We Approximating?

Multilayer Perceptron Learning Utilizing Singular Regions and Search Pruning

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Multilayer Neural Networks

Ch.6 Deep Feedforward Networks (2/3)

Radial-Basis Function Networks. Radial-Basis Function Networks

Optimal transfer function neural networks

Artificial Neural Networks. Edward Gatt

Feed-forward Network Functions

Model of a Biological Neuron as a Temporal Neural Network

INTERNATIONAL COMPUTER SCIENCE INSTITUTE

Comparing linear and non-linear transformation of speech

Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks

A Fast Stochastic Error-Descent Algorithm for Supervised Learning and Optimization

Radial Basis Functions for Process Control Lyle H. Ungar Tom Johnson Richard D. De Veaux University ofpennsylvania Voice Processing Corp. Princeton Un

Multilayer Feedforward Networks. Berlin Chen, 2002

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Modification of Hard-Limiting Multilayer Neural Networks for Confidence Evaluation

Neural Networks Lecture 4: Radial Bases Function Networks

ENTROPIC ANALYSIS AND INCREMENTAL SYNTHESIS OF MULTILAYERED FEEDFORWARD NEURAL NETWORKS

Deep Learning. Convolutional Neural Network (CNNs) Ali Ghodsi. October 30, Slides are partially based on Book in preparation, Deep Learning

Multi-Layer Boosting for Pattern Recognition

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Lecture 13 - Handling Nonlinearity

Artificial Neural Networks 2

Different Criteria for Active Learning in Neural Networks: A Comparative Study

Relating Real-Time Backpropagation and. Backpropagation-Through-Time: An Application of Flow Graph. Interreciprocity.

Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

A Course in Machine Learning

Iterative Construction of Sparse Polynomial Approximations

Neural networks. Chapter 20. Chapter 20 1

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1

LECTURE NOTE #NEW 6 PROF. ALAN YUILLE

An Adaptive Neural Network Scheme for Radar Rainfall Estimation from WSR-88D Observations

10 NEURAL NETWORKS Bio-inspired Multi-Layer Networks. Learning Objectives:

Upper Bounds on the Number of Hidden Neurons in Feedforward Networks with Arbitrary Bounded Nonlinear Activation Functions

Abstract. In this paper we propose recurrent neural networks with feedback into the input

On Inverse Sigmoid Functions

Introduction to Machine Learning

Neural Network Weight Space Symmetries Can Speed up Genetic Learning

On Optimal Physical Synthesis of Sleep Transistors

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

Transcription:

I Implementing Radial Basis Functions Using Bump-Resistor Networks John G. Harris University of Florida EE Dept., 436 CSE Bldg 42 Gainesville, FL 3261 1 harris@j upit er.ee.ufl.edu Abstract- Radial Basis Function (RBF) networks provide a powerful learning architecture for neural networks [6]. We have implemented a RBF network in analog VLSI using the concept of bump-resistors. A bump-resistor is a nonlinear resistor whose conductance is a Gaussian-like function of the difference of two other voltages. The width of the Gaussian basis functions may be continuously varied so that the aggregate interpolating function varies from a nearest-neighbor lookup, piecewise constant function to a globally smooth function. The bump-resis tor methodology extends to arbitrary dimensions while still preserving the radiality of the basis functions. The feedforward network architecture needs no additional circuitry other than voltage sources and the 1D bump-resistors. A nine-transistor variation of the Delbruck bump circuit is used to compute the Gaussian-like basis functions [2]. Below threshold the current output fits a Gaussian extremely well, see Figure 1. Figure 3 shows that the shape of the function deviates from the Gaussian shape above threshold. The width of the bump can to be varied by almost an order of magnitude (see Figure 4). The Delbruck bump circuit is shown in Figure A follower aggregation network shown in Figure 5 computes an average of the inputs voltages ci weighted by conductance values gi [4]: The bump current is used to control the conductance of the resistors in Figure 5 such that gi = G(x - ti) where ti is the voltage representing the center location of each bump-resistor. The circuit now computes: This normalized RBF or partition of unity form has been used by Moody and Darken in learning simula- BT Gaussian C iit 1 4 : : : : : : : : : 1 43 0.4 4.3 a1 4.1 0.0 0.1 0.2 03 0.4 OS Y-TO Figure 1: Measured current from the Delbruck bump circuit (circles) and Gaussian fit (straightline). tions [5]. Some researchers have claimed that RBF networks show improved performance using this formulation [SI. Anderson, Platt and Kirk previously demonstrated the use of follower aggregation in an analog RBF chip [l]. The standard RBF form (without normalization) can be computed by holding the output to virtual ground and measuring the current. A one-dimensional eight-node circuit has been successfully fabricated and tested. Results from fitting 4 points using two different values of U are shown in Figure 6. The left plot shows the small U response which approximates a nearest neighbor lookup and the right plot shows the much smoother result for a large U. There are several strategies for extending this network to multiple dimensions. One is simply to construct a multi-dimensional bump function to use as the conductance function G(x) in Figure 5. For example, Kirk has designed a suitable multidimensional bump function by multiplying together several 1D Delbruck bump functions [3]. An alternate strategy used by Anderson, Platt and Kirk [l] 0-7803-1901-X/94 $4.00 01994 IEEE 1894

2 1 1 4 : : : : : : : : : I 110 11 21 23 1.) U 26 21 21 19 3.0 X M 2 1 1 4 : : : : : : : ; : I 1.0 11 11 13 2.b U 16 21 18 29 3.0 X M Figure 6: Measured chip curva from fitting four data points Figure 7: Two-dimensional network 1895

o.m 4 I 0.4 0.6 Od la 1.2 1.4 1.6 1.8 20 W O Figure 4: Plot of the measured standard deviation of the bump function vs. voltage control knob. Figure 5: Follower aggregation network 1896

Vb = 0.8 1.2 1.6 2.0 X-T (V) Figure 3: Normalized current from the bump circuit (circles) fit to Gaussians with identical variance. without worry of any problems with local minima. Learning the center locations (ti) and the widths of the Gaussians (ui) is a more difficult problem. REFERENCES [l] J. Anderson, J. C. Platt, and D. Kirk. An analog VLSI chip for radial basis functions. In J. Hanson, J. Cowan, and C. L. Giles, editors, Neural Information Processing Systems, pages 765-772. Morgan-Kaufman, Palo Alto, 1993. [2] T. Delbruck. Bump circuits for computing similarity and disimilarity of analog voltages. In International Joint Conference on Neural Networks, Seattle, WA., July 1991. [3] D. Kirk, D. Kerns, K. Fleischer, and A. Barr. Analog VLSI implementation of multi-dimensional gradient descent. In J. Hanson, J. Cowan, and C. L. Giles, editors, Neural Information Processing Systems, pages 765-772. Morgan-Kaufman, Palo Alto, 1993. [4] C. Mead. Analog VLSI and Neural Systems. Addison-Wesley, 1989. [5] J. Moody and C. Darken. Fast learning in net- works of locally-tuned processing units. Neural Computation, 1(2):281-294, 1989. [6] T. Poggio and F. Girosi. Regularization algorithms for learning that are equivalent to multilayer networks. Science, 247:978-982, 1990. [7] D. Scott. Multivariate Density Estimation. Wiley, 1992. [8] H.W. Werntges. Partitions of unity improve neural function approximators. In Proc. IEEE Intl. Conf. on Neural Networks, pages 914-918, San Francisco, CA,, Feb 1993. vol 2,. 1897

Figure 2: The Delbruck bump circuit. The voltage VB controls the width of the bump. is to use current summation to sum the squares of the voltage differences. A circuit must then be used to exponentiate the current and convert the result to a voltage. Rather than build more complex circuitry, we choose to use the physics of resistors in series to combine the dimensions. Figure 7 shows the the network in two dimensions (extension to further dimensions is straightforward). In two dimensions, each branch has two resistors with conductances of G(x) and G(y). The effective conductance H(z,y) is: (3) Typically we choose G(x) to be a Gaussian function since the Gaussian is the only function that can create a radial function by multiplying 1D functions. However, in this case the Gaussian is the wrong choice since H(x,y) will be nonradial. Instead, we choose G with the following form: Now the series combination is: 1 G(x) = (4) 1 + (x/c)2 which is radial. In many applications, radial functions are not strictly necessary. In fact, the density estimation literature reveals that some nonradial functions have been shown to be optimal under certain criteria [7]. We are also studying resistors that are controlled by sigmoidal functions. A feedforward RBF network is sufficient for many applications. A microprocessor can be used to implement a learning network and download the weights to the chip. On-chip learning will drastically speed up the learning process. Learning the coefficients ci is a straightforward process that can be performed with gradient descent 1898