Sample Exam COMP 9444 NEURAL NETWORKS Solutions

Similar documents
Artificial Neural Networks Examination, March 2004

Artificial Neural Networks Examination, June 2005

Artificial Neural Networks Examination, June 2004

Introduction to Neural Networks

Learning Vector Quantization

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Learning Vector Quantization (LVQ)

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

CHALMERS, GÖTEBORGS UNIVERSITET. EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Artificial Neural Networks Examination, March 2002

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

CSC Neural Networks. Perceptron Learning Rule

Christian Mohr

Artificial Intelligence Hopfield Networks

Lecture 5: Recurrent Neural Networks

LINEAR ALGEBRA KNOWLEDGE SURVEY

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

CSC321 Lecture 20: Autoencoders

Using a Hopfield Network: A Nuts and Bolts Approach

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Neural networks. Chapter 19, Sections 1 5 1

EEE 241: Linear Systems

Recurrent and Recursive Networks

Machine Learning. Neural Networks

Lecture 5: Logistic Regression. Neural Networks

Introduction to Neural Networks: Structure and Training

Geometric View of Machine Learning Nearest Neighbor Classification. Slides adapted from Prof. Carpuat

Information Dynamics Foundations and Applications

Neural Networks DWML, /25

Strength-based and time-based supervised networks: Theory and comparisons. Denis Cousineau Vision, Integration, Cognition lab. Université de Montréal

Machine Learning Basics

Fall 2016 MATH*1160 Final Exam

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

T Machine Learning and Neural Networks

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

PV021: Neural networks. Tomáš Brázdil

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

A gradient descent rule for spiking neurons emitting multiple spikes

Artificial Neural Network and Fuzzy Logic

Neural networks. Chapter 20. Chapter 20 1

Artificial Neural Network

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

GCSE MATHEMATICS 8300/2H PRACTICE PAPER SET 4. Exam Date Morning Time allowed: 1 hour 30 minutes. Please write clearly, in block capitals.

Integer weight training by differential evolution algorithms

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

CS534 Machine Learning - Spring Final Exam

2- AUTOASSOCIATIVE NET - The feedforward autoassociative net considered in this section is a special case of the heteroassociative net.

Recurrent Neural Networks

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

CSC321 Lecture 4 The Perceptron Algorithm

Neural networks. Chapter 20, Section 5 1

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

Introduction to Artificial Neural Networks

Lecture 7 Artificial neural networks: Supervised learning

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A.

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

9 Competitive Neural Networks

In the Name of God. Lectures 15&16: Radial Basis Function Networks

Hopfield Network Recurrent Netorks

Neural Networks and Machine Learning research at the Laboratory of Computer and Information Science, Helsinki University of Technology

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

Deep Feedforward Networks. Sargur N. Srihari

Simple Neural Nets for Pattern Classification: McCulloch-Pitts Threshold Logic CS 5870

Introduction to Machine Learning Midterm Exam

Machine Learning Techniques

Part-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287

Radial-Basis Function Networks

Learning and Memory in Neural Networks

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Machine Learning for Physicists Lecture 1

CSCI 315: Artificial Intelligence through Deep Learning

Final Exam. 1 True or False (15 Points)

CS 730/730W/830: Intro AI

Neuroscience Introduction

CMSC 422 Introduction to Machine Learning Lecture 4 Geometry and Nearest Neighbors. Furong Huang /

Symmetry Adapted Orbitals

ECE521 Lectures 9 Fully Connected Neural Networks

Instruction Sheet for SOFT COMPUTING LABORATORY (EE 753/1)

(i) The optimisation problem solved is 1 min

Machine Learning (CS 567) Lecture 2

Jakub Hajic Artificial Intelligence Seminar I

Lecture 15: Exploding and Vanishing Gradients

Associative Neural Networks using Matlab

Course Structure. Psychology 452 Week 12: Deep Learning. Chapter 8 Discussion. Part I: Deep Learning: What and Why? Rufus. Rufus Processed By Fetch

Chapter 15. Dynamically Driven Recurrent Networks

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

Neural Networks Task Sheet 2. Due date: May

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Machine Learning (CS 567) Lecture 3

Machine Learning for Computer Vision 8. Neural Networks and Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

A-level PHYSICS A PHYA5/2A. Unit 5A Astrophysics. Section B. Tuesday 28 June 2016

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

The wake-sleep algorithm for unsupervised neural networks

UNSUPERVISED LEARNING

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)

CSCI 252: Neural Networks and Graphical Models. Fall Term 2016 Prof. Levy. Architecture #7: The Simple Recurrent Network (Elman 1990)

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Transcription:

FAMILY NAME OTHER NAMES STUDENT ID SIGNATURE Sample Exam COMP 9444 NEURAL NETWORKS Solutions (1) TIME ALLOWED 3 HOURS (2) TOTAL NUMBER OF QUESTIONS 12 (3) STUDENTS SHOULD ANSWER ALL QUESTIONS (4) QUESTIONS ARE NOT OF EQUAL VALUE (5) THIS PAPER MAY NOT BE RETAINED BY THE CANDIDATE (6) THERE ARE 180 MARKS ON THE PAPER. CANDIDATES ARE ADVISED TO ALLOCATE THEIR TIME AT ONE MINUTE PER MARK. PLEASE ANSWER PART A IN YOUR EXAMINATION BOOKLET PLEASE ANSWER PART B IN THE SPACES ON THIS QUESTION PAPER ANSWERS MUST BE WRITTEN CLEARLY IN INK. EXCEPT WHERE THEY ARE EXPRESSLY REQUIRED, PENCILS MAY ONLY BE USED FOR DRAWING, SKETCHING OR GRAPHICAL WORK.

(Ask Achim) PART A - Answer in your exam booklet.

PART B - Answer in the spaces on this question paper. Question 8 Temporal Processing (18 marks) (a) With a supervised learning algorithm, we can specify target output values, but we may never get close to those targets at the end of learning. Give two reasons why this might happen. (i) (ii) data may be valid, and inconsistency results from a stochastic aspect of the task (or some aspect of the task is not modelled by the input data collected); the data may contain errors - e.g. measurement errors or typographical errors 8(b) Describe the architecture and the computational task of the NetTalk neural network. architecture: 80 hidden neurons 203 input units (= 7*29)...!...!... one output neuron per English phoneme, + 1 NETtalk architecture Each group of 29 input units represents a letter, so inputs together represent seven letters computational task is to output the representation of the phoneme corresponding to the middle letter of the seven 8(c) Why does a time-delay neural network (TDNN) have the same set of incoming weights for each column of hidden units? To provide temporal translation invariance. or So that the TDNN will be able to identify the input sound, no matter which frame the input sound begins in. 8(d) Distinguish between a feedforward network and a recurrent network. A feedforward network has no cyclic activation flows.

Question 9 Elman and Jordan nets, etc. (18 marks) (a) Draw the weight matrix for a feedforward network, showing the partitioning. You can assume that the weight matrix for connections from the input layer to the hidden layer is W ih, and that the weight matrix for connections from the hidden layer to the output layer is W ho. ho ih ih ho 9(b) In a Jordan network with i input neurons, h hidden layer neurons, and o output neurons: (a) how many neurons will there be in the state vector, and (b) if i = 4, h = 3, and o = 2, draw a diagram showing the connectivity of the network. Do not forget the bias unit. (a) o neurons in state vector (same as output vector thatʼs letter o, not zero) (b) outputs bias 1;? hiddens inputs state 9(c) Draw a diagram illustrating the architecture of Elman s simple recurrent network that performs a temporal version of the XOR task. How are the two inputs to XOR provided to this network? 5 3 4 0 1 2 The inputs are passed sequentially to the single input unit (0) of the temporal XOR net. 9(d) Briefly describe the use of cluster analysis in Elman s lexical class discovery experiments, and one of his conclusions from this. Elman clustered hidden unit activation patterns corresponding to different input vectors and different sequences of input units. He found that the clusters corresponded well to the grammatical contexts in which the inputs (or input sequences) occurred, and thus concluded that the network had in effect learned the grammar.

Question 10 Tensor Product Networks (18 marks) (a) Draw an architectural diagram of a rank 2 tensor product network where the dimensions of the input/output vectors are 3 and 4. You do not need to show the detailed internal structure of the binding units. Binding Units Units representing outputs (fillers/ values) 10(b) Units representing inputs (variables/roles) Draw a diagram of a single binding unit in a rank 2 tensor product network illustrating the internal operation of the binding unit in teaching mode. activation memory * + 10(c) Define the concepts of dense and sparse random representations. How do their properties compare with those of an orthonormal set of representation vectors. In a dense random representation, each vector component is chosen at random from a uniform distribution over say [ 1, +1]. In a sparse random representation, the non-zero components are chosen in this way, but most components are chosen (at random) to be zero. In both cases, the vectors are normalised so that they have length 1. Members of orthonormal sets of vectors have length one, and are orthogonal to one another. Vectors in dense and sparse random representations are orthogonal on average their inner products have a mean of zero. 10(d) What is a Hadamard matrix? Describe how a Hadamard matrix can be used to produce suitable distributed concept representation vectors for a tensor product network. What are the properties of the Hadamard matrix that makes the associated vectors suitable? A Hadamard matrix H is a square matrix of side n, all of whose entries are ±1, which satisfies HH T = I n the identity matrix of side n. The rows of a Hadamard matrix, once normalised, can be used as distributed representation vectors in a tensor product network. This is because the rows are orthogonal to each other, and have no zero-components.

Question 11 - Self-Organising Systems (18 marks) (a) In a 2-D self-organising map with input vectors of dimension m, and k neurons in the map, how many weights will there be? mk 11(b) Describe the competitive process of the Self-Organising Map algorithm. Let m denote the dimension of the input pattern x = [x 1, x 2,..., x m ] T The weight vector for each of the neurons in the SOM also has dimension m. So for neuron j, the weight vector will be: w j = [w j1, w j2,..., w jm ] T For an input pattern x, compute the inner product w j x for each neuron, and choose the largest inner product. Let i(x) denote the index of the winning neuron (and also the output of a trained SOM). 11(c) Briefly explain the concept of a Voronoi cell. Given a set of vectors X, the Voronoi cells about those vectors are the ones that partition the space they lie in, according to the nearest-neighbour rule. That is, the Voronoi cell that a vector lies in is that belonging to the x X to which it is closest. 11(d) Briefly explain the term code book in the context of learning vector quantisation. When compressing data by representing vectors by the labels of a relatively small set of reconstruction vectors, the set of reconstruction vectors is called the code book. 11(e) Describe the relationship between the Self-Organising Map algorithm, and the Learning Vector Quantisation algorithm. In order to use Learning Vector Quantisation (LVQ), a set of approximate reconstruction vectors is first found using the unsupervised SOM algorithm. The supervised LVQ algorithm is then used to fine-tune the vectors found using SOM.

Question 12 Hopfield Nets, BSB (18 marks) (a) Write down the energy function of a discrete Hopfield net. 12(b) E = "0.5# N i=1 # N j=1 w ji x j x i Compute the weight matrix for a 4-neuron Hopfield net with the single fundamental memory ξ 1 = [1, 1, 1,1] stored in it. $ 0 #1 #1 1 ' & ) W = " T #1 0 1 #1 1 " 1 # I 4 = & ) &#1 1 0 #1) & ) % 1 #1 #1 0 ( 12(c) Briefly describe two types of attractor in a dynamical system. An attractor is a bounded subset of space to which non-trivial regions of initial conditions converge at time passes. Pick two of point attractor: system converges to a single point limit cycle: system converges to a cyclic path chaotic attractor: stays within a bounded region of space, but no predictable cyclic path 12(d) Write down the energy function of a BSB network with weight matrix W, feedback constant β, and activation vector x. or $ E = " # ' & )* w % 2 ( ij x i x j ij $ E = " # ' & ) x T Wx % 2 (