Biologically Inspired Compu4ng: Neural Computa4on. Lecture 5. Patricia A. Vargas
|
|
- Peter Sharp
- 6 years ago
- Views:
Transcription
1 Biologically Inspired Compu4ng: Neural Computa4on Lecture 5 Patricia A. Vargas
2 Lecture 5 I. Lecture 4 Revision II. Ar4ficial Neural Networks (Part IV) I. Recurrent Ar4ficial Networks I. GasNet models II. Evolving Ar4ficial Neural Networks I. MLPs II. GasNet models F21BC2 BIC Neural Computa4on
3 Ar4ficial Neural Networks Recurrent Neural Networks The Hopfield Equations Hopfield Neural Network Training performed in one pass: where, w 1 n k k ij =! pi p j N k = 1 w ij is the weight between nodes i & j " N # N is the number of nodes in the network! si = sign $ wij s j % & j= 1 ' n is the number of patterns to be learnt p ik is the value required for the i-th node in pattern k totalmente realimentada (exceto auto-realim Pattern Completion (III) 64 pixel image of an H Same image with 10 pixels altered (I.e. approximately 16% noise added) Execution performed iteratively: " N # s = sign $! w s % Pattern Completion (IV) Food for thought - flight of fancy? Memories of a deceased dog named Tanya...
4 4
5 5
6 6
7 7
8 8
9 9
10 Gas dispersion NOT centred on the node 10
11 Gas Concentration around emitter 11
12 12
13 Nodes do NOT have a spatial relation? [0,1] is called the Mbias or Modulator Bias 13
14 Ar4ficial Neural Networks I. Evolving Ar4ficial Neural Networks I. MLPs I. Topology and weights II. Example: Evolving (training) MLPs to learn some func4ons <Toolbox> from Cazangi & Moioli (2005) VISION x 0 x 1 y 1 x 2 y 2 y o x m Camada de
15 Ar4ficial Neural Networks I. Evolving Ar4ficial Neural Networks I. GasNet models I. Topology + all network parameters + task dependent parameters < genotype >:: (< genes >) < gene >=< node >
16 I. Evolving GasNets models Original de variable in the phenotype space. Node variables Description [Range]DiscreteValues Gene locus Description Coordinates Node coordinates on the [0, 99] < x > < x value > Euclidean plane (100x100) [0, 99] < y > < y value > Electrical Defines the parameters of the [0, 50] <r p > < radius > connectivity two segments of circle centred <r p > on the node that will determine [0, 2π] < θ e > < angular extent > the excitatory and inhibitory < θ e > links < θ p > < orientation > < θ p > Recurrence Determines whether the node {-1,0,1} < rec > < recurrent status > status has an inhibitory, none or excitatory recurrent connection Emitting Determines the circumstances {0,1,2} <E s > < emitting status > status under which the node will emit gas {none, electrical, gas} Type of gas Determines which gas {1, 2} <G t > < gas type > the node will emit Rate of Determines the rate of gas [1, 11] < s > < build up/ build up/ build up and decay decay rate > decay Radius of Maximum radius of [10%,60%]* <G r > < gas radius > emission gas emission * of plane dimension 100x100 Transfer Used in (2) to determine the [1, 11] <K 0 > < transfer function function transfer parameter value K n i default value > parameter default value K 0 i Bias The b i term? on 1 [ 1.0, 1.0] <b> < bias value > Task Parameters which depend on? <? >? parameters the task, e.g. a robot vision sensors input area ([1])
17 I. phenotype Evolving GasNets models NSGasNet space. Node Description [Range] Gene Description variables DiscreteValues locus Recurrence Determines whether the node {-1,0,1} < rec > < recurrent status > status has an inhibitory, none or excitatory recurrent connection Emitting Determines the circumstances {0,1,2} <E s > < emitting status > status under which the node will emit gas {none, electrical, gas} Type of gas Determines which gas {1, 2} <G t > < gas type > the node will emit Rate of Determines the rate of gas [1, 11] < s > < build up/ build up/ build up and decay decay rate > decay Transfer Used in (2) to determine the [1, 11] <K 0 > < transfer function function transfer parameter value K n i default value > parameter default value K 0 i Bias The term b i on 1 [ 1.0, 1.0] <b> < bias value > Task Parameters which depend on? <? >? parameters the task, e.g. a robot vision sensors input area ([1]) Modulator Parameter which depend on [0, 1] < Mbias n > < Modulator bias Bias the number of nodes, for node n > (Mbias n ) i.e. there will be as many Mbias as the number of network nodes
18 1) Initialise population of 100 genotypes in 10x10 grid 2) Evaluate initial genotype fitnesses 3) Repeat until termination criteria met: Repeat 100 times for one generation: Select genotype at random Create mating pool of genotype plus eight nearest grid neighbours One parent P selected proportional to fitness rank in mating pool Offspring O mutated Offspring O evaluated Place O in mating pool, replacing genotype selected proportional to inverse of fitness rank in mating pool The steady-state genetic algorithm works as 18
19 19
20 Original X NSGasNet 20
21 Evolu4on of: All nodes parameters 21
22 50 runs 200 evalua-ons per genera-on lutionary regime parameters employed on Parameter CPG task Mutation rate 8% Fitness function F itness CP G = ( C0 n0 )+( C1 n1 ) 2 Number of runs 50 Maximum number 1000 of generations Population size 100 Genotype size 42 (Original) 46 (NSGasNet) Trials 1 Number of evaluations 200 per trial 22
23 23
24 24
25 25
26 26
27 Evolu4on of: Number of nodes All nodes parameters 27
28 28
29 29
30 30
31 31
32 32
33 40 runs 10 trials per fitness evalua-on 33
34 34
35 2d simulation 3d simulation 35
36 3d simulation 36
37 Lecture 5 References: Husbands,P., Smith, T., Jakobi, N. and O Shea, M. (1998). Beier living through chemistry: Evolving GasNets for robot control, Connec4on Science, vol. 10(3 4), pp , Philippides, P. Husbands, T. Smith, and M. O Shea (1999) Flexible couplings: Diffusing neuromodulators and adap4ve robo4cs, Ar4ficial Life, vol. 11, pp , 2005.ciples and individual variability, Journal of Computa4onal Neuroscience, vol. 7, pp Smith, T. M. S. (2002). The Evolvability of Ar4ficial Neural Net works for Robot Control. University of Sussex, Brighton, England, UK, 2002, PhD thesis. Vargas, P.A., Di Paolo, E. A. & Husbands, P. (2007). "Preliminary Inves4ga4ons on the Evolvability of a Non spa4al GasNet Model". In Proceedings of the 9th European Conference on Ar4ficial life, ECAL'2007, Springer Verlag, Lisbon, Portugal, September 2007, pp: Vargas, P.A., Di Paolo, E. A., & Husbands, P. (2008). "A study of GasNet spa4al embedding in a delayed response task". In the Proc. of ALIFE XI'2008, England, UK, pp: Husbands, P., Philippides, A., Vargas, P. A., Buckley, C, Fine, P., Di Paolo, E. & O Shea, M. (2010). Spa4al, Temporal and Modulatory Factors affec4ng GasNet Evolvability, to appear in the Complexity Journal, Wiley Periodicals, Inc. (invited paper). F21BC2 BIC Neural Computa4on
38 Lecture 5 I. Lecture 4 Revision II. Ar4ficial Neural Networks (Part IV) I. Recurrent Neural Networks I. GasNet Models II. Evolving ANN I. MLP II. GasNet models F21BC2 BIC Neural Computa4on
39 Lecture 6 What s next? Ar4ficial Neural Networks (Part V) F21BC2 BIC Neural Computa4on
40 hip:// (please note that this link only works in Mozilla Firefox or Safari web VISION F21BC2 BIC Neural Computa4on
Artificial Neural Networks Examination, March 2004
Artificial Neural Networks Examination, March 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationFlexible Couplings: Diffusing Neuromodulators and Adaptive Robotics
: Diffusing Neuromodulators and Adaptive Robotics Abstract Recent years have seen the discovery of freely diffusing gaseous neurotransmitters, such as nitric oxide (NO), in biological nervous systems.
More informationChapter 1 Introduction
Chapter 1 Introduction 1.1 Introduction to Chapter This chapter starts by describing the problems addressed by the project. The aims and objectives of the research are outlined and novel ideas discovered
More informationEvolutionary computation
Evolutionary computation Andrea Roli andrea.roli@unibo.it DEIS Alma Mater Studiorum Università di Bologna Evolutionary computation p. 1 Evolutionary Computation Evolutionary computation p. 2 Evolutionary
More informationArtificial Neural Networks Examination, June 2005
Artificial Neural Networks Examination, June 2005 Instructions There are SIXTY questions. (The pass mark is 30 out of 60). For each question, please select a maximum of ONE of the given answers (either
More informationEvolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)
Evolutionary Computation DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) andrea.roli@unibo.it Evolutionary Computation Inspiring principle: theory of natural selection Species face
More informationArtificial Neural Networks Examination, June 2004
Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationArtificial Neural Networks Examination, March 2002
Artificial Neural Networks Examination, March 2002 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationUnicycling Helps Your French: Spontaneous Recovery of Associations by. Learning Unrelated Tasks
Unicycling Helps Your French: Spontaneous Recovery of Associations by Learning Unrelated Tasks Inman Harvey and James V. Stone CSRP 379, May 1995 Cognitive Science Research Paper Serial No. CSRP 379 The
More informationLearning and Memory in Neural Networks
Learning and Memory in Neural Networks Guy Billings, Neuroinformatics Doctoral Training Centre, The School of Informatics, The University of Edinburgh, UK. Neural networks consist of computational units
More informationUsing a Hopfield Network: A Nuts and Bolts Approach
Using a Hopfield Network: A Nuts and Bolts Approach November 4, 2013 Gershon Wolfe, Ph.D. Hopfield Model as Applied to Classification Hopfield network Training the network Updating nodes Sequencing of
More informationARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD
ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided
More informationArtificial Intelligence Hopfield Networks
Artificial Intelligence Hopfield Networks Andrea Torsello Network Topologies Single Layer Recurrent Network Bidirectional Symmetric Connection Binary / Continuous Units Associative Memory Optimization
More informationRevision: Neural Network
Revision: Neural Network Exercise 1 Tell whether each of the following statements is true or false by checking the appropriate box. Statement True False a) A perceptron is guaranteed to perfectly learn
More informationGenetic Engineering and Creative Design
Genetic Engineering and Creative Design Background genes, genotype, phenotype, fitness Connecting genes to performance in fitness Emergent gene clusters evolved genes MIT Class 4.208 Spring 2002 Evolution
More information(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann
(Feed-Forward) Neural Networks 2016-12-06 Dr. Hajira Jabeen, Prof. Jens Lehmann Outline In the previous lectures we have learned about tensors and factorization methods. RESCAL is a bilinear model for
More informationSupporting Online Material for
www.sciencemag.org/cgi/content/full/319/5869/1543/dc1 Supporting Online Material for Synaptic Theory of Working Memory Gianluigi Mongillo, Omri Barak, Misha Tsodyks* *To whom correspondence should be addressed.
More informationbiologically-inspired computing lecture 18
Informatics -inspired lecture 18 Sections I485/H400 course outlook Assignments: 35% Students will complete 4/5 assignments based on algorithms presented in class Lab meets in I1 (West) 109 on Lab Wednesdays
More informationData Mining Part 5. Prediction
Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,
More informationV. Evolutionary Computing. Read Flake, ch. 20. Genetic Algorithms. Part 5A: Genetic Algorithms 4/10/17. A. Genetic Algorithms
V. Evolutionary Computing A. Genetic Algorithms 4/10/17 1 Read Flake, ch. 20 4/10/17 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified model of genetics
More informationToward a Dynamical Systems Analysis of Neuromodulation
Toward a Dynamical Systems Analysis of Neuromodulation Chris Buckley, Seth Bullock and Netta Cohen Biosystems Group, School of Computing University of Leeds, UK christb seth netta@comp.leeds.ac.uk Abstract
More informationArtificial Neural Networks. Q550: Models in Cognitive Science Lecture 5
Artificial Neural Networks Q550: Models in Cognitive Science Lecture 5 "Intelligence is 10 million rules." --Doug Lenat The human brain has about 100 billion neurons. With an estimated average of one thousand
More informationNon-Adaptive Evolvability. Jeff Clune
Non-Adaptive Evolvability! Jeff Clune Assistant Professor Computer Science Evolving Artificial Intelligence Laboratory Evolution Fails to Optimize Mutation Rates (though it would improve evolvability)
More informationShigetaka Fujita. Rokkodai, Nada, Kobe 657, Japan. Haruhiko Nishimura. Yashiro-cho, Kato-gun, Hyogo , Japan. Abstract
KOBE-TH-94-07 HUIS-94-03 November 1994 An Evolutionary Approach to Associative Memory in Recurrent Neural Networks Shigetaka Fujita Graduate School of Science and Technology Kobe University Rokkodai, Nada,
More informationArtificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino
Artificial Neural Networks Data Base and Data Mining Group of Politecnico di Torino Elena Baralis Politecnico di Torino Artificial Neural Networks Inspired to the structure of the human brain Neurons as
More informationArtificial Neural Networks
Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples
More informationFundamentals of Genetic Algorithms
Fundamentals of Genetic Algorithms : AI Course Lecture 39 40, notes, slides www.myreaders.info/, RC Chakraborty, e-mail rcchak@gmail.com, June 01, 2010 www.myreaders.info/html/artificial_intelligence.html
More informationA Novel Activity Detection Method
A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of
More informationThe Role of Nearly Neutral Mutations in the Evolution of Dynamical Neural Networks
The Role of Nearly Neutral Mutations in the Evolution of Dynamical Neural Networks Eduardo Izquierdo-Torres Department of Informatics University of Sussex, BN 9QH. UK. e.j.izquierdo-torres@sussex.ac.uk
More informationArtificial Neural Network and Fuzzy Logic
Artificial Neural Network and Fuzzy Logic 1 Syllabus 2 Syllabus 3 Books 1. Artificial Neural Networks by B. Yagnanarayan, PHI - (Cover Topologies part of unit 1 and All part of Unit 2) 2. Neural Networks
More informationArtificial Intelligence
Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement
More informationSample Exam COMP 9444 NEURAL NETWORKS Solutions
FAMILY NAME OTHER NAMES STUDENT ID SIGNATURE Sample Exam COMP 9444 NEURAL NETWORKS Solutions (1) TIME ALLOWED 3 HOURS (2) TOTAL NUMBER OF QUESTIONS 12 (3) STUDENTS SHOULD ANSWER ALL QUESTIONS (4) QUESTIONS
More information[Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4]
1 EVOLUTIONARY ALGORITHMS [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Schema theorem Genetic Programming Individual learning and population
More informationLecture 5: Recurrent Neural Networks
1/25 Lecture 5: Recurrent Neural Networks Nima Mohajerin University of Waterloo WAVE Lab nima.mohajerin@uwaterloo.ca July 4, 2017 2/25 Overview 1 Recap 2 RNN Architectures for Learning Long Term Dependencies
More informationIn biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.
In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required. In humans, association is known to be a prominent feature of memory.
More informationAn artificial neural networks (ANNs) model is a functional abstraction of the
CHAPER 3 3. Introduction An artificial neural networs (ANNs) model is a functional abstraction of the biological neural structures of the central nervous system. hey are composed of many simple and highly
More informationV. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA
Part 5A: Genetic Algorithms V. Evolutionary Computing A. Genetic Algorithms Read Flake, ch. 20 1 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified
More informationEvolutionary Design I
Evolutionary Design I Jason Noble jasonn@comp.leeds.ac.uk Biosystems group, School of Computing Evolutionary Design I p.1/29 This lecture Harnessing evolution in a computer program How to construct a genetic
More informationGENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS
GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS A genetic algorithm is a random search technique for global optimisation in a complex search space. It was originally inspired by an
More informationStatistical Machine Learning from Data
January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Other Artificial Neural Networks Samy Bengio IDIAP Research Institute, Martigny, Switzerland,
More informationExplorations in Evolutionary Visualisation
Explorations in Evolutionary Visualisation Lionel Barnett Centre for Computational Neuroscience and Robotics (CCNR) School of Cognitive and Computing Sciences, University of Sussex, Brighton, UK. Email:
More informationIV. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA
IV. Evolutionary Computing A. Genetic Algorithms Read Flake, ch. 20 2014/2/26 1 2014/2/26 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified model
More informationNeuroevolution for sound event detection in real life audio: A pilot study
Neuroevolution for sound event detection in real life audio: A pilot study Christian Kroos & Mark D. Plumbley Centre for Vision, Speech and Signal Processing (CVSSP), University of Surrey, UK Background
More informationImproving Coordination via Emergent Communication in Cooperative Multiagent Systems: A Genetic Network Programming Approach
Proceedings of the 2009 IEEE International Conference on Systems, Man, and Cybernetics San Antonio, TX, USA - October 2009 Improving Coordination via Emergent Communication in Cooperative Multiagent Systems:
More informationNeural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21
Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural
More informationThe Evolution of Gene Dominance through the. Baldwin Effect
The Evolution of Gene Dominance through the Baldwin Effect Larry Bull Computer Science Research Centre Department of Computer Science & Creative Technologies University of the West of England, Bristol
More informationNeural networks. Chapter 20. Chapter 20 1
Neural networks Chapter 20 Chapter 20 1 Outline Brains Neural networks Perceptrons Multilayer networks Applications of neural networks Chapter 20 2 Brains 10 11 neurons of > 20 types, 10 14 synapses, 1ms
More informationSimple Neural Nets for Pattern Classification: McCulloch-Pitts Threshold Logic CS 5870
Simple Neural Nets for Pattern Classification: McCulloch-Pitts Threshold Logic CS 5870 Jugal Kalita University of Colorado Colorado Springs Fall 2014 Logic Gates and Boolean Algebra Logic gates are used
More informationParallel Genetic Algorithms
Parallel Genetic Algorithms for the Calibration of Financial Models Riccardo Gismondi June 13, 2008 High Performance Computing in Finance and Insurance Research Institute for Computational Methods Vienna
More informationNeural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome. Evolving Modular Architectures for Neural Networks
Neural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome Evolving Modular Architectures for Neural Networks Andrea Di Ferdinando, Raffaele Calabretta and Domenico
More informationPart 8: Neural Networks
METU Informatics Institute Min720 Pattern Classification ith Bio-Medical Applications Part 8: Neural Netors - INTRODUCTION: BIOLOGICAL VS. ARTIFICIAL Biological Neural Netors A Neuron: - A nerve cell as
More informationArtificial Neural Networks. Part 2
Artificial Neural Netorks Part Artificial Neuron Model Folloing simplified model of real neurons is also knon as a Threshold Logic Unit x McCullouch-Pitts neuron (943) x x n n Body of neuron f out Biological
More informationNeural networks. Chapter 20, Section 5 1
Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of
More informationInteger weight training by differential evolution algorithms
Integer weight training by differential evolution algorithms V.P. Plagianakos, D.G. Sotiropoulos, and M.N. Vrahatis University of Patras, Department of Mathematics, GR-265 00, Patras, Greece. e-mail: vpp
More informationNeural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28
1 / 28 Neural Networks Mark van Rossum School of Informatics, University of Edinburgh January 15, 2018 2 / 28 Goals: Understand how (recurrent) networks behave Find a way to teach networks to do a certain
More informationForecasting & Futurism
Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning
More informationEvolutionary computation
Evolutionary computation Andrea Roli andrea.roli@unibo.it Dept. of Computer Science and Engineering (DISI) Campus of Cesena Alma Mater Studiorum Università di Bologna Outline 1 Basic principles 2 Genetic
More informationAnalysis of Multilayer Neural Network Modeling and Long Short-Term Memory
Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory Danilo López, Nelson Vera, Luis Pedraza International Science Index, Mathematical and Computational Sciences waset.org/publication/10006216
More informationNegatively Correlated Echo State Networks
Negatively Correlated Echo State Networks Ali Rodan and Peter Tiňo School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {a.a.rodan, P.Tino}@cs.bham.ac.uk
More informationNeural networks. Chapter 19, Sections 1 5 1
Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10
More informationApplication of Fully Recurrent (FRNN) and Radial Basis Function (RBFNN) Neural Networks for Simulating Solar Radiation
Bulletin of Environment, Pharmacology and Life Sciences Bull. Env. Pharmacol. Life Sci., Vol 3 () January 04: 3-39 04 Academy for Environment and Life Sciences, India Online ISSN 77-808 Journal s URL:http://www.bepls.com
More informationARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92
ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000
More informationChapter 17: Population Genetics and Speciation
Chapter 17: Population Genetics and Speciation Section 1: Genetic Variation Population Genetics: Normal Distribution: a line graph showing the general trends in a set of data of which most values are near
More informationDynamic Working Memory in Recurrent Neural Networks
Dynamic Working Memory in Recurrent Neural Networks Alexander Atanasov Research Advisor: John Murray Physics 471 Fall Term, 2016 Abstract Recurrent neural networks (RNNs) are physically-motivated models
More informationIntroduction to Artificial Neural Networks
Facultés Universitaires Notre-Dame de la Paix 27 March 2007 Outline 1 Introduction 2 Fundamentals Biological neuron Artificial neuron Artificial Neural Network Outline 3 Single-layer ANN Perceptron Adaline
More informationIndividual learning and population evolution
Genetic Algorithms [Read Chapter 9] [Exercises 9.1, 9.2, 9.3, 9.4] Evolutionary computation Prototypical GA An example: GABIL Genetic Programming Individual learning and population evolution 168 lecture
More informationTowards Synthesizing Artificial Neural Networks that Exhibit Cooperative Intelligent Behavior: Some Open Issues in Artificial Life Michael G.
Towards Synthesizing Artificial Neural Networks that Exhibit Cooperative Intelligent Behavior: Some Open Issues in Artificial Life Michael G. Dyer Computer Science Department, UCLA Overview Introduction
More informationKeywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm
Volume 4, Issue 5, May 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Huffman Encoding
More informationAssociative Neural Networks using Matlab
Associative Neural Networks using Matlab Example 1: Write a matlab program to find the weight matrix of an auto associative net to store the vector (1 1-1 -1). Test the response of the network by presenting
More informationCMSC 421: Neural Computation. Applications of Neural Networks
CMSC 42: Neural Computation definition synonyms neural networks artificial neural networks neural modeling connectionist models parallel distributed processing AI perspective Applications of Neural Networks
More informationArtificial Neural Network
Artificial Neural Network Contents 2 What is ANN? Biological Neuron Structure of Neuron Types of Neuron Models of Neuron Analogy with human NN Perceptron OCR Multilayer Neural Network Back propagation
More informationMicroevolution Changing Allele Frequencies
Microevolution Changing Allele Frequencies Evolution Evolution is defined as a change in the inherited characteristics of biological populations over successive generations. Microevolution involves the
More informationEvolutionary Algorithms
Evolutionary Algorithms a short introduction Giuseppe Narzisi Courant Institute of Mathematical Sciences New York University 31 January 2008 Outline 1 Evolution 2 Evolutionary Computation 3 Evolutionary
More informationIncremental Stochastic Gradient Descent
Incremental Stochastic Gradient Descent Batch mode : gradient descent w=w - η E D [w] over the entire data D E D [w]=1/2σ d (t d -o d ) 2 Incremental mode: gradient descent w=w - η E d [w] over individual
More informationEvolving more efficient digital circuits by allowing circuit layout evolution and multi-objective fitness
Evolving more efficient digital circuits by allowing circuit layout evolution and multi-objective fitness Tatiana Kalganova Julian Miller School of Computing School of Computing Napier University Napier
More informationHebbian Learning using Fixed Weight Evolved Dynamical Neural Networks
Proceedings of the 27 IEEE Symposium on Artificial Life (CI-ALife 27) Hebbian Learning using Fixed Weight Evolved Dynamical Neural Networks Eduardo Izquierdo-Torres Centre for Computational Neuroscience
More informationEvolutionary Robotics
Evolutionary Robotics Previously on evolutionary robotics Evolving Neural Networks How do we evolve a neural network? Evolving Neural Networks How do we evolve a neural network? One option: evolve the
More informationTable S1 shows the SOM parameters that were used in the main manuscript. These correspond to the default set of options of the SOM Toolbox.
Supplemental file 1: SOM training parameters and sensitivity analysis. Table S1 shows the SOM parameters that were used in the main manuscript. These correspond to the default set of options of the SOM
More informationMechanisms of Evolution Microevolution. Key Concepts. Population Genetics
Mechanisms of Evolution Microevolution Population Genetics Key Concepts 23.1: Population genetics provides a foundation for studying evolution 23.2: Mutation and sexual recombination produce the variation
More informationAn Adaptive Clustering Method for Model-free Reinforcement Learning
An Adaptive Clustering Method for Model-free Reinforcement Learning Andreas Matt and Georg Regensburger Institute of Mathematics University of Innsbruck, Austria {andreas.matt, georg.regensburger}@uibk.ac.at
More informationIntroduction to Neural Networks: Structure and Training
Introduction to Neural Networks: Structure and Training Professor Q.J. Zhang Department of Electronics Carleton University, Ottawa, Canada www.doe.carleton.ca/~qjz, qjz@doe.carleton.ca A Quick Illustration
More informationLecture 7 Artificial neural networks: Supervised learning
Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works The neuron as a simple computing element The perceptron Multilayer neural networks Accelerated learning in
More informationMachine Learning. Neural Networks
Machine Learning Neural Networks Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 Biological Analogy Bryan Pardo, Northwestern University, Machine Learning EECS 349 Fall 2007 THE
More informationDevelopment of a Deep Recurrent Neural Network Controller for Flight Applications
Development of a Deep Recurrent Neural Network Controller for Flight Applications American Control Conference (ACC) May 26, 2017 Scott A. Nivison Pramod P. Khargonekar Department of Electrical and Computer
More informationTemplate-Based Representations. Sargur Srihari
Template-Based Representations Sargur srihari@cedar.buffalo.edu 1 Topics Variable-based vs Template-based Temporal Models Basic Assumptions Dynamic Bayesian Networks Hidden Markov Models Linear Dynamical
More informationUsing Variable Threshold to Increase Capacity in a Feedback Neural Network
Using Variable Threshold to Increase Capacity in a Feedback Neural Network Praveen Kuruvada Abstract: The article presents new results on the use of variable thresholds to increase the capacity of a feedback
More informationShort Term Memory Quantifications in Input-Driven Linear Dynamical Systems
Short Term Memory Quantifications in Input-Driven Linear Dynamical Systems Peter Tiňo and Ali Rodan School of Computer Science, The University of Birmingham Birmingham B15 2TT, United Kingdom E-mail: {P.Tino,
More informationMultiobjective Optimization of an Extremal Evolution Model
Multiobjective Optimization of an Extremal Evolution Model Mohamed Fathey Elettreby Mathematics Department, Faculty of Science, Mansoura University, Mansoura 35516, Egypt Reprint requests to M. F. E.;
More informationNEUROEVOLUTION. Contents. Evolutionary Computation. Neuroevolution. Types of neuro-evolution algorithms
Contents Evolutionary Computation overview NEUROEVOLUTION Presenter: Vlad Chiriacescu Neuroevolution overview Issues in standard Evolutionary Computation NEAT method Complexification in competitive coevolution
More informationCISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II)
CISC 889 Bioinformatics (Spring 24) Hidden Markov Models (II) a. Likelihood: forward algorithm b. Decoding: Viterbi algorithm c. Model building: Baum-Welch algorithm Viterbi training Hidden Markov models
More informationNeural Networks Based on Competition
Neural Networks Based on Competition In some examples of pattern classification we encountered a situation in which the net was trained to classify the input signal into one of the output categories, while
More informationSP-CNN: A Scalable and Programmable CNN-based Accelerator. Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay
SP-CNN: A Scalable and Programmable CNN-based Accelerator Dilan Manatunga Dr. Hyesoon Kim Dr. Saibal Mukhopadhyay Motivation Power is a first-order design constraint, especially for embedded devices. Certain
More informationComputational Intelligence Lecture 6: Associative Memory
Computational Intelligence Lecture 6: Associative Memory Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Fall 2011 Farzaneh Abdollahi Computational Intelligence
More informationNeural Networks Learning the network: Backprop , Fall 2018 Lecture 4
Neural Networks Learning the network: Backprop 11-785, Fall 2018 Lecture 4 1 Recap: The MLP can represent any function The MLP can be constructed to represent anything But how do we construct it? 2 Recap:
More informationDEVS Simulation of Spiking Neural Networks
DEVS Simulation of Spiking Neural Networks Rene Mayrhofer, Michael Affenzeller, Herbert Prähofer, Gerhard Höfer, Alexander Fried Institute of Systems Science Systems Theory and Information Technology Johannes
More informationGENETICS - CLUTCH CH.22 EVOLUTIONARY GENETICS.
!! www.clutchprep.com CONCEPT: OVERVIEW OF EVOLUTION Evolution is a process through which variation in individuals makes it more likely for them to survive and reproduce There are principles to the theory
More informationAn Evolution Strategy for the Induction of Fuzzy Finite-state Automata
Journal of Mathematics and Statistics 2 (2): 386-390, 2006 ISSN 1549-3644 Science Publications, 2006 An Evolution Strategy for the Induction of Fuzzy Finite-state Automata 1,2 Mozhiwen and 1 Wanmin 1 College
More informationStability of backpropagation learning rule
Stability of backpropagation learning rule Petr Krupanský, Petr Pivoňka, Jiří Dohnal Department of Control and Instrumentation Brno University of Technology Božetěchova 2, 612 66 Brno, Czech republic krupan,
More informationRestarting a Genetic Algorithm for Set Cover Problem Using Schnabel Census
Restarting a Genetic Algorithm for Set Cover Problem Using Schnabel Census Anton V. Eremeev 1,2 1 Dostoevsky Omsk State University, Omsk, Russia 2 The Institute of Scientific Information for Social Sciences
More information3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield
3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield (1982, 1984). - The net is a fully interconnected neural
More information