A Robot that Learns an Evaluation Function for Acquiring of Appropriate Motions

Similar documents
Effect of number of hidden neurons on learning in large-scale layered neural networks

STATE GENERALIZATION WITH SUPPORT VECTOR MACHINES IN REINFORCEMENT LEARNING. Ryo Goto, Toshihiro Matsui and Hiroshi Matsuo

THE MOVING MAN: DISTANCE, DISPLACEMENT, SPEED & VELOCITY

Lab 6 Forces Part 2. Physics 225 Lab

Learning in State-Space Reinforcement Learning CIS 32

LAB 2 - ONE DIMENSIONAL MOTION

Reinforcement Learning and Control

Constant velocity and constant acceleration

PHY 221 Lab 2. Acceleration and Uniform Motion

5-Sep-15 PHYS101-2 GRAPHING

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

8. Lecture Neural Networks

) (d o f. For the previous layer in a neural network (just the rightmost layer if a single neuron), the required update equation is: 2.

Recurrent Neural Networks

Computational Intelligence Winter Term 2017/18

Higher Unit 6a b topic test

Computational Intelligence

Advanced Placement Physics C Summer Assignment

Introduction to Machine Learning Prof. Sudeshna Sarkar Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Second-order Learning Algorithm with Squared Penalty Term

Address for Correspondence

6.036: Midterm, Spring Solutions

Unit 8: Introduction to neural networks. Perceptrons

Name: Date: Partners: LAB 2: ACCELERATED MOTION

Neural Networks. William Cohen [pilfered from: Ziv; Geoff Hinton; Yoshua Bengio; Yann LeCun; Hongkak Lee - NIPs 2010 tutorial ]

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

EXPERIMENT 7: ANGULAR KINEMATICS AND TORQUE (V_3)

Worksheet for Exploration 6.1: An Operational Definition of Work

Major Ideas in Calc 3 / Exam Review Topics

Introducing the Potential Energy Diagram as a Model of the Atom

LAB 4: FORCE AND MOTION

ECE 471/571 - Lecture 17. Types of NN. History. Back Propagation. Recurrent (feedback during operation) Feedforward

Learning and Memory in Neural Networks

Artificial Intelligence

Impulse, Momentum, and Energy

Lesson 6-1: Relations and Functions

Sin, Cos and All That

Artificial Neural Networks 2

epochs epochs

Artificial Neural Networks Examination, June 2005

Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

PHY221 Lab 2 - Experiencing Acceleration: Motion with constant acceleration; Logger Pro fits to displacement-time graphs

Neural Networks. Hopfield Nets and Auto Associators Fall 2017

Putting the Shot. 3. Now describe the forces that you exerted on the weight. Is it constant? Explain why you think so.

Questions Q1. Select one answer from A to D and put a cross in the box ( )

Designing Information Devices and Systems I Spring 2018 Homework 10

Introduction to Reinforcement Learning

Study on Material Recognition of Artificial Finger for Pneumatic Robot Arm

Kinematics Lab. 1 Introduction. 2 Equipment. 3 Procedures

Which one of the following graphs correctly shows the relationship between potential difference (V) and current (I) for a filament lamp?

Artificial Neural Networks Examination, June 2004

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Similar Shapes and Gnomons

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

On-line Learning of Robot Arm Impedance Using Neural Networks

What s the Average? Stu Schwartz

9.4 Light: Wave or Particle?

AP Calculus. Derivatives.

Chapter 9: The Perceptron

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Master Recherche IAC TC2: Apprentissage Statistique & Optimisation

Motion in 1 Dimension. By Prof. Massimiliano Galeazzi, University of Miami

Neural network modelling of reinforced concrete beam shear capacity

Lab Partner(s) TA Initials (on completion) EXPERIMENT 7: ANGULAR KINEMATICS AND TORQUE

Mr. Harshit K. Dave 1, Dr. Keyur P. Desai 2, Dr. Harit K. Raval 3

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

Mathematics Practice Test 2

MITOCW ocw-18_02-f07-lec17_220k

Feedforward Neural Nets and Backpropagation

Lecture 23: Reinforcement Learning

Introduction. Pre-Lab Questions: Physics 1CL PERIODIC MOTION - PART II Spring 2009

Neural Networks for Protein Structure Prediction Brown, JMB CS 466 Saurabh Sinha

Introduction. Pre-Lab Questions: Physics 1CL PERIODIC MOTION - PART II Fall 2009

COMP6053 lecture: Sampling and the central limit theorem. Jason Noble,

Energy Diagrams --- Attraction

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

Convergence of Hybrid Algorithm with Adaptive Learning Parameter for Multilayer Neural Network

How generative models develop in predictive processing

A shape which can be divided by a straight line, so that each part is the mirror image of the other has reflective symmetry.

ENLARGING AREAS AND VOLUMES

DATA LAB. Data Lab Page 1

Sin, Cos and All That

Basic Principles of Unsupervised and Unsupervised

Name: SOLUTIONS Date: 11/9/2017. M20550 Calculus III Tutorial Worksheet 8

Final Project Physics 590. Mary-Kate McGlinchey MISEP Summer 2005

Akinori Sekiguchi and Yoshihiko Nakamura. Dept. of Mechano-Informatics, University of Tokyo Hongo, Bunkyo-Ku, Tokyo , Japan

Lecture - 24 Radial Basis Function Networks: Cover s Theorem

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja

Neural Nets Supervised learning

The Time-Scaled Trapezoidal Integration Rule for Discrete Fractional Order Controllers

(n 3) 3 + n 3 = (n + 3) 3, ( ) (n 6) 3 + n 3 = (n + 6) 3, ( ) then n is even. Deduce that there is no integer n which satisfies the equation ( ).

Reinforcement Learning for Continuous. Action using Stochastic Gradient Ascent. Hajime KIMURA, Shigenobu KOBAYASHI JAPAN

Motion II. Goals and Introduction

Partner s Name: EXPERIMENT MOTION PLOTS & FREE FALL ACCELERATION

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Physics 8.01 Physics Fall Term = # v x. t " =0. are the values at t = 0.

To factor an expression means to write it as a product of factors instead of a sum of terms. The expression 3x

In other words, we are interested in what is happening to the y values as we get really large x values and as we get really small x values.

Least Mean Squares Regression. Machine Learning Fall 2018

Constructing Potential Energy Diagrams

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Transcription:

A Robot that Learns an Evaluation Function for Acquiring of Appropriate Motions Katsunari Shibata and Yoich Okabe Research Center for Advanced Science and Technology, Univ. of Tokyo -6-1 Komaba, Meguro-ku, 15 Tokyo JAPAN shibata@okabe.rcast.u-tokyo.ac.jp Summary An unsupervised learning method of an is proposed. By this function, a can learn a series of motion to achieve a given purpose. The is calculated by a neural network and time derivative of this function is reduced during motions with random trials. We confirmed by simulation that a with asymmetric motion feature could obtain an appropriate asymmetric and become to achieve a given purpose along an almost optimal path. 1. Introduction In famous Skinner's experiment, a mouse in Skinner's box becomes to be able to push the button to get a piece of food after some trials as shown in Fig. 1 [1]. However, the mouse could not know that it could get a piece of food by pushing the button before the mouse was put into the box. At early stage, the mouse moves randomly, and when it pushes the button unexpectedly, it gets a piece of food. From the experience, the possibility that the mouse pushes the button, slightly becomes large. Then after many experiences, the mouse has an intention to push the button. This means that the mouse in Skinner's experiment did not know only the way to push the button, but also were not given any evaluation functions about getting the food. From this fact, a also can obtain an through its experiences. Then we fig.1 Skinner's experiment assumed a locomotive as shown in Fig.. The has only a purpose to get a. The mouse in Skinner's experiment, can learn the and motion by the interaction between the mouse and the environment without any supervisors. When we make a learn an appropriate series of motion using the interaction, we usually give the an that sensor wheel we made [][]. The evaluation value of the present state is calculated from the sensor signals through the. locomotive If the continues to move towards the gradient direction of the, it will be able to achieve the given fig. An assumed that has a purpose to get a purpose. For example, in the case of the in Fig., an can be the inverse of the between the and the. As shown in Fig., the moves and change gets sensor signals. Then the makes the environment supervisor signal for the motion creator from the sensor signals. The motion creator, that is made by a layered type neural network, sense move learns by the supervisor signal, and become to create an appropriate motion signal. motion creator Then we tried to make a learn both an evaluation function and a series of motion to achieve a given purpose. We evaluation cannot know if the, that we give to a, is function supervisor the best for the and the environment. Furthermore, if the environment changes, the also has to change. For these reasons, the has to learn its for purpose (auto-create) a good adaptation to the environment. fig. Learning system and In this paper, we describe the way to learn the evaluation interaction to the environment

function without any supervisors, and the way to learn the and the motion in parallel. Finally we show the result for comparing which series of motion was closer to the optimal route when the was given a fixed or when the learned the by itself.. Learning Method of a Series of Motion Some ways to learn a series of motion from an evaluation function has been already proposed [][]. We indicate the way that we employed. Here, the motion signal is calculated by a layered type neural network. Let us suppose the dimensions of space as shown Fig.. X, Y axes show sensor signals respectively. The other axis shows the evaluation value. An evaluation value is decided for each set of sensor signals and the surface in this figure is an surface. The moves according to the sum of the output of a neural network and a small uniform random number as follows, M = m + r (1) learning direction A ms = m + r ms = m + k contour of supervising vector m ms motion vector B movable range fig. Motion learning 90 random vector r where M : actual motion vector, m : motion signal vector and r : uniform random number vector. If the has two or more actuator, the neural network has two or more output neurons and each output is added by a different random number. The circle filled with oblique line in Fig. shows the movable range of the by the random vector r. The moves and can get a new evaluation value using the. Then the makes the following supervisor signal, ms = m + r () where ms : supervisor signal vector for motion signal, (t) : and = (t) - (t-1). The neural network learns at every time unit by this supervisor signal, using a conventional supervised learning algorithm like Back Propagation Method []. By the learning through many experiences, the direction of the motion vector becomes close to the gradient direction of the in average as m s in Fig.. Then the can learn a series of motion close to the optimal route under the.. Unsupervised Learning Method of an Evaluation Function Let us think how should the be made. Supposing the dimensions of space in which there is a and a as shown in Fig. 5 (a), we can think of an evaluation method using the between the and the. However, how to normalize on each axis cannot be known easily. Furthermore if there is an area that the cannot move through, like an obstacle as shown in Fig. 5 (a), it is meaningless to evaluate only by the. Then we propose to evaluate by necessary time for arriving at the. If there is an obstacle, the can move beside the area and arrive at the. If we project the route on the time axis as shown in Fig. 5 (b), we can evaluate under only one criterion and there are no necessity to normalize each axis any longer. However, if the modifies the by tracing back the route to the starting point when the arrives at the, the must remember all the states it has existed. In order to avoid wasting memory, we propose a real-time learning method of the. The evaluation function should be as follows. As shown in Fig. 6, the output of the function is small in an initial state, the output becomes larger as time goes by, and finally the output becomes the largest value when the gets the. Here, the is calculated by a layered type neural network and the output function of each neuron is sigmoid between and. The learning method has three phases. When the is in an initial state, to decrease the output of the, the Z obstacle 1 (a) 7 6 5 time 1 5 6 7 (b) fig.5 State evaluation method (a) evaluation on space (b) evaluation on time axis Y X

network learns by the supervisor signal as s (x(t)) = (x(t)) - δ () where s(x(t)) : supervisor signal for the and δ : small constant value. Here we employ 1 as δ. When the arrives at the, the network learns by the supervisor signal as s (x(t)) = max () where max : the maximum value of the (different from the maximum of the output function). Here we employ 0.9 as the maximum value max. Since it is the best state when the arrives at the, the supervisor signal can be a constant value. However, since the initial state is not always the furthest state from the purpose, we cannot set up a constant value, and we set up a slightly small value than the actual output as the supervisor signal. Otherwise from the initial state to the, in order to increase the output smoothly, the supervisor signal is as s (x(t)) = (x(t)) + ξ d (x(t)) dt (5) error output 0.9 Evaluation Function Real Ideal Evaluation Evaluation Learning Target State Initial Time State fig.6 Learning method that realizes an evaluation on time d dt A A neural network for 0 B route (a) 1 5 6 7 8 1 10 11 Learning 9 8 5 6 7 fig.7 An example of optimizing route where ξ : a smoothing constant. Here we employ as the smoothing constant ξ. One thing we have to take care is that the learning is on job learning and the network learns only one iteration using conventional supervised learning algorithm like Back Propagation Method. The output does not become to the value of the supervisor signal, but it becomes closer to the supervisor signal slightly. Let us assume that a moves along the route (b) in Fig. 7 when the goes from a point A to a point B. It takes 11 time unit, but if it moves along the route (a) it can arrive at the point B in 8 time unit. By the learning of the, it becomes smooth around the route (b). On the other hand, d/dt along the route (a) is larger than that along the route (b) because the can arrive at the point B faster. By the learning of motion, the motion changes to the gradient direction of the. The route, that moves along, becomes slightly close to the route (a). By repeating the learning of the and the learning of motion, the becomes to move along the route (a). We can see that the combination of the two learning processes makes the route close to the optimal one.. Structure and Flow of the Learning To realize the combination of the two learning processes, we employ the structure as shown in Fig. 8. The large rectangle shows the brain and there are two neural networks. One of them is to calculate the and the other is to calculate the motion signal. Each of them calculates the output from the sensor signals, and so we can use only one neural network for both purposes. The moves according to the sum of the output of the motion network and a random number. The supervisor signal for the evaluation neural network is usually made from the second order deviation of the output of itself as Eq.(5). The supervisor signal for the motion network is made from the product of the first order deviation of the evaluation output and the random number that was used for creating a motion. Then the each BP d dt input w error BP random + + output input A neural network for motion signal actuator sensor fig.8 Network structure and learning method change route (b) move environ ment state

network learns only one iteration by the supervisor signal respectively at every time unit. Fig. 9 shows the flow of the learning. First the is placed somewhere, and the evaluation network learns by the supervisor signal as Eq.(). Then the repeats a motion as Eq.(1) and learning of both neural networks by the supervisor signals as Eq.() and Eq.(5) respectively until it arrives at the. When it arrives, the estimation network learns by the supervisor signal as Eq.(). Then the process is repeated many times. 5. Simulation In this section, we show the simulation result. We assumed the environment as shown in Fig. 10. The locomotive has two wheels and can move them independently. The rotate angle of each wheel is decided by the sum of each output of the motion network and a random number. In order to make the environment in which we can make the easily, the ratio of the rotation angle against the motion signal is three times larger in the right wheel than in the left wheel. The range of rotating angle of the right wheel is also three times wider that of the left wheel as shown in Fig. 11. The also has two sensors. One of them sensed the from the to the and the other one sensed the. Initially a was placed randomly at the edge of the rectangle except for the edge that the existed. It is drawn by a thick line in Fig. 10. The moved and if the went backward passing under the, we Initial Target Position (randomly selected) repeat initial state change initial point learn to 0.1 unit motion including random factor learn to decrease irregularities motion signal learn to the direction to get more evaluation value state Fail Succeed Fail fig.10 Environment in the simulation arrival to? y n learn to 0.9 fig.9 Learning procedure assumed that the could catch the. If the went backward passing beside the, we assumed that the failed to catch the. In this simulation, when the failed to catch the, the evaluation neural network learned by the supervisor signal 0.1. We defined one experience from placing the on the initial position to the success or failure to catch the. However, the can move only randomly at first phase and cannot arrive at the in a finite time. Then we moved the close to the when the could not arrive after some time. In order to compare, we do another simulation. We gave the an that we can think of easily, and make the learn only the motion signal. Then we show the change of the in Fig. 1. Each graph in Fig. 1 is drawn on centered coordinates. The was placed at the center of the coordinates and corresponding evaluation value was plotted on Z axis when the was placed on each lattice. The contour of the is also drawn. Figure 1 (d) shows the that we gave the in the comparison simulation. Figure 1 shows the contour of the and the motion between the and the when the was placed on each of 5 places. They are also drawn on the centered coordinates but on the dimensional surface. Though the moved and did not move actually, the was fixed and the moved in these figures because of the centered coordinates. On the locus of the, the points were plotted at every 10 time unit. In these graphs except for the graph (a), the moved more when the was far from the and moved to the tangential direction. The reason is that when the rotates actually, the relatively moves according to the product of the rotation angle and the on the centered coordinates. We also show the loci on the absolute coordinates in Fig. 1. Figure 1(a) shows the comparison between the locus after 5000 experiences and that after 0000 experiences. Figure 1(b) shows the comparison to the case of the fixed after 0000 experiences. On the locus of the, the points are also plotted at every 10 time unit. After 1000 experiences, the did not have an enough expansion of value range and the could not get the as shown in Fig. 1(a). After 5000 experiences, the had a sharp shape and the could arrive at the from any initial positions. The got the at the right half of the. The reason is that the right wheel could rotate times more than left wheel and the could get the soon when the was in the right. After 0000 experiences, Target 1-1 - Fig.11 Asymmetric motion

0. 0. (a) 1000 experiences - (a) 1000 experiences (b) 5000 experiences - 0.7 (b) 5000 experiences (c) 0000 experiences - 0.7 (c) 0000 experiences 0.8 (d) fixed (0000 experiences) fig.1 Evaluation function (on centered coordinates) - (d) fixed (0000 experiences) fig.1 Evaluation function and the locus (on centered coordinates) the

ridge of the moved to the left and the moved along the ridge. It means that the approached the with looking at it in the left. In the absolute coordinates as shown in Fig. 1(a), the locus of the became like an arc after 0000 experiences. Furthermore, the went back with rotating itself to the right when the was seen in the right direction. So, the locus after 5000 experiences seems better than that after 0000 experiences at a glance. However, the necessary time was more after 5000 experiences. The and the locus could be reflected the characteristic of the. On the other hand, the in a comparison simulation is symmetric shape as shown in Fig. 1(d), and the locus of the expanded in front of the as shown in Fig. 1(d). When we see the Fig. 1(b), the locus in the case of fixed evaluation function is close to a straight line and it seems better than that in the case of self-organized evaluation function at a glance. Also in this case, the could arrive at the faster in the case of self-organized. Then we show the change of the necessary time for the to arrive at the in Fig. 15. The time is the average of the 1 trials in which the was placed at a regular interval on the edge of the rectangle. If the failed in a trial, the time is set 1000. In this figure, we show the results of simulations. We did two simulations in which the initial weight value in the neural network was different from each other in the case of the fixed evaluation function and in the case of the self-organized evaluation function respectively. We can see that the, using the fixed, could become to get the faster than in the case of the self-organized in an early stage. The reason is that it took some time to form the in the case of the self-organized. We can also see from the Fig. 15 that after many experiences the and the route of the were optimized in the case of the self-organized evaluation function and realized the better route than in the case of the fixed. Y 7 6 5 5000 experiences 0000 experiences 1 (start) 0-5 - - - -1 0 1 5 X (a) fixed self-organized Y 7 5 1 0 (start) -5 - - - -1 0 1 5 X (b) fig.1 Locus of the (on absolute coordinates) 1000 Necessary time to achieve (log) fixed evaluation No.1 fixed evaluation No. self-organized evaluation No.1 self-organized evaluation No. 6. Conclusion We have proposed an unsupervised learning method by which a machine forms an appropriate evaluation function and a series of motion through its experiences. We confirmed in a simulation that the evaluation function obtained by the learning method was reflected the characteristic of the environment. Then the became to be able to move along with the optimal route. 100 0 10000 0000 0000 Experiences fig. 15 Change of the necessary time to get the Reference [1]Skinner B.F., "Cumulative Record", Appleton-Century-Crofts, 1961 []Yoda,H., Miyatake,T. and Matsushima, H., : "A Kinematic Model with a Self-Learning Capability of a Series of Motions Based on Instinctive Incentive", Trans. of IEICE, J7-D-II, 7,107-10, 1990 []Nakano, K., Doya, K., SICE'8, S1-, 198 (in japanese) []Rumelhart D.E., Hinton G.E. and Williams R.J.: "Learning representations by back-propagating errors", Nature,, 9, 5-56, 1986 170 10