Planning & Decision-making in Robotics Search Algorithms: Markov Property, Dependent vs. Independent variables, Dominance relationship
|
|
- Christian Wilkins
- 5 years ago
- Views:
Transcription
1 6-782 Planning & Decision-making in Robotics Search Algorithms: Markov Property, Dependent vs. Independent variables, Dominance relationship Maxim Likhachev Robotics Institute Carnegie Mellon University
2 Consider Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L What should be the variables defining each state (i.e., dimensions of the search)? Carnegie Mellon University 2
3 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Planning needs to be in (x,y,l), where l is the remaining battery level x,y,l =l start - S x 2,y 2,l 2 =l start - x i,y i,l i =0 Carnegie Mellon University 3 S i states with battery level 0 have no successors
4 Markov Property Cost and Set of Successors needs to depend ONLY on the current state (no dependence on the history of the path leading up to it!) c(s 0,s ) S for all states s: succ(s) = function of s for all s in succ(s): c(s,s )= function of s, s S 0 c(s 0,s 2 ) Carnegie Mellon University 4
5 Markov Property Cost and Set of Successors needs to depend ONLY on the current state (no dependence on the history of the path leading up to it!) c(s 0,s ) S for all states s: succ(s) = function of s for all s in succ(s): c(s,s )= function of s, s S 0 c(s 0,s 2 ) Clearly true in an explicit (given) graph Can be violated in implicit (dynamically generated) graphs, where succ(s) and c(s,s ) are computed on-the-fly as a function of s, when using dependent variables Carnegie Mellon University 5
6 Independent vs. Dependent Variables X(s) variables associated with s X(s) = {X ind (s), X dep (s)} X ind (s) independent variables X dep (s) dependent variables X dep (s) = end-effector pose X ind (s) =joint angles Independent Variables are used to define state s two states s and s are considered to be the same state if and only if X ind (s) = X ind (s ) Dependent Variables often used to help with computing cost or list of successor states if for all s, X dep (s) = f(x ind (s)) (that is, only depends on independent variables, then Markov Property holds true) Often however, developers suggest to compute X dep (s) based on the path leading up to X ind (s) Carnegie Mellon University 6
7 Consider Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Carnegie Mellon University 7
8 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level 20 x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 S goal What is l(s 4 ) at the time s 4 gets expanded? What does it break? Carnegie Mellon University 8
9 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level 20 x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 S goal How does this graph look if X ind = (x,y,l)? Carnegie Mellon University 9
10 Consider Planning with Constraints on Rate of Turning Suppose we are planning 2D (x,y) path for a ground robot and constraining its heading to change by at most 45 degrees at each timestep based on the previous transition - Consider X ind =(x,y), X dep =(Ѳ), where Ѳ is robot s heading Example of incompleteness? Carnegie Mellon University 0
11 Consider Planning with Continuous (x,y,ѳ) Suppose we are planning 3D (x,y,ѳ) path for a ground robot but we don t have motion primitives (lattice) that move the robot exactly between the centers of 3D cells - Consider X ind =(x disc,y disc,ѳ disc ), X dep =(x cont,y cont,ѳ cont ), where X dep keeps track of the continuous robot pose along its path [Barraquand, J. & Latombe, 93] Example of incompleteness? Carnegie Mellon University
12 Consider Planning in Dynamic Environments Suppose we are planning a path among moving obstacles - want a collision-free path to s goal - want to minimize some cost function associated with each transition - Consider X ind =(robot pose), X dep =(t), where t is time Example of incompleteness? Carnegie Mellon University 2
13 Consider Planning in Dynamic Environments Suppose we are planning a path among moving obstacles - want a collision-free path to s goal - assume cost function is time - Consider X ind =(robot pose), X dep =(t), where t is time Is it incomplete? Carnegie Mellon University 3
14 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Is it incomplete? Carnegie Mellon University 4
15 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - x 3,y 3,l 3 =l start -2 x 4,y 4,l 4 =? S S 3 x 2,y 2,l 2 =l start - S 4 S goal Carnegie Mellon University 5
16 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - x 3,y 3,l 3 =l start -2 x 4,y 4,l 4 =l start -2 S x 2,y 2,l 2 =l start - S 3 S 4 In fact, don t even need to have variable l since it is the same as g(s)! You can just plan in (x,y) S goal Carnegie Mellon University 6
17 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S 3 x 4,y 4,l 4 =l start -2 S 4 In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 7
18 Back to Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 8
19 Back to Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free Assuming we path are to running s goal = (x optimal goal, y goal search ) - assume cost function is battery (such as consumption A*). - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 9
20 Back to Planning with Battery Constraint Suppose we are What planning happens 2D if we (x,y) are running path for UAV x start,y start,l start =L S start - want a collision-free Assuming we path are to running s goal = (x optimal goal, y goal search ) - assume cost function is battery (such as consumption A*). - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - suboptimal search such as weighted A*? In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 20
21 Dominance Relationship x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject What to the are trajectory the general being conditions feasible given for pruning the UAV dominated battery level states? L - Consider X ind =(x,y,l) x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 Do we really need two copies of s 4? One with l 4 =l start -2 and one with l 4 =l start -3 S goal Carnegie Mellon University 2
22 Dominance Relationship x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject What to the are trajectory the general being conditions feasible given for pruning the UAV dominated battery level states? L - Consider X ind =(x,y,l) if (g(s) g(s ) and s dominates s ), then s can be pruned by search s dominates s implies s cannot be part of a solution that is better than the solution from s x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 Do we really need two copies of s 4? One with l 4 =l start -2 and one with l 4 =l start -3 S goal Carnegie Mellon University 22
23 A* Search with Dominance Check Main function g(s start ) = 0; all other g-values are infinite; OPEN = {s start }; ComputePath(); publish solution; ComputePath function while(s goal is not expanded and OPEN 0) remove s with the smallest [f(s) = g(s)+h(s)] from OPEN; insert s into CLOSED; for every successor s of s such that s not in CLOSED if g(s ) > g(s) + c(s,s ) g(s ) = g(s) + c(s,s ); if there exists state s such that (g(s ) g(s ) AND s dominates s ) continue; //skip inserting state s into OPEN, i.e., prune insert s into OPEN; Carnegie Mellon University 23
24 What You Should Know Dependent vs. Independent variables. Definition of Markov Property The definition and the use of Dominance relationship Carnegie Mellon University 24
Incremental Path Planning
Incremental Path Planning Dynamic and Incremental A* Prof. Brian Williams (help from Ihsiang Shu) 6./6.8 Cognitive Robotics February 7 th, Outline Review: Optimal Path Planning in Partially Known Environments.
More informationOptimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes
Optimization Charles J. Geyer School of Statistics University of Minnesota Stat 8054 Lecture Notes 1 One-Dimensional Optimization Look at a graph. Grid search. 2 One-Dimensional Zero Finding Zero finding
More informationNavigation. Global Pathing. The Idea. Diagram I: Overall Navigation
Navigation Diagram I: Overall Navigation Global Pathing The Idea (focus on 2D coordinates Takes advantage of prior info: navigation space/ dimensions target destination location Risks: Map Resolution too
More information1 Implicit Differentiation
1 Implicit Differentiation In logarithmic differentiation, we begin with an equation y = f(x) and then take the logarithm of both sides to get ln y = ln f(x). In this equation, y is not explicitly expressed
More information16.410/413 Principles of Autonomy and Decision Making
6.4/43 Principles of Autonomy and Decision Making Lecture 8: (Mixed-Integer) Linear Programming for Vehicle Routing and Motion Planning Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute
More informationCS Algorithms and Complexity
CS 50 - Algorithms and Complexity Linear Programming, the Simplex Method, and Hard Problems Sean Anderson 2/15/18 Portland State University Table of contents 1. The Simplex Method 2. The Graph Problem
More informationDD* Lite: Efficient Incremental Search with State Dominance
DD* Lite: Efficient Incremental Search with State Dominance G. Ayorkor Mills-Tettey Anthony Stentz M. Bernardine Dias CMU-RI-TR-07-12 May 2007 Robotics Institute Carnegie Mellon University Pittsburgh,
More informationReinforcement Learning
Reinforcement Learning Ron Parr CompSci 7 Department of Computer Science Duke University With thanks to Kris Hauser for some content RL Highlights Everybody likes to learn from experience Use ML techniques
More informationDecision Trees.
. Machine Learning Decision Trees Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de
More informationSolutions Definition 2: a solution
Solutions As was stated before, one of the goals in this course is to solve, or find solutions of differential equations. In the next definition we consider the concept of a solution of an ordinary differential
More informationProcedure for sketching bode plots (mentioned on Oct 5 th notes, Pg. 20)
Procedure for sketching bode plots (mentioned on Oct 5 th notes, Pg. 20) 1. Rewrite the transfer function in proper p form. 2. Separate the transfer function into its constituent parts. 3. Draw the Bode
More informationPlanning in Markov Decision Processes
Carnegie Mellon School of Computer Science Deep Reinforcement Learning and Control Planning in Markov Decision Processes Lecture 3, CMU 10703 Katerina Fragkiadaki Markov Decision Process (MDP) A Markov
More informationPartially Observable Markov Decision Processes (POMDPs)
Partially Observable Markov Decision Processes (POMDPs) Sachin Patil Guest Lecture: CS287 Advanced Robotics Slides adapted from Pieter Abbeel, Alex Lee Outline Introduction to POMDPs Locally Optimal Solutions
More informationInference' Sampling'Methods' Probabilis5c' Graphical' Models' Inference'In' Template' Models' Daphne Koller
Probabilis5c' Graphical' Models' Inference' Sampling'Methods' Inference'In' Template' Models' DBN Template Specification Weather 0 Velocity 0 Location 0 Weather Velocity Location Weather Velocity Location
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
More informationSufficient Conditions for the Existence of Resolution Complete Planning Algorithms
Sufficient Conditions for the Existence of Resolution Complete Planning Algorithms Dmitry Yershov and Steve LaValle Computer Science niversity of Illinois at rbana-champaign WAFR 2010 December 15, 2010
More informationDecision Trees.
. Machine Learning Decision Trees Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de
More informationCS221 Practice Midterm
CS221 Practice Midterm Autumn 2012 1 ther Midterms The following pages are excerpts from similar classes midterms. The content is similar to what we ve been covering this quarter, so that it should be
More informationPart III: A Simplex pivot
MA 3280 Lecture 31 - More on The Simplex Method Friday, April 25, 2014. Objectives: Analyze Simplex examples. We were working on the Simplex tableau The matrix form of this system of equations is called
More informationData Structures for Efficient Inference and Optimization
Data Structures for Efficient Inference and Optimization in Expressive Continuous Domains Scott Sanner Ehsan Abbasnejad Zahra Zamani Karina Valdivia Delgado Leliane Nunes de Barros Cheng Fang Discrete
More informationThe Mean Value Theorem. Oct
The Mean Value Theorem Oct 14 2011 The Mean Value Theorem Theorem Suppose that f is defined and continuous on a closed interval [a, b], and suppose that f exists on the open interval (a, b). Then there
More informationFundamentals of Metaheuristics
Fundamentals of Metaheuristics Part I - Basic concepts and Single-State Methods A seminar for Neural Networks Simone Scardapane Academic year 2012-2013 ABOUT THIS SEMINAR The seminar is divided in three
More informationCS357: CTL Model Checking (two lectures worth) David Dill
CS357: CTL Model Checking (two lectures worth) David Dill 1 CTL CTL = Computation Tree Logic It is a propositional temporal logic temporal logic extended to properties of events over time. CTL is a branching
More informationMotivation for introducing probabilities
for introducing probabilities Reaching the goals is often not sufficient: it is important that the expected costs do not outweigh the benefit of reaching the goals. 1 Objective: maximize benefits - costs.
More informationA Gentle Introduction to Reinforcement Learning
A Gentle Introduction to Reinforcement Learning Alexander Jung 2018 1 Introduction and Motivation Consider the cleaning robot Rumba which has to clean the office room B329. In order to keep things simple,
More informationCS 4100 // artificial intelligence. Recap/midterm review!
CS 4100 // artificial intelligence instructor: byron wallace Recap/midterm review! Attribution: many of these slides are modified versions of those distributed with the UC Berkeley CS188 materials Thanks
More informationWhat if There Were No Law of Large Numbers?
What if There Were No Law of Large Numbers? We said that the Law of Large Numbers applies whenever we make independent observations on a random variable X that has an expected value. In those cases the
More informationAn Explicit Characterization of Minimum Wheel-Rotation Paths for Differential-Drives
An Explicit Characterization of Minimum Wheel-Rotation Paths for Differential-Drives Hamidreza Chitsaz 1, Steven M. LaValle 1, Devin J. Balkcom, and Matthew T. Mason 3 1 Department of Computer Science
More informationAn Adaptive Clustering Method for Model-free Reinforcement Learning
An Adaptive Clustering Method for Model-free Reinforcement Learning Andreas Matt and Georg Regensburger Institute of Mathematics University of Innsbruck, Austria {andreas.matt, georg.regensburger}@uibk.ac.at
More informationImplicitely and Densely Discrete Black-Box Optimization Problems
Implicitely and Densely Discrete Black-Box Optimization Problems L. N. Vicente September 26, 2008 Abstract This paper addresses derivative-free optimization problems where the variables lie implicitly
More informationReasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence
Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence 1 Outline Reasoning under uncertainty over time Hidden Markov Models Dynamic Bayes Nets 2 Introduction So far we
More informationCMU Lecture 4: Informed Search. Teacher: Gianni A. Di Caro
CMU 15-781 Lecture 4: Informed Search Teacher: Gianni A. Di Caro UNINFORMED VS. INFORMED Uninformed Can only generate successors and distinguish goals from non-goals Informed Strategies that can distinguish
More informationReinforcement Learning
CS7/CS7 Fall 005 Supervised Learning: Training examples: (x,y) Direct feedback y for each input x Sequence of decisions with eventual feedback No teacher that critiques individual actions Learn to act
More informationExample I: Capital Accumulation
1 Example I: Capital Accumulation Time t = 0, 1,..., T < Output y, initial output y 0 Fraction of output invested a, capital k = ay Transition (production function) y = g(k) = g(ay) Reward (utility of
More informationIntroduction to Reinforcement Learning
CSCI-699: Advanced Topics in Deep Learning 01/16/2019 Nitin Kamra Spring 2019 Introduction to Reinforcement Learning 1 What is Reinforcement Learning? So far we have seen unsupervised and supervised learning.
More informationf (1 0.5)/n Z =
Math 466/566 - Homework 4. We want to test a hypothesis involving a population proportion. The unknown population proportion is p. The null hypothesis is p = / and the alternative hypothesis is p > /.
More informationWhere are we in CS 440?
Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning robability: Review of main concepts Chapter 3 Making decisions under uncertainty
More informationRobotics. Path Planning. University of Stuttgart Winter 2018/19
Robotics Path Planning Path finding vs. trajectory optimization, local vs. global, Dijkstra, Probabilistic Roadmaps, Rapidly Exploring Random Trees, non-holonomic systems, car system equation, path-finding
More informationDistributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London
Distributed Data Fusion with Kalman Filters Simon Julier Computer Science Department University College London S.Julier@cs.ucl.ac.uk Structure of Talk Motivation Kalman Filters Double Counting Optimal
More informationSince the two-sided limits exist, so do all one-sided limits. In particular:
SECTION 3.6 IN-SECTION EXERCISES: EXERCISE 1. The Intermediate Value Theorem 1. There are many correct graphs possible. A few are shown below. Since f is continuous on [a, b] and π is between f(a) = 3
More informationAnytime Planning for Decentralized Multi-Robot Active Information Gathering
Anytime Planning for Decentralized Multi-Robot Active Information Gathering Brent Schlotfeldt 1 Dinesh Thakur 1 Nikolay Atanasov 2 Vijay Kumar 1 George Pappas 1 1 GRASP Laboratory University of Pennsylvania
More informationDefinition: A "system" of equations is a set or collection of equations that you deal with all together at once.
System of Equations Definition: A "system" of equations is a set or collection of equations that you deal with all together at once. There is both an x and y value that needs to be solved for Systems
More informationAbstractions and Decision Procedures for Effective Software Model Checking
Abstractions and Decision Procedures for Effective Software Model Checking Prof. Natasha Sharygina The University of Lugano, Carnegie Mellon University Microsoft Summer School, Moscow, July 2011 Lecture
More informationFocused Real-Time Dynamic Programming for MDPs: Squeezing More Out of a Heuristic
Focused Real-Time Dynamic Programming for MDPs: Squeezing More Out of a Heuristic Trey Smith and Reid Simmons Robotics Institute, Carnegie Mellon University {trey,reids}@ri.cmu.edu Abstract Real-time dynamic
More informationCS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University
CS 322: (Social and Inormation) Network Analysis Jure Leskovec Stanord University Initially some nodes S are active Each edge (a,b) has probability (weight) p ab b 0.4 0.4 0.2 a 0.4 0.2 0.4 g 0.2 Node
More informationReinforcement Learning. Spring 2018 Defining MDPs, Planning
Reinforcement Learning Spring 2018 Defining MDPs, Planning understandability 0 Slide 10 time You are here Markov Process Where you will go depends only on where you are Markov Process: Information state
More informationMonochromatic triangles in three-coloured graphs
Charles University, Prague 10th September 2012 Joint work with James Cummings (Carnegie Mellon), Daniel Král (Charles University), Florian Pfender (Colorado), Konrad Sperfeld (Universität Rostock) and
More informationAn Application to Growth Theory
An Application to Growth Theory First let s review the concepts of solution function and value function for a maximization problem. Suppose we have the problem max F (x, α) subject to G(x, β) 0, (P) x
More informationProgramming Robots in ROS Slides adapted from CMU, Harvard and TU Stuttgart
Programming Robots in ROS Slides adapted from CMU, Harvard and TU Stuttgart Path Planning Problem Given an initial configuration q_start and a goal configuration q_goal,, we must generate the best continuous
More information1 Trajectory Generation
CS 685 notes, J. Košecká 1 Trajectory Generation The material for these notes has been adopted from: John J. Craig: Robotics: Mechanics and Control. This example assumes that we have a starting position
More informationDynamic location, euclidean center problem, global optimization.
Annals of Operations Research 6(1986)3 13-3 19 N. MEGIDDO IBM Almaden Research Center, San Jose, California 95120, USA, and Tel-Aviv University, Tel-A viv, Israel Abstract A class of dynamic location problems
More informationMachine Learning I Reinforcement Learning
Machine Learning I Reinforcement Learning Thomas Rückstieß Technische Universität München December 17/18, 2009 Literature Book: Reinforcement Learning: An Introduction Sutton & Barto (free online version:
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University
More informationLOCAL NAVIGATION. Dynamic adaptation of global plan to local conditions A.K.A. local collision avoidance and pedestrian models
LOCAL NAVIGATION 1 LOCAL NAVIGATION Dynamic adaptation of global plan to local conditions A.K.A. local collision avoidance and pedestrian models 2 LOCAL NAVIGATION Why do it? Could we use global motion
More informationMath 310 Introduction to Ordinary Differential Equations Final Examination August 9, Instructor: John Stockie
Make sure this exam has 15 pages. Math 310 Introduction to Ordinary Differential Equations inal Examination August 9, 2006 Instructor: John Stockie Name: (Please Print) Student Number: Special Instructions
More informationSection Handout 5 Solutions
CS 188 Spring 2019 Section Handout 5 Solutions Approximate Q-Learning With feature vectors, we can treat values of states and q-states as linear value functions: V (s) = w 1 f 1 (s) + w 2 f 2 (s) +...
More informationWhere are we in CS 440?
Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning robability: Review of main concepts Chapter 3 Motivation: lanning under uncertainty
More informationMath Day at the Beach 2016
Multiple Choice Write your name and school and mark your answers on the answer sheet. You have 30 minutes to work on these problems. No calculator is allowed. 1. What is the median of the following five
More informationCourse on Automated Planning: Transformations
Course on Automated Planning: Transformations Hector Geffner ICREA & Universitat Pompeu Fabra Barcelona, Spain H. Geffner, Course on Automated Planning, Rome, 7/2010 1 AI Planning: Status The good news:
More informationIntroduction to Spring 2009 Artificial Intelligence Midterm Exam
S 188 Introduction to Spring 009 rtificial Intelligence Midterm Exam INSTRUTINS You have 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators
More informationImplicit Differentiation, Related Rates. Goals: Introduce implicit differentiation. Study problems involving related rates.
Unit #5 : Implicit Differentiation, Related Rates Goals: Introduce implicit differentiation. Study problems involving related rates. Tangent Lines to Relations - Implicit Differentiation - 1 Implicit Differentiation
More informationSymbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad
Symbolic Variable Elimination in Discrete and Continuous Graphical Models Scott Sanner Ehsan Abbasnejad Inference for Dynamic Tracking No one previously did this inference exactly in closed-form! Exact
More informationPotential Field Methods
Randomized Motion Planning Nancy Amato Fall 04, Univ. of Padova Potential Field Methods [ ] Potential Field Methods Acknowledgement: Parts of these course notes are based on notes from courses given by
More information1 The Existence and Uniqueness Theorem for First-Order Differential Equations
1 The Existence and Uniqueness Theorem for First-Order Differential Equations Let I R be an open interval and G R n, n 1, be a domain. Definition 1.1 Let us consider a function f : I G R n. The general
More informationThe Implicit Function Theorem
The Implicit Function Theorem Case 1: A linear equation with m = n = 1 (We ll say what m and n are shortly.) Suppose we know that x and y must always satisfy the equation ax + by = c. (1) Let s write the
More informationReinforcement Learning and Control
CS9 Lecture notes Andrew Ng Part XIII Reinforcement Learning and Control We now begin our study of reinforcement learning and adaptive control. In supervised learning, we saw algorithms that tried to make
More informationGibbs Fields & Markov Random Fields
Statistical Techniques in Robotics (16-831, F10) Lecture#7 (Tuesday September 21) Gibbs Fields & Markov Random Fields Lecturer: Drew Bagnell Scribe: Bradford Neuman 1 1 Gibbs Fields Like a Bayes Net, a
More informationLesson 3: Using Linear Combinations to Solve a System of Equations
Lesson 3: Using Linear Combinations to Solve a System of Equations Steps for Using Linear Combinations to Solve a System of Equations 1. 2. 3. 4. 5. Example 1 Solve the following system using the linear
More informationLecture 11. Linear Momentum and Impulse. Collisions.
Lecture 11 Linear Momentum and Impulse. Collisions. Momentum and Newton s Second Law F net = m a= m Δ v Δ t = Δ (m v ) Δ t = Δ p Δ t Linear momentum p = m v Newton s second law in terms of linear momentum:
More informationArtificial Intelligence
Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Summary of last lecture We know how to do probabilistic reasoning over time transition model P(X t
More informationRobotics. Path Planning. Marc Toussaint U Stuttgart
Robotics Path Planning Path finding vs. trajectory optimization, local vs. global, Dijkstra, Probabilistic Roadmaps, Rapidly Exploring Random Trees, non-holonomic systems, car system equation, path-finding
More informationBayesian Interpretations of Regularization
Bayesian Interpretations of Regularization Charlie Frogner 9.50 Class 15 April 1, 009 The Plan Regularized least squares maps {(x i, y i )} n i=1 to a function that minimizes the regularized loss: f S
More informationPGSS Discrete Math Solutions to Problem Set #4. Note: signifies the end of a problem, and signifies the end of a proof.
PGSS Discrete Math Solutions to Problem Set #4 Note: signifies the end of a problem, and signifies the end of a proof. 1. Prove that for any k N, there are k consecutive composite numbers. (Hint: (k +
More informationEcon Slides from Lecture 10
Econ 205 Sobel Econ 205 - Slides from Lecture 10 Joel Sobel September 2, 2010 Example Find the tangent plane to {x x 1 x 2 x 2 3 = 6} R3 at x = (2, 5, 2). If you let f (x) = x 1 x 2 x3 2, then this is
More informationInterpolation APPLIED PROBLEMS. Reading Between the Lines FLY ROCKET FLY, FLY ROCKET FLY WHAT IS INTERPOLATION? Figure Interpolation of discrete data.
WHAT IS INTERPOLATION? Given (x 0,y 0 ), (x,y ), (x n,y n ), find the value of y at a value of x that is not given. Interpolation Reading Between the Lines Figure Interpolation of discrete data. FLY ROCKET
More informationCourse 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016
Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the
More informationSummary of the simplex method
MVE165/MMG631,Linear and integer optimization with applications The simplex method: degeneracy; unbounded solutions; starting solutions; infeasibility; alternative optimal solutions Ann-Brith Strömberg
More informationEcon 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers
ECO 504 Spring 2009 Chris Sims Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers Christopher A. Sims Princeton University sims@princeton.edu February 4, 2009 0 Example: LQPY The ordinary
More informationSequential decision making under uncertainty. Department of Computer Science, Czech Technical University in Prague
Sequential decision making under uncertainty Jiří Kléma Department of Computer Science, Czech Technical University in Prague https://cw.fel.cvut.cz/wiki/courses/b4b36zui/prednasky pagenda Previous lecture:
More informationOn the concrete complexity of the successor function
On the concrete complexity of the successor function M. Rigo joint work with V. Berthé, Ch. Frougny, J. Sakarovitch http://www.discmath.ulg.ac.be/ http://orbi.ulg.ac.be/handle/2268/30094 Let s start with
More informationUnit 7: Part 1: Sketching the Root Locus
Root Locus Unit 7: Part 1: Sketching the Root Locus Engineering 5821: Control Systems I Faculty of Engineering & Applied Science Memorial University of Newfoundland March 14, 2010 ENGI 5821 Unit 7: Root
More informationHW9 Concepts. Alex Alemi November 1, 2009
HW9 Concepts Alex Alemi November 1, 2009 1 24.28 Capacitor Energy You are told to consider connecting a charged capacitor together with an uncharged one and told to compute (a) the original charge, (b)
More informationData Flow Analysis. Lecture 6 ECS 240. ECS 240 Data Flow Analysis 1
Data Flow Analysis Lecture 6 ECS 240 ECS 240 Data Flow Analysis 1 The Plan Introduce a few example analyses Generalize to see the underlying theory Discuss some more advanced issues ECS 240 Data Flow Analysis
More informationAndrew/CS ID: Midterm Solutions, Fall 2006
Name: Andrew/CS ID: 15-780 Midterm Solutions, Fall 2006 November 15, 2006 Place your name and your andrew/cs email address on the front page. The exam is open-book, open-notes, no electronics other than
More information2.7 Implicit Differentiation and Related Rates Math 125
.7 IMPLICIT DIFFERENTIATION AND RELATED RATES Implicit and Explicit Functions Suppose your boss says, I have had it with your incompetence. You ve screwed up everything we ve ever given you to do! The
More informationdy dx dx dx as a BC Calculus 1 The Chain Rule is notation for a which says that we have the
2.4 2.6 BC Calculus 1 The Chain Rule dy is notation for a which says that we have the for an expression set equal to (the dependent variable), where the variable is x. This is read dee why, dee or the
More informationThe -Calculus & Enumeration Operators
The -Calculus & Enumeration Operators Dana S. Scott University Professor Emeritus Carnegie Mellon University Visiting Scholar University of California, Berkeley dana.scott@cs.cmu.edu Version of 2 October
More informationReview of Optimization Methods
Review of Optimization Methods Prof. Manuela Pedio 20550 Quantitative Methods for Finance August 2018 Outline of the Course Lectures 1 and 2 (3 hours, in class): Linear and non-linear functions on Limits,
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 4, 2015 Today: Generative discriminative classifiers Linear regression Decomposition of error into
More informationCS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C.
CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring 2006 Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C. Latombe Scribe: Neda Nategh How do you update the energy function during the
More informationThe exam is closed book, closed calculator, and closed notes except your one-page crib sheet.
CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark
More informationLecture Support Vector Machine (SVM) Classifiers
Introduction to Machine Learning Lecturer: Amir Globerson Lecture 6 Fall Semester Scribe: Yishay Mansour 6.1 Support Vector Machine (SVM) Classifiers Classification is one of the most important tasks in
More informationCSE 331 Winter 2018 Homework 1
Directions: - Due Wednesday, January 10 by 11 pm. - Turn in your work online using gradescope. You should turn in a single pdf file. You can have more than one answer per page, but please try to avoid
More informationA Decentralized Approach to Multi-agent Planning in the Presence of Constraints and Uncertainty
2011 IEEE International Conference on Robotics and Automation Shanghai International Conference Center May 9-13, 2011, Shanghai, China A Decentralized Approach to Multi-agent Planning in the Presence of
More informationb) since the remainder is 0 I need to factor the numerator. Synthetic division tells me this is true
Section 5.2 solutions #1-10: a) Perform the division using synthetic division. b) if the remainder is 0 use the result to completely factor the dividend (this is the numerator or the polynomial to the
More informationEmbedded Systems - FS 2018
Institut für Technische Informatik und Kommunikationsnetze Prof. L. Thiele Embedded Systems - FS 2018 Sample solution to Exercise 3 Discussion Date: 11.4.2018 Aperiodic Scheduling Task 1: Earliest Deadline
More informationMachine Learning
Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 13, 2011 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting
More informationIntroduction to Root Locus. What is root locus?
Introduction to Root Locus What is root locus? A graphical representation of the closed loop poles as a system parameter (Gain K) is varied Method of analysis and design for stability and transient response
More information14.1 Finding frequent elements in stream
Chapter 14 Streaming Data Model 14.1 Finding frequent elements in stream A very useful statistics for many applications is to keep track of elements that occur more frequently. It can come in many flavours
More informationReinforcement Learning II
Reinforcement Learning II Andrea Bonarini Artificial Intelligence and Robotics Lab Department of Electronics and Information Politecnico di Milano E-mail: bonarini@elet.polimi.it URL:http://www.dei.polimi.it/people/bonarini
More information