Planning & Decision-making in Robotics Search Algorithms: Markov Property, Dependent vs. Independent variables, Dominance relationship

Size: px
Start display at page:

Download "Planning & Decision-making in Robotics Search Algorithms: Markov Property, Dependent vs. Independent variables, Dominance relationship"

Transcription

1 6-782 Planning & Decision-making in Robotics Search Algorithms: Markov Property, Dependent vs. Independent variables, Dominance relationship Maxim Likhachev Robotics Institute Carnegie Mellon University

2 Consider Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L What should be the variables defining each state (i.e., dimensions of the search)? Carnegie Mellon University 2

3 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Planning needs to be in (x,y,l), where l is the remaining battery level x,y,l =l start - S x 2,y 2,l 2 =l start - x i,y i,l i =0 Carnegie Mellon University 3 S i states with battery level 0 have no successors

4 Markov Property Cost and Set of Successors needs to depend ONLY on the current state (no dependence on the history of the path leading up to it!) c(s 0,s ) S for all states s: succ(s) = function of s for all s in succ(s): c(s,s )= function of s, s S 0 c(s 0,s 2 ) Carnegie Mellon University 4

5 Markov Property Cost and Set of Successors needs to depend ONLY on the current state (no dependence on the history of the path leading up to it!) c(s 0,s ) S for all states s: succ(s) = function of s for all s in succ(s): c(s,s )= function of s, s S 0 c(s 0,s 2 ) Clearly true in an explicit (given) graph Can be violated in implicit (dynamically generated) graphs, where succ(s) and c(s,s ) are computed on-the-fly as a function of s, when using dependent variables Carnegie Mellon University 5

6 Independent vs. Dependent Variables X(s) variables associated with s X(s) = {X ind (s), X dep (s)} X ind (s) independent variables X dep (s) dependent variables X dep (s) = end-effector pose X ind (s) =joint angles Independent Variables are used to define state s two states s and s are considered to be the same state if and only if X ind (s) = X ind (s ) Dependent Variables often used to help with computing cost or list of successor states if for all s, X dep (s) = f(x ind (s)) (that is, only depends on independent variables, then Markov Property holds true) Often however, developers suggest to compute X dep (s) based on the path leading up to X ind (s) Carnegie Mellon University 6

7 Consider Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Carnegie Mellon University 7

8 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level 20 x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 S goal What is l(s 4 ) at the time s 4 gets expanded? What does it break? Carnegie Mellon University 8

9 Consider Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level 20 x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 S goal How does this graph look if X ind = (x,y,l)? Carnegie Mellon University 9

10 Consider Planning with Constraints on Rate of Turning Suppose we are planning 2D (x,y) path for a ground robot and constraining its heading to change by at most 45 degrees at each timestep based on the previous transition - Consider X ind =(x,y), X dep =(Ѳ), where Ѳ is robot s heading Example of incompleteness? Carnegie Mellon University 0

11 Consider Planning with Continuous (x,y,ѳ) Suppose we are planning 3D (x,y,ѳ) path for a ground robot but we don t have motion primitives (lattice) that move the robot exactly between the centers of 3D cells - Consider X ind =(x disc,y disc,ѳ disc ), X dep =(x cont,y cont,ѳ cont ), where X dep keeps track of the continuous robot pose along its path [Barraquand, J. & Latombe, 93] Example of incompleteness? Carnegie Mellon University

12 Consider Planning in Dynamic Environments Suppose we are planning a path among moving obstacles - want a collision-free path to s goal - want to minimize some cost function associated with each transition - Consider X ind =(robot pose), X dep =(t), where t is time Example of incompleteness? Carnegie Mellon University 2

13 Consider Planning in Dynamic Environments Suppose we are planning a path among moving obstacles - want a collision-free path to s goal - assume cost function is time - Consider X ind =(robot pose), X dep =(t), where t is time Is it incomplete? Carnegie Mellon University 3

14 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Is it incomplete? Carnegie Mellon University 4

15 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - x 3,y 3,l 3 =l start -2 x 4,y 4,l 4 =? S S 3 x 2,y 2,l 2 =l start - S 4 S goal Carnegie Mellon University 5

16 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - x 3,y 3,l 3 =l start -2 x 4,y 4,l 4 =l start -2 S x 2,y 2,l 2 =l start - S 3 S 4 In fact, don t even need to have variable l since it is the same as g(s)! You can just plan in (x,y) S goal Carnegie Mellon University 6

17 Back to Planning with Battery Constraint - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level x start,y start,l start =L S start x,y,l =l start - S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S 3 x 4,y 4,l 4 =l start -2 S 4 In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 7

18 Back to Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - assume cost function is battery consumption - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 8

19 Back to Planning with Battery Constraint x start,y start,l start =L S start - want a collision-free Assuming we path are to running s goal = (x optimal goal, y goal search ) - assume cost function is battery (such as consumption A*). - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 9

20 Back to Planning with Battery Constraint Suppose we are What planning happens 2D if we (x,y) are running path for UAV x start,y start,l start =L S start - want a collision-free Assuming we path are to running s goal = (x optimal goal, y goal search ) - assume cost function is battery (such as consumption A*). - subject to the trajectory being feasible given the UAV battery level L - Consider X ind =(x,y), X dep =(l), where l is the remaining battery level Whenever you can guarantee that for any state s: if we have two paths π (s start,s) and π 2 (s start,s) s.t. c(π ) c(π 2 ), then it implies that c (s,s ) c 2 (s,s ), x where,y,l =l c i (s,s ) start - cost x 3,y of 3,l a 3 =l least-cost start -2 path x 4,y 4 from,l 4 =l start s to -2 s after s is reached from s S S start via path π 3 S i 4 x 2,y 2,l 2 =l start - suboptimal search such as weighted A*? In general, when is it OK to use X dep (s) in determining succ(s) and edgecosts? S goal Carnegie Mellon University 20

21 Dominance Relationship x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject What to the are trajectory the general being conditions feasible given for pruning the UAV dominated battery level states? L - Consider X ind =(x,y,l) x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 Do we really need two copies of s 4? One with l 4 =l start -2 and one with l 4 =l start -3 S goal Carnegie Mellon University 2

22 Dominance Relationship x start,y start,l start =L S start - want a collision-free path to s goal = (x goal, y goal ) - want to minimize some cost function associated with each transition (for example, minimize the risk of flying close to people) - subject What to the are trajectory the general being conditions feasible given for pruning the UAV dominated battery level states? L - Consider X ind =(x,y,l) if (g(s) g(s ) and s dominates s ), then s can be pruned by search s dominates s implies s cannot be part of a solution that is better than the solution from s x,y,l =l start - 2 S x 2,y 2,l 2 =l start - x 3,y 3,l 3 =l start -2 S x 4,y 4,l 4 =? S 4 Do we really need two copies of s 4? One with l 4 =l start -2 and one with l 4 =l start -3 S goal Carnegie Mellon University 22

23 A* Search with Dominance Check Main function g(s start ) = 0; all other g-values are infinite; OPEN = {s start }; ComputePath(); publish solution; ComputePath function while(s goal is not expanded and OPEN 0) remove s with the smallest [f(s) = g(s)+h(s)] from OPEN; insert s into CLOSED; for every successor s of s such that s not in CLOSED if g(s ) > g(s) + c(s,s ) g(s ) = g(s) + c(s,s ); if there exists state s such that (g(s ) g(s ) AND s dominates s ) continue; //skip inserting state s into OPEN, i.e., prune insert s into OPEN; Carnegie Mellon University 23

24 What You Should Know Dependent vs. Independent variables. Definition of Markov Property The definition and the use of Dominance relationship Carnegie Mellon University 24

Incremental Path Planning

Incremental Path Planning Incremental Path Planning Dynamic and Incremental A* Prof. Brian Williams (help from Ihsiang Shu) 6./6.8 Cognitive Robotics February 7 th, Outline Review: Optimal Path Planning in Partially Known Environments.

More information

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes Optimization Charles J. Geyer School of Statistics University of Minnesota Stat 8054 Lecture Notes 1 One-Dimensional Optimization Look at a graph. Grid search. 2 One-Dimensional Zero Finding Zero finding

More information

Navigation. Global Pathing. The Idea. Diagram I: Overall Navigation

Navigation. Global Pathing. The Idea. Diagram I: Overall Navigation Navigation Diagram I: Overall Navigation Global Pathing The Idea (focus on 2D coordinates Takes advantage of prior info: navigation space/ dimensions target destination location Risks: Map Resolution too

More information

1 Implicit Differentiation

1 Implicit Differentiation 1 Implicit Differentiation In logarithmic differentiation, we begin with an equation y = f(x) and then take the logarithm of both sides to get ln y = ln f(x). In this equation, y is not explicitly expressed

More information

16.410/413 Principles of Autonomy and Decision Making

16.410/413 Principles of Autonomy and Decision Making 6.4/43 Principles of Autonomy and Decision Making Lecture 8: (Mixed-Integer) Linear Programming for Vehicle Routing and Motion Planning Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute

More information

CS Algorithms and Complexity

CS Algorithms and Complexity CS 50 - Algorithms and Complexity Linear Programming, the Simplex Method, and Hard Problems Sean Anderson 2/15/18 Portland State University Table of contents 1. The Simplex Method 2. The Graph Problem

More information

DD* Lite: Efficient Incremental Search with State Dominance

DD* Lite: Efficient Incremental Search with State Dominance DD* Lite: Efficient Incremental Search with State Dominance G. Ayorkor Mills-Tettey Anthony Stentz M. Bernardine Dias CMU-RI-TR-07-12 May 2007 Robotics Institute Carnegie Mellon University Pittsburgh,

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning Ron Parr CompSci 7 Department of Computer Science Duke University With thanks to Kris Hauser for some content RL Highlights Everybody likes to learn from experience Use ML techniques

More information

Decision Trees.

Decision Trees. . Machine Learning Decision Trees Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de

More information

Solutions Definition 2: a solution

Solutions Definition 2: a solution Solutions As was stated before, one of the goals in this course is to solve, or find solutions of differential equations. In the next definition we consider the concept of a solution of an ordinary differential

More information

Procedure for sketching bode plots (mentioned on Oct 5 th notes, Pg. 20)

Procedure for sketching bode plots (mentioned on Oct 5 th notes, Pg. 20) Procedure for sketching bode plots (mentioned on Oct 5 th notes, Pg. 20) 1. Rewrite the transfer function in proper p form. 2. Separate the transfer function into its constituent parts. 3. Draw the Bode

More information

Planning in Markov Decision Processes

Planning in Markov Decision Processes Carnegie Mellon School of Computer Science Deep Reinforcement Learning and Control Planning in Markov Decision Processes Lecture 3, CMU 10703 Katerina Fragkiadaki Markov Decision Process (MDP) A Markov

More information

Partially Observable Markov Decision Processes (POMDPs)

Partially Observable Markov Decision Processes (POMDPs) Partially Observable Markov Decision Processes (POMDPs) Sachin Patil Guest Lecture: CS287 Advanced Robotics Slides adapted from Pieter Abbeel, Alex Lee Outline Introduction to POMDPs Locally Optimal Solutions

More information

Inference' Sampling'Methods' Probabilis5c' Graphical' Models' Inference'In' Template' Models' Daphne Koller

Inference' Sampling'Methods' Probabilis5c' Graphical' Models' Inference'In' Template' Models' Daphne Koller Probabilis5c' Graphical' Models' Inference' Sampling'Methods' Inference'In' Template' Models' DBN Template Specification Weather 0 Velocity 0 Location 0 Weather Velocity Location Weather Velocity Location

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Fall 2015 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

More information

Sufficient Conditions for the Existence of Resolution Complete Planning Algorithms

Sufficient Conditions for the Existence of Resolution Complete Planning Algorithms Sufficient Conditions for the Existence of Resolution Complete Planning Algorithms Dmitry Yershov and Steve LaValle Computer Science niversity of Illinois at rbana-champaign WAFR 2010 December 15, 2010

More information

Decision Trees.

Decision Trees. . Machine Learning Decision Trees Prof. Dr. Martin Riedmiller AG Maschinelles Lernen und Natürlichsprachliche Systeme Institut für Informatik Technische Fakultät Albert-Ludwigs-Universität Freiburg riedmiller@informatik.uni-freiburg.de

More information

CS221 Practice Midterm

CS221 Practice Midterm CS221 Practice Midterm Autumn 2012 1 ther Midterms The following pages are excerpts from similar classes midterms. The content is similar to what we ve been covering this quarter, so that it should be

More information

Part III: A Simplex pivot

Part III: A Simplex pivot MA 3280 Lecture 31 - More on The Simplex Method Friday, April 25, 2014. Objectives: Analyze Simplex examples. We were working on the Simplex tableau The matrix form of this system of equations is called

More information

Data Structures for Efficient Inference and Optimization

Data Structures for Efficient Inference and Optimization Data Structures for Efficient Inference and Optimization in Expressive Continuous Domains Scott Sanner Ehsan Abbasnejad Zahra Zamani Karina Valdivia Delgado Leliane Nunes de Barros Cheng Fang Discrete

More information

The Mean Value Theorem. Oct

The Mean Value Theorem. Oct The Mean Value Theorem Oct 14 2011 The Mean Value Theorem Theorem Suppose that f is defined and continuous on a closed interval [a, b], and suppose that f exists on the open interval (a, b). Then there

More information

Fundamentals of Metaheuristics

Fundamentals of Metaheuristics Fundamentals of Metaheuristics Part I - Basic concepts and Single-State Methods A seminar for Neural Networks Simone Scardapane Academic year 2012-2013 ABOUT THIS SEMINAR The seminar is divided in three

More information

CS357: CTL Model Checking (two lectures worth) David Dill

CS357: CTL Model Checking (two lectures worth) David Dill CS357: CTL Model Checking (two lectures worth) David Dill 1 CTL CTL = Computation Tree Logic It is a propositional temporal logic temporal logic extended to properties of events over time. CTL is a branching

More information

Motivation for introducing probabilities

Motivation for introducing probabilities for introducing probabilities Reaching the goals is often not sufficient: it is important that the expected costs do not outweigh the benefit of reaching the goals. 1 Objective: maximize benefits - costs.

More information

A Gentle Introduction to Reinforcement Learning

A Gentle Introduction to Reinforcement Learning A Gentle Introduction to Reinforcement Learning Alexander Jung 2018 1 Introduction and Motivation Consider the cleaning robot Rumba which has to clean the office room B329. In order to keep things simple,

More information

CS 4100 // artificial intelligence. Recap/midterm review!

CS 4100 // artificial intelligence. Recap/midterm review! CS 4100 // artificial intelligence instructor: byron wallace Recap/midterm review! Attribution: many of these slides are modified versions of those distributed with the UC Berkeley CS188 materials Thanks

More information

What if There Were No Law of Large Numbers?

What if There Were No Law of Large Numbers? What if There Were No Law of Large Numbers? We said that the Law of Large Numbers applies whenever we make independent observations on a random variable X that has an expected value. In those cases the

More information

An Explicit Characterization of Minimum Wheel-Rotation Paths for Differential-Drives

An Explicit Characterization of Minimum Wheel-Rotation Paths for Differential-Drives An Explicit Characterization of Minimum Wheel-Rotation Paths for Differential-Drives Hamidreza Chitsaz 1, Steven M. LaValle 1, Devin J. Balkcom, and Matthew T. Mason 3 1 Department of Computer Science

More information

An Adaptive Clustering Method for Model-free Reinforcement Learning

An Adaptive Clustering Method for Model-free Reinforcement Learning An Adaptive Clustering Method for Model-free Reinforcement Learning Andreas Matt and Georg Regensburger Institute of Mathematics University of Innsbruck, Austria {andreas.matt, georg.regensburger}@uibk.ac.at

More information

Implicitely and Densely Discrete Black-Box Optimization Problems

Implicitely and Densely Discrete Black-Box Optimization Problems Implicitely and Densely Discrete Black-Box Optimization Problems L. N. Vicente September 26, 2008 Abstract This paper addresses derivative-free optimization problems where the variables lie implicitly

More information

Reasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence

Reasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence 1 Outline Reasoning under uncertainty over time Hidden Markov Models Dynamic Bayes Nets 2 Introduction So far we

More information

CMU Lecture 4: Informed Search. Teacher: Gianni A. Di Caro

CMU Lecture 4: Informed Search. Teacher: Gianni A. Di Caro CMU 15-781 Lecture 4: Informed Search Teacher: Gianni A. Di Caro UNINFORMED VS. INFORMED Uninformed Can only generate successors and distinguish goals from non-goals Informed Strategies that can distinguish

More information

Reinforcement Learning

Reinforcement Learning CS7/CS7 Fall 005 Supervised Learning: Training examples: (x,y) Direct feedback y for each input x Sequence of decisions with eventual feedback No teacher that critiques individual actions Learn to act

More information

Example I: Capital Accumulation

Example I: Capital Accumulation 1 Example I: Capital Accumulation Time t = 0, 1,..., T < Output y, initial output y 0 Fraction of output invested a, capital k = ay Transition (production function) y = g(k) = g(ay) Reward (utility of

More information

Introduction to Reinforcement Learning

Introduction to Reinforcement Learning CSCI-699: Advanced Topics in Deep Learning 01/16/2019 Nitin Kamra Spring 2019 Introduction to Reinforcement Learning 1 What is Reinforcement Learning? So far we have seen unsupervised and supervised learning.

More information

f (1 0.5)/n Z =

f (1 0.5)/n Z = Math 466/566 - Homework 4. We want to test a hypothesis involving a population proportion. The unknown population proportion is p. The null hypothesis is p = / and the alternative hypothesis is p > /.

More information

Where are we in CS 440?

Where are we in CS 440? Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning robability: Review of main concepts Chapter 3 Making decisions under uncertainty

More information

Robotics. Path Planning. University of Stuttgart Winter 2018/19

Robotics. Path Planning. University of Stuttgart Winter 2018/19 Robotics Path Planning Path finding vs. trajectory optimization, local vs. global, Dijkstra, Probabilistic Roadmaps, Rapidly Exploring Random Trees, non-holonomic systems, car system equation, path-finding

More information

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London

Distributed Data Fusion with Kalman Filters. Simon Julier Computer Science Department University College London Distributed Data Fusion with Kalman Filters Simon Julier Computer Science Department University College London S.Julier@cs.ucl.ac.uk Structure of Talk Motivation Kalman Filters Double Counting Optimal

More information

Since the two-sided limits exist, so do all one-sided limits. In particular:

Since the two-sided limits exist, so do all one-sided limits. In particular: SECTION 3.6 IN-SECTION EXERCISES: EXERCISE 1. The Intermediate Value Theorem 1. There are many correct graphs possible. A few are shown below. Since f is continuous on [a, b] and π is between f(a) = 3

More information

Anytime Planning for Decentralized Multi-Robot Active Information Gathering

Anytime Planning for Decentralized Multi-Robot Active Information Gathering Anytime Planning for Decentralized Multi-Robot Active Information Gathering Brent Schlotfeldt 1 Dinesh Thakur 1 Nikolay Atanasov 2 Vijay Kumar 1 George Pappas 1 1 GRASP Laboratory University of Pennsylvania

More information

Definition: A "system" of equations is a set or collection of equations that you deal with all together at once.

Definition: A system of equations is a set or collection of equations that you deal with all together at once. System of Equations Definition: A "system" of equations is a set or collection of equations that you deal with all together at once. There is both an x and y value that needs to be solved for Systems

More information

Abstractions and Decision Procedures for Effective Software Model Checking

Abstractions and Decision Procedures for Effective Software Model Checking Abstractions and Decision Procedures for Effective Software Model Checking Prof. Natasha Sharygina The University of Lugano, Carnegie Mellon University Microsoft Summer School, Moscow, July 2011 Lecture

More information

Focused Real-Time Dynamic Programming for MDPs: Squeezing More Out of a Heuristic

Focused Real-Time Dynamic Programming for MDPs: Squeezing More Out of a Heuristic Focused Real-Time Dynamic Programming for MDPs: Squeezing More Out of a Heuristic Trey Smith and Reid Simmons Robotics Institute, Carnegie Mellon University {trey,reids}@ri.cmu.edu Abstract Real-time dynamic

More information

CS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University

CS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University CS 322: (Social and Inormation) Network Analysis Jure Leskovec Stanord University Initially some nodes S are active Each edge (a,b) has probability (weight) p ab b 0.4 0.4 0.2 a 0.4 0.2 0.4 g 0.2 Node

More information

Reinforcement Learning. Spring 2018 Defining MDPs, Planning

Reinforcement Learning. Spring 2018 Defining MDPs, Planning Reinforcement Learning Spring 2018 Defining MDPs, Planning understandability 0 Slide 10 time You are here Markov Process Where you will go depends only on where you are Markov Process: Information state

More information

Monochromatic triangles in three-coloured graphs

Monochromatic triangles in three-coloured graphs Charles University, Prague 10th September 2012 Joint work with James Cummings (Carnegie Mellon), Daniel Král (Charles University), Florian Pfender (Colorado), Konrad Sperfeld (Universität Rostock) and

More information

An Application to Growth Theory

An Application to Growth Theory An Application to Growth Theory First let s review the concepts of solution function and value function for a maximization problem. Suppose we have the problem max F (x, α) subject to G(x, β) 0, (P) x

More information

Programming Robots in ROS Slides adapted from CMU, Harvard and TU Stuttgart

Programming Robots in ROS Slides adapted from CMU, Harvard and TU Stuttgart Programming Robots in ROS Slides adapted from CMU, Harvard and TU Stuttgart Path Planning Problem Given an initial configuration q_start and a goal configuration q_goal,, we must generate the best continuous

More information

1 Trajectory Generation

1 Trajectory Generation CS 685 notes, J. Košecká 1 Trajectory Generation The material for these notes has been adopted from: John J. Craig: Robotics: Mechanics and Control. This example assumes that we have a starting position

More information

Dynamic location, euclidean center problem, global optimization.

Dynamic location, euclidean center problem, global optimization. Annals of Operations Research 6(1986)3 13-3 19 N. MEGIDDO IBM Almaden Research Center, San Jose, California 95120, USA, and Tel-Aviv University, Tel-A viv, Israel Abstract A class of dynamic location problems

More information

Machine Learning I Reinforcement Learning

Machine Learning I Reinforcement Learning Machine Learning I Reinforcement Learning Thomas Rückstieß Technische Universität München December 17/18, 2009 Literature Book: Reinforcement Learning: An Introduction Sutton & Barto (free online version:

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

LOCAL NAVIGATION. Dynamic adaptation of global plan to local conditions A.K.A. local collision avoidance and pedestrian models

LOCAL NAVIGATION. Dynamic adaptation of global plan to local conditions A.K.A. local collision avoidance and pedestrian models LOCAL NAVIGATION 1 LOCAL NAVIGATION Dynamic adaptation of global plan to local conditions A.K.A. local collision avoidance and pedestrian models 2 LOCAL NAVIGATION Why do it? Could we use global motion

More information

Math 310 Introduction to Ordinary Differential Equations Final Examination August 9, Instructor: John Stockie

Math 310 Introduction to Ordinary Differential Equations Final Examination August 9, Instructor: John Stockie Make sure this exam has 15 pages. Math 310 Introduction to Ordinary Differential Equations inal Examination August 9, 2006 Instructor: John Stockie Name: (Please Print) Student Number: Special Instructions

More information

Section Handout 5 Solutions

Section Handout 5 Solutions CS 188 Spring 2019 Section Handout 5 Solutions Approximate Q-Learning With feature vectors, we can treat values of states and q-states as linear value functions: V (s) = w 1 f 1 (s) + w 2 f 2 (s) +...

More information

Where are we in CS 440?

Where are we in CS 440? Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning robability: Review of main concepts Chapter 3 Motivation: lanning under uncertainty

More information

Math Day at the Beach 2016

Math Day at the Beach 2016 Multiple Choice Write your name and school and mark your answers on the answer sheet. You have 30 minutes to work on these problems. No calculator is allowed. 1. What is the median of the following five

More information

Course on Automated Planning: Transformations

Course on Automated Planning: Transformations Course on Automated Planning: Transformations Hector Geffner ICREA & Universitat Pompeu Fabra Barcelona, Spain H. Geffner, Course on Automated Planning, Rome, 7/2010 1 AI Planning: Status The good news:

More information

Introduction to Spring 2009 Artificial Intelligence Midterm Exam

Introduction to Spring 2009 Artificial Intelligence Midterm Exam S 188 Introduction to Spring 009 rtificial Intelligence Midterm Exam INSTRUTINS You have 3 hours. The exam is closed book, closed notes except a one-page crib sheet. Please use non-programmable calculators

More information

Implicit Differentiation, Related Rates. Goals: Introduce implicit differentiation. Study problems involving related rates.

Implicit Differentiation, Related Rates. Goals: Introduce implicit differentiation. Study problems involving related rates. Unit #5 : Implicit Differentiation, Related Rates Goals: Introduce implicit differentiation. Study problems involving related rates. Tangent Lines to Relations - Implicit Differentiation - 1 Implicit Differentiation

More information

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad

Symbolic Variable Elimination in Discrete and Continuous Graphical Models. Scott Sanner Ehsan Abbasnejad Symbolic Variable Elimination in Discrete and Continuous Graphical Models Scott Sanner Ehsan Abbasnejad Inference for Dynamic Tracking No one previously did this inference exactly in closed-form! Exact

More information

Potential Field Methods

Potential Field Methods Randomized Motion Planning Nancy Amato Fall 04, Univ. of Padova Potential Field Methods [ ] Potential Field Methods Acknowledgement: Parts of these course notes are based on notes from courses given by

More information

1 The Existence and Uniqueness Theorem for First-Order Differential Equations

1 The Existence and Uniqueness Theorem for First-Order Differential Equations 1 The Existence and Uniqueness Theorem for First-Order Differential Equations Let I R be an open interval and G R n, n 1, be a domain. Definition 1.1 Let us consider a function f : I G R n. The general

More information

The Implicit Function Theorem

The Implicit Function Theorem The Implicit Function Theorem Case 1: A linear equation with m = n = 1 (We ll say what m and n are shortly.) Suppose we know that x and y must always satisfy the equation ax + by = c. (1) Let s write the

More information

Reinforcement Learning and Control

Reinforcement Learning and Control CS9 Lecture notes Andrew Ng Part XIII Reinforcement Learning and Control We now begin our study of reinforcement learning and adaptive control. In supervised learning, we saw algorithms that tried to make

More information

Gibbs Fields & Markov Random Fields

Gibbs Fields & Markov Random Fields Statistical Techniques in Robotics (16-831, F10) Lecture#7 (Tuesday September 21) Gibbs Fields & Markov Random Fields Lecturer: Drew Bagnell Scribe: Bradford Neuman 1 1 Gibbs Fields Like a Bayes Net, a

More information

Lesson 3: Using Linear Combinations to Solve a System of Equations

Lesson 3: Using Linear Combinations to Solve a System of Equations Lesson 3: Using Linear Combinations to Solve a System of Equations Steps for Using Linear Combinations to Solve a System of Equations 1. 2. 3. 4. 5. Example 1 Solve the following system using the linear

More information

Lecture 11. Linear Momentum and Impulse. Collisions.

Lecture 11. Linear Momentum and Impulse. Collisions. Lecture 11 Linear Momentum and Impulse. Collisions. Momentum and Newton s Second Law F net = m a= m Δ v Δ t = Δ (m v ) Δ t = Δ p Δ t Linear momentum p = m v Newton s second law in terms of linear momentum:

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Summary of last lecture We know how to do probabilistic reasoning over time transition model P(X t

More information

Robotics. Path Planning. Marc Toussaint U Stuttgart

Robotics. Path Planning. Marc Toussaint U Stuttgart Robotics Path Planning Path finding vs. trajectory optimization, local vs. global, Dijkstra, Probabilistic Roadmaps, Rapidly Exploring Random Trees, non-holonomic systems, car system equation, path-finding

More information

Bayesian Interpretations of Regularization

Bayesian Interpretations of Regularization Bayesian Interpretations of Regularization Charlie Frogner 9.50 Class 15 April 1, 009 The Plan Regularized least squares maps {(x i, y i )} n i=1 to a function that minimizes the regularized loss: f S

More information

PGSS Discrete Math Solutions to Problem Set #4. Note: signifies the end of a problem, and signifies the end of a proof.

PGSS Discrete Math Solutions to Problem Set #4. Note: signifies the end of a problem, and signifies the end of a proof. PGSS Discrete Math Solutions to Problem Set #4 Note: signifies the end of a problem, and signifies the end of a proof. 1. Prove that for any k N, there are k consecutive composite numbers. (Hint: (k +

More information

Econ Slides from Lecture 10

Econ Slides from Lecture 10 Econ 205 Sobel Econ 205 - Slides from Lecture 10 Joel Sobel September 2, 2010 Example Find the tangent plane to {x x 1 x 2 x 2 3 = 6} R3 at x = (2, 5, 2). If you let f (x) = x 1 x 2 x3 2, then this is

More information

Interpolation APPLIED PROBLEMS. Reading Between the Lines FLY ROCKET FLY, FLY ROCKET FLY WHAT IS INTERPOLATION? Figure Interpolation of discrete data.

Interpolation APPLIED PROBLEMS. Reading Between the Lines FLY ROCKET FLY, FLY ROCKET FLY WHAT IS INTERPOLATION? Figure Interpolation of discrete data. WHAT IS INTERPOLATION? Given (x 0,y 0 ), (x,y ), (x n,y n ), find the value of y at a value of x that is not given. Interpolation Reading Between the Lines Figure Interpolation of discrete data. FLY ROCKET

More information

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016 Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the

More information

Summary of the simplex method

Summary of the simplex method MVE165/MMG631,Linear and integer optimization with applications The simplex method: degeneracy; unbounded solutions; starting solutions; infeasibility; alternative optimal solutions Ann-Brith Strömberg

More information

Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers

Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers ECO 504 Spring 2009 Chris Sims Econ 504, Lecture 1: Transversality and Stochastic Lagrange Multipliers Christopher A. Sims Princeton University sims@princeton.edu February 4, 2009 0 Example: LQPY The ordinary

More information

Sequential decision making under uncertainty. Department of Computer Science, Czech Technical University in Prague

Sequential decision making under uncertainty. Department of Computer Science, Czech Technical University in Prague Sequential decision making under uncertainty Jiří Kléma Department of Computer Science, Czech Technical University in Prague https://cw.fel.cvut.cz/wiki/courses/b4b36zui/prednasky pagenda Previous lecture:

More information

On the concrete complexity of the successor function

On the concrete complexity of the successor function On the concrete complexity of the successor function M. Rigo joint work with V. Berthé, Ch. Frougny, J. Sakarovitch http://www.discmath.ulg.ac.be/ http://orbi.ulg.ac.be/handle/2268/30094 Let s start with

More information

Unit 7: Part 1: Sketching the Root Locus

Unit 7: Part 1: Sketching the Root Locus Root Locus Unit 7: Part 1: Sketching the Root Locus Engineering 5821: Control Systems I Faculty of Engineering & Applied Science Memorial University of Newfoundland March 14, 2010 ENGI 5821 Unit 7: Root

More information

HW9 Concepts. Alex Alemi November 1, 2009

HW9 Concepts. Alex Alemi November 1, 2009 HW9 Concepts Alex Alemi November 1, 2009 1 24.28 Capacitor Energy You are told to consider connecting a charged capacitor together with an uncharged one and told to compute (a) the original charge, (b)

More information

Data Flow Analysis. Lecture 6 ECS 240. ECS 240 Data Flow Analysis 1

Data Flow Analysis. Lecture 6 ECS 240. ECS 240 Data Flow Analysis 1 Data Flow Analysis Lecture 6 ECS 240 ECS 240 Data Flow Analysis 1 The Plan Introduce a few example analyses Generalize to see the underlying theory Discuss some more advanced issues ECS 240 Data Flow Analysis

More information

Andrew/CS ID: Midterm Solutions, Fall 2006

Andrew/CS ID: Midterm Solutions, Fall 2006 Name: Andrew/CS ID: 15-780 Midterm Solutions, Fall 2006 November 15, 2006 Place your name and your andrew/cs email address on the front page. The exam is open-book, open-notes, no electronics other than

More information

2.7 Implicit Differentiation and Related Rates Math 125

2.7 Implicit Differentiation and Related Rates Math 125 .7 IMPLICIT DIFFERENTIATION AND RELATED RATES Implicit and Explicit Functions Suppose your boss says, I have had it with your incompetence. You ve screwed up everything we ve ever given you to do! The

More information

dy dx dx dx as a BC Calculus 1 The Chain Rule is notation for a which says that we have the

dy dx dx dx as a BC Calculus 1 The Chain Rule is notation for a which says that we have the 2.4 2.6 BC Calculus 1 The Chain Rule dy is notation for a which says that we have the for an expression set equal to (the dependent variable), where the variable is x. This is read dee why, dee or the

More information

The -Calculus & Enumeration Operators

The -Calculus & Enumeration Operators The -Calculus & Enumeration Operators Dana S. Scott University Professor Emeritus Carnegie Mellon University Visiting Scholar University of California, Berkeley dana.scott@cs.cmu.edu Version of 2 October

More information

Review of Optimization Methods

Review of Optimization Methods Review of Optimization Methods Prof. Manuela Pedio 20550 Quantitative Methods for Finance August 2018 Outline of the Course Lectures 1 and 2 (3 hours, in class): Linear and non-linear functions on Limits,

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 4, 2015 Today: Generative discriminative classifiers Linear regression Decomposition of error into

More information

CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C.

CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C. CS 273 Prof. Serafim Batzoglou Prof. Jean-Claude Latombe Spring 2006 Lecture 12 : Energy maintenance (1) Lecturer: Prof. J.C. Latombe Scribe: Neda Nategh How do you update the energy function during the

More information

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet.

The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. CS 188 Spring 2017 Introduction to Artificial Intelligence Midterm V2 You have approximately 80 minutes. The exam is closed book, closed calculator, and closed notes except your one-page crib sheet. Mark

More information

Lecture Support Vector Machine (SVM) Classifiers

Lecture Support Vector Machine (SVM) Classifiers Introduction to Machine Learning Lecturer: Amir Globerson Lecture 6 Fall Semester Scribe: Yishay Mansour 6.1 Support Vector Machine (SVM) Classifiers Classification is one of the most important tasks in

More information

CSE 331 Winter 2018 Homework 1

CSE 331 Winter 2018 Homework 1 Directions: - Due Wednesday, January 10 by 11 pm. - Turn in your work online using gradescope. You should turn in a single pdf file. You can have more than one answer per page, but please try to avoid

More information

A Decentralized Approach to Multi-agent Planning in the Presence of Constraints and Uncertainty

A Decentralized Approach to Multi-agent Planning in the Presence of Constraints and Uncertainty 2011 IEEE International Conference on Robotics and Automation Shanghai International Conference Center May 9-13, 2011, Shanghai, China A Decentralized Approach to Multi-agent Planning in the Presence of

More information

b) since the remainder is 0 I need to factor the numerator. Synthetic division tells me this is true

b) since the remainder is 0 I need to factor the numerator. Synthetic division tells me this is true Section 5.2 solutions #1-10: a) Perform the division using synthetic division. b) if the remainder is 0 use the result to completely factor the dividend (this is the numerator or the polynomial to the

More information

Embedded Systems - FS 2018

Embedded Systems - FS 2018 Institut für Technische Informatik und Kommunikationsnetze Prof. L. Thiele Embedded Systems - FS 2018 Sample solution to Exercise 3 Discussion Date: 11.4.2018 Aperiodic Scheduling Task 1: Earliest Deadline

More information

Machine Learning

Machine Learning Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 13, 2011 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting

More information

Introduction to Root Locus. What is root locus?

Introduction to Root Locus. What is root locus? Introduction to Root Locus What is root locus? A graphical representation of the closed loop poles as a system parameter (Gain K) is varied Method of analysis and design for stability and transient response

More information

14.1 Finding frequent elements in stream

14.1 Finding frequent elements in stream Chapter 14 Streaming Data Model 14.1 Finding frequent elements in stream A very useful statistics for many applications is to keep track of elements that occur more frequently. It can come in many flavours

More information

Reinforcement Learning II

Reinforcement Learning II Reinforcement Learning II Andrea Bonarini Artificial Intelligence and Robotics Lab Department of Electronics and Information Politecnico di Milano E-mail: bonarini@elet.polimi.it URL:http://www.dei.polimi.it/people/bonarini

More information