Vision for Mobile Robot Navigation: A Survey
|
|
- Clarence Pearson
- 5 years ago
- Views:
Transcription
1 Vision for Mobile Robot Navigation: A Survey (February 2002) Guilherme N. DeSouza & Avinash C. Kak presentation by: Job Zondag 27 February 2009
2 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
3 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
4 Indoor Navigation: Map-Based Navigation Vision system needs incorporation of some knowledge of what the robot is supposed to see CAD (geometrical maps) occupancy maps VFF: Virtual Force Fields topological maps sequences of images Vision based localization steps: Acquire sensory information Detect landmarks Establish matches between observation and expectation Calculate position
5 Indoor Navigation: Map-Based Navigation Absolute or global localization: Robot's initial pose is unknown. Incremental localization: Robot's initial pose is proximately known. Goal is to refine the location coordinates. Landmark tracking: Keep track of landmarks in the consecutive images that are recorded as the robot moves.
6 Absolute or Global Localization Atiya and Hager (1993)
7 Incremental Localization: Geometrical Representation of Space Initial position known proximately Keep updating the (uncertainties in the) position of the robot FINALE Kosaka & Kak (1992) Geometrical representation of space Statistical model of uncertainty in the location of the robot (Gaussian distribution)
8 Incremental Localization: Geometrical Representation of Space Using Geometrical Representation of Space Propagation of Positional Uncertainty trough Commanded Motions
9 Incremental Localization: Geometrical Representation of Space Projecting Robot's Positional Uncertainty into Camera Image Kalman Filtering
10 Incremental Localization: Topological Representation of Space NEURO-NAV Meng & Kak (1992) Graph representation of the layout of the hallway 2 modules (using neural networks) Hallway Follower Landmark Detector Supervisory RuleBased Controller
11 Incremental Localization: Topological Representation of Space
12 Incremental Localization: Topological Representation of Space Corridor-following: Neural Networks trained using backpropagation when a Human Supervisior module takes control of the navigation Results (1993): 86 % correct steering 10 % incorrect steering 4 % no decision FUZZY-NAV Kak et al. (1995)
13 Landmark Tracking Possible when known: Approximate location of the robot Identity of the landmarks Landmarks Artificial (circles, barcodes, tape) Natural (doors, windows, trees etc.) Most often: template matching
14 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
15 Map-Building Model of the world not always easy to generate First attempt: Moravec (1981) Stanford Cart World representation: 3D features plotted in a grid of 2 m2 cells 20 meters in 5 hours Moravec & Elfes (1985): occupancy grid
16 Map-Building Occupancy-grid-based approaches: cells with a probability of being occupied
17 Map-Building Occupancy-grid-based approaches: cells with a probability of being occupied Rich in geometrical detail Reliability depends on accuracy of the robot's odometry and sensor uncertainties Not computationally efficient for large or complex spaces Topological approaches Difficult to recognize previously visited nodes
18 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
19 Mapless Navigation: Optical Flow
20 Mapless Navigation: Optical Flow Santos-Victor et al. (1993) Robot: Robee Mimics visual behavior of bees: centering reflex (when flying trough hallway) Lateral position of the eyes: Motion derived features in stead of depth information
21 Mapless Navigation: Optical Flow Sustained behavior: it is desirable that when the robot runs into a section of the corridor deficit in wall texture, the robot drives on.
22 Mapless Navigation: Appearance-Based Matching Store images or templates of the environment and associate those images with commands or controls that will lead the robot to its final destination Gaussier et al. (1997) Neural networks: map perception to action 270 degree image of the environment Local views (subwindows) at x-positions of maximum intensity values
23 Mapless Navigation: Appearance-Based Matching Gaussier et al. (1997) Local views define a place in the environment Each place is asociated with a direction (azimuth) towards the goal A neural network learns to associate views/place with direction
24 Mapless Navigation: Appearance-Based Matching Ohno et al. (1996) VSSR: View-Sequenced Route Representation Correlate video input with database images to determine the position of the robot Use dispacement between the view and template image to compute real world dispacement and required steering actions
25 Mapless Navigation: Object Recognition Kim & Nevatia (1995) Symbolic navigation approach E.g. go to the desk in front of you Establish landmarks from command S-map: squeezed 3D into 2D space map GPS-like path planner
26 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
27 Outdoor Navigation Comparable to indoor navigation: Obstacle-avoidance, landmark detection, map building/updating, position estimation Normally no a priori map of the environment Structured: e.g. Road-following Unstructured: outdoor environment with no regular properties. e.g. Planetary terrain navigation Illumination
28 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
29 Outdoor Navigation: Structured Environments Road following car: NAVLAB 1 3D vision for obstacle detection and avoidance Color vision for road following Pixel classification: determine the probability of every pixel to belong to the representation of the road Color: road reagions tend to appear more blue Texture: road regions tend to appear much smoother compared to non-road regions Hough-like transform: determine the roadvanishingpoint and orientation Reclassify pixels: taking into account the determined road edges.
30 Outdoor Navigation: Structured Environments ALLVIN: Autonomous Land Vehicle In A Neural Network (first reported in 1989) Idea: Learn driving by watching a human driver NN: Back propagation
31 Outdoor Navigation: Structured Environments Gaussian distribution of activations: xi =e d 2i / 10 xi = activation level output node i di = distance ith node and steering angle
32 Outdoor Navigation: Structured Environments Training with synthetic images Training on the fly No experiences of situations that require correction Forgetting due to long strait roads Solution: adding distorted images
33 Outdoor Navigation: Structured Environments ALVINN-VC (Virtual Camera) Allows the system to detect road changes and intersections before they get too close to the vehicle IRRE: Input Reconstruction Reliability Estimation Using the neural network's internal representation to reconstruct the original image Correlate this with the actual input to measure the network's reliability
34 Outline: Types of Navigation Absolute localization (Structured) Map-Based Indoor Map-Building Mapless (Unstructured) Navigation Outdoor Structured Unstructured Incremental localization Landmark tracking Optic flow Appearance based Object recognition
35 Outdoor Navigation: Unstructured Environments Outdoor environment with no regular properties Wandering / exploring Goal position: need for some map building and localization algorithm Vehicle centered coordinate frame External reference (e.g. an external camera) Global positioning reference: (e.g. mountain peaks, the sun)
36 Outdoor Navigation: Unstructured Environments Mars Pathfinder project Launched in December 1996, landed in July 1997
37 Outdoor Navigation: Unstructured Environments Human operators specified waypoints in 3D views of the landing site once a day Deadreckoning-based positioning Moving speed: 15 cm / s Hazard detection every 6.5 cm Maximum travel distance 10 m / day
38 Outdoor Navigation: Illumination Problem: differences in contrast and texture due to variations in illumination Use of color to compensate Lorigo et al. (1997) Exploring robot: Figure out position of obstacles. Vertical slices: histogram of intensity values (RGB, HSV, or BW) Compare with safe window
39 Questions?
Introduction to Machine Learning
Introduction to Machine Learning Neural Networks Varun Chandola x x 5 Input Outline Contents February 2, 207 Extending Perceptrons 2 Multi Layered Perceptrons 2 2. Generalizing to Multiple Labels.................
More informationRobot Localisation. Henrik I. Christensen. January 12, 2007
Robot Henrik I. Robotics and Intelligent Machines @ GT College of Computing Georgia Institute of Technology Atlanta, GA hic@cc.gatech.edu January 12, 2007 The Robot Structure Outline 1 2 3 4 Sum of 5 6
More informationProbability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles
Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles Myungsoo Jun and Raffaello D Andrea Sibley School of Mechanical and Aerospace Engineering Cornell University
More informationAN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009
AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is
More informationIntroduction to Mobile Robotics Information Gain-Based Exploration. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Giorgio Grisetti, Kai Arras
Introduction to Mobile Robotics Information Gain-Based Exploration Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Giorgio Grisetti, Kai Arras 1 Tasks of Mobile Robots mapping SLAM localization integrated
More informationRobotics. Mobile Robotics. Marc Toussaint U Stuttgart
Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand
More informationVision-based navigation around small bodies
Astronet-II, International Final Conference Vision-based navigation around small bodies Pawel Kicman VISION-BASED NAVIGATION IN SPACE Camera LOS (Line-of-sight) sensor Star-horizon measurements Apparent
More informationLarge Scale Environment Partitioning in Mobile Robotics Recognition Tasks
Large Scale Environment in Mobile Robotics Recognition Tasks Boyan Bonev, Miguel Cazorla {boyan,miguel}@dccia.ua.es Robot Vision Group Department of Computer Science and Artificial Intelligence University
More informationParticle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics
Particle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics Recap: State Estimation using Kalman Filter Project state and error
More informationTowards Fully-automated Driving
Towards Fully-automated Driving Challenges and Potential Solutions Dr. Gijs Dubbelman Mobile Perception Systems EE-SPS/VCA Mobile Perception Systems 6 PhDs, postdoc, project manager, software engineer,
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationEXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL IMAGE IN URBAN AREAS. Received September 2015; revised January 2016
International Journal of Innovative Computing, Information and Control ICIC International c 2016 ISSN 1349-4198 Volume 12, Number 2, April 2016 pp. 371 383 EXTRACTION OF PARKING LOT STRUCTURE FROM AERIAL
More informationMulti-Sensor Fusion for Localization of a Mobile Robot in Outdoor Environments
Multi-Sensor Fusion for Localization of a Mobile Robot in Outdoor Environments Thomas Emter, Arda Saltoğlu and Janko Petereit Introduction AMROS Mobile platform equipped with multiple sensors for navigation
More informationVlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro (2016) Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Uncertainty representation Localization Chapter 5 (textbook) What is the course about?
More informationGIS-based Smart Campus System using 3D Modeling
GIS-based Smart Campus System using 3D Modeling Smita Sengupta GISE Advance Research Lab. IIT Bombay, Powai Mumbai 400 076, India smitas@cse.iitb.ac.in Concept of Smart Campus System Overview of IITB Campus
More informationBayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.
Kalman Filter Localization Bayes Filter Reminder Prediction Correction Gaussians p(x) ~ N(µ,σ 2 ) : Properties of Gaussians Univariate p(x) = 1 1 2πσ e 2 (x µ) 2 σ 2 µ Univariate -σ σ Multivariate µ Multivariate
More informationIntroduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping
Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz What is SLAM? Estimate the pose of a robot and the map of the environment
More informationConvolutional Neural Networks
Convolutional Neural Networks Books» http://www.deeplearningbook.org/ Books http://neuralnetworksanddeeplearning.com/.org/ reviews» http://www.deeplearningbook.org/contents/linear_algebra.html» http://www.deeplearningbook.org/contents/prob.html»
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationEKF and SLAM. McGill COMP 765 Sept 18 th, 2017
EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:
More informationUAV Navigation: Airborne Inertial SLAM
Introduction UAV Navigation: Airborne Inertial SLAM Jonghyuk Kim Faculty of Engineering and Information Technology Australian National University, Australia Salah Sukkarieh ARC Centre of Excellence in
More informationSensors Fusion for Mobile Robotics localization. M. De Cecco - Robotics Perception and Action
Sensors Fusion for Mobile Robotics localization 1 Until now we ve presented the main principles and features of incremental and absolute (environment referred localization systems) could you summarize
More informationLane Marker Parameters for Vehicle s Steering Signal Prediction
Lane Marker Parameters for Vehicle s Steering Signal Prediction ANDRIEJUS DEMČENKO, MINIJA TAMOŠIŪNAITĖ, AUŠRA VIDUGIRIENĖ, LEONAS JAKEVIČIUS 3 Department of Applied Informatics, Department of System Analysis
More informationVideo and Motion Analysis Computer Vision Carnegie Mellon University (Kris Kitani)
Video and Motion Analysis 16-385 Computer Vision Carnegie Mellon University (Kris Kitani) Optical flow used for feature tracking on a drone Interpolated optical flow used for super slow-mo optical flow
More informationSynthetic Sensing - Machine Vision: Tracking I MediaRobotics Lab, March 2010
Synthetic Sensing - Machine Vision: Tracking I MediaRobotics Lab, March 2010 References: Forsyth / Ponce: Computer Vision Horn: Robot Vision Schunk: Machine Vision University of Edingburgh online image
More informationUnit 8: Introduction to neural networks. Perceptrons
Unit 8: Introduction to neural networks. Perceptrons D. Balbontín Noval F. J. Martín Mateos J. L. Ruiz Reina A. Riscos Núñez Departamento de Ciencias de la Computación e Inteligencia Artificial Universidad
More informationCITS 4402 Computer Vision
CITS 4402 Computer Vision A/Prof Ajmal Mian Adj/A/Prof Mehdi Ravanbakhsh Lecture 06 Object Recognition Objectives To understand the concept of image based object recognition To learn how to match images
More informationCOS Lecture 16 Autonomous Robot Navigation
COS 495 - Lecture 16 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization
More informationJoint GPS and Vision Estimation Using an Adaptive Filter
1 Joint GPS and Vision Estimation Using an Adaptive Filter Shubhendra Vikram Singh Chauhan and Grace Xingxin Gao, University of Illinois at Urbana-Champaign Shubhendra Vikram Singh Chauhan received his
More informationThe Belief Roadmap: Efficient Planning in Belief Space by Factoring the Covariance. Samuel Prentice and Nicholas Roy Presentation by Elaine Short
The Belief Roadmap: Efficient Planning in Belief Space by Factoring the Covariance Samuel Prentice and Nicholas Roy Presentation by Elaine Short 1 Outline" Motivation Review of PRM and EKF Factoring the
More informationCIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions
CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions December 14, 2016 Questions Throughout the following questions we will assume that x t is the state vector at time t, z t is the
More informationNeural Networks biological neuron artificial neuron 1
Neural Networks biological neuron artificial neuron 1 A two-layer neural network Output layer (activation represents classification) Weighted connections Hidden layer ( internal representation ) Input
More informationMobile Robots Localization
Mobile Robots Localization Institute for Software Technology 1 Today s Agenda Motivation for Localization Odometry Odometry Calibration Error Model 2 Robotics is Easy control behavior perception modelling
More informationMultilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)
Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w x + w 2 x 2 + w 0 = 0 Feature x 2 = w w 2 x w 0 w 2 Feature 2 A perceptron can separate
More informationA RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES
A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES V.Sridevi, P.Malarvizhi, P.Mathivannan Abstract Rain removal from a video is a challenging problem due to random spatial distribution and
More informationMultilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)
Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w 1 x 1 + w 2 x 2 + w 0 = 0 Feature 1 x 2 = w 1 w 2 x 1 w 0 w 2 Feature 2 A perceptron
More informationUSING THE MILITARY LENSATIC COMPASS
USING THE MILITARY LENSATIC COMPASS WARNING This presentation is intended as a quick summary, and not a comprehensive resource. If you want to learn Land Navigation in detail, either buy a book; or get
More informationINTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY
[Gaurav, 2(1): Jan., 2013] ISSN: 2277-9655 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY Face Identification & Detection Using Eigenfaces Sachin.S.Gurav *1, K.R.Desai 2 *1
More informationImproving the travel time prediction by using the real-time floating car data
Improving the travel time prediction by using the real-time floating car data Krzysztof Dembczyński Przemys law Gawe l Andrzej Jaszkiewicz Wojciech Kot lowski Adam Szarecki Institute of Computing Science,
More informationMap Matching Algorithms in GPS Navigating System and Their Functions
Map Matching Algorithms in GPS Navigating System and Their Functions Abstract R.Kamalambal (M.Phil. Student) Department of computer science & Technology Kajamalai Campus, Bharathidasan University, Trichy,
More informationAutonomous Navigation for Flying Robots
Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.2: Kalman Filter Jürgen Sturm Technische Universität München Motivation Bayes filter is a useful tool for state
More informationArtificial Neural Networks Examination, June 2004
Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More information2D Image Processing (Extended) Kalman and particle filter
2D Image Processing (Extended) Kalman and particle filter Prof. Didier Stricker Dr. Gabriele Bleser Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz
More informationUSING THE MILITARY LENSATIC COMPASS
USING THE MILITARY LENSATIC COMPASS WARNING This presentation is intended as a quick summary, and not a comprehensive resource. If you want to learn Land Navigation in detail, either buy a book; or get
More informationDeep Learning (CNNs)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deep Learning (CNNs) Deep Learning Readings: Murphy 28 Bishop - - HTF - - Mitchell
More informationRobust License Plate Detection Using Covariance Descriptor in a Neural Network Framework
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Robust License Plate Detection Using Covariance Descriptor in a Neural Network Framework Fatih Porikli, Tekin Kocak TR2006-100 January 2007
More informationInstrumentation Commande Architecture des Robots Evolués
Instrumentation Commande Architecture des Robots Evolués Program 4a : Automatic Control, Robotics, Signal Processing Presentation General Orientation Research activities concern the modelling and control
More informationEdges and Scale. Image Features. Detecting edges. Origin of Edges. Solution: smooth first. Effects of noise
Edges and Scale Image Features From Sandlot Science Slides revised from S. Seitz, R. Szeliski, S. Lazebnik, etc. Origin of Edges surface normal discontinuity depth discontinuity surface color discontinuity
More informationPseudorandom Coding Techniques for Position Measurement
Pseudorandom Coding Techniques for Position Measurement Emil M. Petriu, Dr. Eng., FIEEE School of Electrical Engineering and Computer Science University of Ottawa, Canada http://www.site.uottawa.ca/~petriu
More informationEECS490: Digital Image Processing. Lecture #26
Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers
More informationUSING THE MILITARY LENSATIC COMPASS
USING THE MILITARY LENSATIC COMPASS WARNING This presentation is intended as a quick summary, and not a comprehensive resource. If you want to learn Land Navigation in detail, either buy a book; or get
More informationWe provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill.
We provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill. 2.3.2. Steering control using point mass model: Open loop commands We consider
More informationCS 3710: Visual Recognition Describing Images with Features. Adriana Kovashka Department of Computer Science January 8, 2015
CS 3710: Visual Recognition Describing Images with Features Adriana Kovashka Department of Computer Science January 8, 2015 Plan for Today Presentation assignments + schedule changes Image filtering Feature
More informationChapter 1 Overview of Maps
Chapter 1 Overview of Maps In this chapter you will learn about: Key points when working with maps General types of maps Incident specific maps Map legend and symbols Map sources A map is a navigational
More informationLego NXT: Navigation and localization using infrared distance sensors and Extended Kalman Filter. Miguel Pinto, A. Paulo Moreira, Aníbal Matos
Lego NXT: Navigation and localization using infrared distance sensors and Extended Kalman Filter Miguel Pinto, A. Paulo Moreira, Aníbal Matos 1 Resume LegoFeup Localization Real and simulated scenarios
More informationFeature extraction: Corners and blobs
Feature extraction: Corners and blobs Review: Linear filtering and edge detection Name two different kinds of image noise Name a non-linear smoothing filter What advantages does median filtering have over
More informationCMSC 421: Neural Computation. Applications of Neural Networks
CMSC 42: Neural Computation definition synonyms neural networks artificial neural networks neural modeling connectionist models parallel distributed processing AI perspective Applications of Neural Networks
More informationArtificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence
Artificial Intelligence (AI) Artificial Intelligence AI is an attempt to reproduce intelligent reasoning using machines * * H. M. Cartwright, Applications of Artificial Intelligence in Chemistry, 1993,
More informationL11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms
L11. EKF SLAM: PART I NA568 Mobile Robotics: Methods & Algorithms Today s Topic EKF Feature-Based SLAM State Representation Process / Observation Models Landmark Initialization Robot-Landmark Correlation
More informationRESTORATION OF VIDEO BY REMOVING RAIN
RESTORATION OF VIDEO BY REMOVING RAIN Sajitha Krishnan 1 and D.Venkataraman 1 1 Computer Vision and Image Processing, Department of Computer Science, Amrita Vishwa Vidyapeetham University, Coimbatore,
More informationVlad Estivill-Castro. Robots for People --- A project for intelligent integrated systems
1 Vlad Estivill-Castro Robots for People --- A project for intelligent integrated systems V. Estivill-Castro 2 Probabilistic Map-based Localization (Kalman Filter) Chapter 5 (textbook) Based on textbook
More informationUsing the Kalman Filter for SLAM AIMS 2015
Using the Kalman Filter for SLAM AIMS 2015 Contents Trivial Kinematics Rapid sweep over localisation and mapping (components of SLAM) Basic EKF Feature Based SLAM Feature types and representations Implementation
More informationWorking Group Cognitive Robotics
Working Group Cognitive Robotics Bernd Krieg-Brückner, Reinhard Moratz, Thomas Röfer, Kai Hübner, Axel Lankenau, Tilman Vierhuff Bremen Institute of Safe Systems Center for Computing Technology Universität
More informationUsing Map and Compass Together
Using Map and Compass Together In situations where you foresee a potential evacuation on foot, where there are no roads, and no indication as to the direction of travel (i.e., road signs), it is recommended
More informationFace Recognition Using Eigenfaces
Face Recognition Using Eigenfaces Prof. V.P. Kshirsagar, M.R.Baviskar, M.E.Gaikwad, Dept. of CSE, Govt. Engineering College, Aurangabad (MS), India. vkshirsagar@gmail.com, madhumita_baviskar@yahoo.co.in,
More informationCLOUD NOWCASTING: MOTION ANALYSIS OF ALL-SKY IMAGES USING VELOCITY FIELDS
CLOUD NOWCASTING: MOTION ANALYSIS OF ALL-SKY IMAGES USING VELOCITY FIELDS Yézer González 1, César López 1, Emilio Cuevas 2 1 Sieltec Canarias S.L. (Santa Cruz de Tenerife, Canary Islands, Spain. Tel. +34-922356013,
More informationOBJECT DETECTION AND RECOGNITION IN DIGITAL IMAGES
OBJECT DETECTION AND RECOGNITION IN DIGITAL IMAGES THEORY AND PRACTICE Bogustaw Cyganek AGH University of Science and Technology, Poland WILEY A John Wiley &. Sons, Ltd., Publication Contents Preface Acknowledgements
More informationParking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition-
Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition- Sanngoen Wanayuth, Akihisa Ohya and Takashi Tsubouchi Abstract The automated
More informationUsing a Hopfield Network: A Nuts and Bolts Approach
Using a Hopfield Network: A Nuts and Bolts Approach November 4, 2013 Gershon Wolfe, Ph.D. Hopfield Model as Applied to Classification Hopfield network Training the network Updating nodes Sequencing of
More informationAdvances in Computer Vision. Prof. Bill Freeman. Image and shape descriptors. Readings: Mikolajczyk and Schmid; Belongie et al.
6.869 Advances in Computer Vision Prof. Bill Freeman March 3, 2005 Image and shape descriptors Affine invariant features Comparison of feature descriptors Shape context Readings: Mikolajczyk and Schmid;
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Last Lecture : Edge Detection Preprocessing of image is desired to eliminate or at least minimize noise effects There is always tradeoff
More informationDP Project Development Pvt. Ltd.
Dear Sir/Madam, Greetings!!! Thanks for contacting DP Project Development for your training requirement. DP Project Development is leading professional training provider in GIS technologies and GIS application
More informationModeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems
Modeling CSC752 Autonomous Robotic Systems Ubbo Visser Department of Computer Science University of Miami February 21, 2017 Outline 1 Modeling and state estimation 2 Examples 3 State estimation 4 Probabilities
More informationSLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada
SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM
More informationToward Online Probabilistic Path Replanning
Toward Online Probabilistic Path Replanning R. Philippsen 1 B. Jensen 2 R. Siegwart 3 1 LAAS-CNRS, France 2 Singleton Technology, Switzerland 3 ASL-EPFL, Switzerland Workshop on Autonomous Robot Motion,
More informationLecture 8: Interest Point Detection. Saad J Bedros
#1 Lecture 8: Interest Point Detection Saad J Bedros sbedros@umn.edu Review of Edge Detectors #2 Today s Lecture Interest Points Detection What do we mean with Interest Point Detection in an Image Goal:
More informationCSE 473: Artificial Intelligence
CSE 473: Artificial Intelligence Hidden Markov Models Dieter Fox --- University of Washington [Most slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials
More informationEPL442: Computational
EPL442: Computational Learning Systems Lab 2 Vassilis Vassiliades Department of Computer Science University of Cyprus Outline Artificial Neuron Feedforward Neural Network Back-propagation Algorithm Notes
More informationCONTROL OF ROBOT CAMERA SYSTEM WITH ACTUATOR S DYNAMICS TO TRACK MOVING OBJECT
Journal of Computer Science and Cybernetics, V.31, N.3 (2015), 255 265 DOI: 10.15625/1813-9663/31/3/6127 CONTROL OF ROBOT CAMERA SYSTEM WITH ACTUATOR S DYNAMICS TO TRACK MOVING OBJECT NGUYEN TIEN KIEM
More informationVEHICLE WHEEL-GROUND CONTACT ANGLE ESTIMATION: WITH APPLICATION TO MOBILE ROBOT TRACTION CONTROL
1/10 IAGNEMMA AND DUBOWSKY VEHICLE WHEEL-GROUND CONTACT ANGLE ESTIMATION: WITH APPLICATION TO MOBILE ROBOT TRACTION CONTROL K. IAGNEMMA S. DUBOWSKY Massachusetts Institute of Technology, Cambridge, MA
More informationA Study of the Kalman Filter applied to Visual Tracking
A Study of the Kalman Filter applied to Visual Tracking Nathan Funk University of Alberta Project for CMPUT 652 December 7, 2003 Abstract This project analyzes the applicability of the Kalman filter as
More informationIntroduction to Mobile Robotics Probabilistic Sensor Models
Introduction to Mobile Robotics Probabilistic Sensor Models Wolfram Burgard 1 Sensors for Mobile Robots Contact sensors: Bumpers Proprioceptive sensors Accelerometers (spring-mounted masses) Gyroscopes
More informationRobots Autónomos. Depto. CCIA. 2. Bayesian Estimation and sensor models. Domingo Gallardo
Robots Autónomos 2. Bayesian Estimation and sensor models Domingo Gallardo Depto. CCIA http://www.rvg.ua.es/master/robots References Recursive State Estimation: Thrun, chapter 2 Sensor models and robot
More informationLearning Lab Seeing the World through Satellites Eyes
Learning Lab Seeing the World through Satellites Eyes ESSENTIAL QUESTION What is a satellite? Lesson Overview: Engage students will share their prior knowledge about satellites and explore what satellites
More informationSimultaneous Localization and Map Building Using Natural features in Outdoor Environments
Simultaneous Localization and Map Building Using Natural features in Outdoor Environments Jose Guivant, Eduardo Nebot, Hugh Durrant Whyte Australian Centre for Field Robotics Department of Mechanical and
More informationAbsolute map-based localization for a planetary rover
Absolute map-based localization for a planetary rover Bach Van Pham, Artur Maligo and Simon Lacroix LAAS/CNRS, Toulouse Work developed within the ESA founded Startiger activity Seeker Outline" On the importance
More informationin Robotics Sebastian Thrun Carnegie Mellon University, Pittsburgh, USA Dieter Fox Wolfram Burgard Institut fur Informatik
Probabilistic Methods for State Estimation in Robotics Sebastian Thrun Computer Science Department and Robotics Institute Carnegie Mellon University, Pittsburgh, USA http://www.cs.cmu.edu/thrun Dieter
More informationTennis player segmentation for semantic behavior analysis
Proposta di Tennis player segmentation for semantic behavior analysis Architettura Software per Robot Mobili Vito Renò, Nicola Mosca, Massimiliano Nitti, Tiziana D Orazio, Donato Campagnoli, Andrea Prati,
More informationA Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems
A Probabilistic Relational Model for Characterizing Situations in Dynamic Multi-Agent Systems Daniel Meyer-Delius 1, Christian Plagemann 1, Georg von Wichert 2, Wendelin Feiten 2, Gisbert Lawitzky 2, and
More informationLocalization. Howie Choset Adapted from slides by Humphrey Hu, Trevor Decker, and Brad Neuman
Localization Howie Choset Adapted from slides by Humphrey Hu, Trevor Decker, and Brad Neuman Localization General robotic task Where am I? Techniques generalize to many estimation tasks System parameter
More informationManipulators. Robotics. Outline. Non-holonomic robots. Sensors. Mobile Robots
Manipulators P obotics Configuration of robot specified by 6 numbers 6 degrees of freedom (DOF) 6 is the minimum number required to position end-effector arbitrarily. For dynamical systems, add velocity
More informationCOMP-4360 Machine Learning Neural Networks
COMP-4360 Machine Learning Neural Networks Jacky Baltes Autonomous Agents Lab University of Manitoba Winnipeg, Canada R3T 2N2 Email: jacky@cs.umanitoba.ca WWW: http://www.cs.umanitoba.ca/~jacky http://aalab.cs.umanitoba.ca
More informationCorners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros
Corners, Blobs & Descriptors With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros Motivation: Build a Panorama M. Brown and D. G. Lowe. Recognising Panoramas. ICCV 2003 How do we build panorama?
More informationUsing Match Uncertainty in the Kalman Filter for a Sonar Based Positioning System
Using atch Uncertainty in the Kalman Filter for a Sonar Based ositioning System Oddbjørn Bergem, Claus Siggaard Andersen, Henrik Iskov Christensen Norwegian Defence Research Establishment, Norway Laboratory
More informationArtificial Intelligence
Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Summary of last lecture We know how to do probabilistic reasoning over time transition model P(X t
More informationMultilayer Perceptrons (MLPs)
CSE 5526: Introduction to Neural Networks Multilayer Perceptrons (MLPs) 1 Motivation Multilayer networks are more powerful than singlelayer nets Example: XOR problem x 2 1 AND x o x 1 x 2 +1-1 o x x 1-1
More informationCS 532: 3D Computer Vision 6 th Set of Notes
1 CS 532: 3D Computer Vision 6 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Lecture Outline Intro to Covariance
More informationXXIII CONGRESS OF ISPRS RESOLUTIONS
XXIII CONGRESS OF ISPRS RESOLUTIONS General Resolutions Resolution 0: Thanks to the Czech Society commends: To congratulate The Czech Society, its president and the Congress Director Lena Halounová, the
More informationConsider a robot lost in space. Within its environment, how can the robot locate where it is?
Artificial Intelligence for Robotics Lesson 1: Localization Intro In this course you will learn how to program self-driving cars! Specifically, in this unit, you will learn how to program a localizer,
More informationPhotometric Redshifts with DAME
Photometric Redshifts with DAME O. Laurino, R. D Abrusco M. Brescia, G. Longo & DAME Working Group VO-Day... in Tour Napoli, February 09-0, 200 The general astrophysical problem Due to new instruments
More information