Safe robot motion planning in dynamic, uncertain environments



Similar documents
Robotics. Chapter 25. Chapter 25 1

Robot Navigation. Johannes Maurer, Institute for Software Technology TEDUSAR Summerschool u

Neuro-Dynamic Programming An Overview

Reinforcement Learning

Online Model Predictive Control of a Robotic System by Combining Simulation and Optimization

Course Outline Department of Computing Science Faculty of Science. COMP Applied Artificial Intelligence (3,1,0) Fall 2015

Nonlinear Model Predictive Control of Hammerstein and Wiener Models Using Genetic Algorithms

Reinforcement Learning

Masters of Engineering in MIE Certificate Program in Financial Engineering

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

An Introduction to Markov Decision Processes. MDP Tutorial - 1

Dynamic Modeling, Predictive Control and Performance Monitoring

Algorithmic Trading Session 1 Introduction. Oliver Steinki, CFA, FRM

Improving proposal evaluation process with the help of vendor performance feedback and stochastic optimal control

Scheduling Algorithms for Downlink Services in Wireless Networks: A Markov Decision Process Approach

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

Computing Near Optimal Strategies for Stochastic Investment Planning Problems

On evaluating performance of exploration strategies for an autonomous mobile robot

Bayesian networks - Time-series models - Apache Spark & Scala

Some Research Directions in Automated Pentesting

CE801: Intelligent Systems and Robotics Lecture 3: Actuators and Localisation. Prof. Dr. Hani Hagras

A Movement Tracking Management Model with Kalman Filtering Global Optimization Techniques and Mahalanobis Distance

C21 Model Predictive Control

Stochastic Gradient Method: Applications

STATISTICAL DATA ANALYSIS COURSE VIA THE MATLAB WEB SERVER

Machine Learning.

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique

Markov Decision Processes for Ad Network Optimization

Appendix A. About RailSys 3.0. A.1 Introduction

Tracking in flussi video 3D. Ing. Samuele Salti

Motion. Complete Table 1. Record all data to three decimal places (e.g., or or 0.000). Do not include units in your answer.

NEURAL NETWORKS A Comprehensive Foundation

A Statistical Framework for Operational Infrasound Monitoring

Gerard Mc Nulty Systems Optimisation Ltd BA.,B.A.I.,C.Eng.,F.I.E.I

TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA

Automatic Train Control based on the Multi-Agent Control of Cooperative Systems

Robotics. Lecture 3: Sensors. See course website for up to date information.

POMPDs Make Better Hackers: Accounting for Uncertainty in Penetration Testing. By: Chris Abbott

Comparing Artificial Intelligence Systems for Stock Portfolio Selection

Robot Perception Continued

E190Q Lecture 5 Autonomous Robot Navigation

UNCERTAINTY IN THE ELECTRIC POWER INDUSTRY Methods and Models for Decision Support

SOLIDWORKS SOFTWARE OPTIMIZATION

Security Optimization of Dynamic Networks with Probabilistic Graph Modeling and Linear Programming

Teaching Introductory Artificial Intelligence with Pac-Man

Operations Research in Production Planning, Scheduling, and Inventory Control

Integrating Artificial Intelligence. Software Testing

Predictive Representations of State

A Log-Robust Optimization Approach to Portfolio Management

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XVI - Fault Accomodation Using Model Predictive Methods - Jovan D. Bošković and Raman K.

Supplement to Call Centers with Delay Information: Models and Insights

Best-response play in partially observable card games

An Environment Model for N onstationary Reinforcement Learning

DATA MINING IN FINANCE

Distributionally Robust Optimization with ROME (part 2)

Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina

Amajor benefit of Monte-Carlo schedule analysis is to

Optimization of warehousing and transportation costs, in a multiproduct multi-level supply chain system, under a stochastic demand

Understanding and Applying Kalman Filtering

Better decision making under uncertain conditions using Monte Carlo Simulation

9700 South Cass Avenue, Lemont, IL URL: fulin

Type-2 fuzzy logic control for a mobile robot tracking a moving target

Financial Mathematics and Simulation MATH Spring 2011 Homework 2

Set-Based Design: A Decision-Theoretic Perspective

Compression algorithm for Bayesian network modeling of binary systems

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006

Tracking of Small Unmanned Aerial Vehicles

Constrained Clustering of Territories in the Context of Car Insurance

Statistics Graduate Courses

Stochastic control for underwater optimal trajectories CQFD & DCNS. Inria Bordeaux Sud Ouest & University of Bordeaux France

CRITICAL CHAIN AND CRITICAL PATH, CAN THEY COEXIST?

BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION

Multiple Network Marketing coordination Model

Unit I. Introduction

Measuring the Performance of an Agent

Agents: Rationality (2)

DIEF, Department of Engineering Enzo Ferrari University of Modena e Reggio Emilia Italy Online Trajectory Planning for robotic systems

NONLINEAR TIME SERIES ANALYSIS

Big Data - Lecture 1 Optimization reminders

SECOND YEAR. Major Subject 3 Thesis (EE 300) 3 Thesis (EE 300) 3 TOTAL 3 TOTAL 6. MASTER OF ENGINEERING IN ELECTRICAL ENGINEERING (MEng EE) FIRST YEAR

Cybersecurity Strategies: The QuERIES Methodology

Transcription:

Safe robot motion planning in dynamic, uncertain environments RSS 2011 Workshop: Guaranteeing Motion Safety for Robots June 27, 2011 Noel du Toit and Joel Burdick California Institute of Technology

Dynamic, Cluttered, Uncertain Env s 6/27/2011 Noel du Toit 2

Conceptual Problem Components: Localization Process Noise Mapping Detection/ tracking Prediction Characterization Occlusions/ drop-outs Problem Formulation? Computation? 6/27/2011 Noel du Toit 3

Safety vs Conservatism Probabilistic uncertainty => conservative Conservatism: incorporate anticipated measurements Previous works: static environments [Prentice 09], [Censi 08], [Yan 05] Enabling capability: complicated agent behaviors, more clutter, agent information gathering Safety: chance constraint conditioning 6/27/2011 Noel du Toit 4

Stochastic Dynamic Programming (SDP) History I-state: state control process noise measurement noise x i xi xi = xi R A 1 A n Belief state: probability distribution State: Transition function: Cost Function Terminal cost Encodes noise properties & and dynamics Stage cost 6/27/2011 Noel du Toit 5

Stochastic Dynamic Programming (SDP) Control Policy: Problem: Stochastic Dynamic Programming (SDP) Feedback over all possible future measurements and the resulting belief states 6/27/2011 Noel du Toit 6

Stochastic Dynamic Programming (SDP) Issues: Computationally intensive [Thrun et al. 05], [Bertsekas 05] No hard constraints Solution: for Linear, quadratic cost, Gaussian noise [Bertsekas 05], [Bar-Shalom 81] Approximations Value & policy iteration [Thrun et al. 05], Roll-out algorithm [Bertsekas 05] Restricted structure approximations [Bertsekas 05] 6/27/2011 Noel du Toit 7

Formulation: Stochastic RHC (SRHC) Stochastic system: Expected cost, chance constraints Belief state: Transition function: Disturbance: 6/27/2011 Noel du Toit 8

Approximation: Most Likely Measurement Effect of measurements Update covariance Update mean Most likely measurement: Same computational benefits of OLRHC approach Approximation: relies on RHC formulation to correct for assumed information to reduce conservatism in solutions Theorem [dutoit 09] : The Most Likely Measurement Approx. introduces no new information Least Informative Approximation 6/27/2011 Noel du Toit 10

Chance Constraints Constrain uncertain state Probabilistic Collision Avoidance Robot and obstacle uncertainty Joint distribution and indicator function 6/27/2011 Noel du Toit 11

Collision Constraints: Evaluation Monte-Carlo Simulation Small level of confidence: ~5ms per evaluation requires ~10000 samples Approximate: small disk/ellipse objects Independent, Gaussian distributions: Dependent, Gaussian distributions 6/27/2011 Noel du Toit 12

Safety: Reaction Horizon Quantify time (# of stages) to react to changes in environment Robot dynamics Environment Reaction horizon, r: react to modeled disturbances Chance constraint conditioning: Use r-stage open-loop predicted distribution Anticipated information: leverage PCL reduction in conservatism 6/27/2011 Noel du Toit 13

Safety vs Conservatism Uncertainty grows over reaction horizon Next stage: new information + re-solve problem (RHC) PCL: leverage new information OL: uncertainty growth results in conservative solutions 6/27/2011 Noel du Toit 14

1-D example: Car following: Collision constraint (maintain some separation distance) Velocity controlled random-walk model Reaction horizon: r=2 (to influence position) 6/27/2011 Noel du Toit 15

1-D Example (cont d) Reaction horizon = 1 Plot separation distance Reaction horizon = 2 6/27/2011 Noel du Toit 16

Dynamic Environment: Oncoming Agents OLRHC PCLRHC 6/27/2011 Noel du Toit 17

Summary Practical systems: trade off conservatism and safety PCLRHC Reduce conservatism through anticipated information RHC: resolve problem to incorporate actual measurements Chance constraint conditioning Allow for modeled disturbances Can still leverage anticipated information See Noel s thesis for various variations on this problem 6/27/2011 Noel du Toit 18

Thank you ndutoit@caltech.edu jwb@robotics.caltech.edu Publications: Questions? Du Toit, N.E. and Burdick, J.W., Robot Motion Planning in Dynamic, Cluttered, Uncertain Environments, accepted to IEEE Transactions on Robotics Du Toit, N.E. and Burdick, J.W., Probabilistic Collision Checking with Chance Constraints, accepted to IEEE Transaction on Robotics Du Toit, N.E., Robot Motion Planning in Dynamic, Cluttered, Uncertain Environments: the Partially Closed-Loop Receding Horizon Control Approach, Ph.D. thesis, Caltech, 2010 Conferences: Workshop on Motion Planning: From Theory to Practice (RSS) 2010 Du Toit, N.E. and Burdick, J.W., Robotic Motion Planning in Dynamic, Cluttered, Uncertain Environments, ICRA 2010 6/27/2011 Noel du Toit 19

Problem Definition Robot: Agent: Constraints: Objective function: 6/27/2011 Noel du Toit 20

Related Work Stochastic systems: Probabilistic vs. non-deterministic Deterministic Probabilistic Static Dynamic Static Dynamic Control OC [Friedland 05] RHC [Mayne 00] OC with augmented states OC with separation [Friedland 05] Stochastic RHC (later) PCLRHC Robotics Graph search, roadmap methods, etc. [LaValle 06], [Choset et al. 07] Dynamic window [Fox et al. 97] Velocity obstacles [Fiorini & Shiller 98] Graph search in extended state space [LaValle 06], [Censi 08] Probabilistic velocity obstacles [Fulgenzi et al. 07] AI MDPs [LaValle 06] Extended state space (time x pose) [LaValle 06] DP [Bertsekas 05] MDPs, POMDPs [Thrun et al. 05] AI: Artificial Intelligence OC: Optimal Control RHC: Receding Horizon Control DP: Dynamic Programming MDP: Markov Decision Process POMDP: Partially Observable MDP 6/27/2011 Noel du Toit 21

Partially Closed-loop RHC Most likely measurement: Restricted information: Resulting belief state: Deterministic in belief state: 6/27/2011 Noel du Toit 22

Simulation Setup Robot: Linear model, linear measurements Velocity constraints: Control constraints: Collision constraints: Agent: Linear constant velocity model, linear measurements 6/27/2011 Noel du Toit 23

PCLRHC Approximation Information gain: relative entropy Baseline: PCL distribution: Executed distribution: 6/27/2011 Noel du Toit 24