Multi-ultrasonic sensor fusion for autonomous mobile robots



Similar documents
ExmoR A Testing Tool for Control Algorithms on Mobile Robots

E190Q Lecture 5 Autonomous Robot Navigation

Synthetic Sensing: Proximity / Distance Sensors

Sensor Modeling for a Walking Robot Simulation. 1 Introduction

Nighttime Vehicle Distance Alarm System

Robot Perception Continued

Microcontrollers, Actuators and Sensors in Mobile Robots

Distributed Sensing for Cooperative Robotics

A Secure Online Reputation Defense System from Unfair Ratings using Anomaly Detections

Topic Dempster-Shafer Theory

Path Tracking for a Miniature Robot

COMBINING THE METHODS OF FORECASTING AND DECISION-MAKING TO OPTIMISE THE FINANCIAL PERFORMANCE OF SMALL ENTERPRISES

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network

Tracking of Small Unmanned Aerial Vehicles

2/26/2008. Sensors For Robotics. What is sensing? Why do robots need sensors? What is the angle of my arm? internal information

Mobile Robot Localization using Range Sensors : Consecutive Scanning and Cooperative Scanning

Real-Time Indoor Mapping for Mobile Robots with Limited Sensing

Contradiction measures and specificity degrees of basic belief assignments

Automatic Labeling of Lane Markings for Autonomous Vehicles

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique

Reactive Agent Technology for Real-Time, Multisensor Target Tracking

CE801: Intelligent Systems and Robotics Lecture 3: Actuators and Localisation. Prof. Dr. Hani Hagras

Probabilistic Evidential Reasoning with Symbolic Argumentation for Space Situation Awareness

Open World Target Identification Using the Transferable Belief Model

3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

Applied Mathematical Sciences, Vol. 7, 2013, no. 112, HIKARI Ltd,

Visual Servoing Methodology for Selective Tree Pruning by Human-Robot Collaborative System

A Genetic Algorithm-Evolved 3D Point Cloud Descriptor

Self Organizing Maps for Visualization of Categories

Recognizing Suspiciousness Degree in Surveillance Area using Interval Type-2 Fuzzy Model in Security Systems

Robotics. Lecture 3: Sensors. See course website for up to date information.

Robot Sensors. Outline. The Robot Structure. Robots and Sensors. Henrik I Christensen

Static Environment Recognition Using Omni-camera from a Moving Vehicle

Automated System for Computationof Burnt Forest Region using Image Processing

Controlling a Mobile Robot with a Personal Digital Assistant

Prediction of DDoS Attack Scheme

The Limits of Human Vision

A Multisensor Multiobject Tracking System for an Autonomous Vehicle Driving in an Urban Environment *

Automated Sewer Pipe Inspection through Image Processing

Face Recognition in Low-resolution Images by Using Local Zernike Moments

A CLASSIFIER FUSION-BASED APPROACH TO IMPROVE BIOLOGICAL THREAT DETECTION. Palaiseau cedex, France; 2 FFI, P.O. Box 25, N-2027 Kjeller, Norway.

Predicting the Risk of Heart Attacks using Neural Network and Decision Tree

A Novel Solution on Alert Conflict Resolution Model in Network Management

Range sensors. Sonar. Laser range finder. Time of Flight Camera. Structured light. 4a - Perception - Sensors. 4a 45

REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING

Development of Docking System for Mobile Robots Using Cheap Infrared Sensors

Flexible mobility management strategy in cellular networks

Credal classification of uncertain data using belief functions

Impact of Reflectors on Solar Energy Systems

A Client-Server Interactive Tool for Integrated Artificial Intelligence Curriculum

Robot Navigation. Johannes Maurer, Institute for Software Technology TEDUSAR Summerschool u

Analecta Vol. 8, No. 2 ISSN

Laser Gesture Recognition for Human Machine Interaction

Topological Mapping with Sensing-Limited Robots

Automated Process for Generating Digitised Maps through GPS Data Compression

Draft dpt for MEng Electronics and Computer Science

Real-Time Power-Efficient Integration of Multi-Sensor Occupancy Grid on Many-Core

Course Outline Department of Computing Science Faculty of Science. COMP Applied Artificial Intelligence (3,1,0) Fall 2015

Simultaneous Gamma Correction and Registration in the Frequency Domain

Calibration of Ultrasonic Sensors of a Mobile Robot*

Human Detection Robot using PIR Sensors

Last Mile Intelligent Driving in Urban Mobility

MUSIC-like Processing of Pulsed Continuous Wave Signals in Active Sonar Experiments

A Simulation Analysis of Formations for Flying Multirobot Systems

On-line novelty detection for autonomous mobile robots

Super-Agent Based Reputation Management with a Practical Reward Mechanism in Decentralized Systems

Planning to Fail - Reliability Needs to Be Considered a Priori in Multirobot Task Allocation

Knowledge Acquisition Approach Based on Rough Set in Online Aided Decision System for Food Processing Quality and Safety

SYSTEMS, CONTROL AND MECHATRONICS

Sensor-Based Robotic Model for Vehicle Accident Avoidance

Forecasting of Economic Quantities using Fuzzy Autoregressive Model and Fuzzy Neural Network

Safe Robot Driving 1 Abstract 2 The need for 360 degree safeguarding

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

Dynamic composition of tracking primitives for interactive vision-guided navigation

How To Test A Robot Platform And Its Components

Appendices master s degree programme Artificial Intelligence

Design of an Intelligent Housing System Using Sensor Data Fusion Approaches

Measuring Software Complexity to Target Risky Modules in Autonomous Vehicle Systems

Using Ultrasonic and Infrared Sensors for Distance Measurement

Autonomous Advertising Mobile Robot for Exhibitions, Developed at BMF

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

STORE VIEW: Pervasive RFID & Indoor Navigation based Retail Inventory Management

Design Report. IniTech for

Master of Science in Computer Science

multisensor integration [7, 8, 9], and distributed sensing [10, 11, 12].

Tracking of Moving Objects from a Moving Vehicle Using a Scanning Laser Rangefinder

Price Prediction of Share Market using Artificial Neural Network (ANN)

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Indoor Positioning Systems WLAN Positioning

A Framework for Developing Mobile, Context-aware Applications

Force/position control of a robotic system for transcranial magnetic stimulation

Test Coverage Criteria for Autonomous Mobile Systems based on Coloured Petri Nets

Design a Low Cost Security Robot Applying in Family

Integrated Modeling for Data Integrity in Product Change Management

Open Access A Facial Expression Recognition Algorithm Based on Local Binary Pattern and Empirical Mode Decomposition

How to preserve the conflict as an alarm in the combination of belief functions?

IMAV: An Intelligent Multi-Agent Model Based on Cloud Computing for Resource Virtualization

WIRELESS BLACK BOX USING MEMS ACCELEROMETER AND GPS TRACKING FOR ACCIDENTAL MONITORING OF VEHICLES

Transcription:

Multi-ultrasonic sensor fusion for autonomous mobile robots Zou Yi *, Ho Yeong Khing, Chua Chin Seng, and Zhou Xiao Wei School of Electrical and Electronic Engineering Nanyang Technological University Nanyang Avenue, Singapore, 639798 ABSTRACT Specular reflections from environments cause uncertainties to ultrasonic sensor range data. In this paper, we examine the application of evidential method for data integration using the specially designed sensor model to overcome the problem. Dempster s rule of combination is used to fuse the sensor data to obtain the map defined on a 2D evidence grid. The sensor model tries to reduce the uncertainties caused by specular reflections with a filtering factor. Experimental results have shown the usefulness of this method. Keywords: Dempster-Shafer, mobile robot, sensor fusion, ultrasonic sensor, specular reflection 1. INTRODUCTION The navigation of a multi-sensor based mobile robot requires a good representation of the environment. An autonomous mobile robot should be able to construct a map of its environment based on the sensory information. To build the map, sensory information from multiple sensors need to be combined to obtain the best view of the surroundings of the mobile robot. However, since there always exist various kinds of incompleteness and uncertainties in sensory information about the real world, the knowledge obtained is not good enough to help the mobile robot perform in an intelligent way. Ultrasonic sensors have been widely used in mobile robots applications as they can produce good range information based on the time of the flight (TOF) principle. This provides an easy and inexpensive way for the robot system to construct and maintain its map for localization, obstacle avoidance and other purposes during its navigation in unknown environments. Applications of ultrasonic sensors in autonomous mobile robots have been discussed by many researchers. 1,2,3,4,9,10 However, there are some fundamental drawbacks 12 which limit the usefulness of ultrasonic devices, of which the most significant one is the possibility of specular reflections 8 from the environment. As shown in Fig. 1b, the specular reflections occur when ultrasonic beam fails to return directly to the receiver since it will be bounced away from the target object. The incidence angle of the sonar signal Figure 1. Example of the specular reflection, α is the half opening angle of sonar beam, θ is the incidence angle and materials of the reflecting surfaces are decisive to the occurrence of specular reflection. 8,12 It happens frequently * Correspondence: E-mail: P144723846@ntu.edu.sg; Telephone: 65-790-4052 ; Fax: 65-790-5471

whenever the robot is moving in a small confined or narrow environment. The consequence of specular reflections is the longer range reading than the real distance from the ultrasonic sensor. False readings caused by specular reflections can be treated as uncertainties in sonar responses from the probabilistic view. The Bayesian updating formula 1 which projects the sensor responses on the evidence grid 13 has been widely used to express the uncertainties in sensory information about whether the area is occupied or empty. Dempster-Shafer s evidence theory has gained advantages over Bayesian for it being able to clearly distinguish the ignorance and the contradiction. In this paper, we present the method of using Dempster-Shafer s evidential method as the basic approach for sensory data fusion. We have applied the Dempster-Shafer s evidence method on the evidence grid using an appropriate ultrasonic sensor model in which a special filtering factor has been used to reduce unreliable ultrasonic readings corrupted by specular reflections from confined environments. 2. FRAMEWORK OF SENSOR FUSION 2.1. Evidence Grid The evidence grid 13 is a framework to integrate multi-sensory information. It is a 2D representation of the real world of the autonomous mobile robot. The Bayesian formula or Dempster s rule of combination can then be easily applied to combine different sources of sensory information 1,13 obtained from the sensor model. Through the sensor fusion algorithm, each cell in the evidence grid is assigned certain values representing its state of either being occupied or being empty. Bayesian theory has its limitation in assigning probabilities only to singletons. Such as in our case, for a cell which is either occupied or empty, the probabilities for these two contradictions must satisfy p({occupied}) +p({empty}) = 1. This means it will not be able to treat the evidence for the world as ignorance, where a cell can have equal evidence for being occupied and being empty. Dempster-Shafer s evidence theory has advantages over Bayesian theory in this aspect since it is able to provide evidence for both singletons and subsets so it can clearly distinguish between ignorance and contradiction. 2.2. Review of Dempster-Shafer s Evidence Theory In applying the Dempster-Shafer s evidence theory, 5,6,7 a sensor is a source of evidence about a set of propositions of interest, called the focal elements. The set of focal elements thatcanbeobservedbyasensoristheframe of discernment (FOD), denoted as θ. In our case, the frame of discernment contains two propositions, θ = {occupied, empty} (1) We can transfer the belief of these propositions to evidence by mass distribution of any element in the power set of θ, denoted as 2 θ, 2 θ = {{occupied}, {empty}, {unknown}} (2) In Equation(2), {unknown} = {occupied} {empty}. Different sources of evidence are combined into a total amount of evidence through the Dempster s rule of combination. 6,7 In our case, the new evidence m n (C) isupdatedbythe two evidence sources, m s from sensors and m o from the old existing evidence. m n (C) iscalledtheorthogonal sum of m s and m o that is calculated as: m n (C) =(m s m o )(C) = Σ A B=C;C6=φm s (A)m o (B) 1 Σ A B=φ m s (A)m o (B) (3) where A, B, C θ. For example, when C= {occupied }, weuseocc to represent occupied, emp for empty and unk for unknown, som n ({occ}) iscalculatedas, m n ({occ}) = m o({occ})m s ({occ})+m o ({occ})m s ({unk})+m o ({unk})m s ({occ}) 1 m o ({occ})m s ({emp}) m o ({emp})m s ({occ}) (4) 2.3. Sensor Model The evidence is obtained by projecting the raw ultrasonic sensor responses onto the evidence grid through the sensor model. The sonar model is a function of the angle and the sonar range reading as shown in Fig. 2. The model converts the range information into probability values. As shown in Fig.2, the model is given in Equation (5) Equation (10) 1, 8

Figure 2. The profile of the ultrasonic sensor model. Region I, where R ²<r<R+ ²: Region II, where R min <r<r ²: m({occupied}) = ( α ω α )2 +( ² R r ) 2 ² 2 (5) m({empty}) = 0.00 (6) m({unknown}) = 1.00 m({occupied}) (7) m({occupied}) = 0.00 (8) m({empty}) = ( α ω α )2 +( R ² r R ² )2 (9) 2 m({unknown}) = 1.00 m({empty}) (10) where R is the range response from the ultrasonic sensor, (r, ω) is the coordinate of a point inside the sonar cone. ² is the range error and it distributes the evidence in Region I. α is the half open beam angle of sonar cone. 3. DEALING WITH SPECULAR REFLECTIONS The sensor model as represented in Equation (5) Equation (10) is effective in converting range information of ultrasonic sensors but is helpless in dealing with the specular reflection problem which is especially obvious in confined environments. To reduce unreliable readings caused by specular reflections in ultrasonic sensory responses, we have applied a specially designed filtering factor the range confidence factor (RCF ). The idea of RCF was first introduced in Ref. 7. However, in Ref. 7, the sensor fusion strategy with RCF applied to the sensor model was based on the Bayesian probability theory, while in this paper, we investigate the effects of RCF in using Dempster- Shafer s evidence method for multi-ultrasonic sensor integration. The RCF used in our case is given in the following equation, 8 when R>R max : when R R max : RCF = RCF min = R th R th +1 (11) RCF = ( Rmax R R max ) τ + R th (12) 1+R th where R is the sonar response and RCF is determined by three parameters, τ, R th and R max. τ is used to reflect the sensitivity of the sensor to the reflected specular signal. 8 Nevertheless, we can use τ to reflect environment

Figure 3. The range confidence factor. R max =6.5m characteristics such as the confined space or the feature of reflective surfaces. In Fig. 3a, with τ < 1, RCF decreases more slowly than it does in Fig. 3b where τ > 1untilRCF reaches the RCF min. So, the more τ is below 1, the less sensitive RCF will be to the range value that is less than R max. RCF min is the lowest RCF value thresholded by R th. R max sets the maximum range detection of sonar, for example, R max can be set as the maximum diagonal length of the environment. As shown in Fig. 3, RCF will be restricted to RCF min when the range detection reaches R max. Therefore, responses affected by specular reflections will produce much less evidence from the sonar model, which consequently avoids accumulating evidence for unreliable sensory information caused by specular reflections. The RCF is applied by simply multiplying it with Equation (5) and Equation (9) as: m({occupied}) =RCF ( α ω α )2 +( ² R r ) 2 ² 2 m({empty}) =RCF ( α ω α )2 +( R ² r R ² )2 (14) 2 The RCF is actually a filtering factor which adjusts the evidence for cells. What really work in RCF are the three values in Equation (12): τ, R th and R max. They are chosen based on heuristic understandings or experimental results. 4. EXPERIMENTS 4.1. Implementation We have tested the effects of applying RCF to sensor model with Dempster-Shafer s evidence method for the multisensor fusion in a confined environment in our laboratory. The overview of the environment is given in Fig. 4. Materials of reflecting surfaces are those that most commonly used in normal laboratory walls. Raw sensor data (13) Figure 4. A typical confined environment

were collected by the Nomad Super Scout II robot platform at different locations. The robot has a sonar ring of 16 Polaroid ultrasonic sensor modules, as shown in Fig. 5. Figure 5. The Nomad Super Scout II mobile robot platform 4.2. Experimental Results Fig. 6 Fig. 10 present all experimental results of fused evidence for occupied cells on the 2D evidence grid. In all figures, * represents m(occ) > 0.8 while represents m(occ) > 0.5. Fig. 6 shows the noisy evidence grid when the RCF was not applied. Uncertainties in this figure are obvious and are particularly pronounced in areas near corners and arcs. However, when the RCF was used to modify the sensor data, these uncertainties were reduced Figure 6. Evidence for being occupied without RCF dramatically as shown in Fig. 7 Fig. 10. These results also show how different selections of the three parameters in Equation (12) could bring different effects. 4.3. Discussions Asmentionedintheabove,theuseofRCF did work very well in reducing the unreliable responses as shown typically in Fig. 7. However we have also noticed that whether RCF can best perform depends highly on the values of the

Figure 7. RCF min =0.1, τ =1.5,R max =500cm Figure 8. RCF min =0.1, τ =0.5,R max =650cm three parameters in Equation (12). RCF acts as a filtering factor in reducing unreliable responses. The price to pay for this improvement is the decreased rate of evidence accumulation. This can be compensated by increasing the number of sets of sensor readings. The sharp difference between Fig. 8 and Fig. 10 are two extreme cases in our experiments. In Fig. 10, the sensor model was assigned the most conservative set of the three values while in Fig. 8 it was just the opposite case. Obviously the robot in Fig. 10 was much better in reducing unreliable responses than in Fig. 8. However as also shown in Fig. 10, a rather great amount of information has been lost as well. In our experiments, we have found the group of values used in Fig. 7 was most satisfactory. Therefore, to find an effective set of values for these three parameters is of great importance in RCF s performance in dealing with the specular reflection problem. The selection of the three parameters is very much a tried and error process.

Figure 9. RCF min =0.1, τ =0.5,R max =325cm Figure 10. RCF min =0.1, τ =2,R max =200cm 5. CONCLUSIONS In this paper, we have investigated the use of the range confidence factor (RCF) proposed in Ref. 7, in ultrasonic sensor model to reduce unreliable responses caused by specular reflections. Dempster-Shafer s evidence theory was used as the basis for multi-ultrasonic sensor fusion strategy on the evidence grid with the appropriate sensor model. Experimental results have confirmed the important role of RCF in reducing unreliable responses. Since the performance of RCF depends highly on the selection of the three values in Equation (12), further work is necessary on developing adaptive methods in sensor model for navigations in unknown environments.

REFERENCES 1. Alberto Elfes Sonar-Based Real-World Mapping and Navigation,IEEE Journal of Robotics and Automation,Vol. RA-3, NO. 3, pp.249-265,1987. 2. A. Howard and L. Kitchen, Generating Sonar Maps in Highly Specular Environments, Proceedings of the Fourth International Conference on Control, Automation, Robotics and Vision, pp.1870-1874,1996. 3. C. C. Chang and Kai-Tai Song, Ultrasonic Sensor Data Integration and Its Application to Environment Perception, Journal of Robotic Systems, pp.663-677, 1996. 4. D. P. Eduardo, M. Nebot and H. Durrant-Whyte, An Evidential Approach to Map-Building for Autonomous Vehicles, IEEE Transaction on Robotics and Automation, Vol. 14, NO.4, pp.623-629, 1998. 5. D. L. Hall, Mathematical Techniques in Multisensor Data Fusion, Artech House, Boston,1992. 6. G. Shafer, A Mathematical Theory of Evidence,Princeton University Press, New York,1976 7. J. W. Guan and D. A. Bell,Evidence Theory And Its Applications, Volume 1, North-Holland, 1992. 8. J. H. Lim and D. W. Cho, Specular Reflection Probability in the Certainty Grid Representation, Transactions of the ASME, Vol. 116, pp.512-520,1994. 9. J. J. Leonard and H. F. Durrrant-Whyte,Directed Sonar Sensing for Mobile Robot Navigation, KluwerAcademic Publishers,1992. 10. K. Hughes and R. Murphy, Ultrasonic Robot Localization Using Dempster-Shafer Theory, Proceedings of SPIE on Neural and Stochastic Methods in Image and Signal Processing, Vol. 1766, pp.2-11,1992. 11. M. Beckerman and E. M. Oblow, Treatment of Systematic Errors in the Processing of Wide-Angle Sonar Sensor Data for Robotic Navigation,IEEE Transactions on Robotics and Automation, Vol. 6, NO.2, pp.137-145,1990. 12. M. Drumheller, Mobile Robot Localization Using Sonar, IEEE Transactions On Pattern Analysis and Machine Intelligence, Vol.PAMI-9,NO.2,March1987 13. M. C. Martin and H. P. Moravec, Robot Evidence Grids, Carnegie Mellon University, 1996.