TACTILE SENSATION FOR ROBOTS



Similar documents
Industrial Robotics. Training Objective

Vibrations can have an adverse effect on the accuracy of the end effector of a

MMGD0203 Multimedia Design MMGD0203 MULTIMEDIA DESIGN. Chapter 3 Graphics and Animations

Scooter, 3 wheeled cobot North Western University. PERCRO Exoskeleton

Optical Digitizing by ATOS for Press Parts and Tools

Unit 1: INTRODUCTION TO ADVANCED ROBOTIC DESIGN & ENGINEERING

CS 6795 Introduction to Cognitive Science Spring 2012 Homework Assignment 3

ARTIFICIAL INTELLIGENCE METHODS IN EARLY MANUFACTURING TIME ESTIMATION

Force measurement. Forces VECTORIAL ISSUES ACTION ET RÉACTION ISOSTATISM

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network

INJECTION MOLDING COOLING TIME REDUCTION AND THERMAL STRESS ANALYSIS

CALIBRATION OF A ROBUST 2 DOF PATH MONITORING TOOL FOR INDUSTRIAL ROBOTS AND MACHINE TOOLS BASED ON PARALLEL KINEMATICS

An Approach for Utility Pole Recognition in Real Conditions

Intelligent Flexible Automation

Human vs. Robotic Tactile Sensing: Detecting Lumps in Soft Tissue

Capacitive Touch Technology Opens the Door to a New Generation of Automotive User Interfaces

CHAPTER 1. Introduction to CAD/CAM/CAE Systems

Analecta Vol. 8, No. 2 ISSN

Introduction to Computer Graphics

CIM Computer Integrated Manufacturing

CATIA V5 Tutorials. Mechanism Design & Animation. Release 18. Nader G. Zamani. University of Windsor. Jonathan M. Weaver. University of Detroit Mercy

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

The Scientific Data Mining Process

A Short Introduction to Computer Graphics

Development of Robotic End-effector Using Sensors for Part Recognition and Grasping

Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA

Simulation Techniques for Tablet and Mobile Phone Design Bill McGinn; Ansys Senior Application Engineer

Deriving the 12-lead Electrocardiogram From Four Standard Leads Based on the Frank Torso Model

ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES

How To Choose The Right End Effector. For Your Application

Synthetic Sensing: Proximity / Distance Sensors

Véronique PERDEREAU ISIR UPMC 6 mars 2013

Improving a Gripper End Effector

A Cognitive Approach to Vision for a Mobile Robot

PARAMETRIC MODELING. David Rosen. December By carefully laying-out datums and geometry, then constraining them with dimensions and constraints,

HAND GESTURE BASEDOPERATINGSYSTEM CONTROL

Design of a six Degree-of-Freedom Articulated Robotic Arm for Manufacturing Electrochromic Nanofilms

VRSPATIAL: DESIGNING SPATIAL MECHANISMS USING VIRTUAL REALITY

DISPLAYING SMALL SURFACE FEATURES WITH A FORCE FEEDBACK DEVICE IN A DENTAL TRAINING SIMULATOR

A Learning Based Method for Super-Resolution of Low Resolution Images

Robots are ready for medical manufacturing

Brain Maps The Sensory Homunculus

5-Axis Test-Piece Influence of Machining Position

An Automatic Optical Inspection System for the Diagnosis of Printed Circuits Based on Neural Networks

Atomic Force Microscope and Magnetic Force Microscope Background Information

Robotics. Lecture 3: Sensors. See course website for up to date information.

Robotic Apple Harvesting in Washington State

A new technology for high current, low insertion force, low resistance and long cycle life power connectors

Robotics and Automation Blueprint

Robot Task-Level Programming Language and Simulation

Super-resolution method based on edge feature for high resolution imaging

Lymon C. Reese & Associates LCR&A Consulting Services Tests of Piles Under Axial Load

Automated Pavement Distress Survey: A Review and A New Direction

Sensory-motor control scheme based on Kohonen Maps and AVITE model

Introduction.

A PAIR OF MEASURES OF ROTATIONAL ERROR FOR AXISYMMETRIC ROBOT END-EFFECTORS

Lecture 6 Scanning Tunneling Microscopy (STM) General components of STM; Tunneling current; Feedback system; Tip --- the probe.

DEFINING AND MEASURING PHYSICAL PARAMETERS OF PC POLISHED FIBER OPTIC CONNECTORS

Application Example: Quality Control of Injection-Molded Parts

Making Accurate Voltage Noise and Current Noise Measurements on Operational Amplifiers Down to 0.1Hz

Colorado School of Mines Computer Vision Professor William Hoff


Design of a Universal Robot End-effector for Straight-line Pick-up Motion

Depth and Excluded Courses

Introduction. C 2009 John Wiley & Sons, Ltd

KALMAN FILTERING FOR 3 D REAL TIME DATA VISUALIZATION OF MULTIBEAM SONAR RECORDS

USING COMPUTER VISION IN SECURITY APPLICATIONS

Robotic assembly. Assembly cost per product. Manual assembly. Automatic assembly using special purpose machines. Annual Production Volume

ExmoR A Testing Tool for Control Algorithms on Mobile Robots

Exploring Computer Science A Freshman Orientation and Exploratory Course

Definitions. A [non-living] physical agent that performs tasks by manipulating the physical world. Categories of robots

Electronic Interface for Pneumatic Grippers Using USB Port

Chapter 3 SYSTEM SCANNING HARDWARE OVERVIEW

The Visual Cortex February 2013

Vision based Vehicle Tracking using a high angle camera

Development of Robotic End-Effector Using Sensors for Part Recognition and Grasping

Frequency, definition Modifiability, existence of multiple operations & strategies

Mobile Robotics I: Lab 2 Dead Reckoning: Autonomous Locomotion Using Odometry

BARE PCB INSPECTION BY MEAN OF ECT TECHNIQUE WITH SPIN-VALVE GMR SENSOR

May 29, Valu Flex TM BIW Geo System A New Concept of Geometry Tooling for Body-In-White Applications

Introduction to Robotics Analysis, systems, Applications Saeed B. Niku

COMP175: Computer Graphics. Lecture 1 Introduction and Display Technologies

Extracting Business. Value From CAD. Model Data. Transformation. Sreeram Bhaskara The Boeing Company. Sridhar Natarajan Tata Consultancy Services Ltd.

TRACKING DRIVER EYE MOVEMENTS AT PERMISSIVE LEFT-TURNS

Periodontology. Digital Art Guidelines JOURNAL OF. Monochrome Combination Halftones (grayscale or color images with text and/or line art)

UOM Robotics and Industrial Automation Laboratory

E190Q Lecture 5 Autonomous Robot Navigation

Submitted By: Submitted To: XYZ ABCD ******** Mechatronics

Design of Multi-camera Based Acts Monitoring System for Effective Remote Monitoring Control

Such As Statements, Kindergarten Grade 8

Agent Simulation of Hull s Drive Theory

Robot coined by Karel Capek in a 1921 science-fiction Czech play

White Paper. "See" what is important

User Guide MTD-3. Motion Lab Systems, Inc.

CHAPTER 6 PRINCIPLES OF NEURAL CIRCUITS.

FUNDAMENTALS OF ROBOTICS

THE STRAIN GAGE PRESSURE TRANSDUCER

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Automatic Detection of PCB Defects

What is Visualization? Information Visualization An Overview. Information Visualization. Definitions

Transcription:

TACTILE SENSATION FOR ROBOTS Kenneth J. Overton** Computer and Information Science Department University of Massachusetts Amherst, Massachusetts Thomas Williams Corporate Research Group Digital Equipment Corporation Maynard, Massachusetts ABSTRACT In this paper we provide a brief discussion of the relationship of tactile sensation to other sensory modalities currently showing prominence in industrial robots. A tactile sensor is presented which provides a matrix of the forces existing between it and the object with which it is in contact. The possible uses of this information in the control of the manipulation of objects by a robot arm/hand system are considered. Examples of the data retrieved from the sensor are given and possible initial processing steps are discussed. A. Introduction To date, the dominant area in robot sensing research has been vision. Notable exceptions include the force feedback or active accommodation work done in connection with the task of peg insertion, e.g. [6] and [11], and the use of vision and force sensing in assembly [1], [11], [12], [18] and [43 (passive accommodation). Other areas of sensor research include range finding techniques, [14] [153, proximity sensing and some limited investigations into the use of tactile force sensor arrays in manipulation (see below). For a comprehensive review of the current state of robot sensor technology the reader is referred to [133. An area receiving limited attention in the literature, but comprising an important area in and of itself is that of the use of tactile sensory data in the control of robot systems. Research in this area has been in progress, and for further discussions, the reader is referred to [17] and recent work by Dixon [33. Research in the use of tactile data by robot manipulators has been generally limited to gross force sensing in the wrist, and in some cases the fingers, for two general reasons. First, until recently, there has been a lack of high resolution sensing arrays which could be utilized in the end effectors of a manipulator. Second, and probably more important, people are far less conscious of the role that the "sense of touch'' plays in everyday activity, as compared to vision. While vision may continue to This work was supported in part by NIH Grant NS 14971-03 (Michael A. Arbib, Principal Investigator) and Digital Equipment Corporation. International Business Machines Graduate Fellow during the 1980-81 academic year. be the primary sense utilized in robots, tactile sensing fills an important role, not only when vision is inapplicable due to inherent ambiguities, but in the very fine control of delicate manipulation. Tactile sensors can be divided into three general categories depending upon the type of information they provide. Contact sensors compose the first class and yield information indicating either the presence or absence of physical stimuli. A robot hand with contact sensing on several sides was developed by Goto [7 3 to handle blocks placed on a table. In latter work, Garrison and Wang [5] constructed a hand containing an array of contact sensors. Recent sensor arrays have utilized woven graphite fibers as the physical transducer [103. The Laboratory for Automation and Systems Analysis at CNRS in Toulouse, France, has developed an "artificial skin" tactile sensor suitable for gross object recognition [23. A high spatial resolution conductive rubber contact sensing array has been developed at the Artificial Intelligence Laboratory at M.I.T. and has been successfully used to identify a bolt from its contact pattern [93. The second category contains the sensors which provide information regarding the magnitude of the force at each sensing point. A hand constructed by Hill and Sword [83 employed analog force sensing arrays. Several of the contact sensing arrays mentioned above can and are being extended to provide analog force images. Depending upon the compliance of the material supporting such sensors, the force information can be related to the amount of deflection of the sensor, thus providing information describing the three dimensional shape of the object being sensed. The final class of tactile sensors provides this information directly. This group consists of sensors which are designed to transduce the three dimensional shape of an object. Page and Pugh [163 presented a sensor which builds a contour map of the object being sensed as the sensor is lowered onto the object. Takeda constructed a hand with arrays of free floating needles as sensors. The needles would conform to the shape of the surface being touched, thereby providing identification information [173. 791

B. The Tactile Sensor This section describes the physical and electrical characteristics of the conductive elastomeric tactile sensor currently being investigated by the authors in the Digital Corporate Research Group. The sensor is designed to provide an analog image of the forces produced when an object comes in contact with the sensor. The columns are arranged in a square matrix with their center to center spacing equal to.1 inches, providing a resolution of 100 detectors per square inch of sensor surface. Along the upper surface of the sensor, the force elements are connected by fine wire in such a way as to form 10 rows. The sensor rests on a printed circuit board which provides connection in the orthogonal direction. The matrix is scanned by an LSI-11/23 computer which collects the data and provides some initial preprocessing. As indicated by Larcombe [10] and Hillis [9], sensing of a single pressure point is possible in an array as just described. However, complex force patterns are difficult to understand due to the electrical interconnection of the individual force detecting elements. This problem is alleviated in the present design by the insertion of a diode in series with each of the elements. The scanning process produces a raster of force values that we call a "force image". Of cource any sampling sequence is possible since the sensor is a random access device. The values returned are proportional to the resistance across the force detector connected to the selected digital line. Since the resistance is proportional to compression of the element, the value returned is an indication of the force being experienced at that detector. The diodes eliminate interference due to alternate current paths in the network. The second function of the computer is to provide some preprocessing of the raw sensor data. This includes the removal of the offsets of the individual detectors, scaling of the data, and the removal of nonlinearities. The change in resistance of the detectors versus the amount of force applied is nonlinear. The degree of nonlinearity and the range over which the detector is applicable is the subject of a series of ongoing tests. Experiments are also being performed to determine the effects of ambient temperature changes and drift of the resistance during prolonged application of force. All of these characteristics have yet to be understood but the initial test results are quite favorable. The most common usage for such sensors is to record the static image produced by an object. Figures 1 through 4 contain grey scale representations of four static force images produced by various stimuli. Each grey scale display Is composed of a square array of 100 "forcels" (force elements), one corresponding to each force detector in the sensor, where the brightness of the forcel is proportional to the force at the corresponding element. Figure 1 is the image produced when a spherical object is pressed onto the sensor. Note that there are three bad forcels in the image: the element in the upper left corner and the two near the center of the pattern of forces produced by the stimulus. Figure 2 resulted when a long thin cylinder was used as the stimulus. The force image of a plate with a hole through it appears in Figure 3 and Figure 14 shows a "V" shaped object depressing the sensor. Figure 1 Grey scale image of the forces sensed when the stimulus is a hemispherical object. C. Sensor Data This section presents some of the force image data produced by the sensor. The sensor was mounted on a circuit board and placed on a table. A stimulus was applied to the sensor by pressing the object to be sensed down onto the sensory surface. The computer was programmed to continually scan the sensor and store each image in a file. The images were later transported to a VAX 11/780*computer where they were displayed on a color graphics terminal ana a plotter. LSI-11/23 and VAX are registered trademarks of the Digital Equipment Corporation Figure 2 Grey scale Force Image of a cylindrical object. thin 792

D. Use of the Data The tactile sensor provides the host computer with an array of numbers, or a "force image", representing the magnitudes of the forces being applied across the sensory surface. The question now becomes one of how the control structures of the robot system are to utilize this information. One such use is to provide accurate information regarding the identity, position, and orientation of the object being touched. Figure 3 Grey scale Image produced by a rigid plate with a hole. Figure 4 Grey scale Force Image produced by the jaws of a pair of pliers. Another form of data can be collected with this sensor and computer system. That is, the sensor can be used to record the sequence of force images resulting when an object moves across the sensory surface. Figure 5 contains four consecutive force images collected as a cylindrical object rolled across the sensor from the upper left corner. The motion of the object is clearly visible and measureable relative to the sensor. Thus, if the sensor had been mounted on a robot gripper, the motion of the stlmuls would be directly available to the computer system controlling the robot. Notice again the bad forcels. The grey scale representation is one possible form for the display of such data. Recall that the forces produced are proportional to the deflection of the sensor. It is then reasonable to plot the data as a surface where the height of the point corresponding to an individual force detector is proportional to the force at that location. Figures 6 through 9 correspond to the grey scale representations shown in Figures 2 through 4. In particular, given the geometric description of the surface of an object, the exact location and orientation in space of the object can be derived from arrays of such lensors. The array of force values may be thought of as comprising a set of control points for a surface patch [193. This patch can be compared with the representation of the surface of the object, thus allowing determination of the location and orientation of the corresponding patch on the object. The accuracy in this case is limited by the possible placements of the surface patch on the object due to symmetry. When two or more fingers are utilized, the force images represent very localized views of the complete object. Each view provides the orientation of a subpatch on the object. Further Information is available from the known spatial relationships between the force sensing arrays at the ends of the fingers. In this case the subpatches are constrained In their relative locations in space. One has very localized samplings of the surface contour detail by the force sensing arrays and global sampling of the contour by knowing the relationships between the finger tips. This information can be used to determine the position of the object, configuration of the grasp, etc. All of the work cited earlier and the approach outlined in the discussion above have one basic idea in common. That is, the information sensed is static. The sensors are employed to take a "snapshot" of the environment which is then processed to yield the desired information. The concept of touch extends beyond this to include the sensing and analyzing of the dynamic characteristics of the tactile information. Consider, for example, the case when one reaches behind a coffee cup and "feels" for the handle. In many cases partial matching in one sensory modality, e.g. vision, may elicit actions which focus sensors in other modalities in the area containing the object in question. Biological evidence of this comes with both foveation (in vision), and exploration with the glabrous skin in hapsis. This allows initial hypotheses to be confirmed by obtaining additional sensory information. Just as static image processing techniques could be applied to static force images, moving image analysis techniques could be applied to dynamic force images. Consider the increase in resolution or reduction in noise that can be obtained by comparison of sequential visual images. One would expect similar improvements with dynamic force sensed images. Although no results are given for such processing, experiments are underway to evaluate its applicability. 793

c E. Conclusion We have briefly reviewed the position of tactile sensory research for robots In light of the other nodalities and have outlined the research to date In this area. A tactile sensor was presented and examples of the data from the sensor were given. The work being conducted by the authors in this area is Just beginning and an extensive program for the investigation of tactile sensation and its applications to industrial robotics Is planned. REFERENCES [1] Blnford, T.O. et al., "Computer Integrated Assembly Systems," Progress Report to NSF, Covering Period March 15. 1974 to September 15, 1974, Stanford AI Lab, September 1974. [2] Briot, M., "The Utilization of an 'Artificial Skin* Sensor for the Identification of Solid Objects," Proceedings of the 9th ISIR, pp. 529-548, Washington, D.C., March 1979. Figure 5 A sequence of images produced by a cylindrical object moving across the sensor represented as grey scale images. d [3] Dixon, J.K., Salazar, S.. ana Slagle, J.R., "Research on Tactile Sensors for an Intelligent Naval Robot," Proceedings of the 9th TSIR, pp. 507-518, Washington, D.C., March 1979. [4] Drake, S.H., Watson, P.C., 8nd Simunovic, S.N., "High Speed Assembly of Precision Parts Using Compliance Instead of Sensory Feedback," Proceedings of 7th ISIR, Tokyo, 1977. [5] Garrison, R.L. and Wang, S.S.W., "Pneumatic Touch Sensor," IBM Technical Disclosure Bulletin, vol. 16, no. 6, November 1973. [6] Goto, T., Inoyaroa, T., and Takeyasu, K., "Precise Insert Operation by Tactile Controlled Robot HI-T-HAND Expert-2," in Proceedings of 4th ISIR, Tokyo, pp. 209-218, 1974. [7] Goto, T., "Compact Packaging by Robot with Tactile Sensors," Proceedings of 2nd ISIR, Chicago, 1972. 7 94

[8] Hill, J.W. and Sword, A.J., "Manipulation Based on Sensor-Directed Control: An Integrated End Effector and Touch Sensing System," Proceedings of the 17th Annual Human Factor Society Convention, Washington, D.C., October 1979. [9] Hillis, W.D., "Active Touch Sensing," MIT AI Lab Memo 629, April, 1981. [10] Larcombe, M.H.E., "Tactile Sensors, Sonar Sensors, and Parallax Sensors for Robot Applications," 3rd Conf. on Robot Tech. and 6th ISIR. University of Nottingham, U.K., March 2M-26, 1976. [11] Nevins, J.L. et al., "Exploratory Research in Industrial Assembly Part MAting," 5th and 6th Reports, C.S. Draper Laboratory, reports R-1111 and R-1276. 1980. [12] Nevins, J.L. et al., "computer Controlled Assembly," Scientific American, vol. 238, no. 2, pp. 62-74. February 1978. [13] Nitzan, D., "Assessment of Robotic Sensors," presented at NSF Robotics Research Workshop, Newport, Rhode Island, April 15-17, 1980 [14] Nitzan, D., Brain, A.E., and Duda, R.O., "The Measurement and Use of Registered Reflections and Range Data in Scene Analysis," Proceedings IEEE, vol. 65, pp. 206-220, February 1977. [15] Nitzan, D., "Scene Analysis Using Range Data," Technical Note 69, Artificial Intelligence Center, SRI, August 1972. [16] Page, C.J., Pugh, A., and Heginbotham, W.B., "Novel Techniques for Tactile Sensing in a Three-Dimensional Environment," 3rd Conf. on Robot Tech. and 6th ISIR, University of Nottingham, March 24-26, 1976. [17] Takeda, S., "Study of Artificial Tactile Sensors for Shape Recognition Algorithm for Tactile Data Input," Proceedings of the 4th International Symposium on Industrial Robots, pp. 199-208. Tokyo, November 1974. [18] Will, P.M., "Computer Controlled Mechanical Assembly," Proceedings on the 5th Intl. Sym. on Ind. Robots, Chicago, 1975. [19] York, B.W. and Riseman, E.M., "A Surface Representation For Computer Vision," Proceedings of the IEEE Workshop on Picture Data Description and Management, August 27-28, 1980. 795