EDROM Humanoid Kid Size 2014



Similar documents
Frequently Asked Questions

Frequently Asked Questions

Rhoban Football Club Team Description Paper

Obstacle Avoidance Design for Humanoid Robot Based on Four Infrared Sensors

An inertial haptic interface for robotic applications

Hand Gestures Remote Controlled Robotic Arm

An experience for teaching humanoid robotics in computer engineering studies

In: Proceedings of RECPAD th Portuguese Conference on Pattern Recognition June 27th- 28th, 2002 Aveiro, Portugal

FRC WPI Robotics Library Overview

Digital Position Control for Analog Servos

Vision-based Walking Parameter Estimation for Biped Locomotion Imitation

Networking Remote-Controlled Moving Image Monitoring System

Solar Cybertech: A Competition of Digitally Controlled Vehicles Poweredby Solar Panels

Human-like Arm Motion Generation for Humanoid Robots Using Motion Capture Database

Effective Use of Android Sensors Based on Visualization of Sensor Information

C.I. La chaîne d information LES CAPTEURS. Page 1 sur 5

How To Program With Adaptive Vision Studio

Massachusetts Institute of Technology

Design of Intelligent Robotics. René van de Molengraft December, 6th, 2010

Figure Cartesian coordinate robot

Introduction to Robotics Analysis, Systems, Applications

REMOTE CONTROL OF REAL EXPERIMENTS VIA INTERNET

How To Run A Factory I/O On A Microsoft Gpu 2.5 (Sdk) On A Computer Or Microsoft Powerbook 2.3 (Powerpoint) On An Android Computer Or Macbook 2 (Powerstation) On

EQUIPMENT AND MATERIALS

Interactive Computer Graphics

CAGILA 2D and 3D software

Self-Balancing Robot Project Proposal Abstract. Strategy. Physical Construction. Spencer Burdette March 9, 2007

Visual Servoing Methodology for Selective Tree Pruning by Human-Robot Collaborative System

Whitepaper. Image stabilization improving camera usability

ENGINEERING METROLOGY

How To Use Trackeye

INTRODUCTION TO SERIAL ARM

EasyC. Programming Tips

Physics 221 Experiment 5: Magnetic Fields

Introduction to Accuracy and Repeatability in Linear Motion Systems

UPS PIco. to be used with. Raspberry Pi B+, A+, B, and A. HAT Compliant. Raspberry Pi is a trademark of the Raspberry Pi Foundation

Indoor Surveillance System Using Android Platform

On-line Robot Adaptation to Environmental Change

Project. Keywords: Hybrid Algorithm, Human-Face Detection, Tracking Unrestricted, Identification of People, Fall, dynamic and kinematic.

Poker Vision: Playing Cards and Chips Identification based on Image Processing

A System for Capturing High Resolution Images

Analytics. CathexisVision. CathexisVision Features. Video Management Solutions

LEGO NXT-based Robotic Arm

Final Year Projects at itm. Topics 2010/2011

SNAPPIN.IO. FWR is a Hardware & Software Factory, which designs and develops digital platforms.

TESTBOX SHAKE TABLE USER MANUAL

Product Information. QUADRA-CHEK 3000 Evaluation Electronics For Metrological Applications

DIGITAL DISPLACEMENT RISING STEP LOAD LRA/RSL TEST EQUIPMENT

TRIMBLE ATS TOTAL STATION ADVANCED TRACKING SYSTEMS FOR HIGH-PRECISION CONSTRUCTION APPLICATIONS

Indoor Surveillance Security Robot with a Self-Propelled Patrolling Vehicle

Dynamic resource management for energy saving in the cloud computing environment

DRO Function guide: Zero setting 1/2 calculation mm/inch conversion INC/ABS conversion

INSTRUCTOR WORKBOOK Quanser Robotics Package for Education for MATLAB /Simulink Users

Lecture Notes - H. Harry Asada Ford Professor of Mechanical Engineering

Z-Knipsers Team Description Paper for RoboCup 2016

Base Station Design and Architecture for Wireless Sensor Networks

INVESTIGATING THE EFFECT OF THE MAGNETIC FIELD FROM A WIRE ON A COMPASS

LAB 7 MOSFET CHARACTERISTICS AND APPLICATIONS

Motion Sensing with mcube igyro Delivering New Experiences for Motion Gaming and Augmented Reality for Android Mobile Devices

Autos Limited Ghana Vehicle Tracking Business Proposal

Research-Grade Research-Grade. Capture

Rules for the IEEE Very Small Competition Version 1.0

Industry First X86-based Single Board Computer JaguarBoard Released

FSI Machine Vision Training Programs

ANDROID LEVERED DATA MONITORING ROBOT

Module 1: Sensor Data Acquisition and Processing in Android

Architecture for Direct Model-to-Part CNC Manufacturing

A Computer Vision System on a Chip: a case study from the automotive domain

Design Report. IniTech for

Automated Recording of Lectures using the Microsoft Kinect

AEO Head Movement Tracker X-GYRO 1000 USER MANUAL(V1.1bata )

Using Received Signal Strength Variation for Surveillance In Residential Areas

T-REDSPEED White paper

Intelligent Flexible Automation

Advantages of Auto-tuning for Servo-motors

RoboCup Advanced 3D Monitor

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

HUR Rehab Line H E A L T H & F I T N E S S E Q U I P M E N T

Creating Smarter, More Interactive Apps and Systems with Computer Vision

Praktikum "Mobile Roboter" mit Simulation und Telerobotik-Zugang (TELEBOTS)

A Method for Image Processing and Distance Measuring Based on Laser Distance Triangulation

Animations in Creo 3.0

How To Control Gimbal

Acceleration levels of dropped objects

Tutorial for Programming the LEGO MINDSTORMS NXT

dspace DSP DS-1104 based State Observer Design for Position Control of DC Servo Motor

Open Architecture Design for GPS Applications Yves Théroux, BAE Systems Canada

telemetry Rene A.J. Chave, David D. Lemon, Jan Buermans ASL Environmental Sciences Inc. Victoria BC Canada I.

A General Framework for Tracking Objects in a Multi-Camera Environment

A Study on M2M-based AR Multiple Objects Loading Technology using PPHT

Synthetic Sensing: Proximity / Distance Sensors

DS1104 R&D Controller Board

Acceleration Introduction: Objectives: Methods:

Acoustic design according to room shape

A new educational tool for Bioloid Kit

Data Sheet. Electro-mechanical counters Hengstler 800 series

AXIS 262+ Network Video Recorder

Rules of Actuator and Guide Alignment in Linear Motion Systems

CS231M Project Report - Automated Real-Time Face Tracking and Blending

VEHICLE DIAGNOSTICS THE AUTOCOM WAY

Transcription:

EDROM Humanoid Kid Size 2014 Ana Patrícia Araújo, Caio Ribeiro, Carlos Eduardo Silva, Carolina Resende, Débora Valverde, Igor Bôas, Iuri Souza, Leonardo Jappe, Leonardo Barbosa, Cassiano Moura, Lucas Gonzaga, Mateus Araújo, Murilo Venâncio, Yuri Sousa, Rogério Gonçalves, João Carlos Carvalho Universidade Federal de Uberlândia, Faculdade de Engenharia Mecânica, Av. João Naves de Ávila 2121, Bloco 1M Uberlândia, Brazil {rsgoncalves@mecanica.ufu.br} Abstract. This paper describes the development, structure and programming of the robot by the group of students named EDROM, from the University of Uberlândia, Minas Gerais, Brazil. This robot is going to the contest on the Robocup 2014 Humanoid Kid Size category. The main ideas for the robot to accomplish its task follow in this article. 1 Introduction The EDROM (Equipe de Desenvolvimento em Robótica Móvel) was created in 2007. Initially, its main objective was to participate in competitions involving war robots. However, any structure built for this purpose was waste just because of the characteristic objective of this modality: the destruction of the enemy robot. Therefore, the team began investing in research in order to participate in nondestructive modalities. Thus, since 2010 it has participated in robotic competitions, such as LARC, CBR and RoboCup. EDROM already won several titles ever since in various categories like Humanoid Race, SEK and Humanoid Soccer [1]. The EDROM team statement committing to participate in the RoboCup 2014 Humanoid League competition, and making a person with sufficient knowledge of the rules available as referee during the competition. In this paper were described the steps necessary to achieve the robotic system that will be exhibited in the Kid Size modality on the Robocup 2014. The first EDROM participation in Robocup was a great experience for all team members on the development and knowledge of methods used to achieve the goal in the modalities in which it participates. Moreover, the learning process due to the difficulties that appeared during the competition triggered a major evolution of the team, thus encouraging further scientific and technical development of more optimized ways now resorted in computer vision and robot movements. Currently, a single-core CPU is used for processing the algorithms that control movement, image recognition and intelligence. Threads were implemented to stratify the activities of the robot and thus achieve a better performance.

The main part of the robot is made of the bioloid kit; there are only two parts that were taken off, the original head, and the Roboard controller. The robot dimensions is in function of the rules [2]. The head was assembled considering its neck movements, which were made with two servo-motors. To lock them it was used a couple of aluminum parts, and in the top of it the Minoru 3D Camera. In the back of the robot, the RB-110 controller was carefully placed, along with all connection cables and PCI mini Graphic card. To protect all of it, it was used an acrylic plate. The robot consists of a camera, servo motors with digital sensors, battery and a friendly user-robot interface composed of keys and LED's. Therefore, the behavior of the robot can be monitored as well as the program inside. Fig. 1. Built robot. 2 Overview Although the robot has been provided with single-core processor, all tasks were divided into four threads. The flags thread simplifies the sensors and camera information easing their interpretation by the intelligence thread, it also monitors the states of the robot; the movement thread recovers the robot from the falls, additionally it performs the movements defined by the intelligence function. The camera thread uses the data obtained by the camera then it recognizes the ball and the goalpost. The camera is capable of capturing 30 frames per second. The CPU used is the Vortex86, single-core, Linux operating system, Lubuntu distribution, which was changed to improve its performance; such as withdrawing unnecessary processes for application such as GUI processes and services relating to printing.

3 Hardware Specifications 3.1 Controller The chosen controller is Roboard RB-110 for its capacity to support Linux, which is used along with several libraries, like OpenCV and RoBoIO. The controller has 1.5 GHz, 250 MB DDR 2 memory, it accepts a SD micro card as the master disk. As previously mentioned above, the Linux operating system was chosen because of its compatibility with OpenCV and it runs smoothly. 3.2 Camera The chosen webcam was Minoru 3D, that can easily calculate distances, thus giving the robot a better knowledge of the environment surrounding it. The camera with two eyes also gives the robot a more friendly face, as the competition is also about making the robots human alike. 3.3 Battery The battery used is a Lithium-Polymer for being lighter than the others with the same capacity, the one that will be used has 2200mAh, 14.8 V. 3.4 Motors The robot has 20 Robotis s motors: 18 servos AX-12A in the legs and arms and 2 analog servos HS-485HB in the neck, whose communication is via TTL for the digital ones and PWM for the analog ones. 3.5 Sensor The sensor that will be used is a Phidget Spatial model 1042, it has a Compass 3- Axis, Gyroscope 3-Axis, Accelerometer 3-Axis. It determines if the robot is falling or has fallen, making it possible to take the necessary action to reverse this situation. 4 Motion About movements, it was made an inverse kinematic modeling of leg s robot. The purpose of solving the inverse kinematics is to find the angle of each joint for a known foot position. The three-dimensional modeling was divided into two twodimensional modeling. The first one was the side view of the robot, whose inputs are the height of the steps and the position of the foot in front of the robot body. The second was the top view. Combining the two projections, kinematic three-dimensional modeling of the foot was achieved. To simplify the model, the plane of the foot was considered

parallel to the floor plane. Besides, the robot body height was considered constant from the floor. Half period of the sine curve was used both on the front view and on the top as the step curves. The sinusoid was adopted because it resembles the trajectory of the human step. Its walking algorithms command the trajectory of the feet to follow sine waves in the x, y and z directions. Once the three directional sine waves are combined, the robot begins to walk. The amplitudes and phases are tweaked empirically. Furthermore, this method was successful because of the relationship between the sinusoidal foot movement and the ZMP (Zero Moment Point) theory [3]. However, the feet trajectories made by only one sinusoidal function are not good enough to make smooth walking pattern. It should be made by combinations of more sinusoidal functions, whose frequencies are two or four times bigger than main sinusoidal function. This modeling allows changing the geometrical parameters input in order to get the output values for different conditions, among them the value of the forward speed, the direction of the step and the orientation of the robot in the field, considering the mechanical constraints of the robot. Therefore, it features an omnidirectional step. Moreover, these parameters can be varied during the execution of the robot controller program enabling control during its movement. The period parameter of the step was chosen as follow: it was observed that at the end of a period time the modeling step of the robot s body did not return to the proper slope to start the next part of the step. Then the natural frequency of the robot was measured in a mechanical test. A force with variable frequency was applied on the side of the robot to determine which frequency gives maximum energy transfer. Knowing the desired frequency, a control interval of the test was determined; thereafter the number of periods within this interval was taken. Thus, the period corresponding to the natural frequency was identified, which was used as the time parameter of the step. 5 Computer Vision There are many difficulties to deal with problems involving image processing because of the high processing power required. It is a big challenge to develop efficient and low-cost techniques, considering the hardware limitations [4]. Therefore, OpenCV library was used because it is an open source computer vision and machine learning software library, designed for computational efficiency and with a strong focus on real-time applications [5]. In this context, the aim of the robot is to identify certain components in images: ball, goalpost, landmarks and others robots. So, the strategy was based on the fact that all these components are always within the limits of the field in the image. Thus, all processing begins by determining the limits of the field, then a region of interest is established and finally external waste components are removed. This was only possible due to low-computational cost routines that give great results. Thereafter, the desired objects are identified in a

selected region that is smaller than the original image, reducing the cost of processing and eliminating external elements that could hinder the better performance. Fig. 2a) Color filter 2b) Field filter Fig. 3a) Ball recognition 3b) Field filter Usually simple operations such as binarization of the images according color limits, morphological transformations like Dilation and Erosion are enough, but

sometimes they are combined with other more advanced algorithms such as the ones for identifying contours and lines (Hough Transform). Algorithms involving histograms and techniques for object tracking have been developed and they are being adapted, improved, and tested in embedded system. Nevertheless, the team is conducting researches on calibration of cameras, 3D computer vision, histogram algorithms, machine learning and other real-time computer vision tools. Using the said operations, data about the location of the object relative to the image are obtained. This data must be processed to achieve their locations relative to the robot. This is achieved from the measures previously obtained from robot dimensions, camera specifications and environmental characteristics, properly related to the positions of the head motors and image data. The team has also been researching the implementation of mapping techniques such as Monte Carlo. Besides, the information from accelerometer-gyroscope-compass, from the judge and from the other robots are the basis for robot to make decisions and react. This work was supported by the FEMEC/UFU, FAPEMIG, CNPq and RADIX. 6 Conclusion This paper presents just a brief description of what was used by this team. There are still a few changes to be made to optimize important functions of the final project. The main goal of this work was to implement a bipedal walking and computational vision method that allows a kid sized humanoid robot to walk and to find the ball and kick it into the goal according to the rules of this competition Robocup 2014 [2]. References 1. EDROM<ufuedrom.wordpress.com> 2. RoboCup Soccer Humanoid League Rules and Setup for the 2014 competition. http://www.informatik.unibremen.de/humanoid/pub/website/downloads/humanoidleague Rules2014-01-10.pdf 3. Bipedal Walking for a Full-sized Humanoid Robot Utilizing Sinusoidal Feet Trajectories and Its Energy Consumption <scholar.lib.vt.edu/theses/available/etd-05042012-155703/unrestricted/han_jaekweon_d_2012.pdf> 4. OPENCV <http://opencv.org/> 5. Learning OpenCV. 2008. BRADSKI, Gary and KAEHLER, Adrian.