Team Description for Humanoid KidSize League of RoboCup Stephen McGill, Seung Joon Yi, Yida Zhang, Aditya Sreekumar, and Professor Dan Lee

Similar documents
Robo-Erectus Jr-2013 KidSize Team Description Paper.

RoboCup TDP Team ZSTT

Kid-Size Humanoid Soccer Robot Design by TKU Team

FUmanoid Team Description Paper 2010

Team TH-MOS. Liu Xingjie, Wang Qian, Qian Peng, Shi Xunlei, Cheng Jiakai Department of Engineering physics, Tsinghua University, Beijing, China

Team Description 2006 for Team RO-PE A

KMUTT Kickers: Team Description Paper

ROBOTICS ENG YOUSEF A. SHATNAWI INTRODUCTION

EROS TEAM. Team Description for Humanoid Kidsize League of Robocup2013

Robo-Erectus Tr-2010 TeenSize Team Description Paper.

Team TH-MOS Abstract. Keywords. 1 Introduction 2 Hardware and Electronics

Team Description Paper: HuroEvolution Humanoid Robot for Robocup 2014 Humanoid League

Team RoBIU. Team Description for Humanoid KidSize League of RoboCup 2014

Optic Flow Based Skill Learning for A Humanoid to Trap, Approach to, and Pass a Ball

Rapid Development System for Humanoid Vision-based Behaviors with Real-Virtual Common Interface

ZJUDancer Team Description Paper Humanoid Kid-Size League of Robocup 2014

Team AcYut Team Description Paper 2018

ZJUDancer Team Description Paper Humanoid Kid-Size League of Robocup 2015

FalconBots RoboCup Humanoid Kid -Size 2014 Team Description Paper. Minero, V., Juárez, J.C., Arenas, D. U., Quiroz, J., Flores, J.A.

The UPennalizers RoboCup Standard Platform League Team Description Paper 2017

WF Wolves & Taura Bots Humanoid Kid Size Team Description for RoboCup 2016

BehRobot Humanoid Adult Size Team

Tsinghua Hephaestus 2016 AdultSize Team Description

Baset Adult-Size 2016 Team Description Paper

Team Description Paper: HuroEvolution Humanoid Robot for Robocup 2010 Humanoid League

NimbRo TeenSize 2014 Team Description

RoboCup 2012 Best Humanoid Award Winner NimbRo TeenSize

Perception. Read: AIMA Chapter 24 & Chapter HW#8 due today. Vision

DEVELOPMENT OF A HUMANOID ROBOT FOR EDUCATION AND OUTREACH. K. Kelly, D. B. MacManus, C. McGinn

AcYut TeenSize Team Description Paper 2017

VATIO UP Team Description Paper for Humanoid KidSize League of RoboCup 2013

A Semi-Minimalistic Approach to Humanoid Design

Active Stabilization of a Humanoid Robot for Impact Motions with Unknown Reaction Forces

KUDOS Team Description Paper for Humanoid Kidsize League of RoboCup 2016

Why Humanoid Robots?*

Team MU-L8 Humanoid League TeenSize Team Description Paper 2014

Sensor system of a small biped entertainment robot

Humanoids. Lecture Outline. RSS 2010 Lecture # 19 Una-May O Reilly. Definition and motivation. Locomotion. Why humanoids? What are humanoids?

Team Description Paper: Darmstadt Dribblers & Hajime Team (KidSize) and Darmstadt Dribblers (TeenSize)

Advanced Distributed Architecture for a Small Biped Robot Control M. Albero, F. Blanes, G. Benet, J.E. Simó, J. Coronel

Humanoid Robot HanSaRam: Recent Development and Compensation for the Landing Impact Force by Time Domain Passivity Approach

YRA Team Description 2011

Bogobots-TecMTY humanoid kid-size team 2009

Humanoid robot. Honda's ASIMO, an example of a humanoid robot

Active Stabilization of a Humanoid Robot for Real-Time Imitation of a Human Operator

Design and Experiments of Advanced Leg Module (HRP-2L) for Humanoid Robot (HRP-2) Development

Adaptive Motion Control with Visual Feedback for a Humanoid Robot

Adaptive Dynamic Simulation Framework for Humanoid Robots

Mechanical Design of Humanoid Robot Platform KHR-3 (KAIST Humanoid Robot - 3: HUBO) *

Korea Humanoid Robot Projects

RoboCup 2013 Humanoid Kidsize League Winner

Nao Devils Dortmund. Team Description for RoboCup Matthias Hofmann, Ingmar Schwarz, and Oliver Urbann

Cost Oriented Humanoid Robots

HUMANOID ROBOT SIMULATOR: A REALISTIC DYNAMICS APPROACH. José L. Lima, José C. Gonçalves, Paulo G. Costa, A. Paulo Moreira

Hanuman KMUTT: Team Description Paper

Development of a Humanoid Biped Walking Robot Platform KHR-1 - Initial Design and Its Performance Evaluation

Current sensing feedback for humanoid stability

SitiK KIT. Team Description for the Humanoid KidSize League of RoboCup 2010

Integration of Manipulation and Locomotion by a Humanoid Robot

SPQR RoboCup 2016 Standard Platform League Qualification Report

MRL Team Description Paper for Humanoid KidSize League of RoboCup 2014

ICHIRO TEAM - Team Description Paper Humanoid TeenSize League of Robocup 2018

ICHIRO TEAM - Team Description Paper Humanoid KidSize League of Robocup 2017

Chapter 1 Introduction

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment

Team KMUTT: Team Description Paper

Autonomous Stair Climbing Algorithm for a Small Four-Tracked Robot

UChile Team Research Report 2009

Darmstadt Dribblers 2005: Humanoid Robot

Introduction to Humanoid Robotics by Dr. Rawichote Chalodhorn (Choppy)

ZJUDancer Team Description Paper

Technique of Standing Up From Prone Position of a Soccer Robot

Shuffle Traveling of Humanoid Robots

MRL Team Description Paper for Humanoid KidSize League of RoboCup 2013

Birth of An Intelligent Humanoid Robot in Singapore

DEVELOPMENT OF THE HUMANOID ROBOT HUBO-FX-1

Design and Implementation of Humanoid Biped Walking Robot Mechanism towards Natural Walking

CONTROL SYSTEM TO BALANCE A BIPED ROBOT BY THE SENSING OF COG TRAJECTORIES

NTU Robot PAL 2009 Team Report

Design and Implementation of a Simplified Humanoid Robot with 8 DOF

sin( x m cos( The position of the mass point D is specified by a set of state variables, (θ roll, θ pitch, r) related to the Cartesian coordinates by:

UKEMI: Falling Motion Control to Minimize Damage to Biped Humanoid Robot

FU-Fighters. The Soccer Robots of Freie Universität Berlin. Why RoboCup? What is RoboCup?

Rhoban Football Club Team Description Paper

The UT Austin Villa 3D Simulation Soccer Team 2008

RoboPatriots: George Mason University 2014 RoboCup Team

Cost Oriented Humanoid Robots

Design and Control of the BUAA Four-Fingered Hand

Chapter 1. Robot and Robotics PP

Using Robot Operating System (ROS) and Single Board Computer to Control Bioloid Robot Motion

Mechanical Design of the Humanoid Robot Platform, HUBO

RoboCup. Presented by Shane Murphy April 24, 2003

Courses on Robotics by Guest Lecturing at Balkan Countries

Biologically Inspired Locomotion Strategies: Novel Ground Mobile Robots at RoMeLa

Engineering Solutions to Build an Inexpensive Humanoid Robot Based on a Distributed Control Architecture

Stabilize humanoid robot teleoperated by a RGB-D sensor

NimbRo 2005 Team Description

MRL Team Description Paper for Humanoid KidSize League of RoboCup 2017

Development and Evaluation of a Centaur Robot

CIT Brains (Kid Size League)

Transcription:

Team DARwIn Team Description for Humanoid KidSize League of RoboCup 2013 Stephen McGill, Seung Joon Yi, Yida Zhang, Aditya Sreekumar, and Professor Dan Lee GRASP Lab School of Engineering and Applied Science, University of Pennsylvania, USA E-mail: ddlee@seas.upenn.edu Web: http://www.seas.upenn.edu/~robocup/ Abstract This paper details the hardware, software, and electrical design of the humanoid robot family, DARwIn (Dynamic Anthropomorphic Robot with Intelligence) a robot family designed as a platform for researching bipedal motions. The DARwIn family was the first US entry into the humanoid division of RoboCup. 1 Introduction The DARwIn (Dynamic Anthropomorphic Robot with Intelligence) series robot is a family of humanoid robots capable of bipedal walking and performing human-like motions (Fig. 1). DARwIn is a research platform developed at the Robotics and Mechanisms Laboratory (RoMeLa) at Virginia Tech for studying robot locomotion and sensing. It was also utilized as the base platform for Virginia Tech s first entry to the humanoid division of RoboCup 2007. [1,2]. The 455 mm tall, 2.8 kg DARwIn-OP (the latest version of DARwIn) has 20 degrees-offreedom (DOF) with each joint actuated by coreless DC motors via distributed control with controllable compliance. DARwIn-OP was developed through collaboration between Virginia Tech, Purdue, University of Pennsylvania, and Robotis under support from the National Science Foundation. Using a computer vision system and accelerometer, DARwIn-OP can implement human-like gaits while navigating obstacles and traverse uneven terrain while implementing complex behaviors such as playing soccer. For RoboCup 2013, Team DARwIn, hopes to replicate our success in the highly competitive RoboCup competition. This year, the team will be comprised of solely the University of Pennsylvania, with more mechanical engineers to provide support at UPenn to deal with required changes to the foot, for example.

(a) DARwIn I (b) DARwIn II (d) DARwIn IV (c) DARwIn III (e) DARwIn-OP Fig. 1. Family of DARwIn Robots Team DARwIn would like to commit to participate in the RoboCup 2013 Humanoid League competition. Team DARwIn is able to provide students who participated in last year s competition to serve as referees, as they have sufficient knowledge of the rules.

2 Research The DARwIn family serves as a research platform used for studying dynamic gaits and walking control algorithms. With few exceptions (i.e. the Honda ASIMO, the Sony QRIO, and the KAIST HUBO [3 7]), most legged robots today walk using what is called the static stability criterion. The static stability criterion is an approach to prevent the robot from falling down by keeping the center of mass of its body over the support polygon by adjusting the position of its links and pose of its body very slowly to minimize dynamic effects [5]. Thus at any given instant in the walk, the robot could pause and not fall over. Static stability walking is generally energy inefficient since the robot must constantly adjust its pose to keep the center of mass of the robot over its support polygon, which generally requires large torques at the joint actuators (similar to a human standing still with one foot off the ground). Humans naturally walk dynamically with the center of mass rarely inside the support polygon. Thus human walking can be considered as a cycle of continuously falling and catching its fall: a cycle of exchanging potential energy and kinetic energy of the system like the motion of an inverted pendulum. Humans fall forward and catch themselves with the swinging foot while continuously progressing forward. This falling motion al- lows the center of mass to continually move forward, minimizing the energy that would reduce the momentum. The lowered potential energy from this forward motion is then increased again by the lifting motion of the supporting leg. One natural question that arises when examining dynamic walking is how to classify the stability of the gait. Dynamic stability is commonly measured using the Zero Moment Point (ZMP), which is defined as the point where the influence of all forces acting on the mechanism can be replaced by one single force without a moment term [8]. If this point remains in the support polygon, then the robot can have some control over the motion of itself by applying force and/or torque to the ground. Once the ZMP moves to the edge of the foot, the robot is on the verge of stability and can do nothing to recover without extending the support polygon (planting another foot or arm). Parameterized gaits can be optimized using the ZMP as a stability criterion or stable hyperbolic gaits can be generated by solving the ZMP equation for a path of the center of mass. Additionally, the ZMP can be measured directly or estimated during walking to give the robot feedback to correct and control its walking. DARwIn is developed and being used for research on such dynamic gaits and control strategies for stability [5, 9].

3 Hardware DARwIn-OP has 20 degrees of freedom (six in each leg, three in each arm, and two in the head). The robot s links are fabricated out of aluminum. The robot uses Robotis Dynamixel RX-28M motors for the joints [10]. The motors operate on a serial TTY network, allowing the motors to be daisy chained together. Each motor has its own built-in optical encoder and position feedback controller, creating distributed control. The computers, sensors, electronics, and computer ports are distributed about DARwIn s upper torso. 4 Electronics DARwIn-OP s electronic system provides power distribution, communication buses, computing platforms, and sensing schemes aimed at making sense of a salient environment. An 11.1V (nominal) lithium polymer battery provides power to the joint actuators and electronics. Thus battery provides 1 Ah, which gives DARwIn approximately 30 minutes of run time. Computing tasks are performed on a Compulabs fit-pc2i computing system that runs the Intel Atom Z530 CPU with 1GB of onboard RAM and built-in WiFi. The PC consumes 8W at full CPU usage. In addition, the computer connects to both a Logitech C905 camera and a Robotis CM-730 microcontroller board. The CM-730 board, featuring the ARM CortexM3 processor, acts as the communication relay between the Dynamixel motors and the fitpc. The microcontroller also provides sensor acquisition and processing. A 6 degree of freedom Inertial Measurement Unit (IMU) will aid in correcting gait cycles in the face of perturbations. A block diagram that outlines the computing relationship is shown in Fig. 2. Fig. 2. Electronics architecture

5 Software The software architecture for the robots is shown in Fig. 3. This architecture is novel in that it uses Lua as a common development platform. Since many of the students do not have strong programming backgrounds, this development platform allows them to participate more fully on the team. Low-level interfaces to the hardware level are implemented as C routines callable from Lua. These routines provide access to the camera and other sensors such as joint encoders and the IMU, and allow the higher-level routines to modify joint angles and stiffnesses. Fig. 3. Software architecture The Lua routines consist of a variety of modules, layered hierarchically: Sensor Module that is responsible for reading joint encoders, IMU, foot sensors, battery status, and button presses on the robot. Camera Interface to the video camera system, including setting parameters, switching cameras, and reading the raw YUYV images. Effector Module to set and vary motor joints and parameters, as well as body and face LED s. Vision Uses acquired camera images to deduce presence and relative location of the ball, goals, lines, and other robots. World Models world state of the robot, including pose and altered ball location. Game StateMch Game state machine to respond to Robocup game

controller and referee button pushes. Head StateMch Head state machine to implement ball tracking, searching, and lookaround behaviors. Body StateMch Body state machine to switch between chasing, ball approach, dribbling, and kicking behaviors. Keyframe motion generator used for scripted motions such as getup and kick motions. Walk Omnidirectional locomotion module. In order to simplify development, all interprocess communications are performed by passing data structures between the various modules, as well as between robots. [11] 6 Vision Fig. 4. Visualization of the color segmentation In each new setting, we may encounter different field conditions such as a change in lighting or the actual color hue of the field objects. In order to account for this, we log a series of images that are then used to train a lookup table. A GUI tool enables us to define the YCbCr values that correspond to green, yellow, white, etc. Once these specific values are selected and defined, the distribution of the points in the color space are spread out and generalized to account for a greater variation. This is done with a Gaussian mixture model that analyzes the probability density function of each of the previously defined pixel values. The boundaries of the color classes are then expanded according to Bayes Theorem. We can then process the individual pixels of the new images by matching their YCbCr values to the broadened definition of the values in the lookup table.

After the image is segmented into its corresponding color classes using the lookup table, the segmentation is bitwise OR-ed in 4x4 blocks. The initial object hypotheses for the ball and goal posts are found by finding connected components in the smaller, bit OR-ed, image, and then using the original image we calculated the statistics of each region. Processing the bit OR-ed image first allowed us to greatly speed up the computation of the system. The bit OR-ed image also produced the set of points that are used in our line detection algorithm. We then check the segmented components for certain attributes like size, shape, and position in order to classify objects, such as the ball and the goal posts. We also compute statistics for the position of detected objects in the world coordinate system using the inverse kinematics of the robot, the centroid, and the bounding box to further filter the object hypotheses. Using these we are able to track the ball and identify the existence and size of goal posts and consequently localize our position on the field. [11] 7 Prior Performance in RoboCup Team DARwIn has been competing since RoboCup 2007 led by Virginia Tech, and in recent years has partnered with the University of Pennsylvania. RoboCup 2010 saw the team play a respectable fourth place, and drove the team to improve its hardware and software. With the introduction of the DARwIn-OP system, Team DARwIn was able to move ahead and enthusiastically took the first place trophy in the kid-size league in Istanbul. With this momentum, we again captured the title in 2012 in Mexico City. This year, we are looking to make our software more portable across the Nao, XOS, and OP teams that we manage. We use our own software, and provide its base at https://github.com/upenn-robocup/upennalizers for the research community, which teams are already using and e-mailing us about setup questions. 8 Conclusion Building on previous research and RoboCup experience, we hope to push DARwIn-OP its highest potential. This year, we are trying to exploit our 3- platform team of the Teen size, SPL, and kid size robots to maintain our primary purpose as a research endeavor for studying robot humanoid motions; and to share this work with the broader community.

References 1. D. Hong. Biologically inspired locomotion strategies: Novel ground mobile robots at romela. 3rd International Conference on Ubiquitous Robots and Ambient Intelligence, Seoul, S. Korea, October 2006. 2. K. Muecke and D. Hong. Darwin s evolution: Development of a humanoid robot. Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, pages 2574 2575, October-November 2007. 3. K. Hirai. The development of honda humanoid robot. IEEE Int. Conf. on Robotics and Automation, pages 1321 1326, 1998. 4. T. Ishida. A small biped entertainment robot sdr-4x ii. Proc. IEEE Symposium on Comp. Intelligence in Robotics and Automation, pages 1046 1051, July 2003. 5. J. Kim. On the Stable Dynamic Walking of Biped Humanoid Robots. PhD Thesis, Korea Advanced Institute of Science and Technology, Daejeon, South Korea, 2006. 6. et al S. H. Collins. A three-dimensional passive-dynamic walking robot with two legs and knees. Int. Journal of Robotics Research, 20(2):607 615, 2001. 7. T. McGeer. Passive dynamic walking. Int. Journal of Robotics Research, 9(2):62 82, April 1990. 8. M. Vukobratovic. Zero-moment point thirty five years of its life. Int. Journal of Humanoid Robotics, 1(1), 2004. 9. Q. Huang, K. Yokoi, and et al S. Kajita. Planning walking patterns for a biped robot. IEEE Trans. on Robotics and Automation, 17(3):280 289, June 2001. 10. K. Kim, Y. Son, and P. Kim. Construction of small humanoids with a new joint actuator module. Proc. 2004 IEEE Int. Conf. on Robotics and Automation, pages 4510 4514, April 2004. 11. J.Brindza, A.Lee, A.Majumdar, B.Scharfman, A.Schneider, R.Shor, and D.Lee. Upennalizers robocup standard platform league team report 2009. Technical report, University of Pennsylvania, 2009.