ELECTROOCULOGRAPHIC GUIDANCE OF A WHEELCHAIR USING EYE MOVEMENTS CODIFICATION

Similar documents
1. INTRODUCTION: 2. EOG: system, handicapped people, wheelchair.

Wheelchair Guidance Strategies Using EOG

Application of LonWorks Technology to Low Level Control of an Autonomous Wheelchair.

Human-Machine Interfaces and Sensory Systems for an Autonomous Wheelchair

An EOG based Human Computer Interface System for Online Control. Carlos A. Vinhais, Fábio A. Santos, Joaquim F. Oliveira

Using Electrooculogram and Electromyogram for powered wheelchair

Review on Eye Movement Controlled Wheelchair

Sensors Based Automated Wheelchair

Design of EOG based Human Machine Interface system

Initial Report on Wheelesley: A Robotic Wheelchair System

Eyeball Motion and Joystick Controlled Wheelchair with Obstacle Detection

Retina Based Mouse Control (RBMC)

Proceedings of the 6th WSEAS International Conference on Applied Computer Science, Tenerife, Canary Islands, Spain, December 16-18,

Human Computer Interaction using Eyes (HCIE)

Towards a Google Glass Based Head Control Communication System for People with Disabilities. James Gips, Muhan Zhang, Deirdre Anderson

SMART Wheelchair by using EMG & EOG

A Study on Gaze Estimation System using Cross-Channels Electrooculogram Signals

Team Autono-Mo. Jacobia. Department of Computer Science and Engineering The University of Texas at Arlington

Using Eye Blinking for EOG-Based Robot Control

LAB 5: Mobile robots -- Modeling, control and tracking

Design and development of a Virtual Instrument for Bio-signal Acquisition and Processing using LabVIEW

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Eye-Tracking Methodolgy

Controlling a Robotic Arm by Brainwaves and Eye Movement

Gaze-controlled Driving

DESIGNING AND CONDUCTING USER STUDIES

Physiology Lessons for use with the Biopac Student Lab

Sensing and Processing of EOG Signals to Control Human Machine Interface System

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment

* Intelli Robotic Wheel Chair for Specialty Operations & Physically Challenged

MEM380 Applied Autonomous Robots I Winter Feedback Control USARSim

DC motor control using arduino

International Journal of Computer Sciences and Engineering. Research Paper Volume-5, Issue-12 E-ISSN:

Non Invasive Brain Computer Interface for Movement Control

NCCT IEEE PROJECTS ADVANCED ROBOTICS SOLUTIONS. Latest Projects, in various Domains. Promise for the Best Projects

Physiology Lessons for use with the BIOPAC Student Lab

Development of Gaze Detection Technology toward Driver's State Estimation

ROBCHAIR - A SEMI-AUTONOMOUS WHEELCHAIR FOR DISABLED PEOPLE. G. Pires, U. Nunes, A. T. de Almeida

Bi-modal human machine interface for controlling an intelligent wheelchair

Voice based Control Signal Generation for Intelligent Patient Vehicle

California University of Pennsylvania Department of Applied Engineering & Technology Electrical Engineering Technology

Fingers Bending Motion Controlled Electrical. Wheelchair by Using Flexible Bending Sensors. with Kalman filter Algorithm

Multi-modality EMG and Visual Based Hands-Free Control of an Intelligent Wheelchair

ADVANCES in NATURAL and APPLIED SCIENCES

CENG 5931 HW 5 Mobile Robotics Due March 5. Sensors for Mobile Robots

Development of an Intuitive Interface for PC Mouse Operation Based on Both Arms Gesture

Multi Robot Navigation and Mapping for Combat Environment

ME 461 Laboratory #5 Characterization and Control of PMDC Motors

An Electooculogram Based Virtual Instrumentation System

High-Level Programming for Industrial Robotics: using Gestures, Speech and Force Control

The Design of Intelligent Wheelchair Based on MSP430

Module 1: Introduction to Experimental Techniques Lecture 2: Sources of error. The Lecture Contains: Sources of Error in Measurement

Wheeled Mobile Robot Obstacle Avoidance Using Compass and Ultrasonic

MOUSE CURSOR CONTROL SYTEM BASED ON SSVEP

BRAINWAVE CONTROLLED WHEEL CHAIR USING EYE BLINKS

Direct gaze based environmental controls

I.1 Smart Machines. Unit Overview:

Implementation of Conventional and Neural Controllers Using Position and Velocity Feedback

Mobile Robots (Wheeled) (Take class notes)

BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE

MULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT

Autonomous Wheelchair for Disabled People

An Example Cognitive Architecture: EPIC

Robotic Vehicle Design

IMU Platform for Workshops

ECE 477 Digital Systems Senior Design Project Rev 8/09. Homework 5: Theory of Operation and Hardware Design Narrative

OBSTACLE DETECTION AND COLLISION AVOIDANCE USING ULTRASONIC DISTANCE SENSORS FOR AN AUTONOMOUS QUADROCOPTER

Mixed Reality technology applied research on railway sector

Shape Memory Alloy Actuator Controller Design for Tactile Displays

Programming PIC Microchips

Performance Improvement of Contactless Distance Sensors using Neural Network

ON HEARING YOUR POSITION THROUGH LIGHT FOR MOBILE ROBOT INDOOR NAVIGATION. Anonymous ICME submission

Path Following and Obstacle Avoidance Fuzzy Controller for Mobile Indoor Robots

MACCS ERP Laboratory ERP Training

A SEMINAR REPORT ON BRAIN CONTROLLED CAR USING ARTIFICIAL INTELLIGENCE

An Autonomous Self- Propelled Robot Designed for Obstacle Avoidance and Fire Fighting

Compensation of a position servo

PupilMouse: Cursor Control by Head Rotation Using Pupil Detection Technique

Robotic Vehicle Design

Laboratory Project 1B: Electromyogram Circuit

A Real-World Experiments Setup for Investigations of the Problem of Visual Landmarks Selection for Mobile Robots

Autonomous Obstacle Avoiding and Path Following Rover

P15051: Robotic Eye for Eye Tracker

SELF-BALANCING MOBILE ROBOT TILTER

CSE Thu 10/22. Nadir Weibel

Laboratory 7: CONTROL SYSTEMS FUNDAMENTALS

Design and Experiment of Electrooculogram (EOG) System and Its Application to Control Mobile Robot

Key-Words: - Fuzzy Behaviour Controls, Multiple Target Tracking, Obstacle Avoidance, Ultrasonic Range Finders

INTERNATIONAL JOURNAL OF ELECTRICAL ENGINEERING & TECHNOLOGY (IJEET) TWO WHEELED SELF BALANCING ROBOT FOR AUTONOMOUS NAVIGATION

Distance Measurement of an Object by using Ultrasonic Sensors with Arduino and GSM Module

POSITIONING AN AUTONOMOUS OFF-ROAD VEHICLE BY USING FUSED DGPS AND INERTIAL NAVIGATION. T. Schönberg, M. Ojala, J. Suomela, A. Torpo, A.

BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE

KINECT CONTROLLED HUMANOID AND HELICOPTER

Biosignal filtering and artifact rejection, Part II. Biosignal processing, S Autumn 2017

intelligent wheelchair

Servo control: Ball on beam

A Virtual Environments Editor for Driving Scenes

Progress Report. Mohammadtaghi G. Poshtmashhadi. Supervisor: Professor António M. Pascoal

Current Rebuilding Concept Applied to Boost CCM for PF Correction

Gesture Identification Using Sensors Future of Interaction with Smart Phones Mr. Pratik Parmar 1 1 Department of Computer engineering, CTIDS

Transcription:

ELECTROOCULOGRAPHIC GUIDANCE OF A WHEELCHAIR USING EYE MOVEMENTS CODIFICATION Rafael Barea, Luciano Boquete, Manuel Mazo, Elena López, L.M. Bergasa. Electronics Department. University of Alcala. Campus Universitario s/n. 28871Alcalá de Henares. Madrid. Spain. T: +34 91 885 65 74 Fax: +34 91 885 65 91 e-mail: barea@depeca.alcala.es http://www.depeca.alcala.es/personal/barea Abstract: This paper presents a new method to control and guide mobile robots. In this case, to send different commands we have used electrooculography (EOG) techniques, so that, control is made by means of the ocular position (eye displacement into its orbit). A inverse eye model is developed based in electrooculography and therefore the saccadic eye movements can be detected and know where user is looking. This control technique can be useful in multiple applications, but in this work it is used to guide a autonomous robot (wheelchair) as a system to help to people with severe disabilities. The system consists of a standard electric wheelchair with an on-board computer, sensors and graphical user interface running on a computer. Keywords: Electrooculographic potential (EOG), eye model, control system, handicapped people, wheelchair. 1 INTRODUCTION Assistive robotics can improve the quality of life for disable people. Nowadays, there are many help systems to control and guide autonomous mobile robots. All this systems allow their users to travel more efficiently and with greater ease [SIAMO, 99]. In the last years, the applications for developing help systems to people with several disabilities are increased, and therefore the traditional systems are not valid. In this new systems, we can see: videooculography systems (VOG) or infrared oculography (IROG) based on detect the eye position using a camera [Lahoud&Cleveland]; there are several techniques based in voice recognition for detecting basic commands to control some instruments or robots; the joystick (sometimes tactil screen) is the most popular technique used to control different applications by people with limited upper body mobility but it requires fine control that the person may be have difficulty to accomplish. All this techniques can be applied to different people according to their disability degree, using always the technique or techniques more efficiently for each person. This paper reports work in the development of a robotic wheelchair system based in electrooculography [Gips et al, 96]. Our system allows the users to tell the robot where to move in gross terms and will then carry out that navigational task using common sensical constraints, such as avoiding collision. This wheelchair system is intented to be a general purpose navigational assistant in environments with accesible features such as ramps and doorways of sufficient width to allow a wheelchair to pass. This work is based on previous research in robot path planing and mobile robotics [Barea et al, 99]; however, a robotic wheelchair must interact with its user, making the robotic system semiautonomous rather than completely autonomous. This paper has been divided into the following sections: section 2 describes the electrooculography technique used to register the eye movement and the eye gaze, in section 3 an eye model based on electrooculography is propused. In section 4, the visual control system and some guidance results are commented and section 5 puts forward the main conclusions and lays down the main lines of work to be followed in the future. 2 ELECTROOCULOGRAPHIC POTENTIAL (EOG) A survey of eye movements recording methods can be seen in [Glenstup&Engell, 95] where are described the main advantages and drawbacks of each one. In this work, the goal is to sense the electrooculographic potential (EOG) because it presents a good face access, good accuracy and resolution, great range of eye displacements, works in real time and is cheap. Our discrete electrooculographic control system (DECS) is based in record the polarization potential or corneal-retinal potential (CRP) [Nicoulau et al., 95]. This potential is commonly known as an electrooculogram. The

EOG ranges from.5 to 3.5 mv in humans and is linearly proportional to eye displacement. The human eye is an electrical dipole with a negative pole at the fundus and a positive pole at the cornea. This system may be used for increasing communication and/or control. The analog signal form the oculographic measurements has been turned into signal suitable for control purposes. The derivation of the EOG is achieved placing two electrodes on the outerside of the eyes to detect horizontal movement and another pair above and below the eye to detect vertical movement. A reference electrode is placed on the forehead. Figure 1 shows the electrode placement. Figure 1. Electrodes placement. The EOG signal changes approximately 2 microvolts for each degree of eye movement. In our system, the signal are sampled 1 times per second. The record of EOG signal have several problems [Jacob, 96]. Firstly, this signal seldom is deterministic, even for same person in different experiments.the EOG signal is a result of a number of factors, including eyeball rotation and movement, eyelid movement, different sources of artifact such as EEG, electrodes placement, head movements, influence of the luminance, etc. For this reasons, it is neccesary to eliminate the shifting resting potential (mean value) because this value changes. To avoid this problem is necesary an ac diferential amplifier where a high pass filter with cutoff at.5 Hz and relatively long time constant is used. The amplifier used have programable gain ranging from 5,1,2 and 5. 3 EYE MODEL BASED IN EOG (BIDIM-EOG) Our aim is to design a system capable of obtaining the gaze direction detecting the eye movements. For this, a model of the ocular motor system based on electrooculography is proposed (figure 2) (Bidimensional dipolar model EOG, BiDiM-EOG). This model allows us to separe saccadic and smooth eye movements and calculate the eye position into its orbit with good accuracy (less than 2º). The filter eliminates the effects due to other biopotentials, just as the blinks over to the EOG signal. The security block detects when the eyes are closed and in this case, the ouput is disabled. After that, the EOG signal is clasified into saccadic or smooth eye movements by means of two detectors. If a saccadic movement is detected, a position control is used, whereas if a smooth movement is detected, a speed control is used to calculate the eye position. The final position (angle) is calculated as the sum of the saccadic and smooth movements. Besides, the model has to adapt itself to the possible variations of adquisition conditions (electrodes placement, electrode-skin contact, etc). To do this, the model parameters are adjusted in accordance with the angle detected. A person, in a voluntary way, only can made saccadic movements unless he tries to follow an object in movement. Therefore, to control some interface it is convenient to focus the study in the detection of saccadic movements (rapid movements). This process can be done processing the derivate of the EOG signal. To avoid problems with the variability of the signal (the isoelectric line varies with time, even though the user keeps the gaze at the same position), a high pass filter with a very small cutoff frecuency (.5 Hz) is used. The process followed can be observed in figure 3 where the results of a process in which the user made a secuence of saccadic movements of ±1º..±4º in horizontal derivation are shown. It is possible to see that the derivate of the electrooculographic signal allows us to determinate when a sudden movement is made in the eye gaze. This variation can be easily translated to angles (figure 3.d). In the following sections, the results reached in the Electronics Department of University of Alcala at the moment are shown. It is possible to use this technique to help disabled people, because we have got an accuracy less than ±2º. Although in this paper we are going to comment the results obtained in the guidance of a wheelchair (help to the mobility), other applications have been developed to increase the facilities in people comunication (help to the comunication) [Barea et al, a].

EOG FILTER SACCADIC MOVEMENTS DETECTOR SMOOTH MOVEMENTS DETECTOR USER SAFETY FEED-BACK PARAMETERS ADJUSTMENT POSITION CONTROL SPEED CONTROL 2S 2 TOTAL OUTPUT CONTROL OUPUT 2 L Figure 5 shows a diagram of the control system. The EOG signal is recorded using Ag-AgCl electrodes and this data, by means of an adquisition system are sent to a PC, in which they are processed to calculate the eye gaze direction. Then, in accordance with the guidance control strategy, the control commands of the wheelchair are sent. The command sent to the wheelchair are the separate linear speed for each wheel. It is possible to see that exists a visual feedback in the system by means of a tactile screen that the user has in front of him. Figure 6 shows the user interface where the commands that the user can generate are: Forward, Backwards, Left, Right and Stop. Figure 2. Bidemensional bipolar model (BiDiM- EOG). a) Eog (horizontal) 6 5 4 1 2 3 4 b) Derivate 1-1 1 2 3 4 1 c)maximum derivate -1 1 2 3 4 5 d) Angle -5 1 2 3 4 Time (s) Figure 4. Wheelchair. EOG Visual feedback Wheelchair Eye Model Figure 3. Process results to detect the eye gaze angle. 4 GUIDANCE OF A WHEELCHAIR USING COMMANDS GENERATED BY EYE MOVEMENTS DETECTED USING ELECTROOCULOGRAPHY. The aim of this control system is to guide an autonomous mobile robot using the positioning of the eye into its orbit by means of EOG signal. In this case, the autonomous vehicle is a wheelchair for disable people. Figure 4 shows the wheelchair used. Eye position Command generator Figure 5. Guidance system. [Command] V cmd Ω cmd

direction of the eye gaze. In this paper, we are going to study the continuous control technique. Figure 6. User interface To control the robot movements there are multiple options: direct access guidance, semiautomatic and automatic sweep (scan) or interpretation of different commands generated by means of eye movements. In former works, we studied the direct access guidance [Barea et al, 99] and automatic and semiautomatic scan [Barea et al, b]. In direct access guidance, the user can see the different guidance commands in a screen (laptop) and select them directly. In this way, when the user looks at somewhere, the cursor is positioned where he is looking, then, the users can select the action to control the wheelchair movements. The actions are validated by time, this is, when a command is selected, it is necessary to stay looking at it for a period of time to validate the action. In scan guidance, it is necessary to do an eye movement (a tick ) to select among the different commands presented in the screen. The actions are validated by time, this is, when a command is selected, if other tick is not generated during a time interval, the command is validated and the guidance action is executed. For this reason, we are going to focus our work in the interpretation of different ocular commands because it allows us to generate simple code for controlling the wheelchair. Figure 7 shows the user interface of the wheelchair. This control attempts to emulate the intuitive control that a non-handicapped person makes when he drives a mobile. This system controls the linear speed as the car accelerator and the angular speed as the steering wheel of a car. For this, we have implemented the following movement commands: UP: Linear speed increase (V++). DOWN: Linear speed decrease (V ). RIGHT: Angular speed increase (W++). LEFT: Angular speed decrease (W--). Commands generation is coded using a state machine that establishes the state (command) where the system is working (figure 8). OFF Off ó Sec On Sec STOP Rep Mov_UP MOV S V Figure 8. State machine UP V++ Mov_DO Mov_LE LEFT S-- Mov_RI DOWN V-- RIGHT S++ Now, some examples of guidance are shown. Figure 9 shows the trajectory obtained using total ocular control. Figure 1 shows EOG horizontal and vertical derivations and figure 11 and 12 show the generation of commands in function of EOG signal detected. Figure 7. User-wheelchair interface. The guidance based on ocular commands has different options, such as continuous guidance, onoff activation commands and tracking of the Figure 9. Trajectory generated.

4 2-2 EOG horizontal Î -4 1 2 3 4 5 6 7 4 2 EOG vertical Ï Ö -2 1 2 3 4 5 6 7 Time (1-1 s) Figure 1. EOG signal Ð Linear speed 1 5 1 2 3 4 5 6 7 Angular speed.1 -.1 1 2 3 4 5 6 7 Right speed 2 1 1 2 3 4 5 6 7 Left speed 2 1 1 2 3 4 5 6 7 Time (1-1 s) Õ Ñ Ò Î Ï Ð Ô Ó Figure 11. Guidance speed Ñ Ò Ó Figure 12. Visual Interface Ô Õ Ö Another example of guidance is shown in figure 13 where the trajectory followed by the wheelchair and different images of the guidance are shown. Figure 13. Trajectory and example of guidance

Nowadays, we are not trying this control with persons with disabilities but we consider that it is not difficult to learn the control commands. Learning to use this system must be done in an acquired skill. Some studies have shown that disable persons usually requires about 15 minutes to learn to use this kind of systems [Gips et al., 96]. Besides, we need several alarm and stop commands for dangerous situations. This codes can be generated by means of the blink and alpha waves in EEG to detect when the eyelids are closed. On the other hand, the robotic wheelchair system must be able to navigate indoor and outdoor environments and should switch automatically between navigations modes for these environment. Therefore, all this system can be apply different navigations modes in function of their disability degree, using always the techniques more efficiently for each people. It is necessary to use different support system to avoid collisions and the robotic system can switch automatically for controlling the system in an autonomous form. For example, if the user lost the control and the system is unstable, the wheelchair should switch and obtain the control system. This work is included in a general purpose navigational assistant in environments with accesible features to allow a wheelchair to pass. This project is known as SIAMO project [SIAMO, 99]. A complete sensory system has been designed made up of ultrasonic, infrared sensors and cameras in order to allow the detection of obstacles, dangerous situations and generated a map of the environment. Then, the control and navigation module has to guarantee a comfortable path tracking and can switch for controlling automatically the wheelchair and made the trajectory o command defined by the user. 5 Conclusions This research project is aimed towards developed a usable, low-cost assistive robotic wheelchair system for disabled people. In this work, we present a system that can be used as a means of control allowing the handicapped, especially those with only eye-motor coordination, to live more independent lives. Eye movements require minimum effort and allow direct selection techniques, and this increase the response time and the rate of information flow. Some of the previous wheelchair robotics research are restricted a particular location and in many areas of robotics, environmental assumptions can be made that simplify the navigation problem. However, a person using a wheelchair and EOG technique should not be limited by the device intended to assist them if the environment have accessible features. The total command control based on EOG permits to wheelchair users to guide it with a enough degree of comfortability. 6 Acknowledgments The autors would like to express their gratitude to the Comision Interministerial de Ciencia y Tecnología (CICYT) for their support through the project TER96-1957-C3-1 and the University of Alcalá for their support through the project Navegación inercial de robots móviles. 7 References [Glenstrup&Engell, 95] A. J. Glenstrup and T. Engell, Eye Controlled Media: Present and Future State. PhD, DIKU (Institute of Computer Science) University of Copenhagen, Denmark, 1995. [Nicolau et al., 95] M.C. Nicolau, J. Burcet, R.V. Rial. "Manual de técnicas de Electrofisiología clínica". University of Islas Baleares. [Jacob, 96] Robert J.K. Jacob. "Eye Movement- Based Human-Computer Interaction Techniques: Toward Non-Command Interfaces". Human- Computer Interaction Lab. Naval Research Laboratory. Washington, D.C. [Gips et al., 96] James Gips, Philip DiMattia, Francis X. "EagleEyes Project". Curran and Peter Olivieri. Computer Science Department, Boston College. Mass. USA. [SIAMO, 99] SIAMO Project (CICYT). Electronics Department. University of Alcala. Madrid. Spain. [Lahoud&Cleveland] Joseph A. Lahoud and Dixon Cleveland. "The Eyegaze Eyetracking System". LC Technologies, Inc. 4 th Anual IEEE Dual-Use Technologies and Aplications Conference. [Barea et al, 99] R. Barea, L. Boquete, M. Mazo, E. López. "Guidance of a wheelchair using electrooculography". Proceeding of the 3rd IMACS International Multiconference on Circuits, Systems, Communications and Computers (CSCC'99). July 1999. [Barea et al, a] R. Barea, L. Boquete, M. Mazo, E. López, L.M. Bergasa. "Help to the mobility and the communication by means of ocular movements". Proceeding of TELEC 2. Santiago de cuba. Cuba. July 2. [Barea et al, b] R. Barea, L. Boquete, M. Mazo, E. López, "EOG technique to guide a wheelchair. Procceding of 16th IMACS World Congress. Laussana, Switzerland. August 2.