BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE

Similar documents
BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE

A SEMINAR REPORT ON BRAIN CONTROLLED CAR USING ARTIFICIAL INTELLIGENCE

BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE

Motor Imagery based Brain Computer Interface (BCI) using Artificial Neural Network Classifiers

Non-Invasive Brain-Actuated Control of a Mobile Robot

1. INTRODUCTION: 2. EOG: system, handicapped people, wheelchair.

BRAIN COMPUTER INTERFACE (BCI) RESEARCH CENTER AT SRM UNIVERSITY

Off-line EEG analysis of BCI experiments with MATLAB V1.07a. Copyright g.tec medical engineering GmbH

Presented by: V.Lakshana Regd. No.: Information Technology CET, Bhubaneswar

from signals to sources asa-lab turnkey solution for ERP research

I. INTRODUCTION MAIN BLOCKS OF ROBOT

Linear Motion Servo Plants: IP01 or IP02. Linear Experiment #0: Integration with WinCon. IP01 and IP02. Student Handout

The Application of Human-Computer Interaction Idea in Computer Aided Industrial Design

MAGNATEST ECM * Low cost eddy current module for non destructive testing using the magnetoinductive

Team Autono-Mo. Jacobia. Department of Computer Science and Engineering The University of Texas at Arlington

MAGNATEST ECM * * Low-cost eddy-current module for non-destructive testing using the magneto-inductive method

ANIMA: Non-conventional Brain-Computer Interfaces in Robot Control through Electroencephalography and Electrooculography, ARP Module

The Study of Methodologies for Identifying the Drowsiness in Smart Traffic System: A Survey Mariya 1 Mrs. Sumana K R 2

Impact of an Energy Normalization Transform on the Performance of the LF-ASD Brain Computer Interface

MASTER SHIFU. STUDENT NAME: Vikramadityan. M ROBOT NAME: Master Shifu COURSE NAME: Intelligent Machine Design Lab

Thanks to Autocheck function, it is possible to perform a complete check-up of the robot thanks to a stepby-step

Human Authentication from Brain EEG Signals using Machine Learning

Human-to-Human Interface

RED TACTON ABSTRACT:

Audio System. Low Line Audio Component Location. Head Unit Audio Control Diagram

Dipartimento di Elettronica Informazione e Bioingegneria Robotics

Classifying the Brain's Motor Activity via Deep Learning

PAP-240 Three Axis Antenna Pedestal and feed drive

Mechatronics Educational Robots Robko PHOENIX

Modelling and Simulation of Tactile Sensing System of Fingers for Intelligent Robotic Manipulation Control

Robot Task-Level Programming Language and Simulation

The Design of Intelligent Wheelchair Based on MSP430

RECOMMENDATION ITU-R BS

Non Invasive Brain Computer Interface for Movement Control

Aztec Micro-grid Power System

Relationship to theory: This activity involves the motion of bodies under constant velocity.

The Perception. Is Reality. Test Bench

GPS Tracking System Using Car Charger

Vision Ques t. Vision Quest. Use the Vision Sensor to drive your robot in Vision Quest!

AutoBench 1.1. software benchmark data book.

RED TACTON.

Development of 24 GHz-band High Resolution Multi-Mode Radar

Emotiv EPOC 3D Brain Activity Map Premium Version User Manual V1.0

Android Phone Based Assistant System for Handicapped/Disabled/Aged People

Pre-Day Questionnaire

Distance Measurement of an Object by using Ultrasonic Sensors with Arduino and GSM Module

A Brain-Computer Interface Based on Steady State Visual Evoked Potentials for Controlling a Robot

Test Booklet. Subject: LA, Grade: 04 LEAP Grade 4 Language Arts Student name:

Azaad Kumar Bahadur 1, Nishant Tripathi 2

Chapter 1. Robots and Programs

Programming PIC Microchips

Smart Phone Accelerometer Sensor Based Wireless Robot for Physically Disabled People

ELECTRICAL ENGINEERING TECHNOLOGY PROGRAM EET 433 CONTROL SYSTEMS ANALYSIS AND DESIGN LABORATORY EXPERIENCES

Fabrication of the kinect remote-controlled cars and planning of the motion interaction courses

INTELLIGENT GUIDANCE IN A VIRTUAL UNIVERSITY

AN AUTONOMOUS SIMULATION BASED SYSTEM FOR ROBOTIC SERVICES IN PARTIALLY KNOWN ENVIRONMENTS

Kongsberg Seatex AS Pirsenteret N-7462 Trondheim Norway POSITION 303 VELOCITY 900 HEADING 910 ATTITUDE 413 HEAVE 888

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment

BRAINWAVE CONTROLLED WHEEL CHAIR USING EYE BLINKS

SIMULATION MODELING WITH ARTIFICIAL REALITY TECHNOLOGY (SMART): AN INTEGRATION OF VIRTUAL REALITY AND SIMULATION MODELING

MEM380 Applied Autonomous Robots I Winter Feedback Control USARSim

M.Sinduja,S.Ranjitha. Department of Electrical & Electronics Engineering, Bharathiyar Institute of Engineering For Women, Deviyakurichi.

An Autonomous Self- Propelled Robot Designed for Obstacle Avoidance and Fire Fighting

Intelligent Bus Tracking and Implementation in FPGA

Development of Gaze Detection Technology toward Driver's State Estimation

The Man-Machine-Man(M 3 ) Interfacing With the Blue Brain Technology

OPERATOR SERIES BLACK TIMELESS DESIGN AND SMART TECHNOLOGY

Session 11 Introduction to Robotics and Programming mbot. >_ {Code4Loop}; Roochir Purani

Evaluation of Connected Vehicle Technology for Concept Proposal Using V2X Testbed

[Kumar, 5(12): December2018] ISSN DOI /zenodo Impact Factor

Hands-free Operation of a Small Mobile Robot*

Mobile Robots (Wheeled) (Take class notes)

Naturalistic Flying Study as a Method of Collecting Pilot Communication Behavior Data

BCI-based Electric Cars Controlling System

acknowledgments...xv introduction...xvii 1 LEGO MINDSTORMS NXT 2.0: people, pieces, and potential getting started with the NXT 2.0 set...

Built-in soft-start feature. Up-Slope and Down-Slope. Power-Up safe start feature. Motor will only start if pulse of 1.5ms is detected.

Teleoperation and System Health Monitoring Mo-Yuen Chow, Ph.D.

Robotics And Remotely Operated Vehicles. P. A. Kulkarni S. G. Karad

Lab 1: Testing and Measurement on the r-one

Design of intelligent vehicle control system based on machine visual

I.1 Smart Machines. Unit Overview:

Laboratory of Advanced Simulations

DESIGN AND DEVELOPMENT OF LIBRARY ASSISTANT ROBOT

COMAND Operator, s Manual

What Is Bluetooth? How Does It Differ from a Wired Connection?

Metrics for Assistive Robotics Brain-Computer Interface Evaluation

Embedded Systems & Robotics (Winter Training Program) 6 Weeks/45 Days

An Introduction to Programming using the NXT Robot:

A Robust Neural Robot Navigation Using a Combination of Deliberative and Reactive Control Architectures

ARTIFICIAL INTELLIGENCE - ROBOTICS

The vehicle needs to receive data from at least four different satellites to give a three dimensional fix on its current position.

145M Final Exam Solutions page 1 May 11, 2010 S. Derenzo R/2. Vref. Address encoder logic. Exclusive OR. Digital output (8 bits) V 1 2 R/2

Information Quality in Critical Infrastructures. Andrea Bondavalli.

Using Simple Force Feedback Mechanisms as Haptic Visualization Tools.

BRAINWAVE RECOGNITION

DAI. Connecting Analog and Frequency Fuel Level Sensors

ME7752: Mechanics and Control of Robots Lecture 1

SCOE SIMULATION. Pascal CONRATH (1), Christian ABEL (1)

5 Lab 5: Position Control Systems - Week 2

E61, E63, E64, E70, E87, E90, E91, E92, E93, R56 BMW AG - TIS

Transcription:

BRAIN CONTROLLED CAR FOR DISABLED USING ARTIFICIAL INTELLIGENCE Presented by V.DIVYA SRI M.V.LAKSHMI III CSE III CSE EMAIL: vds555@gmail.com EMAIL: morampudi.lakshmi@gmail.com Phone No. 9949422146 Of SHRI VISHNU ENGINEERING COLLEGE FOR WOMEN Bhimavaram, W.G.Dist, A.P.

ABSTRACT This paper considers the development of a brain driven car, which would be of great help to the physically disabled people. Since these cars will rely only on what the individual is thinking they will hence not require any physical movement on the part of the individual. The car integrates signals from a variety of sensors like video, weather monitor, anti-collision etc. it also has an automatic navigation system in case of emergency. The car works on the asynchronous mechanism of artificial intelligence. It s a great advance of technology which will make the disabled, abled. INTRODUCTION The video and thermo gram analyzer continuously monitor activities outside the car. soon as the driver is seated the EEG (electroencephalogram) helmet, attached to the top of the seat, is lowered and suitably placed on the driver s head. A wide screen of the computer is placed at an angle aesthetically suitable to the driver. Each program can be controlled either directly by a mouse or by a shortcut. For starting the car, the start button is clicked. Accordingly the computer switches ON the circuit from the battery to the A.C.Series Induction motors. BIOCONTROL SYSTEM The biocontrol system integrates signals from various other systems and compares them with originals in the database. It comprises of the following systems: Brain-computer interface Automatic security system Automatic navigation system Once the driver (disabled) nears the car. The security system of the car is activated. Images as well as thermo graphic results of the driver are previously fed into the database of the computer. If the video images match with the database entries then the security system advances to the next stage. Here the thermo graphic image verification is done with the database. Once the driver passes this stage the door slides to the sides and a ramp is lowered from its floor. The ramp has flip actuators in its lower end. Once the driver enters the ramp, the flip actuates the ramp to be lifted horizontally. Then robotic arms assist the driver to his seat. As Now let us discuss each system in detail. BRAIN COMPUTER INTERFACE Brain-computer interfaces will increase acceptance by offering customized, intelligent help and training, especially for the non-expert user. Development of such a flexible interface paradigm raises several challenges in the areas of machine perception and automatic explanation. The teams doing research in this field have developed a

single-position, brain-controlled switch that responds to specific patterns detected in spatiotemporal electroencephalograms (EEG) measured from the human scalp. We refer to this initial design as the Low- Frequency connection. Data can be fully exported in raw data, FFT & average formats. Ultra low noise balanced DC coupling amplifier. Asynchronous Switch Design (LF-ASD) (Fig.1). Max input 100microV p-p, minimum digital resolution is 100 microv p-p / 256 = 0.390625 micro V p-p. FFT point can select from 128 (0.9375 Hz), 256 (0.46875 Hz), 512 (0.234375 Hz resolution). Support for additional serial ports via plug-in boar; allows extensive serial input & output control. Fig.1 LF- ASD The EEG is then filtered and run through a fast Fourier transform before being displayed as a three dimensional graphic. The data can then be piped into MIDI compatible music programs. Furthermore, MIDI can be adjusted to control other external processes, such as robotics. The experimental control system is configured for the particular task being used in the evaluation. Real Time Workshop generates all the control programs from Simulink models and C/C+ + using MS Visual C++ 6.0. Analysis of data is mostly done within Mat lab environment. FEATURES OF EEG BAND Remote analysis data can be sent and analyzed in real-time over a network or modem Infinite real-time data acquisition (dependent upon hard drive size). Real-time 3-D & 2-D FFT with peak indicator, Raw Data, and Horizontal Bar displays with Quick Draw mode. Full 24 bit color support; data can be analyzed with any standard or user. Customized color palettes; color cycling available in 8 bit mode with QuickDrawmode. Interactive real-time FFT filtering with Quick Draw mode. Real-time 3-D FFT (left, right, coherence and relative coherence), raw wave, sphere frequency and six brain wave switch in one OpenGL display.

Full Brainwave driven Quick Time Movie, Quick Time MIDI control; user configurable Full Brain wave driven sound control, support for 16 bit sound; user configurable Full image capture and playback control; user configurable. accuracies than able-bodied subjects using real movements. 2. Subjects demonstrated activation accuracies in the range of 70-82% with false activations below 2%. 3. Accuracies using actual finger movements were observed in the range 36-83% 4. The average classification accuracy of imaginary movements was over 99% Fig. 2: EEG Transmission Fig.4 Brain-to- Machine Mechanism Fig. 3 EEG TEST RESULTS COMPARING DRIVER ACCURACY WITH/WITHOUT BCI 1. Able-bodied subjects using imaginary movements could attain equal or better control The principle behind the whole mechanism is that the impulse of the human brain can be tracked and even decoded. The Low-Frequency Asynchronous Switch Design traces the motor neurons in the brain. When the driver attempts for a physical movement, he/she sends an impulse to the motor neuron. These motor neurons carry the signal to the physical components such as hands or legs. Hence we decode the message at the motor neuron to obtain maximum accuracy. By observing the sensory neurons we can monitor the eye movement of the driver.

Fig.5 Eyeball Tracking As the eye moves, the cursor on the screen also moves and is also brightened when the driver concentrates on one particular point in his environment. The sensors, which are placed at the front and rear ends of the car, send a live feedback of the environment to the computer. The steering wheel is turned through a specific angle by electromechanical actuators. The angle of turn is calibrated from the distance moved by the dot on the screen. Fig.7 Sensors and Their Range AUTOMATIC SECURITY SYSTEM The EEG of the driver is monitored continually. When it drops less than 4 Hz then the driver is in an unstable state. A message is given to the driver for confirmation to continue the drive. A confirmed reply activates the program automatic drive. The computer prompts the driver for the destination before the drive. AUTOMATIC NAVIGATION SYSTEM Fig.6 Electromechanical Control Unit As the computer is based on artificial intelligence it automatically monitors every route the car travels and stores it in its map database for future use. The map database is analyzed and the shortest route to the destination is chosen. With traffic monitoring system provided by xm satellite radio the computer drives the car automatically. Video

and anti-collision sensors mainly assist this drive by providing continuous live feed of the environment up to 180 m, which is sufficient for the purpose. Fig.8 EEG Analysis Window abler and the disabled vanishes. Thus the integration of bioelectronics with automotive systems is essential to develop efficient and REFERENCE 1. 'Off-line Classification of EEG from the "New York Brain- Computer Interface (BCI)" Flotzinger, D., Kalcher, J., Wolpaw, J.R., McFarland, J.J., and Pfurtscheller, G., Report #378, IIG-Report Series, IIG: Institutes for Information Processing, Graz University of Technology, Austria 1993. 2. "Man-Machine Communications through Brain-Wave Processing" Keirn, Z.A. and Aunon, J.I., IEEE Engineering in Medicine and Biology Magazine, March 1990. CONCLUSION When the above requirements are satisfied and if this car becomes cost effective then we shall witness a revolutionary change in the society where the demarcation between the futuristic vehicles, which shall be witnessed soon helping the disabled in every manner in the field of transportation. 3. Automotive engineering, SAE, June 2005 4. Automotive mechanics, Crouse, tenth edition, 1993 5. "The brain response interface: communication through visually-induced electrical brain responses" Sutter, E.E., Journal of Microcomputer Applications, 1992, 15: 31-45.