International Journal of Research in Computer and Communication Technology, Vol 2, Issue 12, December- 2013
|
|
- Allan Palmer
- 5 years ago
- Views:
Transcription
1 Design Of Virtual Sense Technology For System Interface Mr. Chetan Dhule, Prof.T.H.Nagrare Computer Science & Engineering Department, G.H Raisoni College Of Engineering. ABSTRACT A gesture-based human computer interaction can make computers and devices easier to use, such as by allowing people to control the application on windows by moving their hands through the air. Existing solutions have relied on gesture recognition algorithms they needs exotic hardware, often involving elaborate setups limited to the research lab. Gesture recognition algorithms used so far are not practical or responsive enough for real-world use, partially due to the inadequate data on which the image processing is applied. As existing methods are based on gesture recognition algorithms.it needs ANN training which makes whole process slow and reduce accuracy. Method we proposed is based on real time controlling the motion of mouse in windows according to the motion of hand and fingers by calculating the change in pixels values of RBG colors from a video, without using any ANN training to get exact sequence of motion of hands and fingers. Keywords computer vision, gesture recognition, speech computer human interaction 1. INTRODUCTION Existing solutions have relied on gesture recognition algorithms they needs exotic hardware, often involving elaborate setups limited to the research lab. Existing Gesture recognition algorithms used so far are not practical or responsive enough for real-world use, partially due to the inadequate data on which the image processing is applied. As existing methods are based on gesture recognition algorithms. It needs ANN training which makes whole process slow and reduce accuracy. The main objective of Method we proposed is based on real time controlling the motion of mouse in windows according to the motion of hand and fingers by Id :- chetandhule123@gmail.com calculating the change in pixels values of RBG colors from a video, without using any ANN training to get exact sequence of motion of hands and fingers. 2. PROBLEM DEFINATION Unfortunately, most existing solutions suffer from several shortcomings. Some of the hardware that has been used for processing gestures has required users to wear obtrusive sensors and stand near multiple carefully calibrated cameras. Most cameras used so far rely on color data and are therefore sensitive to environmental factors such as dynamic backgrounds and lighting conditions. The algorithms used to determine gestures from the data returned by the hardware have been unreliable when tested on a wide variety of users, and gestures have generally been limited to basic hand-tracking. Existing solutions have relied on gesture recognition algorithms.since the time needed for the computer to recognize a gesture is usually longer than the time needed to display its result, there is always a lag affecting the practical application of such interfaces. Finally, there have not been any collaborative workspaces or environments that allow users to freely use gestures for completing tasks such as controlling motion and events of mouse. 3. OBJECTIVES Existing solutions have relied on gesture recognition algorithms they needs exotic hardware, often involving elaborate setups limited to the research lab. Existing Gesture recognition algorithms used so far are not practical or responsive enough for real-world use, partially due to the inadequate data on which the image processing is applied. As existing methods are based on gesture recognition algorithms. It needs ANN training which makes whole process slow and reduce accuracy. The main objective of Method we proposed is based on real time Page 1454
2 controlling the motion of mouse in windows according to the motion of hand and fingers by calculating the change in pixels values of RBG colors from a video, without using any ANN training to get exact sequence of motion of hands and fingers. 4. LITERATURE REVIEW The processing of hand gestures has been explored extensively in existing literature. Some of the earlier work by Freeman and Weissman [1] used a video camera and computer vision template matching algorithms to detect a user's hand from across a room and allow the user to control a television set. A user could show an open hand and an on-screen hand icon would appear that could be used to adjust various graphical controls, such as a volume slider. The slider was activated when the user would cover the control for a fixed amount of time. The authors discovered that users enjoyed this alternative to the physical remote control and that the feedback of the on-screen hand was effective in assisting the user. However, users found it tiring to hold their hand up for long amounts of time to activate the different controls. This user fatigue common of gesturebased interfaces has been called gorilla arm. Other approaches have relied on using multiple cameras to produce a 3D image which can be used to detect and track hand motion [2][4]. These systems required an elaborate installation process which had to be completed carefully as calibration parameters such as the distance between the cameras was important in the triangulation algorithms used. These algorithms were also computationally expensive since a large amount of video data needed to be processed in realtime,and stereo-matching typically fails on scenes with little or no texture. Ultimately, such systems would not be useable outside of their special lab environments. In [3], Mistry presented the Sixth Sense wearable gestural interface, which used a camera and projector worn on the user's chest to allow the user to zoom in on projected maps(among other activities) by the use of two-handed gestures. In order for the camera to detect the user's hand, the user had to wear brightly-colored markers on their index fingers and thumbs. The regular webcam worn by the user would also be sensitive to environmental conditions such as bright sunlight or darkness, which would make distinguishing the colored markers much more difficult, if not impossible. Wilson and Oliver [5] aimed to create a Minority Report-like environment that they called G Windows. The user was able to move an onscreen cursor of a Microsoft Windows desktop by pointing with their hand and using voice commands to trigger actions like "close" and "scroll" to affect the underlying application windows. They concluded that users preferred interacting with hand gestures over voice commands and that desktop workspaces designed for gesture interactions were worth pursuing further. When considering collaborative online workspaces, several commercial and academic web-based collaboration solutions have existed for some time. However, interaction with other users in these environments is usually limited to basic sharing of media files, rather than allowing for full real-time collaboration of entire web-based applications and their data between users on distinctly deployed domains, as this paper proposes. Cristian Gadea, Bogdan Ionescu [6] aimed to create Finger-Based Gesture Control of a Collaborative Online Workspace, but system needs continuous internet connectivity, but this is not possible always in India. It needs an online workspace called as UC- IC, the application is within web browser to determine latest hand gesture, but it is not possible always to provide all time high speed connectivity everywhere and every time. Beside this it needs the training to recognize gesture, it slows down the system. In [7,8,9] methods are based on gesture recognition algorithms. It needs ANN training which makes whole process slow and reduce accuracy. Because each time if we are trying to recognize the guestre so the ANN training will be needed, and much of time will be needed. So system will not work or can t match its output speed with exact motion of mouse pointer. 5. SYSTEM ARCHITECTURE In this system we have used different preprocessing techniques, feature extraction a tool for recognizing the pixel based values or coordinates of RBG color by tracking the change in pixel position of different color stickers attached at fingers of user in real time. So accordingly the new updated values will be sent to PC to track motion of mouse. Page 1455
3 In this phase we will get pixel sequence from image without using any ANN training to get exact sequence of motion of hands and fingers. Video Data Capturing acquisition Color Detection : In this phase we will extract color positions of RGB color from pixel sequence to detect the motion of hand and fingures by calculating change in pixel values of RBG colors. Image Processing Pixel Extraction Color Detection Controlling Position of mouse Pointer: Send signals to system to control mouse pointer motion and mouse events. It will give an appropriate command to PC to display the motion of mouse pointer according to motion of users fingers or hand. 6. METHODOLOGY i.hand Position tracking and mouse control Controlling Position of mouse Pointer Fig. 1: Block diagram of the different phases of the system.. Video Capturing: Here continuous video will be given as an input toby our system to the laptop. Image Processing: Image segmentation is done under two phases: 1. Skin Detection Model: To detect hand and fingersfron image. Getting user input virtually is the main aim for this module where user will move his finger in front of camera capture area. This motion will capture and detected by the camera and processed by the system frame by frame. After processing system will try to get the finger co-ordinates and once coordinates get calculated it will operate the cursor position. ii. Laser Pointer Detection 2. Approximate Median model : For subtraction of background.it has been observed that by the use of both methods for segmentation was obtained much better for further process. Pixel Extraction: Page 1456
4 vii. Virtual Sense for file handling. This system will make use of the virtual sense technology in order to copy a file from one system into another within a local area network (LAN)/wify. The user will make an action of picking upon the file that needs to be copied and then move it to the system where the file would be copied and then release it over that system. 7.RESULTS AND DISCUSSION iii. Hand Gesture Based Auto Image Grabbing: (virtual Zoom in/out) The software has provision to control all clicking events of mouse by using a color marker.. After several experiments, it was observed that use of red color marker are more effective in comparison with when other color markers are used. iv. Camera Processing and image capturing: Fig2.Graphical user Interface of application. v. Object based bricks game. vi. Virtual playing of drum by holding drum sticks in hand. Page 1457
5 Fig3:Start camera This application can be very useful for people who want to control computer without actually touching to system or by using wireless mouse which needs always a platform to operate. As a part of future scope the application can be improved to work with mobile phone and play stations. Other mode of human computer interaction like voice recognition, facial expression, eye gaze, etc. can also be combined to make the system more robust and flexible. Acknowledgment We thank the subjects participating in our experiments. 7. REFERENCES Fig: Set the marker color. [1] W. T. Freeman and C. D. Weissman, "Television Control by HandGestures", in Proc. of Int. Workshop on Automatic Face and GestureRecognition. IEEE Computer Society, 1995, pp [2] Z. Jun, Z. Fangwen, W. Jiaqi, Y. Zhengpeng, and C. Jinbo, "3D HandGesture Analysis Based on Multi-Criterion in Multi-Camera System",in ICAL2008: IEEE Int. Conf. on Automation and Logistics. IEEEComputer Society, September 2008, pp Fig:Control motion and clicking events of mouse with the color marker set earlier 8. CONCLUSION By using use of red color marker, there is a significant increase in accuracy and user friendly. The accuracy of use of red color marker system is increased in comparison to the case when other color markers were used individually. The problem of changing lighting condition and color based recognition has been solved in this work by giving the button to set the marker color at starting phase of application. Still there are some problems while recognition speed, where speed of controlling motion of mouse is not 100% which need to be improved for some of the gestures. All mouse movement and keys action has already been mapped that is working well under given circumstances. [3] P. Mistry and P. Maes, "SixthSense: A Wearable Gestural Interface", inacm SIGGRAPH ASIA 2009 Sketches. New York, NY, USA: ACM,2009. [4] A. Utsumi, T. Miyasato, and F. Kishino, "Multi-Camera Hand Pose Recognition System Using Skeleton Image", in RO-MAN'95:Proc. Of4th IEEE Int. Workshop on Robot and Human Communication. IEEE Computer Society, July 1995, pp [5] A. Wilson and N. Oliver, "GWindows: Robust Stereo Vision for Gesture Based Control of Windows", in ICMI '03: Proc. of 5th Int. Con! On Multimodal interfaces. New York, NY, USA: ACM, 2003, pp [6]Cristian Gadea, BogdanIonescu, Dan Ionescu, Shahidul Islam, Bogdan Solomon University of Ottawa, Mgestyk Technologies, Finger-Based Gesture Control of a Collaborative Online Workspace 7th IEEE International Symposium on Applied computational intelligence and Page 1458
6 Informatics May 24-26, 2012 Timisoara, Romania. [7]Manaram Ganasekera, COMPUTER VISION BASED HANDMOVEMENT CAPTURING SYSTEM, The 8th International Conference on Computer Science & Education (ICCSE 2013) April 26-28, Colombo, Sri Lanka [8]Fabrizio Lamberti, Endowing Existing Desktop Applications with Customizable Body Gesturebased Interfaces,IEEE Int l Conference on Consumer Electronics(ICCE), , 2013 [9]Anupam Agrawal, Rohit Raj and Shubha Porwal, Vision-based Multimodal Human- Computer Interaction using Hand and Head Gestures, Proceedings of 2013 IEEE Conference on Information and Communication Technologies (ICT 2013) Page 1459
Research Seminar. Stefano CARRINO fr.ch
Research Seminar Stefano CARRINO stefano.carrino@hefr.ch http://aramis.project.eia- fr.ch 26.03.2010 - based interaction Characterization Recognition Typical approach Design challenges, advantages, drawbacks
More informationGesture Recognition with Real World Environment using Kinect: A Review
Gesture Recognition with Real World Environment using Kinect: A Review Prakash S. Sawai 1, Prof. V. K. Shandilya 2 P.G. Student, Department of Computer Science & Engineering, Sipna COET, Amravati, Maharashtra,
More informationAdvancements in Gesture Recognition Technology
IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 4, Issue 4, Ver. I (Jul-Aug. 2014), PP 01-07 e-issn: 2319 4200, p-issn No. : 2319 4197 Advancements in Gesture Recognition Technology 1 Poluka
More informationINTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY
Ashwini Parate,, 2013; Volume 1(8): 754-761 INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK ROBOT AND HOME APPLIANCES CONTROL USING
More informationMOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device
MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device Enkhbat Davaasuren and Jiro Tanaka 1-1-1 Tennodai, Tsukuba, Ibaraki 305-8577 Japan {enkhee,jiro}@iplab.cs.tsukuba.ac.jp Abstract.
More informationEnabling Cursor Control Using on Pinch Gesture Recognition
Enabling Cursor Control Using on Pinch Gesture Recognition Benjamin Baldus Debra Lauterbach Juan Lizarraga October 5, 2007 Abstract In this project we expect to develop a machine-user interface based on
More informationE90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright
E90 Project Proposal 6 December 2006 Paul Azunre Thomas Murray David Wright Table of Contents Abstract 3 Introduction..4 Technical Discussion...4 Tracking Input..4 Haptic Feedack.6 Project Implementation....7
More informationInternational Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN
International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, www.ijcea.com ISSN 2321-3469 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS ABSTRACT Saniya Zahoor
More informationInteractive Coffee Tables: Interfacing TV within an Intuitive, Fun and Shared Experience
Interactive Coffee Tables: Interfacing TV within an Intuitive, Fun and Shared Experience Radu-Daniel Vatavu and Stefan-Gheorghe Pentiuc University Stefan cel Mare of Suceava, Department of Computer Science,
More informationWhat was the first gestural interface?
stanford hci group / cs247 Human-Computer Interaction Design Studio What was the first gestural interface? 15 January 2013 http://cs247.stanford.edu Theremin Myron Krueger 1 Myron Krueger There were things
More informationA SURVEY ON GESTURE RECOGNITION TECHNOLOGY
A SURVEY ON GESTURE RECOGNITION TECHNOLOGY Deeba Kazim 1, Mohd Faisal 2 1 MCA Student, Integral University, Lucknow (India) 2 Assistant Professor, Integral University, Lucknow (india) ABSTRACT Gesture
More informationGesticulation Based Smart Surface with Enhanced Biometric Security Using Raspberry Pi
www.ijcsi.org https://doi.org/10.20943/01201705.5660 56 Gesticulation Based Smart Surface with Enhanced Biometric Security Using Raspberry Pi R.Gayathri 1, E.Roshith 2, B.Sanjana 2, S. Sanjeev Kumar 2,
More informationCONTROLLING METHODS AND CHALLENGES OF ROBOTIC ARM
CONTROLLING METHODS AND CHALLENGES OF ROBOTIC ARM Aniket D. Kulkarni *1, Dr.Sayyad Ajij D. *2 *1(Student of E&C Department, MIT Aurangabad, India) *2(HOD of E&C department, MIT Aurangabad, India) aniket2212@gmail.com*1,
More informationDevelopment of a telepresence agent
Author: Chung-Chen Tsai, Yeh-Liang Hsu (2001-04-06); recommended: Yeh-Liang Hsu (2001-04-06); last updated: Yeh-Liang Hsu (2004-03-23). Note: This paper was first presented at. The revised paper was presented
More informationVICs: A Modular Vision-Based HCI Framework
VICs: A Modular Vision-Based HCI Framework The Visual Interaction Cues Project Guangqi Ye, Jason Corso Darius Burschka, & Greg Hager CIRL, 1 Today, I ll be presenting work that is part of an ongoing project
More informationHumera Syed 1, M. S. Khatib 2 1,2
A Hand Gesture Recognition Approach towards Shoulder Wearable Computing Humera Syed 1, M. S. Khatib 2 1,2 CSE, A.C.E.T/ R.T.M.N.U, India ABSTRACT: Human Computer Interaction needs computer systems and
More informationDesign a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison
e-issn 2455 1392 Volume 2 Issue 10, October 2016 pp. 34 41 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com Design a Model and Algorithm for multi Way Gesture Recognition using Motion and
More informationGUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer
2010 GUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer By: Abdullah Almurayh For : Dr. Chow UCCS CS525 Spring 2010 5/4/2010 Contents Subject Page 1. Abstract 2 2. Introduction
More informationVisual Interpretation of Hand Gestures as a Practical Interface Modality
Visual Interpretation of Hand Gestures as a Practical Interface Modality Frederik C. M. Kjeldsen Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate
More informationEFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION
EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION 1 Arun.A.V, 2 Bhatath.S, 3 Chethan.N, 4 Manmohan.C.M, 5 Hamsaveni M 1,2,3,4,5 Department of Computer Science and Engineering,
More informationPinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data
Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Hrvoje Benko Microsoft Research One Microsoft Way Redmond, WA 98052 USA benko@microsoft.com Andrew D. Wilson Microsoft
More informationSPY ROBOT CONTROLLING THROUGH ZIGBEE USING MATLAB
SPY ROBOT CONTROLLING THROUGH ZIGBEE USING MATLAB MD.SHABEENA BEGUM, P.KOTESWARA RAO Assistant Professor, SRKIT, Enikepadu, Vijayawada ABSTRACT In today s world, in almost all sectors, most of the work
More informationControlling Humanoid Robot Using Head Movements
Volume-5, Issue-2, April-2015 International Journal of Engineering and Management Research Page Number: 648-652 Controlling Humanoid Robot Using Head Movements S. Mounica 1, A. Naga bhavani 2, Namani.Niharika
More informationSri Shakthi Institute of Engg and Technology, Coimbatore, TN, India.
Intelligent Forms Processing System Tharani B 1, Ramalakshmi. R 2, Pavithra. S 3, Reka. V. S 4, Sivaranjani. J 5 1 Assistant Professor, 2,3,4,5 UG Students, Dept. of ECE Sri Shakthi Institute of Engg and
More informationChallenging areas:- Hand gesture recognition is a growing very fast and it is I. INTRODUCTION
Hand gesture recognition for vehicle control Bhagyashri B.Jakhade, Neha A. Kulkarni, Sadanand. Patil Abstract: - The rapid evolution in technology has made electronic gadgets inseparable part of our life.
More informationA SURVEY ON HAND GESTURE RECOGNITION
A SURVEY ON HAND GESTURE RECOGNITION U.K. Jaliya 1, Dr. Darshak Thakore 2, Deepali Kawdiya 3 1 Assistant Professor, Department of Computer Engineering, B.V.M, Gujarat, India 2 Assistant Professor, Department
More informationFace Recognition Based Attendance System with Student Monitoring Using RFID Technology
Face Recognition Based Attendance System with Student Monitoring Using RFID Technology Abhishek N1, Mamatha B R2, Ranjitha M3, Shilpa Bai B4 1,2,3,4 Dept of ECE, SJBIT, Bangalore, Karnataka, India Abstract:
More informationInteraction via motion observation
Interaction via motion observation M A Foyle 1 and R J McCrindle 2 School of Systems Engineering, University of Reading, Reading, UK mfoyle@iee.org, r.j.mccrindle@reading.ac.uk www.sse.reading.ac.uk ABSTRACT
More informationGESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL
GESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL Darko Martinovikj Nevena Ackovska Faculty of Computer Science and Engineering Skopje, R. Macedonia ABSTRACT Despite the fact that there are different
More informationAvailable online at ScienceDirect. Procedia Computer Science 50 (2015 )
Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 50 (2015 ) 503 510 2nd International Symposium on Big Data and Cloud Computing (ISBCC 15) Virtualizing Electrical Appliances
More informationSIXTH SENSE TECHNOLOGY A STEP AHEAD
SIXTH SENSE TECHNOLOGY A STEP AHEAD B.Srinivasa Ragavan 1, R.Sripathy 2 1 Asst. Professor in Computer Science, 2 Asst. Professor MCA, Sri SRNM College, Sattur, Tamilnadu, (India) ABSTRACT Due to technological
More informationR (2) Controlling System Application with hands by identifying movements through Camera
R (2) N (5) Oral (3) Total (10) Dated Sign Assignment Group: C Problem Definition: Controlling System Application with hands by identifying movements through Camera Prerequisite: 1. Web Cam Connectivity
More informationA Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,
IJCSNS International Journal of Computer Science and Network Security, VOL.11 No.9, September 2011 55 A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang,
More informationHUMAN-COMPUTER INTERACTION: OVERVIEW ON STATE OF THE ART TECHNOLOGY
HUMAN-COMPUTER INTERACTION: OVERVIEW ON STATE OF THE ART TECHNOLOGY *Ms. S. VAISHNAVI, Assistant Professor, Sri Krishna Arts And Science College, Coimbatore. TN INDIA **SWETHASRI. L., Final Year B.Com
More information3D Interaction using Hand Motion Tracking. Srinath Sridhar Antti Oulasvirta
3D Interaction using Hand Motion Tracking Srinath Sridhar Antti Oulasvirta EIT ICT Labs Smart Spaces Summer School 05-June-2013 Speaker Srinath Sridhar PhD Student Supervised by Prof. Dr. Christian Theobalt
More informationVision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab
Vision-based User-interfaces for Pervasive Computing Tutorial Notes Vision Interface Group MIT AI Lab Table of contents Biographical sketch..ii Agenda..iii Objectives.. iv Abstract..v Introduction....1
More informationProjection Based HCI (Human Computer Interface) System using Image Processing
GRD Journals- Global Research and Development Journal for Volume 1 Issue 5 April 2016 ISSN: 2455-5703 Projection Based HCI (Human Computer Interface) System using Image Processing Pankaj Dhome Sagar Dhakane
More informationActive Stereo Vision. COMP 4102A Winter 2014 Gerhard Roth Version 1
Active Stereo Vision COMP 4102A Winter 2014 Gerhard Roth Version 1 Why active sensors? Project our own texture using light (usually laser) This simplifies correspondence problem (much easier) Pluses Can
More informationGesture Recognition Based Mouse Events
Gesture Recognition Based Mouse Events Rachit Puri IM\Web, Multimedia & Services\Web Solutions\Web Engine Samsung Research India-Bangalore Bangalore 560037, India rachit.puri@samsung.com ABSTRACT This
More informationVirtual Touch Human Computer Interaction at a Distance
International Journal of Computer Science and Telecommunications [Volume 4, Issue 5, May 2013] 18 ISSN 2047-3338 Virtual Touch Human Computer Interaction at a Distance Prasanna Dhisale, Puja Firodiya,
More informationDepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface
DepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface Hrvoje Benko and Andrew D. Wilson Microsoft Research One Microsoft Way Redmond, WA 98052, USA
More informationISSN: [Arora * et al., 7(4): April, 2018] Impact Factor: 5.164
IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY REAL TIME SYSTEM CONTROLLING USING A WEB CAMERA BASED ON COLOUR DETECTION Reema Arora *1, Renu Kumari 2 & Shabnam Kumari 3 *1
More informationNavigation of PowerPoint Using Hand Gestures
Navigation of PowerPoint Using Hand Gestures Dnyanada R Jadhav 1, L. M. R. J Lobo 2 1 M.E Department of Computer Science & Engineering, Walchand Institute of technology, Solapur, India 2 Associate Professor
More informationShort Course on Computational Illumination
Short Course on Computational Illumination University of Tampere August 9/10, 2012 Matthew Turk Computer Science Department and Media Arts and Technology Program University of California, Santa Barbara
More informationLive Hand Gesture Recognition using an Android Device
Live Hand Gesture Recognition using an Android Device Mr. Yogesh B. Dongare Department of Computer Engineering. G.H.Raisoni College of Engineering and Management, Ahmednagar. Email- yogesh.dongare05@gmail.com
More informationStudent Attendance Monitoring System Via Face Detection and Recognition System
IJSTE - International Journal of Science Technology & Engineering Volume 2 Issue 11 May 2016 ISSN (online): 2349-784X Student Attendance Monitoring System Via Face Detection and Recognition System Pinal
More informationGetting started 1 System Requirements... 1 Software Installation... 2 Hardware Installation... 2 System Limitations and Tips on Scanning...
Contents Getting started 1 System Requirements......................... 1 Software Installation......................... 2 Hardware Installation........................ 2 System Limitations and Tips on
More informationAn Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi
An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi Department of E&TC Engineering,PVPIT,Bavdhan,Pune ABSTRACT: In the last decades vehicle license plate recognition systems
More informationInteracting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)
Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception
More informationFinger rotation detection using a Color Pattern Mask
Finger rotation detection using a Color Pattern Mask V. Shishir Reddy 1, V. Raghuveer 2, R. Hithesh 3, J. Vamsi Krishna 4,, R. Pratesh Kumar Reddy 5, K. Chandra lohit 6 1,2,3,4,5,6 Electronics and Communication,
More informationMRT: Mixed-Reality Tabletop
MRT: Mixed-Reality Tabletop Students: Dan Bekins, Jonathan Deutsch, Matthew Garrett, Scott Yost PIs: Daniel Aliaga, Dongyan Xu August 2004 Goals Create a common locus for virtual interaction without having
More information3D-Position Estimation for Hand Gesture Interface Using a Single Camera
3D-Position Estimation for Hand Gesture Interface Using a Single Camera Seung-Hwan Choi, Ji-Hyeong Han, and Jong-Hwan Kim Department of Electrical Engineering, KAIST, Gusung-Dong, Yusung-Gu, Daejeon, Republic
More informationBuilding a gesture based information display
Chair for Com puter Aided Medical Procedures & cam par.in.tum.de Building a gesture based information display Diplomarbeit Kickoff Presentation by Nikolas Dörfler Feb 01, 2008 Chair for Computer Aided
More informationDirect gaze based environmental controls
Loughborough University Institutional Repository Direct gaze based environmental controls This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation: SHI,
More informationHand Segmentation for Hand Gesture Recognition
Hand Segmentation for Hand Gesture Recognition Sonal Singhai Computer Science department Medicaps Institute of Technology and Management, Indore, MP, India Dr. C.S. Satsangi Head of Department, information
More informationII. LITERATURE SURVEY
Hand Gesture Recognition Using Operating System Mr. Anap Avinash 1 Bhalerao Sushmita 2, Lambrud Aishwarya 3, Shelke Priyanka 4, Nirmal Mohini 5 12345 Computer Department, P.Dr.V.V.P. Polytechnic, Loni
More informationAUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING
AUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING ABSTRACT Chutisant Kerdvibulvech Department of Information and Communication Technology, Rangsit University, Thailand Email: chutisant.k@rsu.ac.th In
More informationVISUAL FINGER INPUT SENSING ROBOT MOTION
VISUAL FINGER INPUT SENSING ROBOT MOTION Mr. Vaibhav Shersande 1, Ms. Samrin Shaikh 2, Mr.Mohsin Kabli 3, Mr.Swapnil Kale 4, Mrs.Ranjana Kedar 5 Student, Dept. of Computer Engineering, KJ College of Engineering
More informationREBO: A LIFE-LIKE UNIVERSAL REMOTE CONTROL
World Automation Congress 2010 TSI Press. REBO: A LIFE-LIKE UNIVERSAL REMOTE CONTROL SEIJI YAMADA *1 AND KAZUKI KOBAYASHI *2 *1 National Institute of Informatics / The Graduate University for Advanced
More informationReVRSR: Remote Virtual Reality for Service Robots
ReVRSR: Remote Virtual Reality for Service Robots Amel Hassan, Ahmed Ehab Gado, Faizan Muhammad March 17, 2018 Abstract This project aims to bring a service robot s perspective to a human user. We believe
More informationDevelopment of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture
Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture Akira Suganuma Depertment of Intelligent Systems, Kyushu University, 6 1, Kasuga-koen, Kasuga,
More informationA Real Time Static & Dynamic Hand Gesture Recognition System
International Journal of Engineering Inventions e-issn: 2278-7461, p-issn: 2319-6491 Volume 4, Issue 12 [Aug. 2015] PP: 93-98 A Real Time Static & Dynamic Hand Gesture Recognition System N. Subhash Chandra
More informationVarious Calibration Functions for Webcams and AIBO under Linux
SISY 2006 4 th Serbian-Hungarian Joint Symposium on Intelligent Systems Various Calibration Functions for Webcams and AIBO under Linux Csaba Kertész, Zoltán Vámossy Faculty of Science, University of Szeged,
More informationifinger Study of Gesture Recognition Technologies & Its Applications Volume II of II
University of Macau Faculty of Science and Technology ifinger Study of Gesture Recognition Technologies & Its Applications Volume II of II by Chi Ian, Choi, Student No: DB02828 Final Project Report submitted
More informationIntelligent interaction
BionicWorkplace: autonomously learning workstation for human-machine collaboration Intelligent interaction Face to face, hand in hand. The BionicWorkplace shows the extent to which human-machine collaboration
More informationReal Time Hand Gesture Tracking for Network Centric Application
Real Time Hand Gesture Tracking for Network Centric Application Abstract Chukwuemeka Chijioke Obasi 1 *, Christiana Chikodi Okezie 2, Ken Akpado 2, Chukwu Nnaemeka Paul 3, Asogwa, Chukwudi Samuel 1, Akuma
More informationA Very High Level Interface to Teleoperate a Robot via Web including Augmented Reality
A Very High Level Interface to Teleoperate a Robot via Web including Augmented Reality R. Marín, P. J. Sanz and J. S. Sánchez Abstract The system consists of a multirobot architecture that gives access
More informationInternational Journal of Informative & Futuristic Research ISSN (Online):
Reviewed Paper Volume 2 Issue 6 February 2015 International Journal of Informative & Futuristic Research An Innovative Approach Towards Virtual Drums Paper ID IJIFR/ V2/ E6/ 021 Page No. 1603-1608 Subject
More informationCSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2
CSE 165: 3D User Interaction Lecture #7: Input Devices Part 2 2 Announcements Homework Assignment #2 Due tomorrow at 2pm Sony Move check out Homework discussion Monday at 6pm Input Devices CSE 165 -Winter
More informationLabVIEW based Intelligent Frontal & Non- Frontal Face Recognition System
LabVIEW based Intelligent Frontal & Non- Frontal Face Recognition System Muralindran Mariappan, Manimehala Nadarajan, and Karthigayan Muthukaruppan Abstract Face identification and tracking has taken a
More informationDETECTION AND RECOGNITION OF HAND GESTURES TO CONTROL THE SYSTEM APPLICATIONS BY NEURAL NETWORKS. P.Suganya, R.Sathya, K.
Volume 118 No. 10 2018, 399-405 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: 10.12732/ijpam.v118i10.40 ijpam.eu DETECTION AND RECOGNITION OF HAND GESTURES
More informationFOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM
FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM Takafumi Taketomi Nara Institute of Science and Technology, Japan Janne Heikkilä University of Oulu, Finland ABSTRACT In this paper, we propose a method
More informationHuman Computer Interaction by Gesture Recognition
IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 9, Issue 3, Ver. V (May - Jun. 2014), PP 30-35 Human Computer Interaction by Gesture Recognition
More informationA Kinect-based 3D hand-gesture interface for 3D databases
A Kinect-based 3D hand-gesture interface for 3D databases Abstract. The use of natural interfaces improves significantly aspects related to human-computer interaction and consequently the productivity
More informationToward an Augmented Reality System for Violin Learning Support
Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp
More informationKINECT HANDS-FREE. Rituj Beniwal. Department of Electrical Engineering Indian Institute of Technology, Kanpur. Pranjal Giri
KINECT HANDS-FREE Rituj Beniwal Pranjal Giri Agrim Bari Raman Pratap Singh Akash Jain Department of Aerospace Engineering Indian Institute of Technology, Kanpur Atharva Mulmuley Department of Chemical
More informationGlassSpection User Guide
i GlassSpection User Guide GlassSpection User Guide v1.1a January2011 ii Support: Support for GlassSpection is available from Pyramid Imaging. Send any questions or test images you want us to evaluate
More informationNCCT IEEE PROJECTS ADVANCED ROBOTICS SOLUTIONS. Latest Projects, in various Domains. Promise for the Best Projects
NCCT Promise for the Best Projects IEEE PROJECTS in various Domains Latest Projects, 2009-2010 ADVANCED ROBOTICS SOLUTIONS EMBEDDED SYSTEM PROJECTS Microcontrollers VLSI DSP Matlab Robotics ADVANCED ROBOTICS
More informationRelationship to theory: This activity involves the motion of bodies under constant velocity.
UNIFORM MOTION Lab format: this lab is a remote lab activity Relationship to theory: This activity involves the motion of bodies under constant velocity. LEARNING OBJECTIVES Read and understand these instructions
More informationARTIFICIAL ROBOT NAVIGATION BASED ON GESTURE AND SPEECH RECOGNITION
ARTIFICIAL ROBOT NAVIGATION BASED ON GESTURE AND SPEECH RECOGNITION ABSTRACT *Miss. Kadam Vaishnavi Chandrakumar, ** Prof. Hatte Jyoti Subhash *Research Student, M.S.B.Engineering College, Latur, India
More informationTouch & Gesture. HCID 520 User Interface Software & Technology
Touch & Gesture HCID 520 User Interface Software & Technology What was the first gestural interface? Myron Krueger There were things I resented about computers. Myron Krueger There were things I resented
More informationA Novel Approach for Image Cropping and Automatic Contact Extraction from Images
A Novel Approach for Image Cropping and Automatic Contact Extraction from Images Prof. Vaibhav Tumane *, {Dolly Chaurpagar, Ankita Somkuwar, Gauri Sonone, Sukanya Marbade } # Assistant Professor, Department
More informationWadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology
ISSN: 2454-132X Impact factor: 4.295 (Volume 4, Issue 1) Available online at www.ijariit.com Hand Detection and Gesture Recognition in Real-Time Using Haar-Classification and Convolutional Neural Networks
More informationSensor system of a small biped entertainment robot
Advanced Robotics, Vol. 18, No. 10, pp. 1039 1052 (2004) VSP and Robotics Society of Japan 2004. Also available online - www.vsppub.com Sensor system of a small biped entertainment robot Short paper TATSUZO
More informationInterior Design with Augmented Reality
Interior Design with Augmented Reality Ananda Poudel and Omar Al-Azzam Department of Computer Science and Information Technology Saint Cloud State University Saint Cloud, MN, 56301 {apoudel, oalazzam}@stcloudstate.edu
More informationMULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT
MULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT F. TIECHE, C. FACCHINETTI and H. HUGLI Institute of Microtechnology, University of Neuchâtel, Rue de Tivoli 28, CH-2003
More informationBlue Eyes Technology with Electric Imp Explorer Kit Ankita Shaily*, Saurabh Anand I.
ABSTRACT 2018 IJSRST Volume 4 Issue6 Print ISSN: 2395-6011 Online ISSN: 2395-602X National Conference on Smart Computation and Technology in Conjunction with The Smart City Convergence 2018 Blue Eyes Technology
More informationTraining Schedule. Robotic System Design using Arduino Platform
Training Schedule Robotic System Design using Arduino Platform Session - 1 Embedded System Design Basics : Scope : To introduce Embedded Systems hardware design fundamentals to students. Processor Selection
More informationCS295-1 Final Project : AIBO
CS295-1 Final Project : AIBO Mert Akdere, Ethan F. Leland December 20, 2005 Abstract This document is the final report for our CS295-1 Sensor Data Management Course Final Project: Project AIBO. The main
More informationVirtual Grasping Using a Data Glove
Virtual Grasping Using a Data Glove By: Rachel Smith Supervised By: Dr. Kay Robbins 3/25/2005 University of Texas at San Antonio Motivation Navigation in 3D worlds is awkward using traditional mouse Direct
More informationCOMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES
International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3
More informationThe Control of Avatar Motion Using Hand Gesture
The Control of Avatar Motion Using Hand Gesture ChanSu Lee, SangWon Ghyme, ChanJong Park Human Computing Dept. VR Team Electronics and Telecommunications Research Institute 305-350, 161 Kajang-dong, Yusong-gu,
More informationAnalysis of Various Methodology of Hand Gesture Recognition System using MATLAB
Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Komal Hasija 1, Rajani Mehta 2 Abstract Recognition is a very effective area of research in regard of security with the involvement
More informationAir Marshalling with the Kinect
Air Marshalling with the Kinect Stephen Witherden, Senior Software Developer Beca Applied Technologies stephen.witherden@beca.com Abstract. The Kinect sensor from Microsoft presents a uniquely affordable
More informationUser Interface Agents
User Interface Agents Roope Raisamo (rr@cs.uta.fi) Department of Computer Sciences University of Tampere http://www.cs.uta.fi/sat/ User Interface Agents Schiaffino and Amandi [2004]: Interface agents are
More informationSuper resolution with Epitomes
Super resolution with Epitomes Aaron Brown University of Wisconsin Madison, WI Abstract Techniques exist for aligning and stitching photos of a scene and for interpolating image data to generate higher
More informationKINECT CONTROLLED HUMANOID AND HELICOPTER
KINECT CONTROLLED HUMANOID AND HELICOPTER Muffakham Jah College of Engineering & Technology Presented by : MOHAMMED KHAJA ILIAS PASHA ZESHAN ABDUL MAJEED AZMI SYED ABRAR MOHAMMED ISHRAQ SARID MOHAMMED
More informationComparison of Head Movement Recognition Algorithms in Immersive Virtual Reality Using Educative Mobile Application
Comparison of Head Recognition Algorithms in Immersive Virtual Reality Using Educative Mobile Application Nehemia Sugianto 1 and Elizabeth Irenne Yuwono 2 Ciputra University, Indonesia 1 nsugianto@ciputra.ac.id
More information* Intelli Robotic Wheel Chair for Specialty Operations & Physically Challenged
ADVANCED ROBOTICS SOLUTIONS * Intelli Mobile Robot for Multi Specialty Operations * Advanced Robotic Pick and Place Arm and Hand System * Automatic Color Sensing Robot using PC * AI Based Image Capturing
More informationBare PCB Inspection and Sorting System
Bare PCB Inspection and Sorting System Divya C Thomas 1, Jeetendra R Bhandankar 2, Devendra Sutar 3 1, 3 Electronics and Telecommunication Department, Goa College of Engineering, Ponda, Goa, India 2 Micro-
More information