Hand Gesture Recognition for Kinect v2 Sensor in the Near Distance Where Depth Data Are Not Provided

Size: px
Start display at page:

Download "Hand Gesture Recognition for Kinect v2 Sensor in the Near Distance Where Depth Data Are Not Provided"

Transcription

1 , pp Hand Gesture Recognition for Kinect v2 Sensor in the Near Distance Where Depth Data Are Not Provided Min-Soo Kim 1 and Choong Ho Lee 2 1 Dept. of Info. and Comm. Eng., Hanbat National Univ., Daejeon-City, Rep. of Korea 2 Graduate School of Info. and Comm. Eng., Hanbat National Univ., Daejeon-City, Rep. of Korea 1 asdq200@naver.com, 2 chlee@hanbat.ac.kr Abstract Kinect v2 sensor does not provide depth information and skeletal traction function in near distance from the sensor. That is why many researches, to recognize hand gestures, are focused on the skeletal tracking only inside the range of detection. This paper proposes a method which can recognize hand gestures in the distance less than 0.5 meter without conventional skeletal tracking when Kinect v2 sensor is used. The proposed method does not use the information of depth sensor and infrared sensor, but detect hand area and count the number of isolated areas which are generated by drawing a circle in the center of the hand area. This method introduces new detectable gestures without high cost, so that it can be a substitute for the existing mouse-movement controlling and dynamic gesture recognition method such as clicking a mouse, clicking and dragging, rotating an image with two hands, and scaling an image with two hands in near distance. The gestures are appropriate for the user interface of smart devices which employ the interactions based on hand gestures in near distance. Keywords: User interface, Kinect v2 sensor, Hand gesture recognition 1. Introduction In recent years, hand gesture recognition has been actively studied as one of the human computer interactions. Since the gesture recognition can be used for various kinds of digital devices such as smartphones, tablet computers as well as conventional desktop computers and laptop PCs, much of the attention is concentrated on the related works. [1] To detect hand area, the existing method uses various color models such as YCbCr, HSV and RGB. They determine thresholds considering illumination and background objects included in the environments, but they are not consistent and very sensitive to various environmental factors. Since the colors of face area and other skin area are very similar to the hand color, it is very difficult to determine the hand area by color information. For actual application to detect hand area, the existing method shows low performance to discriminate hand area from face area when hand area is overlapped with face area. Further, in order to improve detection performance, it is necessary to wear specific colored gloves [2-4], or uses only depth information without infrared information [5-7]. Without using the color information and special gloves, finger tracking methods are generally used such as in [5, 8, 9]. But these kinds of methods use relatively complicated algorithms [9, 10] such as SVM or Convex Hull or AdaBoost. Moreover, designing hand gesture recognition without high cost has become another important issue and described in [11]. A face recognition technique using Kinect is reported in [12]. On the other hand, for the applications, various sensors are introduced in the market. There are two kinds of sensors that are commonly used. One is used for short distance such as from 0.2 to 1.2 meters and the others are used for relatively long distance for ISSN: IJSEIA Copyright c 2016 SERSC

2 example from 0.8 to 4.0 meters, or from 0.5 to 8.0 meters. These pertain to Kinect v1 and v2 sensor of Microsoft [13] and that one pertains to Intel Realsense SR300 [14]. Further, two kinds of sensors provide the skeletal recognition for hand gesture recognition, and uses depth information in the available distance. However, Kinect v2 sensor neither provides depth information nor detects infrared information in the distance less than 0.5meters. Further, the recognition method in Intel Realsense needs complicated conventional skeletal recognition. Figure 1 shows the depth sensor and IR (infrared sensor) of Kinect v2 sensor. The range of depth sensor and IR sensor of Kinect v2 sensor is 0.5 meter ~ 8.0 meters while it is 0.8 meter ~ 4.0 meters for Kinect v1. As Kinect sensors are developed, many research studies are conducted to recognize human movement activities using the sensors. [9, 15-21] A technique to improve Kinect-skeleton estimation is reported in [15]. To correct radial distortion of RGB camera and find the transformation matrix for the correspondence between the RGB and depth image of Kinect v2 is described in [16]. Intuitive hand gestures for controlling the rotation of 3D digital object is described in [19]. The conventional method which extracts palm region using RGB and depth sensor has been reported by [19]. In addition to that, tracking and evaluation of human motion for medical application is in [20], a depth completion method using Kinect v2 is in [17]. When emphasis is on the real time using Kinect v2 sensor, it is reported in [9, 21]. Figure 1. Kinect for Windows v2 Sensor which has a Color Camera, a Depth Sensor and an IR Sensor However, no research has been reported to utilize Kinect v2 in the near distance outside the detection range but almost all of the research studies were conducted inside the detection range. This paper presents a new simple method to recognize hand gestures in the distance less than 0.5 meter for Kinect v2 sensor without using the data of depth sensor and IR sensor. In other words, this method enables Kinect v2 sensor to detect hand area and recognize hand gestures in the near distance less than 0.5 meter where it does not provide related depth information or infrared information or skeletal recognition method. Furthermore, the newly developed simple gestures will also be presented. This paper extends the research which is described in [22]. Section 2 explains the method of hand area detection which we come up with to detect the hand area in the near distance where the depth sensor cannot give depth information. 408 Copyright c 2016 SERSC

3 Section 3 describes existing hand gesture recognition method used in our method. The proposed method will be presented in Section 4. Further, experimentation and newly developed gestures will be followed in Section 5. It includes hand gestures which use z- direction as well as those of x-y directions with one hand and two hands. The analyses and discussions are made in the same section. Finally, the conclusion and remarks are presented in Section Hand Area Detection in Near Distance Kinect v2 Sensor provides a depth sensor and an infrared sensor to provide depth information and infrared information. But in near distance less than 0.5 meter, it does not give the information. We propose the new method to detect hand area which can be used in the near distance. We used the characteristics of Kinect v2 sensor and infrared sensor which do not give the appropriate information including depth information and infrared radiation information in less than 0.5 meter. Specifically, we used the white areas of two resultant images caused by the side effect of sensor limitations. Figure 2 shows the images. Conducting AND operation of two images, for example (b) and (c), then we can extract the hand area. Another pair of (e) and (f) in Figure 2 can be used to extract hand area in the dark environment. (a) (b) (c) (d) (e) (f) Figure 2. Hand area detection by AND operation of depth image and infrared image. (a) Original Images. (b) Depth Images. (c) Infrared Images. (d) Original Image in darkness. (f) Depth Image in darkness. (g) Infrared Image in darkness Copyright c 2016 SERSC 409

4 Figure 3. The Process of Detecting Hand Region Figure 3 illustrates the process to detect hand region. After extracting the hand area, we conduct morphology opening operation to remove noises in the image. 3. Hand Gesture Recognition After we detected the hand area, we recognize specific gestures. We split the area by drawing a black circle in the center of a hand, then count the number of areas isolated. [6, 23] 410 Copyright c 2016 SERSC

5 3.1. Computation of Weight Centers for Hand Areas To detect the hand area, we should first obtain the center of a hand area, i.e., moment [24], and its coordinates described in [6, 23]. Then the center of a hand m is defined as pq, (1). p q m, x y f ( x, y) dxdy pq (1) And the detected hand area, 0 th -order moment, can be expressed as (2). m0,0 f ( x, y) dxdy (2) Furthermore, the centers of weight, first-order moment, using (3) and (4), respectively as follows: m 1,0 and m 0,1, can be obtained m1,0 dxdyxf ( x, y) (3) m0,1 dxdyyf ( x, y). (4) And, the coordinate x c, y c can be denoted by (5) and (6), respectively. x y c c m 1,0 (5) m m 0,0 0,1 (6) m 0, Division of Hand Areas Figure 4 shows the procedures described in Section 2. Figure 4 (a) shows the area detected, (b) shows the computed locations of hand areas, (c) shows the circles in the centers of hand areas and (d) shows the resulted areas divided by the black circles. Here, we count the number of areas divided and can recognize various gestures. Copyright c 2016 SERSC 411

6 (a) (b) (c) (d) Figure 4. Procedure to Recognize Hand Area Region. (a) Hand Area Obtained. (b) Computation of Centers of Hand Areas. (c) Drawing Virtual Circles. (d) Filling the Circles by Black Color 4. The Proposed Method We propose newly invented hand gestures by counting the number of contours (separate areas) which was isolated by a black circle in the weight center. After detecting the hand areas, we draw a rectangle which includes hand area and draw a black circle whose diameter is 1/3 of the height of the rectangle. Using the number of separate areas and the moving of the areas, we can invent various gestures Assumptions and Advantages The proposed method does not use depth information and infrared information directly, but uses the side effect (error images) outside the detection range of Microsoft Kinect v2 sensor. We assume that the hands are the nearest objects and are moving less than 0.5 meter from the sensor. We propose that this method can be used outside the detection range in the near distance; and simpler than existing method because it does not use complicated algorithms and show more stable detection performance and independent of color of hands, illumination status, and background objects unlike existing methods [1-3, 6, 10] Recognition of Hand Gestures We can discriminate various gestures by counting number of isolated areas. Figure 5 shows how to recognize the specific hand gestures. In Figure 4, (a) to (c) are for singlehand gesture, and (d) to (f) are for two-hand gestures. Here, (a) and (d) mean mouse release ; (b) and (e) mean mouse click ; (c) and (f) designate mode change. Further, (g) expresses zoom in an object but we can zoom out by moving the hands to the opposite direction. To change to the rotation mode, we close two fists like (f), and after that open fists like (e) and move two hands like (h). By counting the number of contours, we discriminate the gestures. For example, (a) has five contours, (b) has two contours, and (c) 412 Copyright c 2016 SERSC

7 has one contour respectively when one hand is used. And (d) has ten contours, (e) has four contours, and (f) has two contours respectively when two hands are used. (a) (b) (c) (d) (e) (f) (g) Figure 5. The Proposed Hand Gestures. (a) Mouse Release. (b) Mouse Click. (c). Mode Change. (d) Mouse Release. (e) Mouse Click. (f) Mode Change. (g) Zoom-in. (h) Rotation 4.3. New Gestures Which Use Z-Axis Information Since Kinect v2 sensor cannot provide depth information and infrared information in near distance, the depth is determined by the size of radius of a circle in the center of hand area. Here, the depth means the distance from the x-y plane of hand area toward the Kinect v2 sensor. Furthermore, we can call the direction of the distance z-axis because it is perpendicular to the x-y plane composed of x, y in (5) and (6). Figure 6 shows the four layers and detected hand areas. Figure 6 illustrates four layers which describe divided distances from hand area toward Kinect v2 sensor. Here, layer 1 to layer 4 are determined by the lengths of radii of the circles in the hand area. It should be noted that the left figures are drawn from the point of human, while right figures are drawn from the point of sensors. For example, the top left figure denotes the layer 1 which is the nearest from the hand area is chosen but the hand area of top right figure is the smallest in the point of sensor s view. Specifically, the thresholds of radii of circles are 38 pixels, 48 pixels, 58 pixels and 68 pixels for layer 1, layer 2, layer 3 and layer 4, respectively. (h) Copyright c 2016 SERSC 413

8 Figure 6. Z-axis values according to the depth which is the distance from x- y plane of the hand area toward the Kinect sensor: left figures are for human s view; right figures are for sensor s view. (a) Layer 1: The radius is larger than or equal to 38 pixels and less than 48 pixels. (b) Layer 2: The radius is larger than or equal to 48 pixels and less than 48 pixels. (c) Layer 3: The radius is larger than or equal to 58 pixels and less than 48 pixels. (d) Layer 4: The radius is larger than or equal to 68 pixels 4. Experimentation We used Kinect v2 sensor and conducted the experimentation in the distance less than 0.5 meter which the sensor does not provide depth information and infrared data. In addition to that, we used openframeworks Kinect v2 in order to make user interface which is based on C++ and opengl. Further, we used various libraries including the addons of the openframeworks, and cross-platforms. The addons include ofxopencv and ofxcv that enable opencv in openframeworks; and ofxkinect2 that is to use Kinect v2 sensor. For the operating system, Microsoft Windows 10 is used and Visual Studio 2015 community is installed. In Figure 7, (a) expresses a right posture to detect hand area for our method while (b) shows a bad posture which gives extra skin area from elbow to wrist incorrectly. In case of (b) the center of the hand area moves to the wrist part, so that the hand area is not extracted. Additionally, we have a palm of one hand tilted in various ways like (c), but have obtained valid results also. In (d), left figure means releasemouse, middle figure means hold the object by mouse-click, and right figure means move an object. When using one hand, we marked a dot on the selected object. Figures (e) and (f) are for the two-hand gestures. In the (e), the left figure denotes mouse release, the center figure denotes hold the object by mouse-click, and the right figure expresses zoom-out. Similarly, in figure (f), the left figure denotes mouse release, the middle figure denotes hold the object by mouse-click, and the right figure expresses rotation. The two dots around an object denote that the focus is on the object. 414 Copyright c 2016 SERSC

9 (a) (b) (c) (d) Copyright c 2016 SERSC 415

10 (e) (f) Figure 7. The Proposed Gestures. (a) A Correct Posture and the Detected Area. (b) An Incorrect Posture and Detected Area. (c) Tilted Palms and the Detected Areas. (e) Clicking an Object with Two Hands and Expanding an Image. (f) Clicking an Object with Two Hands and Rotation We have confirmed that our method is valid in various situations. When the subject person is changed, the hand area is changed. So that, the thresholds to determine the layers are changed according to the lengths of circles which are located at the center of hand areas. We experimented for three persons and confirmed that our method is stable for the ranges in Figure Conclusions This paper has proposed a new simple method to recognize gestures in near distance less than 0.5 meter where Kinect v2 sensor cannot provide depth information and infrared sensor data. The method tracks hand area and counts number of contours, and uses direction of contours. The proposed method is simpler than the existing method which detects finger tracking method because it only checks the number of areas divided by a black circle in the center of hand area and the moving direction. Further, it can be used to develop three-dimensional user interface, since it uses z-axis information using the length of radius of the circle located at the center of a hand area. The proposed hand gestures can be used instead of mouse clicking, dragging and moving, releasing a mouse, rotating an image with two hands, and scaling an image with two hands. The method expands the available ranges of Kinect v2 sensor and can be used also for Kinect v1 sensor. Acknowledgements We thank Hanbat National University. This research was supported by the research fund of Hanbat National University in This paper is a revised and expanded version of a paper entitled A Simple 3D Hand Gesture Interface Based on Hand Area Detection and Tracking" presented at MITA 2016 (The 12 th International Conference on Multimedia Information and Technology and Applications), Luang Prabang, Lao PDR, July 4-6, Copyright c 2016 SERSC

11 References [1] P. Premaratne, Human Computer Interaction Using Hand Gestures: Cognitive Science and Technology, Springer-Verlag New York Inc., (2014). [2] C.-H. Wu and W.-L. Chen and C. H. Lin, Depth-Based Hand Gesture Recognition, vol. 75, no. 12, (2016), pp [3] G. R. S. Murthy and R. S. Jadon, A Review of Vision Based Hand Gesture Recognition, International Journal of Information Technology and Knowledge Management, vol. 2, no. 2, (2009), pp [4] A. Abgottspon, A Hand Gesture Interface for Investigating Real-Time Human-Computer Interaction, ECU098 Informatics, 300CDE Individual Project, Coventry Univ., UK, (2010). [5] A. M. Balazs, Hand and Finger Detection Using JavaCV, (2012). [6] H. Park, A Method for Controlling Mouse Movement Using a Real-Time Camera, Master s Thesis of Brown Univ., Providence, RI, USA, (2010). [7] J. Shotton, A. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore, A. Kipman and A. Blake, Real- Time Human Pose Recognition in Parts from Single Depth Images, Communications of ACM, (2013), pp [8] (2016). [9] R. M. Gurav and P. K. Kadbe, Real Time Tracking and Contour Detection for Gesture Recognition Using OpenCV, 2015 International Conference on Industrial Instrumentation and Control (ICIC), (2015), pp [10] C. Zou, Y. Liu, J. Wang and H. Si, Deformable Part Model Based Hand Detection against Complex Backgrounds, Advances in Images and Graphics Technologies, Springer Link, vol. 634 of the series Comm. in Computer and Info. Science, (2016), pp [11] J. Molina and J. M. Martínez, A Synthetic Training Framework for Providing Gesture Scalability to 2.5D Pose-Based Hand Gesture Recognition Systems, Machine Vision and Applications, vol. 25, issue 5, (2014), pp [12] G. Goswami, M. Vatsa and R. Singh, Face Recognition with RGB-D Images Using Kinect, Face Recognition Across the Imaging Spectrum, Springer Link, (2016), pp [13] (2016). [14] (2016). [15] J. Valcik, J. Sedmidubsk and P. Zezula, Improving Kinect-Skeleton Estimation, Advanced Concepts for Intelligent Vision Systems, Springer Link, vol of Lecture Notes in Computer Science, (2015), pp [16] C. Kim, S. Yun, S.-W. Jung and C. S. Won, Color and Depth Image Correspondence for Kinect v2, Advanced Multimedia and Ubiquitous Engineering, Springer Link, vol. 354 of the series Lecture Notes in Electrical Engineering, (2015), pp [17] W. Song, A.V. Le, S. Yun, S.-W. Jung and C. S. Won, Depth Completion for Kinect v2 Sensor, Multimedia Tools and Applications, Springer Link, (2016), pp [18] L.-C. Chen, Y.-M. Cheng, P.-Y. Chu and F. E. Sandnes, The Common Characteristics of User-Defined and Mid-Air Gestures for Rotating 3D Digital Contents, Universal Access in Human-Computer Interaction Techniques and Environments, Springer Link, vol of the series Lecture Notes in Computer Science, (2016), pp [19] S. Samoil, and S. N. Yanushkevich, Depth Assisted Palm Region Extraction using the Kinect v2 Sensor, 2015 Sixth International Conference on Emerging Security Technologies, (2015), pp [20] H. Alabbasi, A. Gradinaru, F. Moldoveanu and A. Moldoveanu, Human Motion Tracking & Evaluation using Kinect v2 Sensor, The 5th IEEE International Conference on E-Health and Bioengineering, (2015). [21] Y. Lan, J. Li and Z. Ju, Data Fusion-based Real-Time Hand Gesture Recognition with Kinect v2, th International Conference on Human System Interactions (HSI), (2016). [22] M.-S. Kim and C. H. Lee, A Simple 3D Hand Gesture Interface Based on Hand Area Detection and Tracking, Proceedings of MITA 2016 (The 12 th International Conference on Multimedia Information and Technology and Applications), Luang Prabang, Lao PDR, (2016), pp [23] B. Ionescu, D. Coquin, P. Lambert and V. Buzuloiu, Dynamic Hand Gesture Recognition Using the Skeleton of the Hand, EURASIP Journal on Applied Signal Processing, vol. 13, (2005), pp [24] J. Kilian, Simple Image Analysis by Moments, (2001), pp Copyright c 2016 SERSC 417

12 Authors Min-Soo Kim, received his B.E. Degree in Computer and Information Engineering from Hanbat National University, Daejeon, Korea, in His current research interests include pattern recognition, digital image processing and human computer interfaces. Choong Ho Lee, received his B.E. and M.E. Degrees in Electronic Engineering from Yonsei University, Seoul, Korea, in 1985 and 1987, respectively. He also received his Ph.D. in Information Sciences from Tohoku University, Sendai, Japan in March of From 1985 to 2000, he was with KT as a researcher. Since 2000, he has been a professor in Graduate School of Information Communication Engineering of Hanbat National University. His current research interests include pattern recognition, digital image processing and mobile robot control. 418 Copyright c 2016 SERSC

A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,

A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung, IJCSNS International Journal of Computer Science and Network Security, VOL.11 No.9, September 2011 55 A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang,

More information

GESTURE BASED HUMAN MULTI-ROBOT INTERACTION. Gerard Canal, Cecilio Angulo, and Sergio Escalera

GESTURE BASED HUMAN MULTI-ROBOT INTERACTION. Gerard Canal, Cecilio Angulo, and Sergio Escalera GESTURE BASED HUMAN MULTI-ROBOT INTERACTION Gerard Canal, Cecilio Angulo, and Sergio Escalera Gesture based Human Multi-Robot Interaction Gerard Canal Camprodon 2/27 Introduction Nowadays robots are able

More information

The Hand Gesture Recognition System Using Depth Camera

The Hand Gesture Recognition System Using Depth Camera The Hand Gesture Recognition System Using Depth Camera Ahn,Yang-Keun VR/AR Research Center Korea Electronics Technology Institute Seoul, Republic of Korea e-mail: ykahn@keti.re.kr Park,Young-Choong VR/AR

More information

Image Manipulation Interface using Depth-based Hand Gesture

Image Manipulation Interface using Depth-based Hand Gesture Image Manipulation Interface using Depth-based Hand Gesture UNSEOK LEE JIRO TANAKA Vision-based tracking is popular way to track hands. However, most vision-based tracking methods can t do a clearly tracking

More information

GESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL

GESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL GESTURE RECOGNITION SOLUTION FOR PRESENTATION CONTROL Darko Martinovikj Nevena Ackovska Faculty of Computer Science and Engineering Skopje, R. Macedonia ABSTRACT Despite the fact that there are different

More information

3D-Position Estimation for Hand Gesture Interface Using a Single Camera

3D-Position Estimation for Hand Gesture Interface Using a Single Camera 3D-Position Estimation for Hand Gesture Interface Using a Single Camera Seung-Hwan Choi, Ji-Hyeong Han, and Jong-Hwan Kim Department of Electrical Engineering, KAIST, Gusung-Dong, Yusung-Gu, Daejeon, Republic

More information

Toward an Augmented Reality System for Violin Learning Support

Toward an Augmented Reality System for Violin Learning Support Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp

More information

MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device

MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device Enkhbat Davaasuren and Jiro Tanaka 1-1-1 Tennodai, Tsukuba, Ibaraki 305-8577 Japan {enkhee,jiro}@iplab.cs.tsukuba.ac.jp Abstract.

More information

Development of an Intuitive Interface for PC Mouse Operation Based on Both Arms Gesture

Development of an Intuitive Interface for PC Mouse Operation Based on Both Arms Gesture Development of an Intuitive Interface for PC Mouse Operation Based on Both Arms Gesture Nobuaki Nakazawa 1*, Toshikazu Matsui 1, Yusaku Fujii 2 1 Faculty of Science and Technology, Gunma University, 29-1

More information

Enabling Cursor Control Using on Pinch Gesture Recognition

Enabling Cursor Control Using on Pinch Gesture Recognition Enabling Cursor Control Using on Pinch Gesture Recognition Benjamin Baldus Debra Lauterbach Juan Lizarraga October 5, 2007 Abstract In this project we expect to develop a machine-user interface based on

More information

Classification for Motion Game Based on EEG Sensing

Classification for Motion Game Based on EEG Sensing Classification for Motion Game Based on EEG Sensing Ran WEI 1,3,4, Xing-Hua ZHANG 1,4, Xin DANG 2,3,4,a and Guo-Hui LI 3 1 School of Electronics and Information Engineering, Tianjin Polytechnic University,

More information

Gesture Recognition with Real World Environment using Kinect: A Review

Gesture Recognition with Real World Environment using Kinect: A Review Gesture Recognition with Real World Environment using Kinect: A Review Prakash S. Sawai 1, Prof. V. K. Shandilya 2 P.G. Student, Department of Computer Science & Engineering, Sipna COET, Amravati, Maharashtra,

More information

A Real Time Static & Dynamic Hand Gesture Recognition System

A Real Time Static & Dynamic Hand Gesture Recognition System International Journal of Engineering Inventions e-issn: 2278-7461, p-issn: 2319-6491 Volume 4, Issue 12 [Aug. 2015] PP: 93-98 A Real Time Static & Dynamic Hand Gesture Recognition System N. Subhash Chandra

More information

Wadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology

Wadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology ISSN: 2454-132X Impact factor: 4.295 (Volume 4, Issue 1) Available online at www.ijariit.com Hand Detection and Gesture Recognition in Real-Time Using Haar-Classification and Convolutional Neural Networks

More information

Face Detection System on Ada boost Algorithm Using Haar Classifiers

Face Detection System on Ada boost Algorithm Using Haar Classifiers Vol.2, Issue.6, Nov-Dec. 2012 pp-3996-4000 ISSN: 2249-6645 Face Detection System on Ada boost Algorithm Using Haar Classifiers M. Gopi Krishna, A. Srinivasulu, Prof (Dr.) T.K.Basak 1, 2 Department of Electronics

More information

CSE Tue 10/09. Nadir Weibel

CSE Tue 10/09. Nadir Weibel CSE 118 - Tue 10/09 Nadir Weibel Today Admin Teams Assignments, grading, submissions Mini Quiz on Week 1 (readings and class material) Low-Fidelity Prototyping 1st Project Assignment Computer Vision, Kinect,

More information

Implementing RoshamboGame System with Adaptive Skin Color Model

Implementing RoshamboGame System with Adaptive Skin Color Model American Journal of Engineering Research (AJER) e-issn: 2320-0847 p-issn : 2320-0936 Volume-6, Issue-12, pp-45-53 www.ajer.org Research Paper Open Access Implementing RoshamboGame System with Adaptive

More information

R (2) Controlling System Application with hands by identifying movements through Camera

R (2) Controlling System Application with hands by identifying movements through Camera R (2) N (5) Oral (3) Total (10) Dated Sign Assignment Group: C Problem Definition: Controlling System Application with hands by identifying movements through Camera Prerequisite: 1. Web Cam Connectivity

More information

Combined Approach for Face Detection, Eye Region Detection and Eye State Analysis- Extended Paper

Combined Approach for Face Detection, Eye Region Detection and Eye State Analysis- Extended Paper International Journal of Engineering Research and Development e-issn: 2278-067X, p-issn: 2278-800X, www.ijerd.com Volume 10, Issue 9 (September 2014), PP.57-68 Combined Approach for Face Detection, Eye

More information

Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses

Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses Jinki Jung Jinwoo Jeon Hyeopwoo Lee jk@paradise.kaist.ac.kr zkrkwlek@paradise.kaist.ac.kr leehyeopwoo@paradise.kaist.ac.kr Kichan Kwon

More information

Controlling Humanoid Robot Using Head Movements

Controlling Humanoid Robot Using Head Movements Volume-5, Issue-2, April-2015 International Journal of Engineering and Management Research Page Number: 648-652 Controlling Humanoid Robot Using Head Movements S. Mounica 1, A. Naga bhavani 2, Namani.Niharika

More information

Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image

Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image Guided Filtering Using Reflected IR Image for Improving Quality of Depth Image Takahiro Hasegawa, Ryoji Tomizawa, Yuji Yamauchi, Takayoshi Yamashita and Hironobu Fujiyoshi Chubu University, 1200, Matsumoto-cho,

More information

Hand Gesture Recognition System Using Camera

Hand Gesture Recognition System Using Camera Hand Gesture Recognition System Using Camera Viraj Shinde, Tushar Bacchav, Jitendra Pawar, Mangesh Sanap B.E computer engineering,navsahyadri Education Society sgroup of Institutions,pune. Abstract - In

More information

Enhanced Method for Face Detection Based on Feature Color

Enhanced Method for Face Detection Based on Feature Color Journal of Image and Graphics, Vol. 4, No. 1, June 2016 Enhanced Method for Face Detection Based on Feature Color Nobuaki Nakazawa1, Motohiro Kano2, and Toshikazu Matsui1 1 Graduate School of Science and

More information

II. LITERATURE SURVEY

II. LITERATURE SURVEY Hand Gesture Recognition Using Operating System Mr. Anap Avinash 1 Bhalerao Sushmita 2, Lambrud Aishwarya 3, Shelke Priyanka 4, Nirmal Mohini 5 12345 Computer Department, P.Dr.V.V.P. Polytechnic, Loni

More information

Design and Implementation of an Intuitive Gesture Recognition System Using a Hand-held Device

Design and Implementation of an Intuitive Gesture Recognition System Using a Hand-held Device Design and Implementation of an Intuitive Gesture Recognition System Using a Hand-held Device Hung-Chi Chu 1, Yuan-Chin Cheng 1 1 Department of Information and Communication Engineering, Chaoyang University

More information

Robust Hand Gesture Recognition for Robotic Hand Control

Robust Hand Gesture Recognition for Robotic Hand Control Robust Hand Gesture Recognition for Robotic Hand Control Ankit Chaudhary Robust Hand Gesture Recognition for Robotic Hand Control 123 Ankit Chaudhary Department of Computer Science Northwest Missouri State

More information

Real-time AR Edutainment System Using Sensor Based Motion Recognition

Real-time AR Edutainment System Using Sensor Based Motion Recognition , pp. 271-278 http://dx.doi.org/10.14257/ijseia.2016.10.1.26 Real-time AR Edutainment System Using Sensor Based Motion Recognition Sungdae Hong 1, Hyunyi Jung 2 and Sanghyun Seo 3,* 1 Dept. of Film and

More information

Artificial Beacons with RGB-D Environment Mapping for Indoor Mobile Robot Localization

Artificial Beacons with RGB-D Environment Mapping for Indoor Mobile Robot Localization Sensors and Materials, Vol. 28, No. 6 (2016) 695 705 MYU Tokyo 695 S & M 1227 Artificial Beacons with RGB-D Environment Mapping for Indoor Mobile Robot Localization Chun-Chi Lai and Kuo-Lan Su * Department

More information

Development a File Transfer Application by Handover for 3D Video Communication System in Synchronized AR Space

Development a File Transfer Application by Handover for 3D Video Communication System in Synchronized AR Space Development a File Transfer Application by Handover for 3D Video Communication System in Synchronized AR Space Yuki Fujibayashi and Hiroki Imamura Department of Information Systems Science, Graduate School

More information

INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction

INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction Xavier Suau 1,MarcelAlcoverro 2, Adolfo Lopez-Mendez 3, Javier Ruiz-Hidalgo 2,andJosepCasas 3 1 Universitat Politécnica

More information

Robust Hand Gesture Recognition by Using Depth Data and Skeletal Information with Kinect V2 Sensor

Robust Hand Gesture Recognition by Using Depth Data and Skeletal Information with Kinect V2 Sensor Robust Hand Gesture Recognition by Using Depth Data and Skeletal Information with Kinect V Sensor S.Chandrasekhar #, N.N.Mhala * # Ph.D. Scholar, Bapurao Deshmukh College of Engineering Seagram Wardha,

More information

Near Infrared Face Image Quality Assessment System of Video Sequences

Near Infrared Face Image Quality Assessment System of Video Sequences 2011 Sixth International Conference on Image and Graphics Near Infrared Face Image Quality Assessment System of Video Sequences Jianfeng Long College of Electrical and Information Engineering Hunan University

More information

Hand Gesture Recognition System for Daily Information Retrieval Swapnil V.Ghorpade 1, Sagar A.Patil 2,Amol B.Gore 3, Govind A.

Hand Gesture Recognition System for Daily Information Retrieval Swapnil V.Ghorpade 1, Sagar A.Patil 2,Amol B.Gore 3, Govind A. Hand Gesture Recognition System for Daily Information Retrieval Swapnil V.Ghorpade 1, Sagar A.Patil 2,Amol B.Gore 3, Govind A.Pawar 4 Student, Dept. of Computer Engineering, SCS College of Engineering,

More information

A Driver Assaulting Event Detection Using Intel Real-Sense Camera

A Driver Assaulting Event Detection Using Intel Real-Sense Camera , pp.285-294 http//dx.doi.org/10.14257/ijca.2017.10.2.23 A Driver Assaulting Event Detection Using Intel Real-Sense Camera Jae-Gon Yoo 1, Dong-Kyun Kim 2, Seung Joo Choi 3, Handong Lee 4 and Jong-Bae Kim

More information

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison e-issn 2455 1392 Volume 2 Issue 10, October 2016 pp. 34 41 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com Design a Model and Algorithm for multi Way Gesture Recognition using Motion and

More information

A SURVEY ON HAND GESTURE RECOGNITION

A SURVEY ON HAND GESTURE RECOGNITION A SURVEY ON HAND GESTURE RECOGNITION U.K. Jaliya 1, Dr. Darshak Thakore 2, Deepali Kawdiya 3 1 Assistant Professor, Department of Computer Engineering, B.V.M, Gujarat, India 2 Assistant Professor, Department

More information

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces Huidong Bai The HIT Lab NZ, University of Canterbury, Christchurch, 8041 New Zealand huidong.bai@pg.canterbury.ac.nz Lei

More information

Sensory Fusion for Image

Sensory Fusion for Image , pp.34-38 http://dx.doi.org/10.14257/astl.2014.45.07 Sensory Fusion for Image Sungjun Park, Wansik Yun, and Gwanggil Jeon 1 Department of Embedded Systems Engineering, Incheon National University, 119

More information

SCIENCE & TECHNOLOGY

SCIENCE & TECHNOLOGY Pertanika J. Sci. & Technol. 25 (S): 163-172 (2017) SCIENCE & TECHNOLOGY Journal homepage: http://www.pertanika.upm.edu.my/ Performance Comparison of Min-Max Normalisation on Frontal Face Detection Using

More information

Improvement of Accuracy in Remote Gaze Detection for User Wearing Eyeglasses Using Relative Position Between Centers of Pupil and Corneal Sphere

Improvement of Accuracy in Remote Gaze Detection for User Wearing Eyeglasses Using Relative Position Between Centers of Pupil and Corneal Sphere Improvement of Accuracy in Remote Gaze Detection for User Wearing Eyeglasses Using Relative Position Between Centers of Pupil and Corneal Sphere Kiyotaka Fukumoto (&), Takumi Tsuzuki, and Yoshinobu Ebisawa

More information

Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies

Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com A Survey

More information

License Plate Localisation based on Morphological Operations

License Plate Localisation based on Morphological Operations License Plate Localisation based on Morphological Operations Xiaojun Zhai, Faycal Benssali and Soodamani Ramalingam School of Engineering & Technology University of Hertfordshire, UH Hatfield, UK Abstract

More information

Virtual Touch Human Computer Interaction at a Distance

Virtual Touch Human Computer Interaction at a Distance International Journal of Computer Science and Telecommunications [Volume 4, Issue 5, May 2013] 18 ISSN 2047-3338 Virtual Touch Human Computer Interaction at a Distance Prasanna Dhisale, Puja Firodiya,

More information

Face Detector using Network-based Services for a Remote Robot Application

Face Detector using Network-based Services for a Remote Robot Application Face Detector using Network-based Services for a Remote Robot Application Yong-Ho Seo Department of Intelligent Robot Engineering, Mokwon University Mokwon Gil 21, Seo-gu, Daejeon, Republic of Korea yhseo@mokwon.ac.kr

More information

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Hrvoje Benko Microsoft Research One Microsoft Way Redmond, WA 98052 USA benko@microsoft.com Andrew D. Wilson Microsoft

More information

SIMULATION-BASED MODEL CONTROL USING STATIC HAND GESTURES IN MATLAB

SIMULATION-BASED MODEL CONTROL USING STATIC HAND GESTURES IN MATLAB SIMULATION-BASED MODEL CONTROL USING STATIC HAND GESTURES IN MATLAB S. Kajan, J. Goga Institute of Robotics and Cybernetics, Faculty of Electrical Engineering and Information Technology, Slovak University

More information

Face Detection: A Literature Review

Face Detection: A Literature Review Face Detection: A Literature Review Dr.Vipulsangram.K.Kadam 1, Deepali G. Ganakwar 2 Professor, Department of Electronics Engineering, P.E.S. College of Engineering, Nagsenvana Aurangabad, Maharashtra,

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY Ashwini Parate,, 2013; Volume 1(8): 754-761 INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK ROBOT AND HOME APPLIANCES CONTROL USING

More information

Immersive Real Acting Space with Gesture Tracking Sensors

Immersive Real Acting Space with Gesture Tracking Sensors , pp.1-6 http://dx.doi.org/10.14257/astl.2013.39.01 Immersive Real Acting Space with Gesture Tracking Sensors Yoon-Seok Choi 1, Soonchul Jung 2, Jin-Sung Choi 3, Bon-Ki Koo 4 and Won-Hyung Lee 1* 1,2,3,4

More information

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2 CSE 165: 3D User Interaction Lecture #7: Input Devices Part 2 2 Announcements Homework Assignment #2 Due tomorrow at 2pm Sony Move check out Homework discussion Monday at 6pm Input Devices CSE 165 -Winter

More information

Study on Hand Gesture Recognition

Study on Hand Gesture Recognition Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 1, January 2015,

More information

Prediction and Correction Algorithm for a Gesture Controlled Robotic Arm

Prediction and Correction Algorithm for a Gesture Controlled Robotic Arm Prediction and Correction Algorithm for a Gesture Controlled Robotic Arm Pushkar Shukla 1, Shehjar Safaya 2, Utkarsh Sharma 3 B.Tech, College of Engineering Roorkee, Roorkee, India 1 B.Tech, College of

More information

International Journal of Informative & Futuristic Research ISSN (Online):

International Journal of Informative & Futuristic Research ISSN (Online): Reviewed Paper Volume 2 Issue 6 February 2015 International Journal of Informative & Futuristic Research An Innovative Approach Towards Virtual Drums Paper ID IJIFR/ V2/ E6/ 021 Page No. 1603-1608 Subject

More information

3D Human-Gesture Interface for Fighting Games Using Motion Recognition Sensor

3D Human-Gesture Interface for Fighting Games Using Motion Recognition Sensor Wireless Pers Commun (2016) 89:927 940 DOI 10.1007/s11277-016-3294-9 3D Human-Gesture Interface for Fighting Games Using Motion Recognition Sensor Jongmin Kim 1 Hoill Jung 2 MyungA Kang 3 Kyungyong Chung

More information

A Survey on Hand Gesture Recognition and Hand Tracking Arjunlal 1, Minu Lalitha Madhavu 2 1

A Survey on Hand Gesture Recognition and Hand Tracking Arjunlal 1, Minu Lalitha Madhavu 2 1 A Survey on Hand Gesture Recognition and Hand Tracking Arjunlal 1, Minu Lalitha Madhavu 2 1 PG scholar, Department of Computer Science And Engineering, SBCE, Alappuzha, India 2 Assistant Professor, Department

More information

Detection and Verification of Missing Components in SMD using AOI Techniques

Detection and Verification of Missing Components in SMD using AOI Techniques , pp.13-22 http://dx.doi.org/10.14257/ijcg.2016.7.2.02 Detection and Verification of Missing Components in SMD using AOI Techniques Sharat Chandra Bhardwaj Graphic Era University, India bhardwaj.sharat@gmail.com

More information

AR 2 kanoid: Augmented Reality ARkanoid

AR 2 kanoid: Augmented Reality ARkanoid AR 2 kanoid: Augmented Reality ARkanoid B. Smith and R. Gosine C-CORE and Memorial University of Newfoundland Abstract AR 2 kanoid, Augmented Reality ARkanoid, is an augmented reality version of the popular

More information

Air Marshalling with the Kinect

Air Marshalling with the Kinect Air Marshalling with the Kinect Stephen Witherden, Senior Software Developer Beca Applied Technologies stephen.witherden@beca.com Abstract. The Kinect sensor from Microsoft presents a uniquely affordable

More information

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005. Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.

More information

Pupil Detection and Tracking Based on a Round Shape Criterion by Image Processing Techniques for a Human Eye-Computer Interaction System

Pupil Detection and Tracking Based on a Round Shape Criterion by Image Processing Techniques for a Human Eye-Computer Interaction System Pupil Detection and Tracking Based on a Round Shape Criterion by Image Processing Techniques for a Human Eye-Computer Interaction System Tsumoru Ochiai and Yoshihiro Mitani Abstract The pupil detection

More information

Natural Gesture Based Interaction for Handheld Augmented Reality

Natural Gesture Based Interaction for Handheld Augmented Reality Natural Gesture Based Interaction for Handheld Augmented Reality A thesis submitted in partial fulfilment of the requirements for the Degree of Master of Science in Computer Science By Lei Gao Supervisors:

More information

Fingertip Detection: A Fast Method with Natural Hand

Fingertip Detection: A Fast Method with Natural Hand Fingertip Detection: A Fast Method with Natural Hand Jagdish Lal Raheja Machine Vision Lab Digital Systems Group, CEERI/CSIR Pilani, INDIA jagdish@ceeri.ernet.in Karen Das Dept. of Electronics & Comm.

More information

Community Update and Next Steps

Community Update and Next Steps Community Update and Next Steps Stewart Tansley, PhD Senior Research Program Manager & Product Manager (acting) Special Guest: Anoop Gupta, PhD Distinguished Scientist Project Natal Origins: Project Natal

More information

Design of an Interactive Smart Board Using Kinect Sensor

Design of an Interactive Smart Board Using Kinect Sensor Design of an Interactive Smart Board Using Kinect Sensor Supervisor: Dr. Jia Uddin Nasrul Karim Sarker - 13201025 Muhammad Touhidul Islam - 13201021 Md. Shahidul Islam Majumder - 13201022 Department of

More information

International Conference on Advances in Mechanical Engineering and Industrial Informatics (AMEII 2015)

International Conference on Advances in Mechanical Engineering and Industrial Informatics (AMEII 2015) International Conference on Advances in Mechanical Engineering and Industrial Informatics (AMEII 2015) Equipment body feeling maintenance teaching system Research Based on Kinect Fushuan Wu 1, a, Jianren

More information

Mobile Colored Overlays for People with Visual Stress

Mobile Colored Overlays for People with Visual Stress , pp.25-30 http://dx.doi.org/10.14257/ijmue.2014.9.6.04 Mobile Colored Overlays for People with Visual Stress Young Gun Jang Dept. of Computer and Information Engr. Chongju University, Korea ygjang@cju.ac.kr

More information

Research on Hand Gesture Recognition Using Convolutional Neural Network

Research on Hand Gesture Recognition Using Convolutional Neural Network Research on Hand Gesture Recognition Using Convolutional Neural Network Tian Zhaoyang a, Cheng Lee Lung b a Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China E-mail address:

More information

Hand & Upper Body Based Hybrid Gesture Recognition

Hand & Upper Body Based Hybrid Gesture Recognition Hand & Upper Body Based Hybrid Gesture Prerna Sharma #1, Naman Sharma *2 # Research Scholor, G. B. P. U. A. & T. Pantnagar, India * Ideal Institue of Technology, Ghaziabad, India Abstract Communication

More information

ReVRSR: Remote Virtual Reality for Service Robots

ReVRSR: Remote Virtual Reality for Service Robots ReVRSR: Remote Virtual Reality for Service Robots Amel Hassan, Ahmed Ehab Gado, Faizan Muhammad March 17, 2018 Abstract This project aims to bring a service robot s perspective to a human user. We believe

More information

Augmented Reality using Hand Gesture Recognition System and its use in Virtual Dressing Room

Augmented Reality using Hand Gesture Recognition System and its use in Virtual Dressing Room International Journal of Innovation and Applied Studies ISSN 2028-9324 Vol. 10 No. 1 Jan. 2015, pp. 95-100 2015 Innovative Space of Scientific Research Journals http://www.ijias.issr-journals.org/ Augmented

More information

Square Pixels to Hexagonal Pixel Structure Representation Technique. Mullana, Ambala, Haryana, India. Mullana, Ambala, Haryana, India

Square Pixels to Hexagonal Pixel Structure Representation Technique. Mullana, Ambala, Haryana, India. Mullana, Ambala, Haryana, India , pp.137-144 http://dx.doi.org/10.14257/ijsip.2014.7.4.13 Square Pixels to Hexagonal Pixel Structure Representation Technique Barun kumar 1, Pooja Gupta 2 and Kuldip Pahwa 3 1 4 th Semester M.Tech, Department

More information

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment Proceedings of the International MultiConference of Engineers and Computer Scientists 2016 Vol I,, March 16-18, 2016, Hong Kong Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free

More information

3D Interaction using Hand Motion Tracking. Srinath Sridhar Antti Oulasvirta

3D Interaction using Hand Motion Tracking. Srinath Sridhar Antti Oulasvirta 3D Interaction using Hand Motion Tracking Srinath Sridhar Antti Oulasvirta EIT ICT Labs Smart Spaces Summer School 05-June-2013 Speaker Srinath Sridhar PhD Student Supervised by Prof. Dr. Christian Theobalt

More information

II. EXPERIMENTAL SETUP

II. EXPERIMENTAL SETUP J. lnf. Commun. Converg. Eng. 1(3): 22-224, Sep. 212 Regular Paper Experimental Demonstration of 4 4 MIMO Wireless Visible Light Communication Using a Commercial CCD Image Sensor Sung-Man Kim * and Jong-Bae

More information

SLIC based Hand Gesture Recognition with Artificial Neural Network

SLIC based Hand Gesture Recognition with Artificial Neural Network IJSTE - International Journal of Science Technology & Engineering Volume 3 Issue 03 September 2016 ISSN (online): 2349-784X SLIC based Hand Gesture Recognition with Artificial Neural Network Harpreet Kaur

More information

Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture

Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture Akira Suganuma Depertment of Intelligent Systems, Kyushu University, 6 1, Kasuga-koen, Kasuga,

More information

Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL

Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL Yap Hwa Jentl, Zahari Taha 2, Eng Tat Hong", Chew Jouh Yeong" Centre for Product Design and Manufacturing (CPDM).

More information

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY A SURVEY ON GESTURE RECOGNITION TECHNOLOGY Deeba Kazim 1, Mohd Faisal 2 1 MCA Student, Integral University, Lucknow (India) 2 Assistant Professor, Integral University, Lucknow (india) ABSTRACT Gesture

More information

Implementation of Augmented Reality System for Smartphone Advertisements

Implementation of Augmented Reality System for Smartphone Advertisements , pp.385-392 http://dx.doi.org/10.14257/ijmue.2014.9.2.39 Implementation of Augmented Reality System for Smartphone Advertisements Young-geun Kim and Won-jung Kim Department of Computer Science Sunchon

More information

Real Time Hand Gesture Tracking for Network Centric Application

Real Time Hand Gesture Tracking for Network Centric Application Real Time Hand Gesture Tracking for Network Centric Application Abstract Chukwuemeka Chijioke Obasi 1 *, Christiana Chikodi Okezie 2, Ken Akpado 2, Chukwu Nnaemeka Paul 3, Asogwa, Chukwudi Samuel 1, Akuma

More information

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3

More information

An Overview of Hand Gestures Recognition System Techniques

An Overview of Hand Gestures Recognition System Techniques IOP Conference Series: Materials Science and Engineering PAPER OPEN ACCESS An Overview of Hand Gestures Recognition System Techniques To cite this article: Farah Farhana Mod Ma'asum et al 2015 IOP Conf.

More information

Research of an Algorithm on Face Detection

Research of an Algorithm on Face Detection , pp.217-222 http://dx.doi.org/10.14257/astl.2016.141.47 Research of an Algorithm on Face Detection Gong Liheng, Yang Jingjing, Zhang Xiao School of Information Science and Engineering, Hebei North University,

More information

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi Department of E&TC Engineering,PVPIT,Bavdhan,Pune ABSTRACT: In the last decades vehicle license plate recognition systems

More information

OMR Auto Grading System

OMR Auto Grading System OMR Auto Grading System Nithin T. nithint_11484@aitpune.edu.in Md Nasim mdnasim_11720@aitpune.edu.in T. Raj Shekhar t.rajshekhar_11684@aitpune.edu.in Omendra Singh Gautam omendrsinghgautam_11667@aitpune.edu.in

More information

Automatic Licenses Plate Recognition System

Automatic Licenses Plate Recognition System Automatic Licenses Plate Recognition System Garima R. Yadav Dept. of Electronics & Comm. Engineering Marathwada Institute of Technology, Aurangabad (Maharashtra), India yadavgarima08@gmail.com Prof. H.K.

More information

Hand gesture recognition based on concentric circular scan lines and weighted K-nearest neighbor algorithm

Hand gesture recognition based on concentric circular scan lines and weighted K-nearest neighbor algorithm DOI 10.1007/s11042-016-4265-6 Hand gesture recognition based on concentric circular scan lines and weighted K-nearest neighbor algorithm Yanqiu Liu 1 Xiuhui Wang 1 Ke Yan 1 Received: 24 February 2016 /

More information

Gesticulation Based Smart Surface with Enhanced Biometric Security Using Raspberry Pi

Gesticulation Based Smart Surface with Enhanced Biometric Security Using Raspberry Pi www.ijcsi.org https://doi.org/10.20943/01201705.5660 56 Gesticulation Based Smart Surface with Enhanced Biometric Security Using Raspberry Pi R.Gayathri 1, E.Roshith 2, B.Sanjana 2, S. Sanjeev Kumar 2,

More information

Recognizing Gestures on Projected Button Widgets with an RGB-D Camera Using a CNN

Recognizing Gestures on Projected Button Widgets with an RGB-D Camera Using a CNN Recognizing Gestures on Projected Button Widgets with an RGB-D Camera Using a CNN Patrick Chiu FX Palo Alto Laboratory Palo Alto, CA 94304, USA chiu@fxpal.com Chelhwon Kim FX Palo Alto Laboratory Palo

More information

What was the first gestural interface?

What was the first gestural interface? stanford hci group / cs247 Human-Computer Interaction Design Studio What was the first gestural interface? 15 January 2013 http://cs247.stanford.edu Theremin Myron Krueger 1 Myron Krueger There were things

More information

Intelligent Traffic Sign Detector: Adaptive Learning Based on Online Gathering of Training Samples

Intelligent Traffic Sign Detector: Adaptive Learning Based on Online Gathering of Training Samples 2011 IEEE Intelligent Vehicles Symposium (IV) Baden-Baden, Germany, June 5-9, 2011 Intelligent Traffic Sign Detector: Adaptive Learning Based on Online Gathering of Training Samples Daisuke Deguchi, Mitsunori

More information

ifinger Study of Gesture Recognition Technologies & Its Applications Volume II of II

ifinger Study of Gesture Recognition Technologies & Its Applications Volume II of II University of Macau Faculty of Science and Technology ifinger Study of Gesture Recognition Technologies & Its Applications Volume II of II by Chi Ian, Choi, Student No: DB02828 Final Project Report submitted

More information

Estimation of Folding Operations Using Silhouette Model

Estimation of Folding Operations Using Silhouette Model Estimation of Folding Operations Using Silhouette Model Yasuhiro Kinoshita Toyohide Watanabe Abstract In order to recognize the state of origami, there are only techniques which use special devices or

More information

A Vehicle Speed Measurement System for Nighttime with Camera

A Vehicle Speed Measurement System for Nighttime with Camera Proceedings of the 2nd International Conference on Industrial Application Engineering 2014 A Vehicle Speed Measurement System for Nighttime with Camera Yuji Goda a,*, Lifeng Zhang a,#, Seiichi Serikawa

More information

Face Recognition System Based on Infrared Image

Face Recognition System Based on Infrared Image International Journal of Engineering Inventions e-issn: 2278-7461, p-issn: 2319-6491 Volume 6, Issue 1 [October. 217] PP: 47-56 Face Recognition System Based on Infrared Image Yong Tang School of Electronics

More information

Non-Contact Gesture Recognition Using the Electric Field Disturbance for Smart Device Application

Non-Contact Gesture Recognition Using the Electric Field Disturbance for Smart Device Application , pp.133-140 http://dx.doi.org/10.14257/ijmue.2014.9.2.13 Non-Contact Gesture Recognition Using the Electric Field Disturbance for Smart Device Application Young-Chul Kim and Chang-Hyub Moon Dept. Electronics

More information

Implementation of Real Time Hand Gesture Recognition

Implementation of Real Time Hand Gesture Recognition Implementation of Real Time Hand Gesture Recognition Manasa Srinivasa H S, Suresha H S M.Tech Student, Department of ECE, Don Bosco Institute of Technology, Bangalore, Karnataka, India Associate Professor,

More information

Homeostasis Lighting Control System Using a Sensor Agent Robot

Homeostasis Lighting Control System Using a Sensor Agent Robot Intelligent Control and Automation, 2013, 4, 138-153 http://dx.doi.org/10.4236/ica.2013.42019 Published Online May 2013 (http://www.scirp.org/journal/ica) Homeostasis Lighting Control System Using a Sensor

More information

Content Based Image Retrieval Using Color Histogram

Content Based Image Retrieval Using Color Histogram Content Based Image Retrieval Using Color Histogram Nitin Jain Assistant Professor, Lokmanya Tilak College of Engineering, Navi Mumbai, India. Dr. S. S. Salankar Professor, G.H. Raisoni College of Engineering,

More information