International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN

Size: px
Start display at page:

Download "International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN"

Transcription

1 International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS ABSTRACT Saniya Zahoor Srinagar, J&K, India There is a huge diversity in the nature and ways of gestures. Making it almost impossible to formulate and implement a specific gesture recognition algorithm. A good number of attempts are being made to make such applications or devices which act as an aid to the mute people and help them to communicate with the people. One of such device is the electronic gloves which analyze the posture of the human hand in it and sounds a specific message. But it limits the gesture domain to a specific limit and leaves the user with a few of messages. Even some of Android apps have been developed in the market which uses the text to speech converter to sound a specific message. This paper discusses sixth sense and augmented reality; further, it proposes and implements a model based on augmented reality to help mute people by capturing and interpreting gestures and producing sound for mute people. 1. INTRODUCTION Sixth Sense is a gestural interface device developed by Pranav Mistry who developed both the hardware and software [1]. It comprises of a neck-worn pendant that contains both a data projector and camera. Head-worn versions were also built at MIT Media Lab in 1997 that combined cameras and illumination systems for interactive photographic art, and also included gesture recognition. Sixth Sense is a name for extra information supplied by a wearable computer, such as the device called "WuW" (Wear your World). The Sixth Sense technology contains a pocket projector, a mirror and a camera contained in a head-mounted, handheld or pendant-like, wearable device. Both the projector and the camera are connected to a mobile computing device in the user s pocket. The projector projects visual information Saniya Zahoor 372

2 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS enabling surfaces, walls and physical objects around us to be used as interfaces; while the camera recognizes and tracks users' hand gestures and physical objects using computer-vision based techniques. The software program processes the video stream data captured by the camera and tracks the locations of the colored markers at the tips of the user s fingers. The movements and arrangements of these fiducially are interpreted into gestures that act as interaction instructions for the projected application interfaces. Sixth Sense supports multitouch and multi-user interaction [2]. During a 2009 TED talk given by Pranav Mistry and his advisor Professor Pattie Maes, she showed a video demonstrating a number of applications of the Sixth Sense system. In 2010, the inventor Pranav also showed live demos. Four colored cursors are controlled by four fingers wearing different colored markers in real time. The projector displays video feedback to the user on a vertical wall. The projector displaying a map on the wall, and the user controlling it using zoom and pan gestures. The user can make a framing gesture to instruct the camera take a picture. It is hinted that the photo will be automatically cropped to remove the user's hands. Similarly augmented reality (AR) is a live direct or indirect view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data [3]. It is related to a more general concept called mediated reality, in which a view of reality is modified (possibly even diminished rather than augmented) by a computer. As a result, the technology functions by enhancing one s current perception of reality. By contrast, virtual reality replaces the real world with a simulated one. Augmentation is conventionally in real-time and in semantic context with environmental elements, such as sports scores on TV during a match. With the help of advanced AR technology, the information about the surrounding real world of the user becomes interactive and digitally manipulable. Artificial information about the environment and its objects can be overlaid on the real world. 2. SIXTH SENSE TECHNOLOGY AND AUGMENTED REALITY 'Sixth Sense' is a wearable gestural interface that augments the physical world around us with digital information and lets us use natural hand gestures to interact with that information. We've evolved over millions of years to sense the world around us. When we encounter something, someone or some place, we use our five natural senses to perceive information about it; that information helps us make decisions and chose the right actions to take. But arguably the most useful information that can help us make the right decision is not naturally perceivable with our five senses, namely the data, information and knowledge that mankind has accumulated about everything and which is increasingly all available online. Although the miniaturization of computing devices allows us to carry computers in our pockets, keeping us continually connected to the digital world, there is no link between our digital devices and our interactions with the physical world. Information is confined traditionally on paper or digitally on a screen. Sixth Sense bridges this gap, bringing intangible, digital information out into the tangible world, and allowing us to interact with this information via natural hand gestures. Sixth Sense frees information from its confines by seamlessly integrating it with reality, and thus making the entire world your computer. Saniya Zahoor 373

3 International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN The Sixth Sense prototype is comprised of a pocket projector, a mirror, and a camera. The hardware components are coupled in a pendant like mobile wearable device. Both the projector and the camera are connected to the mobile computing device in the user s pocket. The projector projects visual information enabling surfaces, walls and physical objects around us to be used as interfaces; while the camera recognizes and tracks user's hand gestures and physical objects using computer-vision based techniques. The software program processes the video stream data captured by the camera and tracks the locations of the colored markers at the tip of the user s fingers using simple computer-vision techniques. The movements and arrangements of these fiducially are interpreted into gestures that act as interaction instructions for the projected application interfaces. The maximum number of tracked fingers is only constrained by the number of unique fiducially, thus Sixth Sense also supports multi-touch and multi-user interaction. The Sixth Sense prototype implements several applications that demonstrate the usefulness, viability, and flexibility of the system. The map application lets the user navigate a map displayed on a nearby surface using hand gestures, similar to gestures supported by Multi- Touch based systems, letting the user zoom in, zoom out or pan using intuitive hand movements. The drawing application lets the user draw on any surface by tracking the fingertip movements of the user s index finger. Sixth Sense also recognizes user s freehand gestures (postures). For example, the Sixth Sense system implements a gestural camera that takes photos of the scene the user is looking at by detecting the framing gesture. The user can stop by any surface or wall and flick through the photos he/she has taken. Sixth Sense also lets the user draw icons or symbols in the air using the movement of the index finger and recognizes those symbols as interaction instructions. For example, drawing a magnifying glass symbol takes the user to the map application or drawing symbol lets the user check his mail. The Sixth Sense system also augments physical objects the user is interacting with by projecting more information about these objects projected on them. For example, a newspaper can show live video news or dynamic information can be provided on a regular piece of paper. The gesture of drawing a circle on the user s wrist projects an analog watch. Augmented reality (AR) is a live direct or indirect view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data. It is related to a more general concept called mediated reality, in which a view of reality is modified (possibly even diminished rather than augmented) by a computer. As a result, the technology functions by enhancing one s current perception of reality.by contrast, virtual reality replaces the real world with a simulated one. Augmentation is conventionally in real-time and in semantic context with environmental elements, such as sports scores on TV during a match. With the help of advanced AR technology (e.g. adding computer vision and object recognition) the information about the surrounding real world of the user becomes interactive and digitally manipulable. Artificial information about the environment and its objects can be overlaid on the real world [3]. In the world of computer graphics and augmented reality, the idea of dimensionality is not as clear-cut as one might expect. The physical world is three dimensional. Virtual worlds can exist in one, two, three, or more dimensions. However, that world can also be displayed in different dimensions. Stepping back from augmented reality for a moment, consider a movie. The world that the movie is created in is the physical world which is three dimensional. Saniya Zahoor 374

4 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS However, movies are typically displayed on a two-dimensional movie screen. Thus, the virtual world of the movie is three dimensional, but the display of that world is two dimensional. How then do we perceive the world of the movie as being somewhat three dimensional? There are a variety of depth cues that we as humans use to discern the threedimensionality of the physical (and also virtual) world. Indeed, each of our eyes has a twodimensional retina on which images of the three-dimensional world are sensed. 3. ISSUES The product is not yet released into the market. There are no legal obligations as of now because there are some modifications still being made to the product itself. However, there are some security concerns about this product. New inventions in information technology have some kind of security concerns. For example, when a person is taking a picture using the Sixth Sense, he just snaps with his fingers which have little marker caps on them. People don t want to be captured on a camera of a random person [4]. One can never tell when Sixth Sense is taking a picture because it is not a big object and nothing is being helped in your hand in order to take a picture of a camera or a phone. This is invading other s privacy which can be a huge problem [5]. Some of the health issues are regarding Sixth Sense s projection technology. When the device is projecting on a hard surface, it is not private enough for just the user [5]. People around him can see the projection since it is very detailed. Projection is better in the night time and dark areas rather than mornings and bright areas. This is an issue because the vision of the user can be damaged when using this instrument. Sixth Sense should be able to shift its projection techniques during different times of the day. That way it won t be an issue for the vision of the user. Since the device is still being modified and tested, Mistry can try to overcome issues with projection. Concerns about the pricing of this device are also rising among the people. Mistry announced that the present device is costing about $350 for each, on his website (Mistry, 2010). This information was updated in 2009, but ever since then, there is not news about this device. Mistry is working on many other technologies and inventions, but the world doesn t know whether the work on this device is stopped (Doherty, 2009). People and the manufactures who are ready to have this product out in the market do not know when this will be out. Pranav Mistry said, This prototype needs some serious engineering and programming. Everyone is waiting hoping that this will be available in the market once all the modifications are done (Boyd, 2009) [6]. 4. PROPOSED WORK The gesture done by the user is simply tracked in real time and converted to the massage the mute user wants to convey. The domain of such functions are infinite (X-axis is infinite) The range is also the same infinite thus any massage can be sounded (Y-axis is infinite) Saniya Zahoor 375

5 International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN We can use some background on the image capturing window to enhance the interaction process and to make it user-friendly There is the user flexibility in this system and it is natural as the hand gestures are involved. Figure 5.1 Screen short of interactive background Finally, we implemented this method to track the user s gestures. As the user points to any of the coordinates on screen the program will be detecting the specific color of the finger and in the backside, all coordinates are associated to a specific massage, which the mute user wants to convey. Finally, the required massage is sounded which depicts our entire work done and fulfills our purpose. 5. IMPLEMENTATION DETAILS AND RESULTS Implementation has been done in Matlab; the logic of the program is based on marker position recognition and corresponding voice speech production and it has following six steps of processing. Image acquisition: Image is acquired using Image acquisition toolbox in Matlab. Image Acquisition Toolbox is used to acquire images and video from cameras and frame grabbers directly into Matlab We can detect hardware automatically and configure hardware properties using Image Acquiring Hardware Info function. We use Winvideo (Windows video) as an adapter with video constructor 1, 2 or 3 depending on the camera we want to choose. For external webcam, we use 3 as video constructor. A frame rate can be set out using Vid.FrameGrabinterval function. Finally, we can make connected web cam to capture video using preview(vid) function. Saniya Zahoor 376

6 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS Figure 1 - Image Acquisition function output Figure 2 - Video Acquisition code in Matlab Segmentation: Segmentation is used for identifying components of an image. It is the process of dividing an image into multiple parts. This is typically used to identify objects or other relevant information in digital images. There are many different ways to perform image segmentation like Thresholding, Color-based Segmentation, Transform methods, Texture methods. Saniya Zahoor 377

7 International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN Figure 3 - Matlab code for Implementing Bounding box Propriety In our project, we have used color-based segmentation in which the Bounding box returns the smallest rectangle containing the region of a Red color in the captured image frames. Figure 1 Bounding box returning the rectangle with the minimum area containing Red color Background subtraction: Background subtraction, also known as Foreground Detection, is a technique in the fields of image processing and computer vision wherein an image s foreground is extracted for further processing (object recognition etc.). Saniya Zahoor 378

8 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS Figure 5 - Matlab code for background subtraction In our project, we used background subtraction to extract the region of interest i.e the Redcolored marker from image frames. Figure 6 - Result of above Matlab code Background subtraction is a widely used approach for detecting moving objects in videos from static cameras. The rationale in the approach is that of detecting the moving objects from the difference between the current frame and a reference frame, often called background image, or background model Feature Extraction: Feature extraction a type of dimensionality reduction that efficiently represents interesting parts of an image. Common feature extraction techniques in Matlab include Histogram of Oriented Gradients (HOG), color extraction, Speeded Up Robust Features (SURF), Local Binary Patterns (LBP), Haar wavelets, and color histograms. We use Red color detection and extraction as a part of feature extraction to detect the motion of our marker. Saniya Zahoor 379

9 International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN Tracking: The Tracker algorithm identifies the marker contained in a bounding box. The bounding box returns the particular coordinates of the centroid of the rectangle. This is the rectangle with the minimum area but enclosing all the region of interest ( red color in our case). Voice output corresponding to tracking: Based on coordinates of centroid returned, the particular voice output is produced. This is done by using the Sound function of Matlab to output the voice corresponding to the location tracked. Figure 7 - Matlab code for the production of voice output corresponding to the marker location We are to implement our work of recognizing and interpreting gesture for mute people in a real time. We have already designed the algorithm to associate all the coordinates of twodimensional Euclidean plane to specific messages.now as the mute person points to any known coordinate the required message is sounded. We implemented the algorithm in Matlab where we created a video object and called it to take a video. Out of the continuous frames we took a snapshot after a counted number of frames, which provided us as a delay function. After taking the snapshot we processed the image and then retrieved the coordinates of red color. Now we have the coordinates towards which our user is pointing to. These coordinate values are then compared to the cases of a switch where each case is sounding a message. The result of the project is that our work provided an aid to mute people for capturing and interpreting gestures and producing sound for mute people. If we analyze it in black box view then we have a camera to sense the user s gestures and produce the desired sound. In addition to that, it can simply be used as gesture interface for devices. The user is now in a position to select the required choice out of menu by only pointing to it. Saniya Zahoor 380

10 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS 6. CONCLUSIONS After analyzing the way the mute people are doing conversation. We personally came up with a conclusion that the first thing they are doing to communicate is using gestures. Thus we can simply say that in the case of mute people the message they want to deliver is encrypted in the form of a gesture. But it is a natural case so we will never be able to know the algorithm behind this natural encryption and the key that can be used to decrypt it. REFERENCES 1. Carmigniani, J., Furht, B., Anisetti, M., Ceravolo, P., Damiani, E. and Ivkovic, M., Augmented reality technologies, systems, and applications. Multimedia tools and applications, 51(1), pp Mistry, P. and Maes, P., 2009, December. SixthSense: a wearable gestural interface. In ACM SIGGRAPH ASIA 2009 Sketches (p. 11). ACM. 3. Azuma, R.T., A survey of augmented reality. Presence: Teleoperators & Virtual Environments, 6(4), pp Behling, O. and Eckel, N.L., Making sense out of intuition. Academy of Management Perspectives, 5(1), pp Neumann, U. and Majoros, A., 1998, March. Cognitive, performance, and systems issues for augmented reality applications in manufacturing and maintenance. In Virtual Reality Annual International Symposium, Proceedings., IEEE 1998 (pp. 4-11). IEEE. 6. Azuma, R.T., The challenge of making augmented reality work outdoors. Mixed reality: Merging real and virtual worlds, pp Saniya Zahoor 381

Sixth Sense Technology

Sixth Sense Technology Sixth Sense Technology Hima Mohan Ad-Hoc Faculty Carmel College Mala, Abstract Sixth Sense Technology integrates digital information into the physical world and its objects, making the entire world your

More information

SIXTH SENSE TECHNOLOGY A STEP AHEAD

SIXTH SENSE TECHNOLOGY A STEP AHEAD SIXTH SENSE TECHNOLOGY A STEP AHEAD B.Srinivasa Ragavan 1, R.Sripathy 2 1 Asst. Professor in Computer Science, 2 Asst. Professor MCA, Sri SRNM College, Sattur, Tamilnadu, (India) ABSTRACT Due to technological

More information

GUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer

GUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer 2010 GUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer By: Abdullah Almurayh For : Dr. Chow UCCS CS525 Spring 2010 5/4/2010 Contents Subject Page 1. Abstract 2 2. Introduction

More information

Chapter 1 - Introduction

Chapter 1 - Introduction 1 "We all agree that your theory is crazy, but is it crazy enough?" Niels Bohr (1885-1962) Chapter 1 - Introduction Augmented reality (AR) is the registration of projected computer-generated images over

More information

MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device

MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device MOBAJES: Multi-user Gesture Interaction System with Wearable Mobile Device Enkhbat Davaasuren and Jiro Tanaka 1-1-1 Tennodai, Tsukuba, Ibaraki 305-8577 Japan {enkhee,jiro}@iplab.cs.tsukuba.ac.jp Abstract.

More information

R (2) Controlling System Application with hands by identifying movements through Camera

R (2) Controlling System Application with hands by identifying movements through Camera R (2) N (5) Oral (3) Total (10) Dated Sign Assignment Group: C Problem Definition: Controlling System Application with hands by identifying movements through Camera Prerequisite: 1. Web Cam Connectivity

More information

Interior Design using Augmented Reality Environment

Interior Design using Augmented Reality Environment Interior Design using Augmented Reality Environment Kalyani Pampattiwar 2, Akshay Adiyodi 1, Manasvini Agrahara 1, Pankaj Gamnani 1 Assistant Professor, Department of Computer Engineering, SIES Graduate

More information

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Hrvoje Benko Microsoft Research One Microsoft Way Redmond, WA 98052 USA benko@microsoft.com Andrew D. Wilson Microsoft

More information

VICs: A Modular Vision-Based HCI Framework

VICs: A Modular Vision-Based HCI Framework VICs: A Modular Vision-Based HCI Framework The Visual Interaction Cues Project Guangqi Ye, Jason Corso Darius Burschka, & Greg Hager CIRL, 1 Today, I ll be presenting work that is part of an ongoing project

More information

Advancements in Gesture Recognition Technology

Advancements in Gesture Recognition Technology IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 4, Issue 4, Ver. I (Jul-Aug. 2014), PP 01-07 e-issn: 2319 4200, p-issn No. : 2319 4197 Advancements in Gesture Recognition Technology 1 Poluka

More information

AR Tamagotchi : Animate Everything Around Us

AR Tamagotchi : Animate Everything Around Us AR Tamagotchi : Animate Everything Around Us Byung-Hwa Park i-lab, Pohang University of Science and Technology (POSTECH), Pohang, South Korea pbh0616@postech.ac.kr Se-Young Oh Dept. of Electrical Engineering,

More information

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3

More information

ROBOT VISION. Dr.M.Madhavi, MED, MVSREC

ROBOT VISION. Dr.M.Madhavi, MED, MVSREC ROBOT VISION Dr.M.Madhavi, MED, MVSREC Robotic vision may be defined as the process of acquiring and extracting information from images of 3-D world. Robotic vision is primarily targeted at manipulation

More information

International Journal of Research in Computer and Communication Technology, Vol 2, Issue 12, December- 2013

International Journal of Research in Computer and Communication Technology, Vol 2, Issue 12, December- 2013 Design Of Virtual Sense Technology For System Interface Mr. Chetan Dhule, Prof.T.H.Nagrare Computer Science & Engineering Department, G.H Raisoni College Of Engineering. ABSTRACT A gesture-based human

More information

Robust Hand Gesture Recognition for Robotic Hand Control

Robust Hand Gesture Recognition for Robotic Hand Control Robust Hand Gesture Recognition for Robotic Hand Control Ankit Chaudhary Robust Hand Gesture Recognition for Robotic Hand Control 123 Ankit Chaudhary Department of Computer Science Northwest Missouri State

More information

Future Directions for Augmented Reality. Mark Billinghurst

Future Directions for Augmented Reality. Mark Billinghurst Future Directions for Augmented Reality Mark Billinghurst 1968 Sutherland/Sproull s HMD https://www.youtube.com/watch?v=ntwzxgprxag Star Wars - 1977 Augmented Reality Combines Real and Virtual Images Both

More information

Augmented Reality And Ubiquitous Computing using HCI

Augmented Reality And Ubiquitous Computing using HCI Augmented Reality And Ubiquitous Computing using HCI Ashmit Kolli MS in Data Science Michigan Technological University CS5760 Topic Assignment 2 akolli@mtu.edu Abstract : Direct use of the hand as an input

More information

Interactive Coffee Tables: Interfacing TV within an Intuitive, Fun and Shared Experience

Interactive Coffee Tables: Interfacing TV within an Intuitive, Fun and Shared Experience Interactive Coffee Tables: Interfacing TV within an Intuitive, Fun and Shared Experience Radu-Daniel Vatavu and Stefan-Gheorghe Pentiuc University Stefan cel Mare of Suceava, Department of Computer Science,

More information

A Real Time Static & Dynamic Hand Gesture Recognition System

A Real Time Static & Dynamic Hand Gesture Recognition System International Journal of Engineering Inventions e-issn: 2278-7461, p-issn: 2319-6491 Volume 4, Issue 12 [Aug. 2015] PP: 93-98 A Real Time Static & Dynamic Hand Gesture Recognition System N. Subhash Chandra

More information

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations Viviana Chimienti 1, Salvatore Iliano 1, Michele Dassisti 2, Gino Dini 1, and Franco Failli 1 1 Dipartimento di

More information

AR 2 kanoid: Augmented Reality ARkanoid

AR 2 kanoid: Augmented Reality ARkanoid AR 2 kanoid: Augmented Reality ARkanoid B. Smith and R. Gosine C-CORE and Memorial University of Newfoundland Abstract AR 2 kanoid, Augmented Reality ARkanoid, is an augmented reality version of the popular

More information

Technology designed to empower people

Technology designed to empower people Edition July 2018 Smart Health, Wearables, Artificial intelligence Technology designed to empower people Through new interfaces - close to the body - technology can enable us to become more aware of our

More information

Beyond: collapsible tools and gestures for computational design

Beyond: collapsible tools and gestures for computational design Beyond: collapsible tools and gestures for computational design The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published

More information

Enabling Cursor Control Using on Pinch Gesture Recognition

Enabling Cursor Control Using on Pinch Gesture Recognition Enabling Cursor Control Using on Pinch Gesture Recognition Benjamin Baldus Debra Lauterbach Juan Lizarraga October 5, 2007 Abstract In this project we expect to develop a machine-user interface based on

More information

DepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface

DepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface DepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface Hrvoje Benko and Andrew D. Wilson Microsoft Research One Microsoft Way Redmond, WA 98052, USA

More information

Toward an Augmented Reality System for Violin Learning Support

Toward an Augmented Reality System for Violin Learning Support Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp

More information

DESIGN OF AN AUGMENTED REALITY

DESIGN OF AN AUGMENTED REALITY DESIGN OF AN AUGMENTED REALITY MAGNIFICATION AID FOR LOW VISION USERS Lee Stearns University of Maryland Email: lstearns@umd.edu Jon Froehlich Leah Findlater University of Washington Common reading aids

More information

Live Hand Gesture Recognition using an Android Device

Live Hand Gesture Recognition using an Android Device Live Hand Gesture Recognition using an Android Device Mr. Yogesh B. Dongare Department of Computer Engineering. G.H.Raisoni College of Engineering and Management, Ahmednagar. Email- yogesh.dongare05@gmail.com

More information

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY Volume 117 No. 22 2017, 209-213 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY Mrs.S.Hemamalini

More information

TOUCHABLE HOLOGRAMS AND HAPTIC FEEDBACK: REAL EXPERIENCE IN A VIRTUAL WORLD

TOUCHABLE HOLOGRAMS AND HAPTIC FEEDBACK: REAL EXPERIENCE IN A VIRTUAL WORLD TOUCHABLE HOLOGRAMS AND HAPTIC FEEDBACK: REAL EXPERIENCE IN A VIRTUAL WORLD 1 PRAJAKTA RATHOD, 2 SANKET MODI 1 Assistant Professor, CSE Dept, NIRMA University, Ahmedabad, Gujrat 2 Student, CSE Dept, NIRMA

More information

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces Huidong Bai The HIT Lab NZ, University of Canterbury, Christchurch, 8041 New Zealand huidong.bai@pg.canterbury.ac.nz Lei

More information

CSC 2524, Fall 2017 AR/VR Interaction Interface

CSC 2524, Fall 2017 AR/VR Interaction Interface CSC 2524, Fall 2017 AR/VR Interaction Interface Karan Singh Adapted from and with thanks to Mark Billinghurst Typical Virtual Reality System HMD User Interface Input Tracking How can we Interact in VR?

More information

A Survey of Mobile Augmentation for Mobile Augmented Reality System

A Survey of Mobile Augmentation for Mobile Augmented Reality System A Survey of Mobile Augmentation for Mobile Augmented Reality System Mr.A.T.Vasaya 1, Mr.A.S.Gohil 2 1 PG Student, C.U.Shah College of Engineering and Technology, Gujarat, India 2 Asst.Proffesor, Sir Bhavsinhji

More information

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real... v preface Motivation Augmented reality (AR) research aims to develop technologies that allow the real-time fusion of computer-generated digital content with the real world. Unlike virtual reality (VR)

More information

Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane

Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane Journal of Communication and Computer 13 (2016) 329-337 doi:10.17265/1548-7709/2016.07.002 D DAVID PUBLISHING Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane

More information

Practical Image and Video Processing Using MATLAB

Practical Image and Video Processing Using MATLAB Practical Image and Video Processing Using MATLAB Chapter 1 Introduction and overview What will we learn? What is image processing? What are the main applications of image processing? What is an image?

More information

New interface approaches for telemedicine

New interface approaches for telemedicine New interface approaches for telemedicine Associate Professor Mark Billinghurst PhD, Holger Regenbrecht Dipl.-Inf. Dr-Ing., Michael Haller PhD, Joerg Hauber MSc Correspondence to: mark.billinghurst@hitlabnz.org

More information

Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction

Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction Minghao Cai 1(B), Soh Masuko 2, and Jiro Tanaka 1 1 Waseda University, Kitakyushu, Japan mhcai@toki.waseda.jp, jiro@aoni.waseda.jp

More information

Enhanced Virtual Transparency in Handheld AR: Digital Magnifying Glass

Enhanced Virtual Transparency in Handheld AR: Digital Magnifying Glass Enhanced Virtual Transparency in Handheld AR: Digital Magnifying Glass Klen Čopič Pucihar School of Computing and Communications Lancaster University Lancaster, UK LA1 4YW k.copicpuc@lancaster.ac.uk Paul

More information

Gesture Recognition with Real World Environment using Kinect: A Review

Gesture Recognition with Real World Environment using Kinect: A Review Gesture Recognition with Real World Environment using Kinect: A Review Prakash S. Sawai 1, Prof. V. K. Shandilya 2 P.G. Student, Department of Computer Science & Engineering, Sipna COET, Amravati, Maharashtra,

More information

Projection Based HCI (Human Computer Interface) System using Image Processing

Projection Based HCI (Human Computer Interface) System using Image Processing GRD Journals- Global Research and Development Journal for Volume 1 Issue 5 April 2016 ISSN: 2455-5703 Projection Based HCI (Human Computer Interface) System using Image Processing Pankaj Dhome Sagar Dhakane

More information

Introduction to Virtual Reality (based on a talk by Bill Mark)

Introduction to Virtual Reality (based on a talk by Bill Mark) Introduction to Virtual Reality (based on a talk by Bill Mark) I will talk about... Why do we want Virtual Reality? What is needed for a VR system? Examples of VR systems Research problems in VR Most Computers

More information

AUGMENTED REALITY, FEATURE DETECTION Applications on camera phones. Prof. Charles Woodward, Digital Systems VTT TECHNICAL RESEARCH CENTRE OF FINLAND

AUGMENTED REALITY, FEATURE DETECTION Applications on camera phones. Prof. Charles Woodward, Digital Systems VTT TECHNICAL RESEARCH CENTRE OF FINLAND AUGMENTED REALITY, FEATURE DETECTION Applications on camera phones Prof. Charles Woodward, Digital Systems VTT TECHNICAL RESEARCH CENTRE OF FINLAND AUGMENTED REALITY (AR) Mixes virtual objects with view

More information

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Komal Hasija 1, Rajani Mehta 2 Abstract Recognition is a very effective area of research in regard of security with the involvement

More information

Vishnu Nath. Usage of computer vision and humanoid robotics to create autonomous robots. (Ximea Currera RL04C Camera Kit)

Vishnu Nath. Usage of computer vision and humanoid robotics to create autonomous robots. (Ximea Currera RL04C Camera Kit) Vishnu Nath Usage of computer vision and humanoid robotics to create autonomous robots (Ximea Currera RL04C Camera Kit) Acknowledgements Firstly, I would like to thank Ivan Klimkovic of Ximea Corporation,

More information

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception

More information

ThumbsUp: Integrated Command and Pointer Interactions for Mobile Outdoor Augmented Reality Systems

ThumbsUp: Integrated Command and Pointer Interactions for Mobile Outdoor Augmented Reality Systems ThumbsUp: Integrated Command and Pointer Interactions for Mobile Outdoor Augmented Reality Systems Wayne Piekarski and Bruce H. Thomas Wearable Computer Laboratory School of Computer and Information Science

More information

Application Areas of AI Artificial intelligence is divided into different branches which are mentioned below:

Application Areas of AI   Artificial intelligence is divided into different branches which are mentioned below: Week 2 - o Expert Systems o Natural Language Processing (NLP) o Computer Vision o Speech Recognition And Generation o Robotics o Neural Network o Virtual Reality APPLICATION AREAS OF ARTIFICIAL INTELLIGENCE

More information

Mohammad Akram Khan 2 India

Mohammad Akram Khan 2 India ISSN: 2321-7782 (Online) Impact Factor: 6.047 Volume 4, Issue 8, August 2016 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case

More information

Multimodal Interaction Concepts for Mobile Augmented Reality Applications

Multimodal Interaction Concepts for Mobile Augmented Reality Applications Multimodal Interaction Concepts for Mobile Augmented Reality Applications Wolfgang Hürst and Casper van Wezel Utrecht University, PO Box 80.089, 3508 TB Utrecht, The Netherlands huerst@cs.uu.nl, cawezel@students.cs.uu.nl

More information

Multi-Modal User Interaction

Multi-Modal User Interaction Multi-Modal User Interaction Lecture 4: Multiple Modalities Zheng-Hua Tan Department of Electronic Systems Aalborg University, Denmark zt@es.aau.dk MMUI, IV, Zheng-Hua Tan 1 Outline Multimodal interface

More information

Interaction Techniques using Head Mounted Displays and Handheld Devices for Outdoor Augmented Reality

Interaction Techniques using Head Mounted Displays and Handheld Devices for Outdoor Augmented Reality Interaction Techniques using Head Mounted Displays and Handheld Devices for Outdoor Augmented Reality by Rahul Budhiraja A thesis submitted in partial fulfillment of the requirements for the Degree of

More information

CSC 2524, Fall 2018 Graphics, Interaction and Perception in Augmented and Virtual Reality AR/VR

CSC 2524, Fall 2018 Graphics, Interaction and Perception in Augmented and Virtual Reality AR/VR CSC 2524, Fall 2018 Graphics, Interaction and Perception in Augmented and Virtual Reality AR/VR Karan Singh Inspired and adapted from material by Mark Billinghurst What is this course about? Fundamentals

More information

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison e-issn 2455 1392 Volume 2 Issue 10, October 2016 pp. 34 41 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com Design a Model and Algorithm for multi Way Gesture Recognition using Motion and

More information

COLOR IMAGE SEGMENTATION USING K-MEANS CLASSIFICATION ON RGB HISTOGRAM SADIA BASAR, AWAIS ADNAN, NAILA HABIB KHAN, SHAHAB HAIDER

COLOR IMAGE SEGMENTATION USING K-MEANS CLASSIFICATION ON RGB HISTOGRAM SADIA BASAR, AWAIS ADNAN, NAILA HABIB KHAN, SHAHAB HAIDER COLOR IMAGE SEGMENTATION USING K-MEANS CLASSIFICATION ON RGB HISTOGRAM SADIA BASAR, AWAIS ADNAN, NAILA HABIB KHAN, SHAHAB HAIDER Department of Computer Science, Institute of Management Sciences, 1-A, Sector

More information

The use of gestures in computer aided design

The use of gestures in computer aided design Loughborough University Institutional Repository The use of gestures in computer aided design This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation: CASE,

More information

Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture

Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture Akira Suganuma Depertment of Intelligent Systems, Kyushu University, 6 1, Kasuga-koen, Kasuga,

More information

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface Xu Zhao Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan sheldonzhaox@is.ics.saitamau.ac.jp Takehiro Niikura The University

More information

AUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING

AUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING AUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING ABSTRACT Chutisant Kerdvibulvech Department of Information and Communication Technology, Rangsit University, Thailand Email: chutisant.k@rsu.ac.th In

More information

Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience

Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience , pp.150-156 http://dx.doi.org/10.14257/astl.2016.140.29 Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience Jaeho Ryu 1, Minsuk

More information

3D and Sequential Representations of Spatial Relationships among Photos

3D and Sequential Representations of Spatial Relationships among Photos 3D and Sequential Representations of Spatial Relationships among Photos Mahoro Anabuki Canon Development Americas, Inc. E15-349, 20 Ames Street Cambridge, MA 02139 USA mahoro@media.mit.edu Hiroshi Ishii

More information

PUZZLAR, A PROTOTYPE OF AN INTEGRATED PUZZLE GAME USING MULTIPLE MARKER AUGMENTED REALITY

PUZZLAR, A PROTOTYPE OF AN INTEGRATED PUZZLE GAME USING MULTIPLE MARKER AUGMENTED REALITY PUZZLAR, A PROTOTYPE OF AN INTEGRATED PUZZLE GAME USING MULTIPLE MARKER AUGMENTED REALITY Marcella Christiana and Raymond Bahana Computer Science Program, Binus International-Binus University, Jakarta

More information

ISSN: [Arora * et al., 7(4): April, 2018] Impact Factor: 5.164

ISSN: [Arora * et al., 7(4): April, 2018] Impact Factor: 5.164 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY REAL TIME SYSTEM CONTROLLING USING A WEB CAMERA BASED ON COLOUR DETECTION Reema Arora *1, Renu Kumari 2 & Shabnam Kumari 3 *1

More information

Image Processing and Particle Analysis for Road Traffic Detection

Image Processing and Particle Analysis for Road Traffic Detection Image Processing and Particle Analysis for Road Traffic Detection ABSTRACT Aditya Kamath Manipal Institute of Technology Manipal, India This article presents a system developed using graphic programming

More information

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY A SURVEY ON GESTURE RECOGNITION TECHNOLOGY Deeba Kazim 1, Mohd Faisal 2 1 MCA Student, Integral University, Lucknow (India) 2 Assistant Professor, Integral University, Lucknow (india) ABSTRACT Gesture

More information

GLOSSARY for National Core Arts: Media Arts STANDARDS

GLOSSARY for National Core Arts: Media Arts STANDARDS GLOSSARY for National Core Arts: Media Arts STANDARDS Attention Principle of directing perception through sensory and conceptual impact Balance Principle of the equitable and/or dynamic distribution of

More information

Visual Interpretation of Hand Gestures as a Practical Interface Modality

Visual Interpretation of Hand Gestures as a Practical Interface Modality Visual Interpretation of Hand Gestures as a Practical Interface Modality Frederik C. M. Kjeldsen Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate

More information

A SURVEY ON HAND GESTURE RECOGNITION

A SURVEY ON HAND GESTURE RECOGNITION A SURVEY ON HAND GESTURE RECOGNITION U.K. Jaliya 1, Dr. Darshak Thakore 2, Deepali Kawdiya 3 1 Assistant Professor, Department of Computer Engineering, B.V.M, Gujarat, India 2 Assistant Professor, Department

More information

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations Viviana Chimienti, Salvatore Iliano, Michele Dassisti 2, Gino Dini, Franco Failli Dipartimento di Ingegneria Meccanica,

More information

Simulation of Water Inundation Using Virtual Reality Tools for Disaster Study: Opportunity and Challenges

Simulation of Water Inundation Using Virtual Reality Tools for Disaster Study: Opportunity and Challenges Simulation of Water Inundation Using Virtual Reality Tools for Disaster Study: Opportunity and Challenges Deepak Mishra Associate Professor Department of Avionics Indian Institute of Space Science and

More information

Number Plate Recognition Using Segmentation

Number Plate Recognition Using Segmentation Number Plate Recognition Using Segmentation Rupali Kate M.Tech. Electronics(VLSI) BVCOE. Pune 411043, Maharashtra, India. Dr. Chitode. J. S BVCOE. Pune 411043 Abstract Automatic Number Plate Recognition

More information

EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION

EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION EFFICIENT ATTENDANCE MANAGEMENT SYSTEM USING FACE DETECTION AND RECOGNITION 1 Arun.A.V, 2 Bhatath.S, 3 Chethan.N, 4 Manmohan.C.M, 5 Hamsaveni M 1,2,3,4,5 Department of Computer Science and Engineering,

More information

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright E90 Project Proposal 6 December 2006 Paul Azunre Thomas Murray David Wright Table of Contents Abstract 3 Introduction..4 Technical Discussion...4 Tracking Input..4 Haptic Feedack.6 Project Implementation....7

More information

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY. Augmented Reality-An Emerging Technology

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY. Augmented Reality-An Emerging Technology [Lotlikar, 2(3): March, 2013] ISSN: 2277-9655 IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY Augmented Reality-An Emerging Technology Trupti Lotlikar *1, Divya Mahajan 2, Javid

More information

The Mixed Reality Book: A New Multimedia Reading Experience

The Mixed Reality Book: A New Multimedia Reading Experience The Mixed Reality Book: A New Multimedia Reading Experience Raphaël Grasset raphael.grasset@hitlabnz.org Andreas Dünser andreas.duenser@hitlabnz.org Mark Billinghurst mark.billinghurst@hitlabnz.org Hartmut

More information

A Digital Reality Daniel Gilyana & Arielle Pineda

A Digital Reality Daniel Gilyana & Arielle Pineda A Digital Reality Daniel Gilyana & Arielle Pineda Are you really using your iphone to its full potential? A New Reality Daniel Gilyana & Arielle Pineda Augmented reality allows us to see a window to an

More information

Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses

Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses Jinki Jung Jinwoo Jeon Hyeopwoo Lee jk@paradise.kaist.ac.kr zkrkwlek@paradise.kaist.ac.kr leehyeopwoo@paradise.kaist.ac.kr Kichan Kwon

More information

Immersive Training. David Lafferty President of Scientific Technical Services And ARC Associate

Immersive Training. David Lafferty President of Scientific Technical Services And ARC Associate Immersive Training David Lafferty President of Scientific Technical Services And ARC Associate Current Situation Great Shift Change Drive The Need For Training Conventional Training Methods Are Expensive

More information

Theory and Practice of Tangible User Interfaces Tuesday, Week 9

Theory and Practice of Tangible User Interfaces Tuesday, Week 9 Augmented Reality Theory and Practice of Tangible User Interfaces Tuesday, Week 9 Outline Overview Examples Theory Examples Supporting AR Designs Examples Theory Outline Overview Examples Theory Examples

More information

PROPOSED SYSTEM FOR MID-AIR HOLOGRAPHY PROJECTION USING CONVERSION OF 2D TO 3D VISUALIZATION

PROPOSED SYSTEM FOR MID-AIR HOLOGRAPHY PROJECTION USING CONVERSION OF 2D TO 3D VISUALIZATION International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 7, Issue 2, March-April 2016, pp. 159 167, Article ID: IJARET_07_02_015 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=7&itype=2

More information

Augmented Reality Lecture notes 01 1

Augmented Reality Lecture notes 01 1 IntroductiontoAugmentedReality Lecture notes 01 1 Definition Augmented reality (AR) is a live, direct or indirect, view of a physical, real-world environment whose elements are augmented by computer-generated

More information

Abstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction

Abstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction Creating a Collaborative Multi Touch Computer Aided Design Program Cole Anagnost, Thomas Niedzielski, Desirée Velázquez, Prasad Ramanahally, Stephen Gilbert Iowa State University { someguy tomn deveri

More information

International Journal of Informative & Futuristic Research ISSN (Online):

International Journal of Informative & Futuristic Research ISSN (Online): Reviewed Paper Volume 2 Issue 6 February 2015 International Journal of Informative & Futuristic Research An Innovative Approach Towards Virtual Drums Paper ID IJIFR/ V2/ E6/ 021 Page No. 1603-1608 Subject

More information

Virtual Touch Human Computer Interaction at a Distance

Virtual Touch Human Computer Interaction at a Distance International Journal of Computer Science and Telecommunications [Volume 4, Issue 5, May 2013] 18 ISSN 2047-3338 Virtual Touch Human Computer Interaction at a Distance Prasanna Dhisale, Puja Firodiya,

More information

DESIGN STYLE FOR BUILDING INTERIOR 3D OBJECTS USING MARKER BASED AUGMENTED REALITY

DESIGN STYLE FOR BUILDING INTERIOR 3D OBJECTS USING MARKER BASED AUGMENTED REALITY DESIGN STYLE FOR BUILDING INTERIOR 3D OBJECTS USING MARKER BASED AUGMENTED REALITY 1 RAJU RATHOD, 2 GEORGE PHILIP.C, 3 VIJAY KUMAR B.P 1,2,3 MSRIT Bangalore Abstract- To ensure the best place, position,

More information

Chapter 6. [6]Preprocessing

Chapter 6. [6]Preprocessing Chapter 6 [6]Preprocessing As mentioned in chapter 4, the first stage in the HCR pipeline is preprocessing of the image. We have seen in earlier chapters why this is very important and at the same time

More information

Table of Contents 1. Image processing Measurements System Tools...10

Table of Contents 1. Image processing Measurements System Tools...10 Introduction Table of Contents 1 An Overview of ScopeImage Advanced...2 Features:...2 Function introduction...3 1. Image processing...3 1.1 Image Import and Export...3 1.1.1 Open image file...3 1.1.2 Import

More information

Various Calibration Functions for Webcams and AIBO under Linux

Various Calibration Functions for Webcams and AIBO under Linux SISY 2006 4 th Serbian-Hungarian Joint Symposium on Intelligent Systems Various Calibration Functions for Webcams and AIBO under Linux Csaba Kertész, Zoltán Vámossy Faculty of Science, University of Szeged,

More information

Digitizing Color. Place Value in a Decimal Number. Place Value in a Binary Number. Chapter 11: Light, Sound, Magic: Representing Multimedia Digitally

Digitizing Color. Place Value in a Decimal Number. Place Value in a Binary Number. Chapter 11: Light, Sound, Magic: Representing Multimedia Digitally Chapter 11: Light, Sound, Magic: Representing Multimedia Digitally Fluency with Information Technology Third Edition by Lawrence Snyder Digitizing Color RGB Colors: Binary Representation Giving the intensities

More information

Augmented Desk Interface. Graduate School of Information Systems. Tokyo , Japan. is GUI for using computer programs. As a result, users

Augmented Desk Interface. Graduate School of Information Systems. Tokyo , Japan. is GUI for using computer programs. As a result, users Fast Tracking of Hands and Fingertips in Infrared Images for Augmented Desk Interface Yoichi Sato Institute of Industrial Science University oftokyo 7-22-1 Roppongi, Minato-ku Tokyo 106-8558, Japan ysato@cvl.iis.u-tokyo.ac.jp

More information

Head Tracking for Google Cardboard by Simond Lee

Head Tracking for Google Cardboard by Simond Lee Head Tracking for Google Cardboard by Simond Lee (slee74@student.monash.edu) Virtual Reality Through Head-mounted Displays A head-mounted display (HMD) is a device which is worn on the head with screen

More information

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005. Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.

More information

Available online at ScienceDirect. Procedia Computer Science 50 (2015 )

Available online at   ScienceDirect. Procedia Computer Science 50 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 50 (2015 ) 503 510 2nd International Symposium on Big Data and Cloud Computing (ISBCC 15) Virtualizing Electrical Appliances

More information

What is Augmented Reality?

What is Augmented Reality? What is Augmented Reality? Well, this is clearly a good place to start. I ll explain what Augmented Reality (AR) is, and then what the typical applications are. We re going to concentrate on only one area

More information

Image Extraction using Image Mining Technique

Image Extraction using Image Mining Technique IOSR Journal of Engineering (IOSRJEN) e-issn: 2250-3021, p-issn: 2278-8719 Vol. 3, Issue 9 (September. 2013), V2 PP 36-42 Image Extraction using Image Mining Technique Prof. Samir Kumar Bandyopadhyay,

More information

EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON

EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON Josep Amat 1, Alícia Casals 2, Manel Frigola 2, Enric Martín 2 1Robotics Institute. (IRI) UPC / CSIC Llorens Artigas 4-6, 2a

More information

Fig.1 AR as mixed reality[3]

Fig.1 AR as mixed reality[3] Marker Based Augmented Reality Application in Education: Teaching and Learning Gayathri D 1, Om Kumar S 2, Sunitha Ram C 3 1,3 Research Scholar, CSE Department, SCSVMV University 2 Associate Professor,

More information

ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality

ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality The MIT Faculty has made this article openly available. Please share how this access benefits you. Your

More information

Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality

Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality Arindam Dey PhD Student Magic Vision Lab University of South Australia Supervised by: Dr Christian Sandor and Prof.

More information

Human Computer Interaction by Gesture Recognition

Human Computer Interaction by Gesture Recognition IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 9, Issue 3, Ver. V (May - Jun. 2014), PP 30-35 Human Computer Interaction by Gesture Recognition

More information