HamsaTouch: Tactile Vision Substitution with Smartphone and Electro-Tactile Display

Similar documents
Tactile Vision Substitution with Tablet and Electro-Tactile Display

Design of Cylindrical Whole-hand Haptic Interface using Electrocutaneous Display

Tactile Presentation to the Back of a Smartphone with Simultaneous Screen Operation

SmartTouch: Electric Skin to Touch the Untouchable

Combination of Cathodic Electrical Stimulation and Mechanical Damped Sinusoidal Vibration to Express Tactile Softness in the Tapping Process *

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface

SmartTouch - Augmentation of Skin Sensation with Electrocutaneous Display

Wearable Tactile Device using Mechanical and Electrical Stimulation for Fingertip Interaction with Virtual World

Expression of 2DOF Fingertip Traction with 1DOF Lateral Skin Stretch

Simultaneous presentation of tactile and auditory motion on the abdomen to realize the experience of being cut by a sword

Title: A Comparison of Different Tactile Output Devices In An Aviation Application

Comparison of Haptic and Non-Speech Audio Feedback

Virtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback

Hiroyuki Kajimoto Satoshi Saga Masashi Konyo. Editors. Pervasive Haptics. Science, Design, and Application

Reconsideration of Ouija Board Motion in Terms of Haptic Illusions (Ⅲ) -Experiment with 1-DoF Linear Rail Device-

VIRTUAL FIGURE PRESENTATION USING PRESSURE- SLIPPAGE-GENERATION TACTILE MOUSE

Figure 2. Haptic human perception and display. 2.2 Pseudo-Haptic Feedback 2. RELATED WORKS 2.1 Haptic Simulation of Tapping an Object

ITS '14, Nov , Dresden, Germany

Fibratus tactile sensor using reflection image

3D Form Display with Shape Memory Alloy

Wearable Haptic Display to Present Gravity Sensation

Exploring Surround Haptics Displays

ISMCR2004. Abstract. 2. The mechanism of the master-slave arm of Telesar II. 1. Introduction. D21-Page 1

The Effect of Frequency Shifting on Audio-Tactile Conversion for Enriching Musical Experience

Shape Memory Alloy Actuator Controller Design for Tactile Displays

Evaluation of Roller-Type Itch-Relief Device Employing Hot and Cold Alternating Stimuli

Illusion of Surface Changes induced by Tactile and Visual Touch Feedback

Discrimination of Virtual Haptic Textures Rendered with Different Update Rates

Perceptual Force on the Wrist under the Hanger Reflex and Vibration

Blind navigation with a wearable range camera and vibrotactile helmet

Vibrotactile Apparent Movement by DC Motors and Voice-coil Tactors

BRAILLE READING DEVICE BASED IN ELECTROCUTANEOUS STIMULATION

2 (

Facilitation of Affection by Tactile Feedback of False Heartbeat

A Study of Direction s Impact on Single-Handed Thumb Interaction with Touch-Screen Mobile Phones

Sensor system of a small biped entertainment robot

Selective Stimulation to Skin Receptors by Suction Pressure Control

Geo-Located Content in Virtual and Augmented Reality

Figure 1. The game was developed to be played on a large multi-touch tablet and multiple smartphones.

ZeroTouch: A Zero-Thickness Optical Multi-Touch Force Field

An Investigation of Search Behaviour in a Tactile Exploration Task for Sighted and Non-sighted Adults.

Distributed Vision System: A Perceptual Information Infrastructure for Robot Navigation

TapBoard: Making a Touch Screen Keyboard

Substitute eyes for Blind using Android

Touch & Haptics. Touch & High Information Transfer Rate. Modern Haptics. Human. Haptics

Touch Your Way: Haptic Sight for Visually Impaired People to Walk with Independence

Tactile Actuators Using SMA Micro-wires and the Generation of Texture Sensation from Images

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright

MULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT

HAND-SHAPED INTERFACE FOR INTUITIVE HUMAN- ROBOT COMMUNICATION THROUGH HAPTIC MEDIA

An Emotional Tactile Interface Completing with Extremely High Temporal Bandwidth

Haptic presentation of 3D objects in virtual reality for the visually disabled

Haptic Interface using Sensory Illusion Tomohiro Amemiya

Tactile sensing system using electro-tactile feedback

A Fingernail-Mounted Tactile Display for Augmented Reality Systems

R (2) Controlling System Application with hands by identifying movements through Camera

Virtual Tactile Maps

Introduction to Mediated Reality

Haptic-Emoticon: Haptic Content Creation and Sharing System To Enhancing Text-Based Communication

Development of a telepresence agent

Gesture Recognition with Real World Environment using Kinect: A Review

Cutaneous Feedback of Fingertip Deformation and Vibration for Palpation in Robotic Surgery

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

Chapter 2 Introduction to Haptics 2.1 Definition of Haptics

Experience of Immersive Virtual World Using Cellular Phone Interface

Proposal for the Object Oriented Display : The Design and Implementation of the MEDIA 3

Augmentation of Acoustic Shadow for Presenting a Sense of Existence

A Glove Interface with Tactile feeling display for Humanoid Robotics and Virtual Reality systems

A Tactile Display using Ultrasound Linear Phased Array

Haptics for Guide Dog Handlers

CS545 Contents XIV. Components of a Robotic System. Signal Processing. Reading Assignment for Next Class

Jane Li. Assistant Professor Mechanical Engineering Department, Robotic Engineering Program Worcester Polytechnic Institute

Graphical User Interfaces for Blind Users: An Overview of Haptic Devices

THE HEIDELBERG TACTILE VISION SUBSTITUTION SYSTEM

2 Our Hardware Architecture

Electrical stimulation of mechanoreceptors

Electro-tactile Feedback System for a Prosthetic Hand

Multisensory Virtual Environment for Supporting Blind Persons' Acquisition of Spatial Cognitive Mapping a Case Study

Development of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane

Lecture 7: Human haptics

ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality

Do-It-Yourself Object Identification Using Augmented Reality for Visually Impaired People

Advancements in Gesture Recognition Technology

Omni-Directional Catadioptric Acquisition System

Interactive Exploration of City Maps with Auditory Torches

"From Dots To Shapes": an auditory haptic game platform for teaching geometry to blind pupils. Patrick Roth, Lori Petrucci, Thierry Pun

Tactile letter recognition under different modes of stimulus presentation*

Azaad Kumar Bahadur 1, Nishant Tripathi 2

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces

Seeing with the Brain Paul Bach-y-Rita, Mitchell E. Tyler & Kurt A. Kaczmarek Published online: 13 Nov 2009.

Evaluation of Five-finger Haptic Communication with Network Delay

Using Simulation to Design Control Strategies for Robotic No-Scar Surgery

702. Investigation of attraction force and vibration of a slipper in a tactile device with electromagnet

Welcome, Introduction, and Roadmap Joseph J. LaViola Jr.

GUIDED WEAPONS RADAR TESTING

Development of Video Chat System Based on Space Sharing and Haptic Communication

QS Spiral: Visualizing Periodic Quantified Self Data

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

A low resolution image sensor for tactile vision sensory substitution

Feeding human senses through Immersion

Transcription:

HamsaTouch: Tactile Vision Substitution with Smartphone and Electro-Tactile Display Hiroyuki Kajimoto The University of Electro-Communications 1-5-1 Chofugaoka, Chofu, Tokyo 1828585, JAPAN kajimoto@kaji-lab.jp Masaki Suzuki EyePlusPlus, Inc. Shinsagawa Bldg. 201, 1-27-1, Ishihara, Sumida-ku, Tokyo 1300011, JAPAN suzuki@eyeplus2.com Yonezo Kanno EyePlusPlus, Inc. Shinsagawa Bldg. 201, 1-27-1, Ishihara, Sumida-ku, Tokyo 1300011, JAPAN kanno@eyeplus2.com Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the author. Copyright is held by the authors. CHI2014, April 26 May 1, 2014, Toronto, Ontario, Canada. ACM 978-1-4503-2474-8/14/04. http://dx.doi.org/10.1145/2559206.2581164 Abstract This paper documents the development and evaluation of a novel tactile vision substitution system (TVSS) for the people with visual impairments. The system is composed of an electro-tactile display with 512 electrodes, the same number of optical sensors beneath each electrode, and a smartphone with a camera and an LCD. The smartphone acquires the surrounding view, conducts image processing and displays the image on the LCD. The image is captured by the optical sensors and converted to a tactile image by the electro-tactile display. While the concept of the TVSS is classic, combining the commonly available mobile device and electro-tactile display enables a low cost yet powerful and compact system. Furthermore, optical communication architecture enables an open development environment. Author Keywords electro-tactile display, optical communication, reading aid, smartphone, tactile vision substitution system, visually impaired ACM Classification Keywords H5.2 [Information interfaces and presentation]: User Interfaces - Haptic I/O.

Introduction Since the development of the first tactile vision substitution system (TVSS) in the 1960s, there have been numerous attempts to present surrounding visual information to a tactile channel for the people with visual impairments. The system consists of a camera and tactile display. However, the developed systems have not prevailed in the blind community to date, for several reasons, summarized by the following two points. using an LCD and phototransistors, and the third is to use an electro-tactile display. The first is cost. While the number of people with visual impairments is large and increasing [1], it is not large enough to expect economies of scale, which typically leads to expensive systems with non-state-of-the-art device elements. The second issue is the way of presentation. Our surrounding environment is complex and three dimensional (3D). If we simply do image processing such as edge enhancement, we obtain many features that are not comprehensible for a tactile channel that has much less capacity than a visual channel. We need to reduce information; filtering out further objects and concentrating on closer obstacles seems a promising approach, which requires depth sensing. Therefore, this issue is strongly related to sensing capability, which is partly the cost issue. This paper documents the development and evaluation of a novel TVSS named HamsaTouch (Figure 1), which tries to resolve these issues with the following three frameworks. The first is to use a commonly available smartphone as a camera and image processing device, the second is to use simple optical data transmission Figure 1 (Top) Overview of HamsaTouch. Image captured by the smartphone is converted to tactile pattern. (Bottom) Cross-section of HamsaTouch and its information processing procedure.

The cost of hardware is reduced by using a smartphone, and we can always use a state-of-the-art image sensor and processing unit by replacing the smartphone, which could be done by the end users. The software cost is reduced by using optical data transmission, because we only need to process the image and display it on the LCD. The adoption of an electro-tactile display contributes to the reduction in hardware cost and size, because it only requires electrodes and a switching circuit. At present, we do not have a solution to the depth sensing issue. However, by employing a smartphone, we can expect future adoption of a depth camera. Furthermore, we can use simple image processing to address this issue, which will be discussed in the future work section. RELATED WORK The first tactile vision substitution system was developed by Bach-y-Rita et al., and was composed of a camera and vibrators on the stomach or back [2]. Simple image processing, such as edge enhancement was employed. Recently, a few research groups developed a haptic aid for the people with visual impairments using a smartphone as a sensing device [3], [4]. Akhter et al. developed a vision substitution system, using a smartphone as a parallax-based 3D sensing device [3]. Vera et al. developed a virtual white cane using a smartphone as a depth sensor combined with a laser pointer [4]. These studies share the common concept of using a smartphone as a cheap image processing unit. Our work inherits the concept, and attempts to refine it with an electro-tactile display and an optical transmission technique. As the size of the tactile display limited its daily use, numerous attempts were made to use electrical stimulation instead of vibration [5], [6], [7]. Electrical stimulation enables small, flexible and low cost tactile pattern presentation so that the stimulation site was not limited to the stomach or back but could be used on the forehead [5] or tongue [7]. While the sensation elicited by electro-tactile stimulation was considered unstable compared to mechanical vibration, waveform optimization [8], and impedance based control [9], [10] enabled durable and not unpleasant electrical stimulation. The electrical tactile display was also applied as a touch-panel to achieve tactile feedback [11]. Our work is based on these technologies. Optical data transmission itself is quite a common technology, but using a visual display as a parallel optical signal was, as far as we know, first developed by Sugimoto et al., and named Display-Based Computing [12]. They first used a projector or an LCD to control numerous robots with high positional accuracy. Nojima et al. used a similar method to control numerous hairs on a tablet, actuated by shape memory alloys [13]. As far as the speed of communication is concerned, equivalent data transmission is possible using wireless transmission. We adopted this technology for HamsaTouch for three reasons. One is its simple architecture, which clearly divides the image processing part and display part, leading to low software costs.

Another is its low legal barrier. Inspection is required for a wireless device, which is not a significant cost for the huge smartphone market but is actually quite a burden for a smaller market. The third is that we can use the device without a smartphone as a paperreading device, similar to Optacon [14]. Eight 64-channel half-bridge switches (HV507, SuperTex Inc.) drive 512 electrodes (32 by 16), which are located on the top of the device. At any one moment, a single electrode is set as an anode and all the other electrodes work as cathodes. A twodimensional pattern is produced by high speed scanning. One of our authors has developed a finger-mounted tactile vision conversion system named SmartTouch [15], composed of 4 by 4 electrodes and the same number of phototransistors beneath each electrode, which is a technical predecessor to this work. Figure 2 Top and bottom view. The top contains 512 electrodes, and the bottom contains the same number of phototransistors. SYSTEM Hardware Figure 3 illustrates the hardware structure. The system uses a microcontroller (mbed NXP LPC1768, NXP Semiconductors NV), which drives a high-speed D/A and A/D converter for electrical stimulation, and multichannel analog multiplexer and A/D converter for photo sensing. The stimulating pulse is generated by the D/A converter and converted to a current pulse by a voltage/current converter, driven by a high-voltage source (300 V). The current pulse passes through a resistor to measure the voltage and current. This means that the system can measure the electrical impedance of each electrode, which can be used for stabilizing sensation and for touch-sensing. Figure 3 Hardware structure. 512 phototransistors are aligned just beneath the electrodes. A 32-channel analog multiplexer drives each row, and a 16-channel A/D converter acquires row data, sending it to the microcontroller. Figure 2 shows the top and bottom view of the device. The tactile spatial resolution defined as the so-called two-point discrimination threshold is around 1.5 mm at the tip of the finger, 3 mm at the other parts of the finger, and 8 mm on the palm [16]. We set the electrode density to 2.54 mm to cover most parts of the hand. This covers a 81.28 mm by 40.64 mm area, which is close to the size of a 4-inch LCD adopted by numerous smartphones (ex. iphone4, Apple Inc.: LCD size 74.8mm by 49.9 mm). Total sensing and stimulation loop was around 50 to 100 Hz depending on the number of stimulation points.

Image Processing We used an Android-based smartphone (Galaxy SIII mini, 1 GHz Dual-core ARM Cortex-A9, Samsung Electronics) as an imaging unit. An OpenCV4Android [17] was used for image processing. As a first attempt, simple edge detection using Canny filter and dilation operation for thickening the edges was applied. Refresh rate was 25 to 30 fps. Figure 5 shows the environment, processed image, and the 32 by 16 pattern obtained by phototransistors (Figure 4). not particularly ask participants to observe the whole structure. Most participants could quickly and confidently answer the vertical and horizontal bars, while in few exceptional cases they moved the device too large and observed the edges of the LCD. Evaluation We conducted an experiment to see if the whole system works properly as a tactile vision substitution system. Experiment setup is shown in Figure 5. Participants were asked to sit in front of a 27-inch LCD monitor, blindfolded, hold the device with their dominant hands, and contact the electro-tactile display with their palms. They orally answered displayed visual patterns by using the device. The patterns were vertical bar, horizontal bar, cross-shape, and circle, all were fit to 100 mm square size. Distance between the participants body and the LCD was around 700 mm, but they were allowed to move their arms freely. Figure 5 Experiment setup. Figure 4 Environment, processed images by the smartphone, and 32 by 16 electrical stimulation patterns. We recruited five laboratory members as participants excluding authors, all right-handed male aged 22-25. They had basic knowledge of electro-tactile sensation but naïve to this device. 16 trials (4 conditions 4 times) were conducted in random order for each participant. Correct answer rates were 90% for vertical bar, 90% for horizontal bar, 35% for cross-shape, and 65% for circle, 70% in average. Cross-shape was misinterpreted most because it contains bars and corners, and we did Figure 6 Answer rates for the four presented patterns. Conclusions We developed and evaluated a novel tactile vision substitution system for the people with visual impairments, consisting of an electro-tactile display, optical sensors, and a smartphone. The hardware cost is reduced by using the smartphone, and we can adopt a state-of-the-art imager by replacing this part, including a future 3D sensing unit. The software cost is also reduced by using optical data transmission,

because of its simple architecture. In other words, we attempt to interface between the people with visual impairments and ongoing technological advances. Future work Our future work will include more sophisticated image processing, such as depth reconstruction. In our previous attempts to mount a tactile display and a camera on the forehead [5], we observed that users frequently shifted their heads left and right to obtain distance, using motion parallax (i.e. a closer object moves faster). This observation reconfirms the importance of distance information, and at the same time, implies motion analysis such as optical flow calculation might be a good substitute for 3D depth sensing. References [1] WHO, Visual impairment and blindness http://www.who.int/mediacentre/factsheets/fs282/en/ [2] Bach-y Rita, P., Collins, C. C., Saunders, F., White, B., Scadden, L. Vision substitution by tactile image projection. Nature (1969), 221: 963 964. [3] Akhter, S., Mirsalahuddin, J., Marquina, F.B., Islam, S. A smartphone-based haptic vision substitution system for the blind. In Proc. Annual Northeast Bioengineering Conf. (2011), 1-2. [4] Vera, P., Zenteno, D., Salas. J. A smartphonebased virtual white cane, Pattern Anal. Applic. (2013) [5] Kajimoto, H., Kanno, Y., Tachi, S. Forehead electro-tactile display for vision substitution. In Proc. EuroHaptics (2006). [6] Collins, C. C. Tactile television - mechanical and electrical image projection. IEEE Trans. Man-Machine Systems (1970), MMS-11(1): 65 71. [7] Kaczmarek, K. A. Electrotactile and vibrotactile displays for sensory substitution systems. IEEE Trans. Biomed. Eng. (1991), 38:1 16. [8] Kaczmarek, K. A., Webster, J. G., Radwin, R. G. Maximal dynamic range electrotactile stimulation waveforms. IEEE Trans. Biomed. Eng. (1992), 39(7): 701-715. [9] Tachi, S., Tanie, K., Komiyama, K., Abe, M. Electrocutaneous communication in a guide dog robot (MELDOG). IEEE Trans. Biomed. Eng. (1985), 32: 461-469. [10] Kajimoto, H. Electro-tactile display with real-time impedance feedback using pulse width modulation. IEEE Trans. Haptics (2012), 5(2): 184-188. [11] Kajimoto, H. Skeletouch: Transparent electrotactile display for mobile surfaces. In Proc. SIGGRAPH ASIA Emerging Technologies (2012). [12] Sugimoto, M., Kodama, K., Nakamura, A., Kojima M., Inami, M. A display-based tracking system: displaybased computing for measurement systems. In Proc. 17th Int. Conf. Artificial Reality and Telexistence (2007), 31-38. [13] Nojima, T., Ooide, Y., Kawaguchi, H. Hairlytop interface: an interactive surface display comprised of hair-like soft actuators. In Proc. World Haptics Conference (2013), 431-435. [14] Linvill, G., Bliss, J. C. A direct translation reading aid for the blind. In Proc. IEEE, (1966), 54(1): 40-51. [15] Kajimoto, H., Inami, M., Kawakami, N., Tachi, S. SmartTouch: electric skin to touch the untouchable. IEEE Computer Graphics & Applications Magazine (2004), 36-43. [16] Johansson, R. S., Vallbo, A. B. Tactile sensory coding in the glabrous skin of the human hand, Trends in Neurosciences (1983), 6: 27-32. [17] OpenCV4Android SDK http://docs.opencv.org/doc/tutorials/introduction/andro id_binary_package/o4a_sdk.html