Aalborg Universitet. Auditory feedback in a multimodal balancing task: Serafin, Stefania; Turchet, Luca; Nordahl, Rolf

Similar documents
Audio-haptic physically-based simulation of walking on different grounds

A multimodal architecture for simulating natural interactive walking in virtual environments

DO YOU HEAR A BUMP OR A HOLE? AN EXPERIMENT ON TEMPORAL ASPECTS IN THE RECOGNITION OF FOOTSTEPS SOUNDS

Welcome to this course on «Natural Interactive Walking on Virtual Grounds»!

Extraction of ground reaction forces for real-time synthesis of walking sounds Serafin, Stefania; Turchet, Luca; Nordahl, Rolf

PHYSICALLY BASED SOUND SYNTHESIS AND CONTROL OF FOOTSTEPS SOUNDS

Exploring sonic interaction design and presence: Natural Interactive Walking in Porto.

Aalborg Universitet. Published in: Eurohaptics DOI (link to publication from Publisher): / _32. Publication date: 2012

Published in: Proceedings of the 8th International Conference on Tangible, Embedded and Embodied Interaction

Augmented Exercise Biking with Virtual Environments for Elderly Users Bruun-Pedersen, Jon Ram; Serafin, Stefania; Kofoed, Lise

Interactive Simulation: UCF EIN5255. VR Software. Audio Output. Page 4-1

3D sound in the telepresence project BEAMING Olesen, Søren Krarup; Markovic, Milos; Madsen, Esben; Hoffmann, Pablo Francisco F.; Hammershøi, Dorte

Impact of the size of the hearing aid on the mobile phone near fields Bonev, Ivan Bonev; Franek, Ondrej; Pedersen, Gert F.

University of Huddersfield Repository

Low frequency sound reproduction in irregular rooms using CABS (Control Acoustic Bass System) Celestinos, Adrian; Nielsen, Sofus Birkedal

The King-Kong Effects: Improving Sensation of Walking in VR with Visual and Tactile Vibrations at each Step

Waves Nx VIRTUAL REALITY AUDIO

Predicting audio step feedback for real walking in virtual environments

Virtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback

Magnusson, Charlotte; Rassmus-Gröhn, Kirsten; Szymczak, Delphine

Gamescape Principles Basic Approaches for Studying Visual Grammar and Game Literacy Nobaew, Banphot; Ryberg, Thomas

Force versus Frequency Figure 1.

A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems

Comparison of Haptic and Non-Speech Audio Feedback

Figure 2. Haptic human perception and display. 2.2 Pseudo-Haptic Feedback 2. RELATED WORKS 2.1 Haptic Simulation of Tapping an Object

ENHANCED HUMAN-AGENT INTERACTION: AUGMENTING INTERACTION MODELS WITH EMBODIED AGENTS BY SERAFIN BENTO. MASTER OF SCIENCE in INFORMATION SYSTEMS

Simultaneous presentation of tactile and auditory motion on the abdomen to realize the experience of being cut by a sword

Compressibility and Crushability Reproduction through an Amorphous Haptic Interface

Sound rendering in Interactive Multimodal Systems. Federico Avanzini

The Effect of Frequency Shifting on Audio-Tactile Conversion for Enriching Musical Experience

Andersen, Hans Jørgen; Morrison, Ann Judith; Knudsen, Lars Leegaard

Addressing Carrier Aggregation with Narrow-band Tunable Antennas Barrio, Samantha Caporal Del; Morris, Art; Pedersen, Gert F.

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

The Mixed Reality Book: A New Multimedia Reading Experience

Published in: Proceedings of the Workshop on What to Study in HCI at CHI 2015 Conference on Human Factors in Computing Systems

Directional dependence of loudness and binaural summation Sørensen, Michael Friis; Lydolf, Morten; Frandsen, Peder Christian; Møller, Henrik

Antenna Diversity on a UMTS HandHeld Phone Pedersen, Gert F.; Nielsen, Jesper Ødum; Olesen, Kim; Kovacs, Istvan

Predicting audio step feedback for real walking in virtual environments

Geo-Located Content in Virtual and Augmented Reality

Interactive Multimedia Contents in the IllusionHole

Low-Profile Fabry-Pérot Cavity Antenna with Metamaterial SRR Cells for Fifth Generation Systems

Paper Body Vibration Effects on Perceived Reality with Multi-modal Contents

Haptic presentation of 3D objects in virtual reality for the visually disabled

Drumtastic: Haptic Guidance for Polyrhythmic Drumming Practice

Aalborg Universitet. Published in: Acustica United with Acta Acustica. Publication date: Document Version Early version, also known as pre-print

Virtual Reality Calendar Tour Guide

Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances

Citation for published version (APA): Parigi, D. (2013). Performance-Aided Design (PAD). A&D Skriftserie, 78,

Determining Optimal Player Position, Distance, and Scale from a Point of Interest on a Terrain

Perception of room size and the ability of self localization in a virtual environment. Loudspeaker experiment

Published in: Proceedings of NAM 98, Nordic Acoustical Meeting, September 6-9, 1998, Stockholm, Sweden

Velvety Massage Interface (VMI): Tactile Massage System Applied Velvet Hand Illusion

Capacitive Face Cushion for Smartphone-Based Virtual Reality Headsets

Expression of 2DOF Fingertip Traction with 1DOF Lateral Skin Stretch

Perception. Read: AIMA Chapter 24 & Chapter HW#8 due today. Vision

Three-dimensional sound field simulation using the immersive auditory display system Sound Cask for stage acoustics

Combining Subjective and Objective Assessment of Loudspeaker Distortion Marian Liebig Wolfgang Klippel

Feelable User Interfaces: An Exploration of Non-Visual Tangible User Interfaces

The Game Experience Questionnaire

A Practical FPGA-Based LUT-Predistortion Technology For Switch-Mode Power Amplifier Linearization Cerasani, Umberto; Le Moullec, Yannick; Tong, Tian

From Binaural Technology to Virtual Reality

Tele-Nursing System with Realistic Sensations using Virtual Locomotion Interface

High Gain K-Band Patch Antenna for Low Earth Orbit Interlink Between Nanosatellites Squadrito, Paolo; Zhang, Shuai; Pedersen, Gert F.

Distance Protection of Cross-Bonded Transmission Cable-Systems

On the creation of standards for interaction between real robots and virtual worlds

3D AUDIO AR/VR CAPTURE AND REPRODUCTION SETUP FOR AURALIZATION OF SOUNDSCAPES

Augmented Home. Integrating a Virtual World Game in a Physical Environment. Serge Offermans and Jun Hu

Aalborg Universitet. Published in: Acustica United with Acta Acustica. Publication date: Document Version Early version, also known as pre-print

Reach Out and Touch Someone

Aalborg Universitet. Published in: Antennas and Propagation (EuCAP), th European Conference on

Capability for Collision Avoidance of Different User Avatars in Virtual Reality

DBR based passively mode-locked 1.5m semiconductor laser with 9 nm tuning range Moskalenko, V.; Williams, K.A.; Bente, E.A.J.M.

Haplug: A Haptic Plug for Dynamic VR Interactions

Exploring Haptics in Digital Waveguide Instruments

From acoustic simulation to virtual auditory displays

AUDITORY ILLUSIONS & LAB REPORT FORM

PROJECT CONCEPT THE CUBE EXPERIMENT 12/19/2016 ANDREW ALAIN ALEX HERCULE DESJARDINS SAMUEL WALKER. Virtual maquette of installation

Exploring Surround Haptics Displays

Psychoacoustic Cues in Room Size Perception

MELODIOUS WALKABOUT: IMPLICIT NAVIGATION WITH CONTEXTUALIZED PERSONAL AUDIO CONTENTS

Published in: Proceedings of the 16th Conference on Power Electronics and Applications, EPE 14-ECCE Europe

Experimenting with Sound Immersion in an Arts and Crafts Museum

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright

Sound source localization and its use in multimedia applications

Omni-Directional Catadioptric Acquisition System

Toward Principles for Visual Interaction Design for Communicating Weight by using Pseudo-Haptic Feedback

VIRTUAL REALITY Introduction. Emil M. Petriu SITE, University of Ottawa

TOUCH & FEEL VIRTUAL REALITY. DEVELOPMENT KIT - VERSION NOVEMBER 2017

Operational modal analysis applied to a horizontal washing machine: A comparative approach Sichani, Mahdi Teimouri; Mahjoob, Mohammad J.

A Technical Introduction to Audio Cables by Pear Cable

Drexel-SDP GK-12 ACTIVITY

Mixed and Augmented Reality Reference Model as of January 2014

The analysis of multi-channel sound reproduction algorithms using HRTF data

A Java Virtual Sound Environment

Aalborg Universitet. Linderum Electricity Quality - Measurements and Analysis Silva, Filipe Miguel Faria da; Bak, Claus Leth. Publication date: 2013

QS Spiral: Visualizing Periodic Quantified Self Data

EMERGENCE OF COMMUNICATION IN TEAMS OF EMBODIED AND SITUATED AGENTS

Aalborg Universitet. Emulating Wired Backhaul with Wireless Network Coding Thomsen, Henning; Carvalho, Elisabeth De; Popovski, Petar

A 28 GHz FR-4 Compatible Phased Array Antenna for 5G Mobile Phone Applications Parchin, Naser Ojaroudi; Shen, Ming; Pedersen, Gert F.

Transcription:

Aalborg Universitet Auditory feedback in a multimodal balancing task: Serafin, Stefania; Turchet, Luca; Nordahl, Rolf Published in: Proceedings of the SMC Conferences Publication date: 2011 Document Version Accepted author manuscript, peer reviewed version Link to publication from Aalborg University Citation for published version (APA): Serafin, S., Turchet, L., & Nordahl, R. (2011). Auditory feedback in a multimodal balancing task:: Walking on a virtual rope. In Proceedings of the SMC Conferences Universitá di Padova. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights.? Users may download and print one copy of any publication from the public portal for the purpose of private study or research.? You may not further distribute the material or use it for any profit-making activity or commercial gain? You may freely distribute the URL identifying the publication in the public portal? Take down policy If you believe that this document breaches copyright please contact us at vbn@aub.aau.dk providing details, and we will remove access to the work immediately and investigate your claim. Downloaded from vbn.aau.dk on: januar 10, 2018

AUDITORY FEEDBACK IN A MULTIMODAL BALANCING TASK: WALKING ON A VIRTUAL PLANK Stefania Serafin Department of Architecture, Design and Media Technology Aalborg University Copenhagen sts@create.aau.dk Luca Turchet Department of Architecture, Design and Media Technology Aalborg University Copenhagen tur@create.aau.dk Rolf Nordahl Department of Architecture, Design and Media Technology Aalborg University Copenhagen rn@create.aau.dk ABSTRACT We describe a multimodal system which exploits the use of footwear-based interaction in virtual environments. We developed a pair of shoes enhanced with pressure sensors, actuators, and markers. Such shoes control a multichannel surround sound system and drive a physically based sound synthesis engine which simulates the act of walking on different surfaces. We present the system in all its components, and explain its ability to simulate natural interactive walking in virtual environments. The system was used in an experiment whose goal was to assess the ability of subjects to walk blindfolded on a virtual plank. Results show that subjects perform the task slightly better when they are exposed to haptic feedback as opposed to auditory feedback, although no significant differences are measured. The combination of auditory and haptic feedback does not significantly enhances the task performance. 1. INTRODUCTION In the academic community, foot-based interactions have mostly been concerned with the engineering of locomotion interfaces for virtual environments [1]. A notable exception is the work of Paradiso and coworkers, who pioneered the development of shoes enhanced with sensors, able to capture 16 different parameters such as pressure, orientation, acceleration [2]. Such shoes were used for entertainment purpose as well as for rehabilitation studies [3]. The company Nike has also developed an accelerometer which can be attached to running shoes and connected to an ipod, in such a way that, when a person runs, the ipod tracks and reports different information. In this paper we mostly focus on enhancing our awareness of auditory and haptic feedback in foot based devices, topic which is still rather unexplored. We describe a multimodal interactive space which has been developed with the goal of creating audio-haptic-visual simulations of walking-based interactions. The system requires users to walk around a space wearing a pair of shoes enhanced with sensors and actuators. The position of such Copyright: c 2011 Stefania Serafin et al. This is an open-access article distributed under the terms of the Creative Commons Attribution 3.0 Unported License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. shoes is tracked by a motion capture system, and the shoes drive a audio-visual-haptic synthesis engine based on physical models. The idea of enhancing shoes with sensors and actuators is similar to the ones we have been exploring in the context of the Natural Interactive Walking (NIW) FET-Open EU project 1 [7, 8]. The ultimate goal of this project is to provide closed-loop interaction paradigms enabling the transfer of skills that have been previously learned in everyday tasks associated to walking. In the NIW project, several walking scenarios are simulated in a multimodal context, where especially audition and haptic play an important role. As a case study of the developed architecture, we describe an experiment where subjects were asked to walk straight on a virtual plank. The use of audio-haptic augmented footwear for navigation has not been extensively explored in the research community. An exception are the Cabboots [4], a pair of actuated boots which provide information concerning the shape of a path. As another example, recently Takeuchi proposed Gilded Gait, a system which changes the perceived physical texture of the ground [5]. The Gilded Gait system is designed as a pair of insoles with vibrotactile feedback to simulate ground textures. Recently, it has also been demonstrated that walking straight is a hard task also in the physical world. The research was justified by the common belief that people, when getting lost, tend to walk into circles. Subjects were asked to walk straight in two outdoor environments: a forest and a desert. When subjects were not able to see the sun, they walked in circles. It was suggested that veering from a straight course is the result of accumulating noise in the sensorimotor system, which, without an external directional reference to recalibrate the subjective straight ahead [6]. In this paper, we investigate the ability of subjects to walk straight on a narrow virtual plank, with the help of auditory and visual feedback. The results of this research can be applied to the fields of rehabilitation, navigation in virtual and physical worlds as well as entertainment. 2. THE OVERALL ARCHITECTURE Figure 1 shows a schematic representation of the overall architecture developed. The architecture consists of a motion capture system (MoCap)(Optitrack by Naturalpoint), 1 http://www.niwproject.eu/

Control room Experiment room Optitrack ----------------- Physical models Fireface 800 Fireface 800 2 Amplifier Arduino Visual feedback Markers Pressure sensors Markers Markers Pressure sensors Haptic feedback ------------- Haptic feedback Auditory feedback Figure 1. A schematic representation of the multimodal architecture to simulate natural interactive walking. two soundcards (Fireface 800), twelve loudspeakers (Dynaudio), two amplifiers, two haptic shoes and two computers. Such system is placed in an acoustically isolated laboratory which consists of a control room and a larger interaction room where the setup is installed and where the experiments are performed. The control room is used by the experimenters providing the stimuli and collecting the results. It hosts two desktop computers. The first computer runs the motion capture software 2, while the second runs the audio-haptic synthesis engine. The two computers are connected through an ethernet cable and communicate via UDP. The coordinates relative to the motion capture system are sent from the first to the second computer which processes them in order to control the sound engine. A transparent glass divides the two rooms, so it is possible for the experimenters to see the users performing the assigned task. The two rooms are connected by means of a talkback system. The experiment room is 5.45 m large, 5.55 m long, and 2.85 m high, and the walking area available to the users is about 24m 2. 3. SIMULATION HARDWARE 3.1 Tracking the user The user locomotion is tracked by an Optitrack motion capture system 3, composed by 16 infrared cameras 4. The cameras are placed in a configuration optimized for the tracking of the feet and head position simultaneously. In 2 Tracking Tools 2.0 by Naturalpoint 3 http://naturalpoint.com/optitrack/ 4 OptiTrack FLEX:V100R2 order to achieve this goal, markers are placed on the top of each shoe worn by the subjects as well as on top of the head. Users are also tracked by using the pressure sensors embedded in a pair of sandals. Specifically, a pair of lightweight sandals was used (Model Arpenaz-50, Decathlon, Villeneuve d Ascq, France). The sole has two FSR pressure sensors 5 whose aim is to detect the pressure force of the feet during the locomotion of a subject wearing the shoes. The two sensors were placed in correspondence to the heel and toe respectively in each shoe. The analogue values of each of these sensors were digitalized by means of an Arduino Diecimila board 6 and were used to drive the audio and haptic synthesis. 3.2 Actuated shoes In order to provide haptic feedback during the act of walking, a pair of sandals has been recently enhanced with sensors and actuators [9]. The particular model of shoes chosen has light, stiff foam soles that are easy to gouge and fashion. Four cavities were made in the thickness of the sole to accommodate four vibrotactile actuators (Haptuator, Tactile Labs Inc., Deux-Montagnes, Qc, Canada). These electromagnetic recoil-type actuators have an operational, linear bandwidth of 50 500 Hz and can provide up to 3 G of acceleration when connected to light loads. As indicated in Figure 2, two actuators were placed under the heel of the wearer and the other two under the ball of the foot. These were bonded in place to ensure good transmission of the vibrations inside the soles. When activated, vibrations propagated far in the light, stiff foam. In 5 I.E.E. SS-U-N-S-00039 6 http://arduino.cc/

vibration sound haptuator 13 mm force sensor simulation force measurement headphones Figure 2. The developed haptic shoes used in this experiment. the present configuration, the four actuators were driven by the same signal but could be activated separately to emphasize, for instance, the front or back activation, or to realize other effects such as modulating different, back-front signals during heel-toe movements. A cable exits from each shoe, with the function of transporting the signals of the pressure sensors and for the actuators. Such cables were about 5 meters long, and they were connected 7 to two 4TP (twisted pair) cables: one 4TP cable carries the sensor signals to a breakout board 8, which then interfaces to an Arduino board; the other 4TP cable carries the actuator signals from a pair of Pyle Pro PCA1 9 mini 2X15 W stereo amplifiers, driven by outputs from a FireFace 800 soundcard. 10 Each stereo amplifier handles 4 actuators found on a single shoe, each output channel of the amplifier driving two actuators connected in parallel. The PC handles the Arduino through a USB connection, and the FireFace soundcard through a FireWire connection. In our virtual environment the auditory feedback can be delivered by means of headphones (specifically, Sennheiser HD 650) or a set of 16 channels loudspeakers (Dynaudio BM5A speakers). 4. AUDIO-HAPTIC FEEDBACK We developed a multimodal synthesis engine able to reproduce auditory and haptic feedback. Auditory feedback is obtained by the combination of a footstep and a soundscape sound synthesis engine. Haptic feedback is provided by means of the haptic shoes previously described. The haptic synthesis is driven by the same engine used for the synthesis of footstep sounds, and is able to simulate the haptic sensation of walking on different surfaces, as illustrated in [9]. The engine for footstep sounds, based on physical models, is able to render the sounds of footsteps both on solid and aggregate surfaces. Several different materials have been simulated, in particular wood, creaking wood, and metal as concerns the solid surfaces, and gravel, snow, sand, dirt, forest underbrush, dry leaves, and high grass as regards the aggregate surfaces. A complete description of such engine in terms of sound design, implementation and control systems is presented in [10]. Using such engine, we implemented a comprehensive collection of footstep sounds. As solid surfaces, we implemented metal, wood, and creaking wood. In these materials, the impact model 7 through DB9 connectors 8 containing trimmers, that form voltage dividers with the FSRs 9 http://www.pyleaudio.com/manuals/pca1.pdf 10 http://www.rme-audio.com/english/firewire/ff800.htm was used to simulate the act of walking, while the friction model was used to simulate the creaking sounds typical of creaking wood floors. As aggregate surfaces, we implemented gravel, sand, snow, forest underbrush, dry leaves, pebbles and high grass. The simulated metal, wood and creaking wood surfaces were furthermore enhanced by using some reverberation. To control the audio-haptic footsteps synthesizer in our virtual environment, we use the haptic shoes: the audiohaptic synthesis is driven interactively during the locomotion of the subject wearing the shoes. The description of the control algorithms based on the analysis of the values of the pressure sensors coming from the haptic shoes can be found [9]. Such engine has been extensively tested by means of several audio and audio-haptic experiments and results can be found in [11] [12] [13] [14]. 4.1 Movement to sound mapping Figure 3 shows the dimensions of the path the users were asked to walk on. The mapping between feet movement and delivered auditory feedback was designed as follows: zone 1: a narrow band, 15 cm large and 120 cm long, corresponding to the straight direction to be covered. When both the feet stepped inside this zone, a creaking sound corresponding to the ecological sound of stepping on a creaking plank was provided. zones 2 and 3: two narrow bands, 10 cm large and 120 cm long, contiguous to zone 1 and placed at its left and right respectively. When one of the feet was stepping in this zone, no feedback was provided to it. However, subject was still able to continue the task by having the foot move to the correct zone 1. This is analogue to the situation of balancing with only one foot. zone 4 and 5: the areas contiguous to zone 2 and 3 and placed at its left and right respectively. When both the feet were inside such zone this was considered as failure for the task, and the recording of a long scream of a person falling down was triggered. zone 6: the area in front of zones 1, 2 and 3. When one or both the feet were inside this zone this was considered as success for the task, and the recording of a drum roll with an applause was triggered. zone 7: the area beyond of zones 1, 2 and 3. When both the feet were inside this zone, no sound was delivered.

Figure 3. The zones in which the walking space was divided. 5. EXPERIMENT DESIGN We designed an experiment in order to investigate the role of auditory and haptic feedback in facilitating a balancing task on a virtual plank. In such experiment, we asked subjects to walk straight in order not to virtually fall from the plank. Specifically, subjects were give the following instructions: Imagine you are walking on a wooden plank. Your task is to walk from one side to the other. Walk slowly and pay attention to the feedback you receive in order to succeed on your task. If your feet are outside of the plank you will fall. Figure 4 shows a subject performing the experiment. In this particular situation, no visual feedback was provided, and subjects were asked to walk being driven only by auditory and haptic feedback. The same stimuli were provided for the auditory and haptic simulation and designed as follows: when a user is walking on top of the virtual plank, the feet s position is detected by the motion capture system. In this case, the synthesis engine provides as a stimulus the sound and haptic feedback of a creaking wood. 5.1 Participants The experiment was performed by 15 participants, 14 men and 1 women, aged between 22 and 28 (mean=23.8, standard deviation=1.97). All participants reported normal hearing conditions. The participants took on average 6.8 minutes to complete the experiment. Subjects were randomly exposed to the four following conditions: auditory feedback, haptic feedback, audio-haptic feedback and no feedback. Each condition was tried twice, given in total eight trials for each subject. 6. RESULTS OF THE EXPERIMENT Table 1 shows the performance for each subject. The numbers in each row for each conditions indicate whether the subject performed successfully the task ones, twice or never. The results show that feedback helps balance mostly when haptic stimuli are provided. In this case, 46.6% of the tasks Figure 4. A subject performing the experiment of walking on a virtual plank. The tape on the floor represents the area where the plank is located.

Table 1. Summary of the results of the experiment. The number in each element of the matrix represents the times the task was successful (once, twice or never). Condition/ Audio (A) Haptic (H) Audio-haptic (AH) No-feedback (N) Subject number 1 2 1 2 1 2 2 1 1 1 3 1 4 1 1 5 1 2 6 1 1 2 1 7 2 1 8 1 9 1 1 10 11 1 1 12 2 1 1 13 14 1 2 1 15 2 2 1 2 were successfully completed. In the case where a combination of auditory and haptic feedback was provided, 43.3% of the tasks were completed. With only auditory feedback, 40% of the tasks were completed, while with no feedback only 26.6%. These results show how feedback slightly helps the balancing task. Haptic feedback performed better than the combination of auditory and haptic. This can be due to the fact that haptic feedback was provided directly to the feet, so the subjects had a closer spatial connection between the path they had to step on and the corresponding feedback. A post-experimental questionnaire was also performed, where subjects were asked several questions on the ability to freely move in the environment, to adjust to the technology and to which feedback was the most helpful. Indeed, 7 subjects found the haptic feedback to be the most helpful, 6 subjects the auditory feedback and 2 subjects the combination of auditory and haptic feedback. One subject indeed commented that the most useful feedback was when there was background noise (the pink noise used to mask the auditory feedback) and only vibration was provided. All subjects claimed to notice the relationship between actions performed and feedback provided. The subjects also commented that the feedback did not always match their expectations, since sometimes no feedback was provided. It is hard to assess if this was due to a technical fault of the system (for example, due to faulty tracking from the motion capture system), or to the fact that subjects were experiencing the condition with no feedback. Some subjects understood that the condition without any feedback was done on purpose, others confused it with a bug of the system. Some subjects also commented on the fact that shoes were not fitting their size. Moreover, some felt disable without the visual feedback. One subject observed that he simply ignored the feedback and walked straight. This is an indication of his unwillingness of suspending his disbelief, and behave in a way similar to how they would behave when walking on a real narrow plank [15]. Overall, observations of most of the subjects showed that they were walking carefully listening and feeling the feedback in order to successfully complete the task. It is hard to assess whether the lack of feedback was the condition subjects were exposed to, the fact that they were outside the plank or a fault of the system. Afterall, previous mentioned research has shown that subjects do not walk straight even when they think they do. Some of the test subjects were noticeably not walking straight, although in the post-experimental questionnaire they commented on a faulty system. Very few understood that the lack of feedback was provided intentionally. 7. CONCLUSION In this paper, we introduced a multimodal architecture whose goal is to simulate natural interactive walking in virtual environments. We present an experiment which assessed the role of auditory and haptic feedback, together with their combination, in helping subjects to complete the task of walking on a virtual rope. The experiment provided some indications that haptic feedback at the feet level is more useful than auditory feedback when balancing on a virtual plank. Moreover, most subjects behaved in the virtual world as if they would have behaved in the real world, i.e., by walking slowly and carefully to try not to fall from the plank. More experiments, however, are needed to achieve a better understanding of the role of the different modalities in helping navigation and balance control.

8. ACKNOWLEDGMENT The research leading to these results has received funding from the European Community s Seventh Framework Program under FET-Open grant agreement 222107 NIW - Natural Interactive Walking. 11 The authors would like to thank Vincent Hayward, Smilen Dimitrov and Amir Berrezag who built the sandals used in this experiment, and Jon Ram Pedersen and Kristina Daniliauskaite who collaborated in preliminary versions of the described experiment. 9. REFERENCES [1] A. Pelah and J. Koenderink, Editorial: Walking in real and virtual environments, ACM Transactions on Applied Perception (TAP), vol. 4, no. 1, p. 1, 2007. [2] J. Paradiso, K. Hsiao, and E. Hu, Interactive music for instrumented dancing shoes, in Proc. of the 1999 International Computer Music Conference, 1999, pp. 453 456. [3] A. Benbasat, S. Morris, and J. Paradiso, A wireless modular sensor architecture and its application in onshoe gait analysis, in Sensors, 2003. Proceedings of IEEE, vol. 2, 2003. [11] R. Nordahl, S. Serafin, and L. Turchet, Sound synthesis and evaluation of interactive footsteps for virtual reality applications, in Proc. IEEE VR 2010, 2010. [12] R. Nordahl, A. Berrezag, S. Dimitrov, L. Turchet, V. Hayward, and S. Serafin, Preliminary experiment combining virtual reality haptic shoes and audio synthesis, in Proc. Eurohaptics, 2010. [13] S. Serafin, L. Turchet, R. Nordahl,, S. Dimitrov, A. Berrezag, and V. Hayward, Identification of virtual grounds using virtual reality haptic shoes and sound synthesis, in Proc. Eurohaptics symposium on Haptics and Audio-visual environments, 2010. [14] L. Turchet, R. Nordahl, and S. Serafin, Examining the role of context in the recognition of walking sound, in Proc. of Sound and Music Computing Conference, 2010. [15] M. Slater, Place illusion and plausibility can lead to realistic behaviour in immersive virtual environments, Philosophical Transactions of the Royal Society B: Biological Sciences, vol. 364, no. 1535, p. 3549, 2009. [4] M. Frey, CabBoots: shoes with integrated guidance system, in Proceedings of the 1st international conference on Tangible and embedded interaction. ACM, 2007, pp. 245 246. [5] Y. Takeuchi, Gilded gait: reshaping the urban experience with augmented footsteps, in Proceedings of the 23nd annual ACM symposium on User interface software and technology. ACM, 2010, pp. 185 188. [6] J. Souman, I. Frissen, M. Sreenivasa, and M. Ernst, Walking straight into circles, Current Biology, vol. 19, no. 18, pp. 1538 1542, 2009. [7] Y. Visell, F. Fontana, B. Giordano, R. Nordahl, S. Serafin, and R. Bresin, Sound design and perception in walking interactions, International journal of humancomputer studies, vol. 67, no. 11, pp. 947 959, 2009. [8] R. Nordahl, S. Serafin, and L. Turchet, Sound synthesis and evaluation of interactive footsteps for virtual reality applications, in Virtual Reality Conference (VR), 2010 IEEE. IEEE, 2010, pp. 147 153. [9] L. Turchet, R. Nordahl, A. Berrezag, S. Dimitrov, V. Hayward, and S. Serafin, Audio-haptic physically based simulation of walking sounds, in Proc. of IEEE International Workshop on Multimedia Signal Processing, 2010. [10] L. Turchet, S. Serafin, S. Dimitrov, and R. Nordahl, Physically based sound synthesis and control of footsteps sounds, in Proceedings of Digital Audio Effects Conference, 2010. 11 Natural Interactive Walking Project: www.niwproject.eu