Evaluation of Guidance Systems in Public Infrastructures Using Eye Tracking in an Immersive Virtual Environment

Similar documents
Chapter 9. Conclusions. 9.1 Summary Perceived distances derived from optic ow

Simulation of Handicapped People Finding their Way through Transport Infrastructures

Virtual Reality Based Scalable Framework for Travel Planning and Training

AGENT PLATFORM FOR ROBOT CONTROL IN REAL-TIME DYNAMIC ENVIRONMENTS. Nuno Sousa Eugénio Oliveira

A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems

FP7 ICT Call 6: Cognitive Systems and Robotics

Perceptual Characters of Photorealistic See-through Vision in Handheld Augmented Reality

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

Application of 3D Terrain Representation System for Highway Landscape Design

Effective Iconography....convey ideas without words; attract attention...

What will the robot do during the final demonstration?

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

Reinventing movies How do we tell stories in VR? Diego Gutierrez Graphics & Imaging Lab Universidad de Zaragoza

Virtual Reality Calendar Tour Guide

3D User Interfaces. Using the Kinect and Beyond. John Murray. John Murray

Admin. Today: Designing for Virtual Reality VR and 3D interfaces Interaction design for VR Prototyping for VR

Service Cooperation and Co-creative Intelligence Cycle Based on Mixed-Reality Technology

Wi-Fi Fingerprinting through Active Learning using Smartphones

A Study on the Navigation System for User s Effective Spatial Cognition

Human Autonomous Vehicles Interactions: An Interdisciplinary Approach

Interaction Techniques for Immersive Virtual Environments: Design, Evaluation, and Application

Surface Contents Author Index

Immersive Simulation in Instructional Design Studios

Immersive Real Acting Space with Gesture Tracking Sensors

Spatial navigation in humans

Cognitive robots and emotional intelligence Cloud robotics Ethical, legal and social issues of robotic Construction robots Human activities in many

3D display is imperfect, the contents stereoscopic video are not compatible, and viewing of the limitations of the environment make people feel

AUGMENTED REALITY IN URBAN MOBILITY

Tracking. Alireza Bahmanpour, Emma Byrne, Jozef Doboš, Victor Mendoza and Pan Ye

Notes on a method of recording and analyzing sequences of urban space and color

CSE 165: 3D User Interaction. Lecture #11: Travel

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Re-build-ing Boundaries: The Roles of Boundaries in Mixed Reality Play

Mission-focused Interaction and Visualization for Cyber-Awareness!

Mid-term report - Virtual reality and spatial mobility

Waves Nx VIRTUAL REALITY AUDIO

CSC 2524, Fall 2017 AR/VR Interaction Interface

ENHANCED HUMAN-AGENT INTERACTION: AUGMENTING INTERACTION MODELS WITH EMBODIED AGENTS BY SERAFIN BENTO. MASTER OF SCIENCE in INFORMATION SYSTEMS

Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances

HandsIn3D: Supporting Remote Guidance with Immersive Virtual Environments

A Hybrid Immersive / Non-Immersive

3D interaction techniques in Virtual Reality Applications for Engineering Education

This list supersedes the one published in the November 2002 issue of CR.

MSMS Software for VR Simulations of Neural Prostheses and Patient Training and Rehabilitation

EYE MOVEMENT STRATEGIES IN NAVIGATIONAL TASKS Austin Ducworth, Melissa Falzetta, Lindsay Hyma, Katie Kimble & James Michalak Group 1

Immersive Guided Tours for Virtual Tourism through 3D City Models

An Approach to Semantic Processing of GPS Traces

Safe and Efficient Autonomous Navigation in the Presence of Humans at Control Level

Arcaid: Addressing Situation Awareness and Simulator Sickness in a Virtual Reality Pac-Man Game

INTELLIGENT GUIDANCE IN A VIRTUAL UNIVERSITY

AR 2 kanoid: Augmented Reality ARkanoid

Introduction to Foresight

School of Computer Science. Course Title: Introduction to Human-Computer Interaction Date: 8/16/11

Lecture 19: Depth Cameras. Kayvon Fatahalian CMU : Graphics and Imaging Architectures (Fall 2011)

ISCW 2001 Tutorial. An Introduction to Augmented Reality

ABSTRACT. A usability study was used to measure user performance and user preferences for

Distributed Vision System: A Perceptual Information Infrastructure for Robot Navigation

The International Encyclopedia of the Social and Behavioral Sciences, Second Edition

MELODIOUS WALKABOUT: IMPLICIT NAVIGATION WITH CONTEXTUALIZED PERSONAL AUDIO CONTENTS

The Amalgamation Product Design Aspects for the Development of Immersive Virtual Environments

Arup is a multi-disciplinary engineering firm with global reach. Based on our experiences from real-life projects this workshop outlines how the new

Ubiquitous Home Simulation Using Augmented Reality

CS123. Programming Your Personal Robot. Part 3: Reasoning Under Uncertainty

Railway Training Simulators run on ESRI ArcGIS generated Track Splines

Touch Your Way: Haptic Sight for Visually Impaired People to Walk with Independence

How the Geometry of Space controls Visual Attention during Spatial Decision Making

DIFFERENCE BETWEEN A PHYSICAL MODEL AND A VIRTUAL ENVIRONMENT AS REGARDS PERCEPTION OF SCALE

Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback

Mission Space. Value-based use of augmented reality in support of critical contextual environments

Part I Introduction to the Human Visual System (HVS)

SITUATED DESIGN OF VIRTUAL WORLDS USING RATIONAL AGENTS

Digitalisation as day-to-day-business

Omni-Directional Catadioptric Acquisition System

Immersive Interaction Group

A Robust Neural Robot Navigation Using a Combination of Deliberative and Reactive Control Architectures

Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience

How do you teach AI the value of trust?

OBJECTIVE OF THE BOOK ORGANIZATION OF THE BOOK


Developing video games with cultural value at National Library of Lithuania

Interior Design using Augmented Reality Environment

Graz University of Technology (Austria)

Navigation of Transport Mobile Robot in Bionic Assembly System

User Interfaces in Panoramic Augmented Reality Environments

Why interest in visual perception?

Automotive Needs and Expectations towards Next Generation Driving Simulation

AUGMENTED REALITY IN VOLUMETRIC MEDICAL IMAGING USING STEREOSCOPIC 3D DISPLAY

Professor Aljosa Smolic SFI Research Professor of Creative Technologies

Technologies. Philippe Fuchs Ecole des Mines, ParisTech, Paris, France. Virtual Reality: Concepts and. Guillaume Moreau.

AN AUTONOMOUS SIMULATION BASED SYSTEM FOR ROBOTIC SERVICES IN PARTIALLY KNOWN ENVIRONMENTS

Computer Graphics. Spring April Ghada Ahmed, PhD Dept. of Computer Science Helwan University

Virtual Reality to Support Modelling. Martin Pett Modelling and Visualisation Business Unit Transport Systems Catapult

EXECUTIVE SUMMARY. St. Louis Region Emerging Transportation Technology Strategic Plan. June East-West Gateway Council of Governments ICF

iwindow Concept of an intelligent window for machine tools using augmented reality

SMART MANUFACTURING: A Competitive Necessity. SMART MANUFACTURING INDUSTRY REPORT Vol 1 No 1.

From Future Scenarios to Roadmapping A practical guide to explore innovation and strategy

Design and Application of Multi-screen VR Technology in the Course of Art Painting

Scholarly Article Review. The Potential of Using Virtual Reality Technology in Physical Activity Settings. Aaron Krieger.

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.

University of Geneva. Presentation of the CISA-CIN-BBL v. 2.3

Transcription:

Evaluation of Guidance Systems in Public Infrastructures Using Eye Tracking in an Immersive Virtual Environment Helmut Schrom-Feiertag 1, Christoph Schinko 2, Volker Settgast 3, and Stefan Seer 1 1 Austrian Institute of Technology, Vienna, Austria {helmut.schrom-feiertag,stefan.seer}@ait.ac.at 2 Institute of Computer Graphics and Knowledge Visualization, TU Graz, Austria christoph.schinko@fraunhofer.at 3 Fraunhofer Austria, Graz, Austria volker.settgast@fraunhofer.at Abstract. In this paper we present a novel method to evaluate guidance systems and navigation solutions for public infrastructures based on an immersive virtual environment in combination with a mobile eye tracking system. It opens new opportunities for wayfinding studies in public infrastructures already during the planning phase. Our approach embeds an interface for natural locomotion in the virtual environment offering significant advantages related to the user s spatial perception as well as physical and cognitive demands. Accurate measurements of position, locomotion and gaze within the virtual environment enable a great simplification for the analysis of eye tracking data. We conducted a study with participants that included virtual and real world scenarios within a train station. First results exhibit similar behaviour of participants in the real and virtual environment, confirming the comparability and applicability of our method. Keywords: Wayfinding, Immersive Virtual Environment, Eye Tracking, Gaze Analysis, Visual Attention. 1 Introduction Public transport infrastructures are becoming more multi-functional by serving as a short or long distance transportation hub and, for instance, as a shopping mall at the same time. For passengers and visitors, the increasing complexity of such infrastructures can lead to disorientation or subjective uncertainties. These aspects need to be fully taken into account already in the planning phase to prevent wrong design decisions which might result in higher costs in case they have to be changed later. To this end, it is important to understand how people interact with their environment and how different visual information sources influence the orientation and navigation behaviour., September 23, 2014, Vienna, Austria Copyright 2014 for the individual papers by the papers' authors. Copying permitted for private and academic purposes. This volume is published and copyrighted by its editors. 62

For such questions, eye tracking has become a popular method to investigate human wayfinding and to gain insights into perception and processing of visual information. Research on wayfinding has been conducted within lab and field studies, which has provided evidence for comparable eye-movements in experimental and real world scenarios [2]. This study also revealed that the freedom of immersed participants to move their head and body is of high importance. This paper presents our approach of combining an immersive virtual environment with an embedded interface for natural locomotion together with a mobile eye tracking system and demonstrates its applicability for wayfinding studies. 2 Immersive Virtual Environment The immersive virtual environment DAVE [4], as shown in Figure 1, allows to efficiently perform wayfinding studies with enhanced comparability of the results due to controlling and reproducing the testing conditions (e.g. time, passers-by, train and bus schedule, etc.) for each participant. By extending the functionality of this test environment with a mobile eye tracking system from SMI, we are able to test different scenarios with alternating visual information within complex infrastructures. Fig. 1. The DAVE, a four-sided immersive stereoscopic projection environment. Results in [7] show that body orientation and movement have a strong impact on the results of eye tracking studies in spatial navigation tasks. To enhance the awareness of movement and body perception within the virtual environment, it has been equipped with a Microsoft Kinect sensor for tracking the user s motion. The user has to walk in place to move forward and to turn the shoulders to invoke rotations in the virtual world. The outcome of [3] revealed the importance between head alignment and heading. In our observations, the head was consistently aligned with the current walking direction because the participants not only turned the shoulder but rather changed the complete body orientation into the walking direction. In [6] a survey of empirical studies on the perception of egocentric distances in virtual environments is made and the egocentric distance is reported to be 63

shorter with a mean about 74% of the modelled distances. The main influencing factors to facilitate a veridical spatial perception are binocular disparity, high quality graphics, a rich environment and an enhanced user s sense of presence. Of great importance is the participants feeling of their own body in the virtual environment. These requirements are fulfilled by the immersive virtual environment DAVE. The validity of our hands-free steering approach has been explored in [1] by conducting a case study with parallel test groups, exposing individuals to wayfinding exercises in the real world and the corresponding virtual world. The validation results showed that the perceived durations, egocentric distances and directions do not differ statistically significantly between the real and the virtual world. Maybe the use of stereoscopic shutter glasses instead of head mounted devices providing the perception of the own body in the virtual scene, instead of a virtual avatar, provides an enhanced sense of presence and therefore a better metric in the virtual environment leading to better results in the estimation of egocentric distances and directions. But there is still much unknown about presence and a promising field for further research on spatial perception in virtual environments. 3 Analysis Method Within the virtual environment DAVE all activities, movements and user interactions are recorded to be available for detailed analysis. Figure 2 shows a participant performing an experiment in the virtual environment as well as the corresponding eye tracking video indicating the gaze point with an orange circle. Fig. 2. Participant in the virtual environment (left) and the gaze point as an overlay in the eye tracking video (right). Given the highly accurate data on position, body alignment and viewing direction of the participant in the virtual environment together with the gaze point from the eye tracking system, it is possible to compute the intersection between the viewing ray and the triangle mesh of the 3D model for each frame. The overall accuracy is a combination of the eye tracking glasses (according 64

to the manufacturer: 0.5 ) and the optical tracking. This enables to project an attention map onto the virtual 3D environment and furthermore to create saliency maps for visualizing the distribution of fixation in the environment as well as the amount of time one particular object was fixated. The aggregated intersections between the viewing rays of participants and the 3D mesh of the environment are shown in Figure 3. Fig. 3. Fixation points in the 3D model (left) and projected onto the ground plane (right). The evaluations are carried out according to the three levels of metric from [5] which are defined as follows: user task performance (time required, distance travelled, number of wrong decisions), physical behaviour (characteristics of locomotion, looking around, shape of trajectory) and decision making (viewed objects and signs identified via attention maps, thinking aloud, interviews). 4 Findings and Outlook The applicability of the method crucially relies on the similarity between the spatial perception in the virtual and in the real environment. In order to demonstrate the comparability and applicability of the method, we conducted a wayfinding study with two scenarios within the virtual and real world setting of the main train station in Vienna. We recruited different public transport users with specific demographic behaviour and experience. Each participant started on platform 12 and was assigned to walk first to the departure location of the tram line D (scenario 1) and afterwards to the location of the local bus line 13A (scenario 2). Both scenarios are aided wayfinding scenarios, classified after the taxonomy in [8], using external aids like signage and maps. Aided wayfinding using signs is considered to be rather simple since it does not require considerable cognitive effort from the user. Therefore it is important to provide all the relevant information at every decision point. Large public infrastructures such as the main train station in Vienna are very complex which makes the planning of the guidance system more difficult and error-prone. 65

The proposed method represents a new approach to support an effective evaluation of guidance systems in a goal oriented manner. The main advantage of this method lies in evaluating a multitude of different scenarios already in the planning phase avoiding costly mistakes. It greatly simplifies the data analysis, also reducing time and costs to reveal fixation, attention, saliency maps, objects of interest and route choices based on eye tracking measurements. Our approach provides a basis for an efficient evaluation of guidance systems and navigation solutions in public infrastructures and many additional application areas. The analysis of the recorded data from the virtual and the real environment is still ongoing. First results of the field studies already exhibit similar critical areas within the real and virtual world where participants showed similar behaviour facing the same challenges like insufficient signposting or misunderstanding of signs. The final results will provide unique insights into task based visual attention and wayfinding behaviour and will reveal the full potential of the method. Acknowledgments. This work is part of the project MOVING and is funded by the research program line ways2go within the framework of the Austrian strategic initiative IV2Splus Intelligent Transport Systems and Services plus under the project number 835733. The study was performed using SMI Eye Tracking devices. References 1. Bauer, D., Schneckenburger, J., Settgast, V., Millonig, A., Gartner, G.: Hands free steering in a virtual world for the evaluation of guidance systems in pedestrian infrastructures: design and validation. In: Transportation Research Board 92nd Annual Meeting. No. 13-1484 (2013) 2. Foulsham, T., Walker, E., Kingstone, A.: The where, what and when of gaze allocation in the lab and the natural environment. Vision Research 51(17), 1920 1931 (2011) 3. Hollands, M.A., Patla, A.E., Vickers, J.N.: look where you re going! : gaze behaviour associated with maintaining and changing the direction of locomotion. Experimental Brain Research 143(2), 221 230 (2002) 4. Lancelle, M., Settgast, V., Fellner, D.: Definitely affordable virtual environment. Proceedings of IEEE virtual reality. IEEE 1 (2009) 5. Lessels, Ruddle, R.A.: Three levels of metric for evaluating wayfinding. Presence Teleoperators Virtual Environments 15(6), 637 654 (2006) 6. Renner, R.S., Velichkovsky, B.M., Helmert, J.R.: The perception of egocentric distances in virtual environments - a review. ACM Comput. Surv. 46(2), 23:123:40 (2013) 7. Schwarzkopf, S., von Stülpnagel, R., Büchner, S.J., Konieczny, L.: What lab eye tracking tells us about wayfinding a comparison of stationary and mobile eye tracking in a large building scenario. In: Eye Tracking for Spatial Research, Proceedings of the 1st International Workshop (2013) 8. Wiener, J.M., Bchner, S.J., Hlscher, C.: Taxonomy of human wayfinding tasks: A knowledge-based approach. Spatial Cognition & Computation 9(2), 152 165 (2009) 66