Open Archive TOULOUSE Archive Ouverte (OATAO)

Similar documents
X11 in Virtual Environments ARL

Interactions and Applications for See- Through interfaces: Industrial application examples

synchrolight: Three-dimensional Pointing System for Remote Video Communication

Mobile Multi-Display Environments

Ethereal Planes: A Design Framework for 2D Information Spaces in 3D Mixed Reality Environments

VolGrab: Realizing 3D View Navigation by Aerial Hand Gestures

From Room Instrumentation to Device Instrumentation: Assessing an Inertial Measurement Unit for Spatial Awareness

Mario Romero 2014/11/05. Multimodal Interaction and Interfaces Mixed Reality

Enabling Remote Proxemics through Multiple Surfaces

This is an author-deposited version published in: Eprints ID: 9445

DESIGN OF AN AUGMENTED REALITY

NTT DOCOMO Technical Journal. 1. Introduction. 2. Process of Popularizing Glasses-Type Devices

Mixed / Augmented Reality in Action

Augmented reality as an aid for the use of machine tools

Ubiquitous Home Simulation Using Augmented Reality

Reflecting on Domestic Displays for Photo Viewing and Sharing

AUTOM AT ICS: Research activities on Automation

Collaboration on Interactive Ceilings

PhonePaint: Using Smartphones as Dynamic Brushes with Interactive Displays

The Mixed Reality Book: A New Multimedia Reading Experience

Virtual Co-Location for Crime Scene Investigation and Going Beyond

Wayfinding with Simulated Prosthetic Vision: Performance comparison with regular and structure-enhanced renderings

Gradual Engagement: Facilitating Information Exchange between Digital Devices as a Function of Proximity

An Immersive, Interactive and Augmented Classroom: A Proof-of-Concept

Occlusion-Aware Menu Design for Digital Tabletops

VR4D: An Immersive and Collaborative Experience to Improve the Interior Design Process

Geo-Located Content in Virtual and Augmented Reality

ActivityDesk: Multi-Device Configuration Work using an Interactive Desk

New interface approaches for telemedicine

Integration of Hand Gesture and Multi Touch Gesture with Glove Type Device

Paint with Your Voice: An Interactive, Sonic Installation

The whole of science is nothing more than a refinement of everyday thinking. Albert Einstein,

Virtual Reality Calendar Tour Guide

AUGMENTED REALITY AS AN AID FOR THE USE OF MACHINE TOOLS

Social and Spatial Interactions: Shared Co-Located Mobile Phone Use

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

Interior Design with Augmented Reality

COMET: Collaboration in Applications for Mobile Environments by Twisting

PRESS RELEASE EUROSATORY 2018

FlexAR: A Tangible Augmented Reality Experience for Teaching Anatomy

Mobile and Pervasive Game Technologies. Joel Ross ICS 62 05/19/2011

CREATING TOMORROW S SOLUTIONS INNOVATIONS IN CUSTOMER COMMUNICATION. Technologies of the Future Today

INTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces

Enhancing Shipboard Maintenance with Augmented Reality

Determining Optimal Player Position, Distance, and Scale from a Point of Interest on a Terrain

Using Mixed Reality as a Simulation Tool in Urban Planning Project for Sustainable Development

Enhancing industrial processes in the industry sector by the means of service design

EnhancedTable: Supporting a Small Meeting in Ubiquitous and Augmented Environment

DiamondTouch SDK:Support for Multi-User, Multi-Touch Applications

Interactions in a Human-Scale Immersive Environment: the CRAIVE- Lab

Knowledge Acquisition and Representation in Facility Management

Studying Depth in a 3D User Interface by a Paper Prototype as a Part of the Mixed Methods Evaluation Procedure

Future Directions for Augmented Reality. Mark Billinghurst

Beyond: collapsible tools and gestures for computational design

Activity-Centric Configuration Work in Nomadic Computing

Our research seeks to design an office

To link to this article : DOI : /ISTC URL :

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Beyond Actuated Tangibles: Introducing Robots to Interactive Tabletops

immersive visualization workflow

Augmented reality for underwater activities with the use of the DOLPHYN

Chapter 1 - Introduction

Capacitive Face Cushion for Smartphone-Based Virtual Reality Headsets

Towards Wearable Gaze Supported Augmented Cognition

StreetGamez: A Moving Projector Platform for Projected Street Games

COLLABORATION WITH TANGIBLE AUGMENTED REALITY INTERFACES.

Augmented and Virtual Reality 6.S063 Engineering Interaction Technologies. Prof. Stefanie Mueller MIT CSAIL HCI Engineering Group

Augmented Reality And Ubiquitous Computing using HCI

Organic UIs in Cross-Reality Spaces

Occlusion based Interaction Methods for Tangible Augmented Reality Environments

Tutorial: The Web of Things

EMPOWERING THE CONNECTED FIELD FORCE WORKER WITH ADVANCED ANALYTICS MATTHEW SHORT ACCENTURE LABS

Virtual Tactile Maps

Re-build-ing Boundaries: The Roles of Boundaries in Mixed Reality Play

Effective Iconography....convey ideas without words; attract attention...

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /

Virtual Reality for Real Estate a case study

REPORT ON THE CURRENT STATE OF FOR DESIGN. XL: Experiments in Landscape and Urbanism

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data

YOUR PRODUCT IN 3D. Scan and present in Virtual Reality, Augmented Reality, 3D. SCANBLUE.COM

Touch & Gesture. HCID 520 User Interface Software & Technology

Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances

Two-Handed Interactive Menu: An Application of Asymmetric Bimanual Gestures and Depth Based Selection Techniques

This is an author-deposited version published in : Eprints ID : 18192

Visualizing the future of field service

Transporters: Vision & Touch Transitive Widgets for Capacitive Screens

Abstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction

Vocabulary Game Using Augmented Reality Expressing Elements in Virtual World with Objects in Real World

Lab 5: Advanced camera handling and interaction

LOOKING AHEAD: UE4 VR Roadmap. Nick Whiting Technical Director VR / AR

Spatial augmented reality to enhance physical artistic creation.

DESIGN FOR INTERACTION IN INSTRUMENTED ENVIRONMENTS. Lucia Terrenghi*

Extending Ambient Intelligence to the Internet of Things: New Challenges for QoC Management

Development of a telepresence agent

Multi-Modal User Interaction

The Ubiquitous Lab Or enhancing the molecular biology research experience

Promoting citizen-based services through local cultural partnerships

AR-Enhanced Human-Robot-Interaction Methodologies Algorithms

Getting started with AutoCAD mobile app. Take the power of AutoCAD wherever you go

Transcription:

Open Archive TOULOUSE Archive Ouverte (OATAO) OATAO is an open access repository that collects the work of Toulouse researchers and makes it freely available over the web where possible. This is an author-deposited version published in : http://oatao.univ-toulouse.fr/ Eprints ID : 15344 The contribution was presented at MobileHCI 2015 : http://mobilehci.acm.org/2015/ To cite this version : Serrano, Marcos and Ens, Barrett and Yang, Xing-Dong and Irani, Pourang Desktop-Gluey: Augmenting Desktop Environments with Wearable Devices. (2015) In: 17th International Conference on Human- Computer Interaction with Mobile Devices and Services Adjunct (ACM MobileHCI 2015), 24 August 2015 (Copenhagen, Denmark). Any correspondence concerning this service should be sent to the repository administrator: staff-oatao@listes-diff.inp-toulouse.fr

Desktop-Gluey: Augmenting Desktop Environments with Wearable Devices Marcos Serrano University of Toulouse - IRIT Toulouse, France marcos.serrano@irit.fr Barrett Ens University of Manitoba Winnipeg, MB, Canada bens@cs.umanitoba.ca Xing-Dong Yang Dartmouth College Hanover, NH, USA xing-dong.yang@dartmouth.edu Pourang Irani University of Manitoba Winnipeg, MB, Canada irani@cs.umanitoba.ca Abstract Upcoming consumer-ready head-worn displays (HWDs) can play a central role in unifying the interaction experience in Distributed display environments (DDEs). We recently implemented Gluey, a HWD system that glues together the input mechanisms across a display ecosystem to facilitate content migration and seamless interaction across multiple, co-located devices. Gluey can minimize device switching costs, opening new possibilities and scenarios for multi-device interaction. In this paper, we propose Desktop-Gluey, a system to augment situated desktop environments, allowing users to extend the physical displays in their environment, organize information in spatial layouts, and carry desktop content with them. We extend this metaphor beyond the desktop to provide anywhere and anytime support for mobile and collaborative interactions. Figure 1. Copying a sketch from a desktop screen and pasting it to a tablet with Gluey [5] Author Keywords Head-Worn Display, distributed displays, multi-display environments, input redirection, content migration. Introduction A new generation of lightweight, see-through headworn displays (hereafter referred to as HWDs ) is emerging for general purpose use (e.g. Google Glass,

Meta, Microsoft HoloLens). These wearable devices will soon co-exist within a larger environment of distributed displays (i.e. desktop monitors, tablets, smartphones) we depend on for daily information tasks. However, distributed display environments (DDEs) are commonly afflicted by device switching and data transfer costs. To this end we implemented Gluey [5], a system that acts as a glue to facilitate seamless information flow and input redirection across multiple devices (Figure 1). The HWD embedded cameras and spatial sensors allow Gluey to discover and maintain a spatial model of the relative positions of various devices, which can be used to facilitate cross-device interactions. Gluey allows users to move content across devices and to freely interact with other displays using any available input device. We envision that Gluey will soon become the new desktop environment, in which, with only a mouse and a keyboard, a user can create a desktop anywhere and anytime by using available mobile or wearable devices. Gluey Gluey [5] exploits the unique features of HWDs such as view-fixed displays, cameras and inclusion of spatial sensors. From these sensors we can determine a user s head position in relation to the environment and from the camera data, we can reveal what the user is viewing. For example by simply moving a smartphone into view of the HWD s camera, the user can link his desktop keyboard and the smartphone. Further, we can use the device s view-fixed display as an alwaysavailable canvas for showing visual feedback about interactions. For example, this display can show a visible clipboard space to store multiple data objects in transit between copy/paste operations to multiple interleaved destinations. Gluey pairs input devices with displays to provide a unified interaction experience in the DDE: the Gluey user can use any input device, such as a keyboard, mouse, mobile touchscreen or mid-air finger gesture to control multiple displays. We implemented Gluey on an Epson Moverio BT-100 head-worn display equipped with a Logitech C270 HD webcam. Our prototype tracks the position of surrounding devices with the marker-based ARToolkitPlus. We set the HWD background color to black to maintain display transparency. Each input device is associated with only one display at any time. Desktop-Gluey Scenario Our existing Gluey prototype illustrates how HWDs can leverage the interaction experience in DDEs. In our current work, we envision Desktop-Gluey, which extends beyond the physical displays in a desktop environment. In this novel concept, users can organize information in spatial layouts and carry desktop content anywhere. Here we illustrate the capabilities of Desktop-Gluey through the following usage scenario. John, an architect, relies on numerous devices while switching between various spatial locations in his daily work: office, constructions, client s workplace, public transports and home. With Desktop-Gluey, John can seamlessly carry his virtual desktop with him, extending available displays (tablets, smartphones) and utilizing any available input device for interaction. Extending Physical Displays In a traditional desktop setup, Gluey can extend the physical displays with virtual windows, which can

display additional contents (Figure 2). As device resolution and FoV continue to increase, these virtual windows can eventually replace the real monitors and reduce desktop clutter. The virtual windows can even have larger virtual footprints than their physical counterparts. More interestingly, the user can carry such displays and use the desktops wherever needed. Figure 3. Spatial arrangement of virtual windows. Figure 2. Extending physical displays with virtual windows. Spatial Information Layouts Using Desktop-Gluey, virtual content can be spread out beyond the boundaries of desktop screens in specific spatial arrangements (Figure 3), similar to the Personal Cockpit interface [1]. For instance, the user can always choose to place a calendar, email client, and a contact list in three different spatial locations, each endowed with specific semantic association to minimize the efforts in content relocation and window management. Of course, one aspect that needs further investigation concerns placing such windows ergonomically to allow the user to view them with great conform and striking a balance between display size, location and reachability to facilitate direct interaction with the displays. Mobile Desktop With Desktop-Gluey, the physical desktop now becomes mobile. All the user needs is a keyboard and mouse. In situations where no input device is available, the user can employ any device with him/her, such as a tablet or a smartphone (Figure 4). Even if there is no personal device at hand, the user could still interact with Desktop-Gluey using on-body gestures (e.g. [4]). Figure 4. A mobile desktop allows the user to take their applications and content wherever they may go. Collocated Desktops Current desktop environment are meant to be used by a single user: collocated interaction can only take place around one person s desktop. As such, only one user can interact with the desktop with others having only

minor participation in any collaborative work. Desktop- Gluey could provide a unified collocated experience by allowing multiple users to interact with a single shared virtual desktop. For instance, a professor and a student working on a paper could share documents seamlessly by using virtual windows residing on the desk surface. Moreover, each user could interact with these documents using their own input devices (Figure 5). other (corner-to-corner), the windows could be arranged in an L layout (Figure 6). As such, windows would be oriented in the most optimal locations for both users to interact with content. Figure 6. F-formations, modified from [3] to show corresponding spatial layouts for virtual desktop windows. Conclusion Our proposal explores how to augment current desktop environments using wearable ecosystems. We propose Desktop-Gluey to expand beyond physical desktop displays, to organize information in spatial layouts, and to support mobility and collaboration. Figure 5. Collocated interaction around a virtual desktop. The proximity [3] of the user can be used to define the spatial properties and the content of the collocated virtual windows. For instance, as a user exits from the public space (1.2m) to personal (0.45m), the virtual windows become smaller and display personal content. In the same manner the f-formations, i.e. the relative body orientation among users [3], could be mapped to different spatial layouts of a virtual desktop [1]. For instance, when two users are facing each other (faceto-face), virtual windows could be placed on their side. When the users are facing the same direction (side-byside), the windows could be situated in front of them. Finally, when the users are facing perpendicular to each References [1] Ens, B., Finnegan, R, and Irani, P. 2014. The personal cockpit: a spatial interface for effective task switching on head-worn displays. Proc. CHI '14. [2] Ens, B., Hincapié-Ramos, J-D., and Irani, P. 2014. Ethereal planes: a design framework for 2D information space in 3D mixed reality environments. Proc. SUI '14. [3] Marquardt, Nicolai, Ken Hinckley, and Saul Greenberg. "Cross-device interaction via micro-mobility and f-formations." Proc. UIST 13. [4] Serrano, M., Ens, B., and Irani, P. 2014. Exploring the use of hand-to-face input for interacting with headworn displays. Proc. CHI '14. [5] Serrano, M., Ens, B., Yang, X-D., and Irani, P. 2015. Gluey: Developing a Head-Worn Display Interface to Unify the Interaction Experience in Distributed Display Environments. Proc. MobileHCI 15.