3D and Sequential Representations of Spatial Relationships among Photos

Similar documents
synchrolight: Three-dimensional Pointing System for Remote Video Communication

Re-build-ing Boundaries: The Roles of Boundaries in Mixed Reality Play

Augmented Reality And Ubiquitous Computing using HCI

Tangible Bits: Towards Seamless Interfaces between People, Bits and Atoms

FOCAL LENGTH CHANGE COMPENSATION FOR MONOCULAR SLAM

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

Steady Steps and Giant Leap Toward Practical Mixed Reality Systems and Applications

A Survey of Mobile Augmentation for Mobile Augmented Reality System

Enhanced Virtual Transparency in Handheld AR: Digital Magnifying Glass

Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience

Head-Movement Evaluation for First-Person Games

ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality

Gaze informed View Management in Mobile Augmented Reality

Novel Hemispheric Image Formation: Concepts & Applications

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations

Distributed Vision System: A Perceptual Information Infrastructure for Robot Navigation

Tableau Machine: An Alien Presence in the Home

Paint with Your Voice: An Interactive, Sonic Installation

MOBILE AUGMENTED REALITY FOR SPATIAL INFORMATION EXPLORATION

Lifelog-Style Experience Recording and Analysis for Group Activities

Annotation Overlay with a Wearable Computer Using Augmented Reality

AR Tamagotchi : Animate Everything Around Us

DESIGN STYLE FOR BUILDING INTERIOR 3D OBJECTS USING MARKER BASED AUGMENTED REALITY

Guidelines for Implementing Augmented Reality Procedures in Assisting Assembly Operations

An Efficient Nonlinear Filter for Removal of Impulse Noise in Color Video Sequences

Application of 3D Terrain Representation System for Highway Landscape Design

AUGMENTED REALITY FOR COLLABORATIVE EXPLORATION OF UNFAMILIAR ENVIRONMENTS

Introduction to Virtual Reality (based on a talk by Bill Mark)

Touch Your Way: Haptic Sight for Visually Impaired People to Walk with Independence

The Mixed Reality Book: A New Multimedia Reading Experience

INTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT

AR 2 kanoid: Augmented Reality ARkanoid

Geo-Located Content in Virtual and Augmented Reality

X11 in Virtual Environments ARL

High Performance Imaging Using Large Camera Arrays

Realistic Visual Environment for Immersive Projection Display System

Interior Design using Augmented Reality Environment

Indexed Color. A browser may support only a certain number of specific colors, creating a palette from which to choose

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface

Gesture Recognition with Real World Environment using Kinect: A Review

Attention Meter: A Vision-based Input Toolkit for Interaction Designers

Multi-Modal User Interaction

Dynamic Designs of 3D Virtual Worlds Using Generative Design Agents

Movie 10 (Chapter 17 extract) Photomerge

Usability and Playability Issues for ARQuake

INTERIOR DESIGN USING AUGMENTED REALITY

Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data

DiamondTouch SDK:Support for Multi-User, Multi-Touch Applications

Light-Field Database Creation and Depth Estimation

INTERACTIVE 3D VIRTUAL HYDRAULICS Using virtual reality environments in teaching and research of fluid power systems and components

EnSight in Virtual and Mixed Reality Environments

The use of gestures in computer aided design

Future Directions for Augmented Reality. Mark Billinghurst

Image Extraction using Image Mining Technique

Automated Virtual Observation Therapy

Figure 1 HDR image fusion example

Mohammad Akram Khan 2 India

White Paper High Dynamic Range Imaging

International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN

Recent Progress on Wearable Augmented Interaction at AIST

The principles of CCTV design in VideoCAD

BLIPS perfectly aligned cleaned by any impurity Gently press the lens

Extremes of Social Visualization in Art

MULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT

Augmented Reality Applications for Nuclear Power Plant Maintenance Work

Augmented Reality: Its Applications and Use of Wireless Technologies

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

Computer simulator for training operators of thermal cameras

A Mixed Reality Approach to HumanRobot Interaction

Beyond Actuated Tangibles: Introducing Robots to Interactive Tabletops

Psychophysics of night vision device halo

An augmented-reality (AR) interface dynamically

A TELE-INSTRUCTION SYSTEM FOR ULTRASOUND PROBE OPERATION BASED ON SHARED AR TECHNOLOGY

Multi-User Interaction in Virtual Audio Spaces

rainbottles: gathering raindrops of data from the cloud

Content Based Image Retrieval Using Color Histogram

Intelligent interaction

PHYSICS. Chapter 35 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT

User Study on a Position- and Direction-aware Museum Guide using 3-D Maps and Animated Instructions

Forest Inventory System. User manual v.1.2

INTELLIGENT GUIDANCE IN A VIRTUAL UNIVERSITY

Augmented Reality From Science to Mass-Market Stefan Misslinger, metaio, Inc.

A Quality Watch Android Based Application for Monitoring Robotic Arm Statistics Using Augmented Reality

Augmented reality as an aid for the use of machine tools

The Control of Avatar Motion Using Hand Gesture

10.2 Images Formed by Lenses SUMMARY. Refraction in Lenses. Section 10.1 Questions

CSC 170 Introduction to Computers and Their Applications. Lecture #3 Digital Graphics and Video Basics. Bitmap Basics

Social Editing of Video Recordings of Lectures

Product Requirements Document

SIMGRAPH - A FLIGHT SIMULATION DATA VISUALIZATION WORKSTATION. Joseph A. Kaplan NASA Langley Research Center Hampton, Virginia

VR/AR Concepts in Architecture And Available Tools

Constructing Representations of Mental Maps

Indoor Positioning with a WLAN Access Point List on a Mobile Device

Photography is everywhere

Immersive Augmented Reality Display System Using a Large Semi-transparent Mirror

The Seamless Localization System for Interworking in Indoor and Outdoor Environments

DESIGN COLLABORATION FOR INTELLIGENT CONSTRUCTION MANAGEMENT IN MOBILIE AUGMENTED REALITY

Be aware that there is no universal notation for the various quantities.

Learning Based Interface Modeling using Augmented Reality

A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,

Transcription:

3D and Sequential Representations of Spatial Relationships among Photos Mahoro Anabuki Canon Development Americas, Inc. E15-349, 20 Ames Street Cambridge, MA 02139 USA mahoro@media.mit.edu Hiroshi Ishii MIT Media Laboratory E15-328, 20 Ames Street Cambridge, MA 02139 USA ishii@media.mit.edu Abstract This paper proposes automatic representations of spatial relationships among photos for structure analysis and review of a photographic subject. Based on camera tracking, photos are shown in a 3D virtual reality space to represent global spatial relationships. At the same time, the spatial relationships between two of the photos are represented in slide show sequences. This proposal allows people to organize photos quickly in spatial representations with qualitative meaning. Keywords 3D visualization, slide show, structure analysis, photo organization, photomicrograph, camera tracking ACM Classification Keywords H5.1. [Information Interfaces and Presentation]: Multimedia Information Systems artificial, augmented, and virtual realities. Copyright is held by the author/owner(s). CHI 2006, April 22 27, 2006, Montréal, Québec, Canada. ACM 1-59593-298-4/06/0004. Introduction When people take several photos of a subject, they take them from different points of view. For example, a material researcher takes photomicrographs of one material with several magnification levels to analyze its structure. An architectural photographer takes still

photos of a house in every room in order to record its floor plan. Spatial relationships among such photos have important meaning for structure analysis and review of the photographic subject. However, these relationships are not always clear by viewing the photos. Photomicrographs with totally different magnification levels provide few visual cues to reveal the corresponding areas. Still photos in different rooms rarely include the same object. Instead of taking several photos, creating a video seems better for recording a structure of a subject because adjacent video images of one sequence would include the same visual area. Therefore additional effort is sometimes required for efficient analysis and review of the structure. When use of a video is not convenient, people take just several photos of one subject and make annotations or manual layouts to show the spatial relationships. However, these manual tasks are time-consuming, and the results cannot be easily converted to other formats. They also usually show only qualitative relationships. In this paper, we propose automatic representations of spatial relationships among photos based on camera tracking. In this proposal, spatial relationships are calculated from camera parameters (position, orientation, and focal length) at the time of shooting. Based on the calculation, various representations can be generated without time-consuming manual tasks. Since multiple representations allow for multiperspective analysis and review, we propose two simultaneous contrasting representations for one set of photos. One representation is a 3D virtual space that shows global spatial relationships among the photos Figure 1. 3D virtual space representing the global spatial relationship among photos. Zoom In Figure 2. Slide show (zoom in) representing the spatial relationship between two photos. (Figure 1). The other is a slide show sequence of two (or more) photos showing local spatial relationships among the photos (Figure 2). These representations show spatial relationships quantitatively so that they help people to analyze and review the subject in different ways from qualitative manually-created representations. For example, empty spaces in the 3D space can remind people of non-captured areas of a material that should be captured. The slide show will provide a sense of speed and direction, for example, a sense of walking in a house.

Related Work There are some previous works that represent spatial relationships among photos. Salient Stills is a technique that merges frame images of a movie into a high resolution still image using optical flow [4]. If the camera is moving, the resulting still image will represent a wider view than the camera image. If an object is moving, a trajectory of the object s motion will be represented in the resulting still image. This technique allows people to see multiple points of view and/or one sequence at a glance, so it has been used for surveillance. STAMP is a technique for constructing a pseudo-3d virtual space [3]. With STAMP, several photos are shown with corresponding areas overlapping one another. A user can switch a main photo with morphing animations that maintain the overlaps so as to provide pseudo spatial movements, which are helpful for understanding the spatial relationships among photos. This technique has been used for sequential directions on a web site. Salient Stills and STAMP have a common characteristic, which is that corresponding visual areas are required between photos, while the corresponding areas are not required in our proposal. Smith et al. proposed an educational application using a digital camera with a global positioning system (GPS) and digital compass to record its position and orientation [2]. A photo taken with this camera is shown with a past photo at the same location so that it allows students to see the historical changes. While they focused on the relationships between photos shot at different times at the same location, we are focusing on the relationships between photos shot at different locations. System Design Our proposal can be applied to various types of cameras, such as microscopes and still cameras. This section describes common system design issues. Camera parameter tracking For any camera, its parameters (position, orientation, point of focus, etc.) are tracked to calculate spatial relationships among the captured photos. In the case of a microscope, the system tracks the camera position relative to a microscope slide and magnification value. Because the camera (or the slide) might change its position slightly when high powered magnification is used, accurate camera position tracking is required. In the case of a still camera, we assume that the photographic subject is the point of focus of the camera. Therefore, the system tracks the camera position and orientation in real space, as well as its focal length. Because some photos may be taken at locations near each other, accurate camera tracking is needed for still cameras, as well as microscopes. At the same time, wide area camera tracking is also required because some photos may be taken at locations far from one another. Photo arranging in a 3D virtual space Captured photos are placed in a 3D virtual space automatically, based on camera tracking. A user can browse global spatial relationships among the photos by walking through the space. In the case of a microscope, one axis of the space is the magnification value. The photograph size in the space is decided based on the magnification value (Figure 3). All photos have the same orientation and show a kind of tree structure. In the case of a still camera, the 3D space

Magnification camera: p 1 and p 2 are position vectors and o 1, o 2 are orientation vectors of two photos. Figure 3. An example of the 3D virtual space in the case of microscope. corresponds to real space. The size of photos is decided according to the field of view. Selection of slide show effects When a user selects one photo and then selects another photo, the system represents the spatial relationship between the first one and the second one with a slide show effect. For example, when the second photo is a part of the first one, the second one appears from the corresponding point in the first one with a zoom in effect (Figure 2). When the second photo is next to the first one, the second one appears from the side and the first one disappears to the opposite side (Figure 4). The slide show continues if more photos are selected. Below are examples of calculations for selecting a slide show effect in the case of a still If o 1 o 2 and o 1 (p 2 - p 1 ) 0 and p 1, height p 2, height, the system assumes that two photos line up side by side and then slide in from side (left/right) effect is selected. If p 1 + k o 1 p 2 (k > 0) and o 1 o 2 / o 1 o 2 1, the system assumes that the first photo is in front of the second one and then box out effect is selected. Prototype Implementation For the first prototype, we implemented a system for still cameras. As we mentioned in the system design section, high quality camera tracking in real space is required. Since such tracking technologies have been proposed in the augmented reality (AR) research area [1], we decided to use a camera tracking method developed for AR systems. We used an AR software development kit (SDK) called MR Platform SDK [5] for camera tracking. This SDK allows us to track a camera in real space in real time Figure 4. Slide show (moving to right) representing the spatial relationship between two photos.

using a six-degree-of-freedom sensor (POLHEMUS FASTRAK). The sensor has a limited sensing area, but it has enough accuracy to distinguish spatial differences between shooting locations even if they are very close to one another. Because the SDK is designed for easy enhancement by users, other tracking methods, such as GPS, ultrasound sensor, inertia sensor, and computer vision technique can be used with the SDK. We will use these methods for camera tracking in future prototype systems. We attached a sensor receiver with a button to a CCD camera (WATEC WAT 221S + M96001l). The button is used for taking a photo. Because this camera has a fixed point of focus, we did not implement focus length tracking, using a fixed value as the focus length parameter in this prototype. The camera and the sensor are connected to a Linux PC. When the button is pressed, a photo is taken and sent to the PC. The sensor data is also sent to the PC at the same time. The photo and its position and orientation that are calculated using the sensor data are related to each other. Our software has four views on a PC display (Figure 5). The top left view shows the real-time camera view. The top right view shows a 3D virtual space that represents the global spatial relationship among the captured photos. The bottom left view shows the captured photo thumbnails in the order in which they were photographed. The bottom right view shows a selected photo. When another photo is selected, this view shows the spatial relationship between both the first and the second selected photos with a slide show effect. The prototype system has six effects. A suitable effect is selected through the flow in Figure 6. Figure 5. Our software has four views: the real-time camera view, the captured photo thumbnails view, the 3D virtual space view, and the selected photo view with slide show effects. Figure 6. A suitable slide show effect is selected through this flow.

Discussion The automatic photo indexing using camera parameters would be beneficial to those who organize several photos in spatial representations. Some material and biological researchers are looking forward to using our proposed system as they expect it will reduce the time it takes to make presentation materials with several photomicrographs. Tight coupling of photo shooting and photo indexing enables real-time representations of spatial relationships among photos. We expect real-time representations to indicate which areas should be captured at the next shooting. Spatial relationships can be represented in various formats. Each format has its own advantages and disadvantages. Our proposal attempts to enhance the advantages and reduce the disadvantages by computer-aided multi-format data representations. Our proposed representations include empty areas/time in some cases. When the captured photos are far from one another, our proposed 3D virtual space is sparse. When a user selects two photos that are far from each other, to represent the distance between them the system may show nothing for a while in a sequence. But this emptiness has some meaning because it is based on actual qualitative spatial relationships. We believe this allows people to feel the space beyond the captured photos. We think it is a kind of information extension. Future Work As we develop this project, we will expand the first prototype system for still cameras and implement another prototype system for microscopes. It is also our goal to investigate better representations. After implementation, we will evaluate how people interpret the proposed representations, as well as utilize our system. In addition, we will determine if our proposed representations of spatial representation among photos make the effects described in the discussion section. Acknowledgements This project has been funded by the Media Lab s Things That Think consortium. References [1] Azuma, R., Baillot, Y., Behringer, R., Feiner, S., Julier, S., and MacIntyre, B. Recent advances in augmented reality. IEEE Computer Graphics and Applications 21, 6 (2001), 34-47. [2] Smith, B. K., Blankinship, E., Ashford, A., Baker, M., and Hirzel, T. Inquiry with imagery: Historical archive retrieval with digital cameras. In Proc. MM 1999, ACM Press (1999), 405-408. [3] Spatio-Temporal Association with Multiple Photos (STAMP). http://home.csis.utokyo.ac.jp/~tanaka/stamp/stamp.html [4] Teodosio, L. and Bender, W. Salient stills. In ACM Trans. Multimedia Comput. Commun. Appl. 1, 1 (Feb. 2005), 16-36. [5] Uchiyama, S., Takemoto, K., Satoh, K., Yamamoto, H., and Tamura, H., MR Platform: a basic body on which mixed reality applications are built, In Proc. ISMAR 2002, IEEE Computer Society Press (2002), 246-253.