ITS '14, Nov , Dresden, Germany
|
|
- Deborah Atkinson
- 5 years ago
- Views:
Transcription
1 3D Tabletop User Interface Using Virtual Elastic Objects Figure 1: 3D Interaction with a virtual elastic object Hiroaki Tateyama Graduate School of Science and Engineering, Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan, Takumi Kusano Graduate School of Science and Engineering, Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan, Takashi Komuro Graduate School of Science and Engineering, Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan, komuro@mail.saitama-u.ac.jp Abstractt In this paper, we propose a method to reduce the inconsistency between virtual and real spaces in manipulating a 3D virtual object with users fingers. When a user tries to hold a virtual object, fingers do not stop on the surface of the object and thrust into the object since virtual objects cannot give reaction force. We therefore try to prevent fingers from thrusting into a virtual object by letting the object deform or glide through the fingers. A virtual object is deformed by using a spring-based model and solving the equation of equilibrium. Whether the object glides through the fingers or not is determinedd by calculating resultant force added to the object and resultant force of static friction when the fingers touch the object. Based on these methods, we constructed a 3D tabletop interface that enables interaction with virtual objects with a greater sense of reality. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-partuses, contact the Owner/Author. Copyright is held by the owner/author(s). ITS '14, Nov , Dresden, Germany components of this work must be honored. For all other ACM /14/ /11. Author Keywords Virtual reality; 3D interaction; object manipulation; vision-based UI. ACM Classification Keywords H.5.2 [Information interfaces and presentation]: User Interface - Interaction styles Introduction With the progress of 3D displaying technology, research of interaction systems that allows users to manipulate 283
2 stereoscopic 3D virtual objects has been actively conducted. For example, an interaction system that allows the users to move a 3D virtual object on an auto-stereoscopic display by putting their hand close to the table [1], but the users cannot manipulate the object directly. On the other hand, systems that enable direct interaction with a virtual object have also been developed. Yoshida et al. has developed a system that can present 42 different viewpoint images without glasses and that enables interaction with a sense of reality [2]. Interaction with a greater sense of reality in a virtual space has been realized by projecting proper 3D images according to the users real-time viewpoint position and by moving virtual objects according to the users motion following physical laws [3]. In another way, a high-speed stereo camera can be used to enhance a sense of reality in interaction with virtual objects. A system that detects users fingers position with a high-speed stereo camera and that moves a virtual object linked to the position with little latency has been developed [4]. These systems enable interaction with virtual objects using users hand, but interaction is limited to simple manipulations such as pushing. On the other hand, there are systems that realize various manipulations such as holding and lifting [5, 6], which enables flexible and realistic interaction. However, there is a problem that virtual objects cannot give reaction force. When a user tries to hold a virtual object, for example, fingers do not stop on the surface of the object and thrust into the object. This causes inconsistency between the real space and the virtual space, and impairs the sense of reality. A solution for this problem has been proposed in the case of VR systems using a head-mounted display (HMD) [7, 8]. Users can control a virtual hand seen through the HMD with the movement of the users real hand. The systems stop the virtual fingers on the surface of an object before the fingers thrust into the object. However, this method can be used when the users real hand is invisible and only the virtual hand is visible to users. In this paper, we elasticize the object and let it deform or glide through the fingers depending on the force added to the objects to prevent the fingers from thrusting into a virtual object. Thereby, we reduce the inconsistency between virtual and real spaces in manipulating a 3D virtual object with users fingers. Also, we constructed a 3D tabletop interface based on the proposed method which enables interaction with virtual objects with a greater sense of reality. Virtual Object Deformation Virtual object deformation is a method to elasticize a virtual object and to let it deform along the shape of users fingers. In this study, we use cylindrical objects and apply a spring-based model. We consider only the top-view shape of the objects and ignore the height of the objects. Figure 2(a) shows the spring-based model that is used to represent a virtual object. A certain number of control points are placed on the circumference uniformly and each control points are connected to the neighboring points and to the central point with a spring. Since control points that compose a virtual object are closely-arranged when seen locally, they can be regarded as linearly connected points. Figure 2(b) shows approximated control points that are linearly connected. There are two cases when control points move. 284
3 (a) Cylindrical object (b) Linear approximation Figure 2: Spring-based model Control points are moved by a finger. Control points are moved by forces from neighboring springs. Let be the horizontal position of a control point ( = 1,2 ). When a control point is moved by a finger to the position, is determined by the following equation. = (1) Figure 3 shows the equilibrium of forces when control points are moved by forces from neighboring springs. Let be the natural length of vertical springs, the length of a vertical spring after deformation, the angle of a vertical spring, and, the spring constants of vertical and horizontal springs. The equation of the equilibrium of forces is written as follows. Figure 3: Equilibrium of forces Using the following relations, sin = (3) sin = (4) the equation can be transformed as follows. (2 + ) =0 (5) The positions of all control points can be calculated by solving a system of equations constructed from Eq. (1) and Eq. (5). It is necessary to deform a virtual object along the shape of fingers. Figure 4 shows the flow of the object deformation algorithm. ( sin sin ) + =0 (2) 285
4 4) Other control points are again moved based on the equation of the equilibrium of forces. 5) By repeating it until all the control points are outside fingers. 6) The result is presented to the user. With this algorithm, users can manipulate virtual objects without the fingers thrusting into the object. Figure 4: Object deformation algorithm The details of the algorithm are as follows. 1) When the system detects that some fingers are inside a virtual object, the nearest control point to each fingertip is moved to the position of the fingertip. 2) Other control points are moved based on the equation of the equilibrium of forces. 3) If the destination of the control points is inside the finger, the points are fixed to the nearest contour of the finger. Virtual Object Gliding Virtual object gliding is a method to let a virtual object glides through the fingers when the object is deformed to some extent. This prevents the object from overdeforming even when fingers thrust into a virtual object deeply. Whether the object glides through the fingers or not is determined by calculating resultant force added to the object and resultant force of static friction when the fingers touch the object. Let be the number of contact points between fingers and the object, the force added to a contact point, the resultant force of, the angle between and, and the friction coefficient. The condition to let the object glides through the fingers is as follows. > cos (6) Using this method, pushing manipulation and holding manipulation can be treated in a unified way. Figure 5 shows the forces applied to a virtual object when =2. 3D Tabletop Interface Based on the proposed methods, we constructed a 3D tabletop interface system that enables interaction with virtual objects. We modified the system developed by 286
5 Figure 5: Forces applied to a virtual object Kusano et al. [9] to construct our system. The system consists of an upwardly-placed multi-view autostereoscopic display and a camera installed above the display. The appearance of the system is shown in Figure 6. Users can interact with 3D virtual objects which are presented on the display. Hand regions are extracted from a camera image using color information and the system recognizes whether fingers are inside a virtual object. We implemented virtual object deformation and gliding methods to the system. The result of deformation and gliding is illustrated in Figure 7. The processing time was short enough and the system ran at 60 fps in real time. Conclusions In this study, we proposed a method to reduce the inconsistency between virtual and real spaces in manipulating 3D virtual objects. By elasticizing a virtual object, the system prevents users fingers from Figure 6: 3D tabletop interface Figure 7: Object deformation and gliding thrusting into the virtual object. Based on the approach, we constructed a 3D tabletop interface that enables interaction with virtual objects with a greater sense of reality. In future works, we will increase shapes of virtual objects such as cube and triangular pyramid, realize 287
6 more various operations such as pushing and lifting, and construct practical application systems. References 1. Kobayashi, K., Oikawa, M., Koike, T., Utsugi, K., Yamasaki, M., and Kitagawa, S. Character Interaction System with Autostereoscopic Display and Range Sensor. In Proc. 3DUI 2007, Yoshida, T., Kamuro, S., Minamizawa, K., Nii, H., and Tachi, S. RePro3D: Full-parallax 3D Display with Haptic Feedback using Retro-reflective Projection Technology. In Proc. ISVRI 2011, Benko, H., Jota, R., and Wilson, A. MirageTable: Freehand Interaction on a Projected Augmented Reality Tabletop. In Proc. CHI 2012, Niikura, T., and Komuro, T. 3D Touch Panel Interface Using an Autostereoscopic Display. In Proc. ITS 2012, Hilliges, O., Kim, D., Izadi, S., Weiss, M., and Wilson, A. HoloDesk: Direct 3D Interactions with a Situated See-Through Display. In Proc. CHI 2012, Lee, J., Olwal, A., Ishii, H., and Boulanger, C. SpaceTop: Integrating 2D and Spatial 3D Interactions in a See-through Desktop Environment. In Proc. CHI 2013, Prachyabured, M. and Borst, C.W. Dropping the ball: Releasing a virtual grasp. In Proc. 3DUI 2011, Burns, E., Razzaque, S., Panter, A.T., Whitton, M.C., McCallus, M.R. and Brooks Jr, F.P. The Hand is Slower than the Eye: A Quantitative Exploration of Visual Dominance over Proprioception. In Proc. VR 2005, Kusano, T., Niikura, T., and Komuro, T. A Virtually Tangible 3D Interaction System using an Autostereoscopic Display. In Proc. SUI 2013,
Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface
Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface Xu Zhao Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan sheldonzhaox@is.ics.saitamau.ac.jp Takehiro Niikura The University
More informationVolGrab: Realizing 3D View Navigation by Aerial Hand Gestures
VolGrab: Realizing 3D View Navigation by Aerial Hand Gestures Figure 1: Operation of VolGrab Shun Sekiguchi Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, 338-8570, Japan sekiguchi@is.ics.saitama-u.ac.jp
More informationLocalized Space Display
Localized Space Display EE 267 Virtual Reality, Stanford University Vincent Chen & Jason Ginsberg {vschen, jasong2}@stanford.edu 1 Abstract Current virtual reality systems require expensive head-mounted
More informationDevelopment of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane
Journal of Communication and Computer 13 (2016) 329-337 doi:10.17265/1548-7709/2016.07.002 D DAVID PUBLISHING Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane
More informationArcaid: Addressing Situation Awareness and Simulator Sickness in a Virtual Reality Pac-Man Game
Arcaid: Addressing Situation Awareness and Simulator Sickness in a Virtual Reality Pac-Man Game Daniel Clarke 9dwc@queensu.ca Graham McGregor graham.mcgregor@queensu.ca Brianna Rubin 11br21@queensu.ca
More information2 Outline of Ultra-Realistic Communication Research
2 Outline of Ultra-Realistic Communication Research NICT is conducting research on Ultra-realistic communication since April in 2006. In this research, we are aiming at creating natural and realistic communication
More informationInvestigating Gestures on Elastic Tabletops
Investigating Gestures on Elastic Tabletops Dietrich Kammer Thomas Gründer Chair of Media Design Chair of Media Design Technische Universität DresdenTechnische Universität Dresden 01062 Dresden, Germany
More informationRear-screen and kinesthetic vision 3D manipulator
Yang et al. Visualization in Engineering (217) 5:9 DOI 1.1186/s4327-17-47- RESEARCH Open Access Rear-screen and kinesthetic vision 3D manipulator Chao-Chung Yang *, Shih-Chung Jessy Kang, Hsiang-Wen Yang
More informationDo Stereo Display Deficiencies Affect 3D Pointing?
Do Stereo Display Deficiencies Affect 3D Pointing? Mayra Donaji Barrera Machuca SIAT, Simon Fraser University Vancouver, CANADA mbarrera@sfu.ca Wolfgang Stuerzlinger SIAT, Simon Fraser University Vancouver,
More informationHaptic Media Construction and Utilization of Human-harmonized "Tangible" Information Environment
Haptic Media Construction and Utilization of Human-harmonized "Tangible" Information Environment Susumu Tachi *1,*2, Kouta Minamizawa *1, Masahiro Furukawa *1, Charith Lasantha Fernando *1 *1 Keio University,
More informationTangible Lenses, Touch & Tilt: 3D Interaction with Multiple Displays
SIG T3D (Touching the 3rd Dimension) @ CHI 2011, Vancouver Tangible Lenses, Touch & Tilt: 3D Interaction with Multiple Displays Raimund Dachselt University of Magdeburg Computer Science User Interface
More informationApplication of 3D Terrain Representation System for Highway Landscape Design
Application of 3D Terrain Representation System for Highway Landscape Design Koji Makanae Miyagi University, Japan Nashwan Dawood Teesside University, UK Abstract In recent years, mixed or/and augmented
More informationT(ether): spatially-aware handhelds, gestures and proprioception for multi-user 3D modeling and animation
T(ether): spatially-aware handhelds, gestures and proprioception for multi-user 3D modeling and animation The MIT Faculty has made this article openly available. Please share how this access benefits you.
More informationOutput Devices - Visual
IMGD 5100: Immersive HCI Output Devices - Visual Robert W. Lindeman Associate Professor Department of Computer Science Worcester Polytechnic Institute gogo@wpi.edu Overview Here we are concerned with technology
More informationDevelopment of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane
Development of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane Makoto Yoda Department of Information System Science Graduate School of Engineering Soka University, Soka
More informationBeyond: collapsible tools and gestures for computational design
Beyond: collapsible tools and gestures for computational design The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published
More informationIllusion of Surface Changes induced by Tactile and Visual Touch Feedback
Illusion of Surface Changes induced by Tactile and Visual Touch Feedback Katrin Wolf University of Stuttgart Pfaffenwaldring 5a 70569 Stuttgart Germany katrin.wolf@vis.uni-stuttgart.de Second Author VP
More informationFeelable User Interfaces: An Exploration of Non-Visual Tangible User Interfaces
Feelable User Interfaces: An Exploration of Non-Visual Tangible User Interfaces Katrin Wolf Telekom Innovation Laboratories TU Berlin, Germany katrin.wolf@acm.org Peter Bennett Interaction and Graphics
More informationDepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface
DepthTouch: Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface Hrvoje Benko and Andrew D. Wilson Microsoft Research One Microsoft Way Redmond, WA 98052, USA
More informationThe Mixed Reality Book: A New Multimedia Reading Experience
The Mixed Reality Book: A New Multimedia Reading Experience Raphaël Grasset raphael.grasset@hitlabnz.org Andreas Dünser andreas.duenser@hitlabnz.org Mark Billinghurst mark.billinghurst@hitlabnz.org Hartmut
More informationSublimate: State-Changing Virtual and Physical Rendering to Augment Interaction with Shape Displays
Sublimate: State-Changing Virtual and Physical Rendering to Augment Interaction with Shape Displays Daniel Leithinger Sean Follmer Alex Olwal Samuel Luescher Akimitsu Hogge Jinha Lee Hiroshi Ishii MIT
More informationUngrounded Kinesthetic Pen for Haptic Interaction with Virtual Environments
The 18th IEEE International Symposium on Robot and Human Interactive Communication Toyama, Japan, Sept. 27-Oct. 2, 2009 WeIAH.2 Ungrounded Kinesthetic Pen for Haptic Interaction with Virtual Environments
More informationVirtual Experiments as a Tool for Active Engagement
Virtual Experiments as a Tool for Active Engagement Lei Bao Stephen Stonebraker Gyoungho Lee Physics Education Research Group Department of Physics The Ohio State University Context Cues and Knowledge
More informationCSC 2524, Fall 2017 AR/VR Interaction Interface
CSC 2524, Fall 2017 AR/VR Interaction Interface Karan Singh Adapted from and with thanks to Mark Billinghurst Typical Virtual Reality System HMD User Interface Input Tracking How can we Interact in VR?
More informationAir-filled type Immersive Projection Display
Air-filled type Immersive Projection Display Wataru HASHIMOTO Faculty of Information Science and Technology, Osaka Institute of Technology, 1-79-1, Kitayama, Hirakata, Osaka 573-0196, Japan whashimo@is.oit.ac.jp
More informationA Study of Direction s Impact on Single-Handed Thumb Interaction with Touch-Screen Mobile Phones
A Study of Direction s Impact on Single-Handed Thumb Interaction with Touch-Screen Mobile Phones Jianwei Lai University of Maryland, Baltimore County 1000 Hilltop Circle, Baltimore, MD 21250 USA jianwei1@umbc.edu
More informationOcclusion based Interaction Methods for Tangible Augmented Reality Environments
Occlusion based Interaction Methods for Tangible Augmented Reality Environments Gun A. Lee α Mark Billinghurst β Gerard J. Kim α α Virtual Reality Laboratory, Pohang University of Science and Technology
More informationFuture Directions for Augmented Reality. Mark Billinghurst
Future Directions for Augmented Reality Mark Billinghurst 1968 Sutherland/Sproull s HMD https://www.youtube.com/watch?v=ntwzxgprxag Star Wars - 1977 Augmented Reality Combines Real and Virtual Images Both
More informationImage Manipulation Interface using Depth-based Hand Gesture
Image Manipulation Interface using Depth-based Hand Gesture UNSEOK LEE JIRO TANAKA Vision-based tracking is popular way to track hands. However, most vision-based tracking methods can t do a clearly tracking
More informationFigure 1. The game was developed to be played on a large multi-touch tablet and multiple smartphones.
Capture The Flag: Engaging In A Multi- Device Augmented Reality Game Suzanne Mueller Massachusetts Institute of Technology Cambridge, MA suzmue@mit.edu Andreas Dippon Technische Universitat München Boltzmannstr.
More informationInteractive Multimedia Contents in the IllusionHole
Interactive Multimedia Contents in the IllusionHole Tokuo Yamaguchi, Kazuhiro Asai, Yoshifumi Kitamura, and Fumio Kishino Graduate School of Information Science and Technology, Osaka University, 2-1 Yamada-oka,
More informationTransporters: Vision & Touch Transitive Widgets for Capacitive Screens
Transporters: Vision & Touch Transitive Widgets for Capacitive Screens Florian Heller heller@cs.rwth-aachen.de Simon Voelker voelker@cs.rwth-aachen.de Chat Wacharamanotham chat@cs.rwth-aachen.de Jan Borchers
More informationFrom Room Instrumentation to Device Instrumentation: Assessing an Inertial Measurement Unit for Spatial Awareness
From Room Instrumentation to Device Instrumentation: Assessing an Inertial Measurement Unit for Spatial Awareness Alaa Azazi, Teddy Seyed, Frank Maurer University of Calgary, Department of Computer Science
More informationAutomated Virtual Observation Therapy
Automated Virtual Observation Therapy Yin-Leng Theng Nanyang Technological University tyltheng@ntu.edu.sg Owen Noel Newton Fernando Nanyang Technological University fernando.onn@gmail.com Chamika Deshan
More informationDevelopment of Informal Communication Environment Using Interactive Tiled Display Wall Tetsuro Ogi 1,a, Yu Sakuma 1,b
Development of Informal Communication Environment Using Interactive Tiled Display Wall Tetsuro Ogi 1,a, Yu Sakuma 1,b 1 Graduate School of System Design and Management, Keio University 4-1-1 Hiyoshi, Kouhoku-ku,
More informationAugmented Keyboard: a Virtual Keyboard Interface for Smart glasses
Augmented Keyboard: a Virtual Keyboard Interface for Smart glasses Jinki Jung Jinwoo Jeon Hyeopwoo Lee jk@paradise.kaist.ac.kr zkrkwlek@paradise.kaist.ac.kr leehyeopwoo@paradise.kaist.ac.kr Kichan Kwon
More informationIntroduction to Virtual Reality (based on a talk by Bill Mark)
Introduction to Virtual Reality (based on a talk by Bill Mark) I will talk about... Why do we want Virtual Reality? What is needed for a VR system? Examples of VR systems Research problems in VR Most Computers
More informationMixed Reality Approach and the Applications using Projection Head Mounted Display
Mixed Reality Approach and the Applications using Projection Head Mounted Display Ryugo KIJIMA, Takeo OJIKA Faculty of Engineering, Gifu University 1-1 Yanagido, GifuCity, Gifu 501-11 Japan phone: +81-58-293-2759,
More informationPractical Data Visualization and Virtual Reality. Virtual Reality VR Display Systems. Karljohan Lundin Palmerius
Practical Data Visualization and Virtual Reality Virtual Reality VR Display Systems Karljohan Lundin Palmerius Synopsis Virtual Reality basics Common display systems Visual modality Sound modality Interaction
More informationUsing Real Objects for Interaction Tasks in Immersive Virtual Environments
Using Objects for Interaction Tasks in Immersive Virtual Environments Andy Boud, Dr. VR Solutions Pty. Ltd. andyb@vrsolutions.com.au Abstract. The use of immersive virtual environments for industrial applications
More informationLOOKING AHEAD: UE4 VR Roadmap. Nick Whiting Technical Director VR / AR
LOOKING AHEAD: UE4 VR Roadmap Nick Whiting Technical Director VR / AR HEADLINE AND IMAGE LAYOUT RECENT DEVELOPMENTS RECENT DEVELOPMENTS At Epic, we drive our engine development by creating content. We
More informationWearable Haptic Display to Present Gravity Sensation
Wearable Haptic Display to Present Gravity Sensation Preliminary Observations and Device Design Kouta Minamizawa*, Hiroyuki Kajimoto, Naoki Kawakami*, Susumu, Tachi* (*) The University of Tokyo, Japan
More informationHCI Outlook: Tangible and Tabletop Interaction
HCI Outlook: Tangible and Tabletop Interaction multiple degree-of-freedom (DOF) input Morten Fjeld Associate Professor, Computer Science and Engineering Chalmers University of Technology Gothenburg University
More informationUsing Scalable, Interactive Floor Projection for Production Planning Scenario
Using Scalable, Interactive Floor Projection for Production Planning Scenario Michael Otto, Michael Prieur Daimler AG Wilhelm-Runge-Str. 11 D-89013 Ulm {michael.m.otto, michael.prieur}@daimler.com Enrico
More informationFlexible Active Touch Using 2.5D Display Generating Tactile and Force Sensations
This is the accepted version of the following article: ICIC Express Letters 6(12):2995-3000 January 2012, which has been published in final form at http://www.ijicic.org/el-6(12).htm Flexible Active Touch
More informationVR based HCI Techniques & Application. November 29, 2002
VR based HCI Techniques & Application November 29, 2002 stefan.seipel@hci.uu.se What is Virtual Reality? Coates (1992): Virtual Reality is electronic simulations of environments experienced via head mounted
More informationIntegrated Driving Aware System in the Real-World: Sensing, Computing and Feedback
Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback Jung Wook Park HCI Institute Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA, USA, 15213 jungwoop@andrew.cmu.edu
More informationDevelopment of Video Chat System Based on Space Sharing and Haptic Communication
Sensors and Materials, Vol. 30, No. 7 (2018) 1427 1435 MYU Tokyo 1427 S & M 1597 Development of Video Chat System Based on Space Sharing and Haptic Communication Takahiro Hayashi 1* and Keisuke Suzuki
More informationRealistic Visual Environment for Immersive Projection Display System
Realistic Visual Environment for Immersive Projection Display System Hasup Lee Center for Education and Research of Symbiotic, Safe and Secure System Design Keio University Yokohama, Japan hasups@sdm.keio.ac.jp
More informationHaptic Reproduction and Interactive Visualization of a Beating Heart Based on Cardiac Morphology
MEDINFO 2001 V. Patel et al. (Eds) Amsterdam: IOS Press 2001 IMIA. All rights reserved Haptic Reproduction and Interactive Visualization of a Beating Heart Based on Cardiac Morphology Megumi Nakao a, Masaru
More informationHamsaTouch: Tactile Vision Substitution with Smartphone and Electro-Tactile Display
HamsaTouch: Tactile Vision Substitution with Smartphone and Electro-Tactile Display Hiroyuki Kajimoto The University of Electro-Communications 1-5-1 Chofugaoka, Chofu, Tokyo 1828585, JAPAN kajimoto@kaji-lab.jp
More informationTouch Feedback in a Head-Mounted Display Virtual Reality through a Kinesthetic Haptic Device
Touch Feedback in a Head-Mounted Display Virtual Reality through a Kinesthetic Haptic Device Andrew A. Stanley Stanford University Department of Mechanical Engineering astan@stanford.edu Alice X. Wu Stanford
More informationThe Control of Avatar Motion Using Hand Gesture
The Control of Avatar Motion Using Hand Gesture ChanSu Lee, SangWon Ghyme, ChanJong Park Human Computing Dept. VR Team Electronics and Telecommunications Research Institute 305-350, 161 Kajang-dong, Yusong-gu,
More informationHead-Movement Evaluation for First-Person Games
Head-Movement Evaluation for First-Person Games Paulo G. de Barros Computer Science Department Worcester Polytechnic Institute 100 Institute Road. Worcester, MA 01609 USA pgb@wpi.edu Robert W. Lindeman
More informationBeyond Actuated Tangibles: Introducing Robots to Interactive Tabletops
Beyond Actuated Tangibles: Introducing Robots to Interactive Tabletops Sowmya Somanath Department of Computer Science, University of Calgary, Canada. ssomanat@ucalgary.ca Ehud Sharlin Department of Computer
More informationOmni-Directional Catadioptric Acquisition System
Technical Disclosure Commons Defensive Publications Series December 18, 2017 Omni-Directional Catadioptric Acquisition System Andreas Nowatzyk Andrew I. Russell Follow this and additional works at: http://www.tdcommons.org/dpubs_series
More informationSmartCanvas: A Gesture-Driven Intelligent Drawing Desk System
SmartCanvas: A Gesture-Driven Intelligent Drawing Desk System Zhenyao Mo +1 213 740 4250 zmo@graphics.usc.edu J. P. Lewis +1 213 740 9619 zilla@computer.org Ulrich Neumann +1 213 740 0877 uneumann@usc.edu
More informationStereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.
Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.
More informationMulti touch Vector Field Operation for Navigating Multiple Mobile Robots
Multi touch Vector Field Operation for Navigating Multiple Mobile Robots Jun Kato The University of Tokyo, Tokyo, Japan jun.kato@ui.is.s.u tokyo.ac.jp Figure.1: Users can easily control movements of multiple
More informationComparison of Haptic and Non-Speech Audio Feedback
Comparison of Haptic and Non-Speech Audio Feedback Cagatay Goncu 1 and Kim Marriott 1 Monash University, Mebourne, Australia, cagatay.goncu@monash.edu, kim.marriott@monash.edu Abstract. We report a usability
More informationEvaluation of a Soft-Surfaced Multi-touch Interface
Evaluation of a Soft-Surfaced Multi-touch Interface Anna Noguchi, Toshifumi Kurosawa, Ayaka Suzuki, Yuichiro Sakamoto, Tatsuhito Oe, Takuto Yoshikawa, Buntarou Shizuki, and Jiro Tanaka University of Tsukuba,
More informationExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality
ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality The MIT Faculty has made this article openly available. Please share how this access benefits you. Your
More informationA Multimodal Locomotion User Interface for Immersive Geospatial Information Systems
F. Steinicke, G. Bruder, H. Frenz 289 A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems Frank Steinicke 1, Gerd Bruder 1, Harald Frenz 2 1 Institute of Computer Science,
More informationVR-programming. Fish Tank VR. To drive enhanced virtual reality display setups like. Monitor-based systems Use i.e.
VR-programming To drive enhanced virtual reality display setups like responsive workbenches walls head-mounted displays boomes domes caves Fish Tank VR Monitor-based systems Use i.e. shutter glasses 3D
More informationToward an Augmented Reality System for Violin Learning Support
Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp
More informationHMD based VR Service Framework. July Web3D Consortium Kwan-Hee Yoo Chungbuk National University
HMD based VR Service Framework July 31 2017 Web3D Consortium Kwan-Hee Yoo Chungbuk National University khyoo@chungbuk.ac.kr What is Virtual Reality? Making an electronic world seem real and interactive
More informationHigh-Speed Video Analysis of Two-Dimensional Movement of Objects onto Fine Beads
High-Speed Video Analysis of Two-Dimensional Movement of Objects onto Fine Beads Minoru SATO minoru@tokai.ac.jp School of Science, Tokai University, 4-1-1 Kitakaname, Hiratsuka, Kanagawa, 259-1292, Japan.
More informationVIRTUAL FIGURE PRESENTATION USING PRESSURE- SLIPPAGE-GENERATION TACTILE MOUSE
VIRTUAL FIGURE PRESENTATION USING PRESSURE- SLIPPAGE-GENERATION TACTILE MOUSE Yiru Zhou 1, Xuecheng Yin 1, and Masahiro Ohka 1 1 Graduate School of Information Science, Nagoya University Email: ohka@is.nagoya-u.ac.jp
More informationTheory and Practice of Tangible User Interfaces Tuesday, Week 9
Augmented Reality Theory and Practice of Tangible User Interfaces Tuesday, Week 9 Outline Overview Examples Theory Examples Supporting AR Designs Examples Theory Outline Overview Examples Theory Examples
More informationPinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data
Pinch-the-Sky Dome: Freehand Multi-Point Interactions with Immersive Omni-Directional Data Hrvoje Benko Microsoft Research One Microsoft Way Redmond, WA 98052 USA benko@microsoft.com Andrew D. Wilson Microsoft
More informationINTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT
INTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT TAYSHENG JENG, CHIA-HSUN LEE, CHI CHEN, YU-PIN MA Department of Architecture, National Cheng Kung University No. 1, University Road,
More informationVirtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback
Virtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback Taku Hachisu The University of Electro- Communications 1-5-1 Chofugaoka, Chofu, Tokyo 182-8585, Japan +81 42 443 5363
More informationA Gesture-Based Interface for Seamless Communication between Real and Virtual Worlds
6th ERCIM Workshop "User Interfaces for All" Long Paper A Gesture-Based Interface for Seamless Communication between Real and Virtual Worlds Masaki Omata, Kentaro Go, Atsumi Imamiya Department of Computer
More informationToward Principles for Visual Interaction Design for Communicating Weight by using Pseudo-Haptic Feedback
Toward Principles for Visual Interaction Design for Communicating Weight by using Pseudo-Haptic Feedback Kumiyo Nakakoji Key Technology Laboratory SRA Inc. 2-32-8 Minami-Ikebukuro, Toshima, Tokyo, 171-8513,
More informationTest of pan and zoom tools in visual and non-visual audio haptic environments. Magnusson, Charlotte; Gutierrez, Teresa; Rassmus-Gröhn, Kirsten
Test of pan and zoom tools in visual and non-visual audio haptic environments Magnusson, Charlotte; Gutierrez, Teresa; Rassmus-Gröhn, Kirsten Published in: ENACTIVE 07 2007 Link to publication Citation
More informationSubjective Image Quality Assessment of a Wide-view Head Mounted Projective Display with a Semi-transparent Retro-reflective Screen
Subjective Image Quality Assessment of a Wide-view Head Mounted Projective Display with a Semi-transparent Retro-reflective Screen Duc Nguyen Van 1 Tomohiro Mashita 1,2 Kiyoshi Kiyokawa 1,2 and Haruo Takemura
More informationColumn-Parallel Architecture for Line-of-Sight Detection Image Sensor Based on Centroid Calculation
ITE Trans. on MTA Vol. 2, No. 2, pp. 161-166 (2014) Copyright 2014 by ITE Transactions on Media Technology and Applications (MTA) Column-Parallel Architecture for Line-of-Sight Detection Image Sensor Based
More informationX11 in Virtual Environments ARL
COMS W4172 Case Study: 3D Windows/Desktops 2 Steven Feiner Department of Computer Science Columbia University New York, NY 10027 www.cs.columbia.edu/graphics/courses/csw4172 February 8, 2018 1 X11 in Virtual
More informationVIRTUAL REALITY Introduction. Emil M. Petriu SITE, University of Ottawa
VIRTUAL REALITY Introduction Emil M. Petriu SITE, University of Ottawa Natural and Virtual Reality Virtual Reality Interactive Virtual Reality Virtualized Reality Augmented Reality HUMAN PERCEPTION OF
More informationRobot Control Using Natural Instructions Via Visual and Tactile Sensations
Journal of Computer Sciences Original Research Paper Robot Control Using Natural Instructions Via Visual and Tactile Sensations Takuya Ikai, Shota Kamiya and Masahiro Ohka Department of Complex Systems
More informationExpression of 2DOF Fingertip Traction with 1DOF Lateral Skin Stretch
Expression of 2DOF Fingertip Traction with 1DOF Lateral Skin Stretch Vibol Yem 1, Mai Shibahara 2, Katsunari Sato 2, Hiroyuki Kajimoto 1 1 The University of Electro-Communications, Tokyo, Japan 2 Nara
More informationMulti-touch Interface for Controlling Multiple Mobile Robots
Multi-touch Interface for Controlling Multiple Mobile Robots Jun Kato The University of Tokyo School of Science, Dept. of Information Science jun.kato@acm.org Daisuke Sakamoto The University of Tokyo Graduate
More informationNavigating the Virtual Environment Using Microsoft Kinect
CS352 HCI Project Final Report Navigating the Virtual Environment Using Microsoft Kinect Xiaochen Yang Lichuan Pan Honor Code We, Xiaochen Yang and Lichuan Pan, pledge our honor that we have neither given
More informationAUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING
AUGMENTED REALITY APPLICATIONS USING VISUAL TRACKING ABSTRACT Chutisant Kerdvibulvech Department of Information and Communication Technology, Rangsit University, Thailand Email: chutisant.k@rsu.ac.th In
More informationChapter 2 Introduction to Haptics 2.1 Definition of Haptics
Chapter 2 Introduction to Haptics 2.1 Definition of Haptics The word haptic originates from the Greek verb hapto to touch and therefore refers to the ability to touch and manipulate objects. The haptic
More informationRecognizing Gestures on Projected Button Widgets with an RGB-D Camera Using a CNN
Recognizing Gestures on Projected Button Widgets with an RGB-D Camera Using a CNN Patrick Chiu FX Palo Alto Laboratory Palo Alto, CA 94304, USA chiu@fxpal.com Chelhwon Kim FX Palo Alto Laboratory Palo
More informationTelecommunication and remote-controlled
Spatial Interfaces Editors: Frank Steinicke and Wolfgang Stuerzlinger Telexistence: Enabling Humans to Be Virtually Ubiquitous Susumu Tachi The University of Tokyo Telecommunication and remote-controlled
More informationDouble-side Multi-touch Input for Mobile Devices
Double-side Multi-touch Input for Mobile Devices Double side multi-touch input enables more possible manipulation methods. Erh-li (Early) Shen Jane Yung-jen Hsu National Taiwan University National Taiwan
More informationSimplifying Remote Collaboration through Spatial Mirroring
Simplifying Remote Collaboration through Spatial Mirroring Fabian Hennecke 1, Simon Voelker 2, Maximilian Schenk 1, Hauke Schaper 2, Jan Borchers 2, and Andreas Butz 1 1 University of Munich (LMU), HCI
More informationHaptics in Remote Collaborative Exercise Systems for Seniors
Haptics in Remote Collaborative Exercise Systems for Seniors Hesam Alizadeh hesam.alizadeh@ucalgary.ca Richard Tang richard.tang@ucalgary.ca Permission to make digital or hard copies of part or all of
More informationInformation Layout and Interaction on Virtual and Real Rotary Tables
Second Annual IEEE International Workshop on Horizontal Interactive Human-Computer System Information Layout and Interaction on Virtual and Real Rotary Tables Hideki Koike, Shintaro Kajiwara, Kentaro Fukuchi
More informationKissenger: A Kiss Messenger
Kissenger: A Kiss Messenger Adrian David Cheok adriancheok@gmail.com Jordan Tewell jordan.tewell.1@city.ac.uk Swetha S. Bobba swetha.bobba.1@city.ac.uk ABSTRACT In this paper, we present an interactive
More informationInteracting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)
Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception
More informationThe Design of Internet-Based RobotPHONE
The Design of Internet-Based RobotPHONE Dairoku Sekiguchi 1, Masahiko Inami 2, Naoki Kawakami 1 and Susumu Tachi 1 1 Graduate School of Information Science and Technology, The University of Tokyo 7-3-1
More informationEnhancedTable: Supporting a Small Meeting in Ubiquitous and Augmented Environment
EnhancedTable: Supporting a Small Meeting in Ubiquitous and Augmented Environment Hideki Koike 1, Shin ichiro Nagashima 1, Yasuto Nakanishi 2, and Yoichi Sato 3 1 Graduate School of Information Systems,
More informationCSE 165: 3D User Interaction. Lecture #14: 3D UI Design
CSE 165: 3D User Interaction Lecture #14: 3D UI Design 2 Announcements Homework 3 due tomorrow 2pm Monday: midterm discussion Next Thursday: midterm exam 3D UI Design Strategies 3 4 Thus far 3DUI hardware
More informationVirtual/Augmented Reality (VR/AR) 101
Virtual/Augmented Reality (VR/AR) 101 Dr. Judy M. Vance Virtual Reality Applications Center (VRAC) Mechanical Engineering Department Iowa State University Ames, IA Virtual Reality Virtual Reality Virtual
More informationMohammad Akram Khan 2 India
ISSN: 2321-7782 (Online) Impact Factor: 6.047 Volume 4, Issue 8, August 2016 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case
More informationUsing Hands and Feet to Navigate and Manipulate Spatial Data
Using Hands and Feet to Navigate and Manipulate Spatial Data Johannes Schöning Institute for Geoinformatics University of Münster Weseler Str. 253 48151 Münster, Germany j.schoening@uni-muenster.de Florian
More informationGeo-Located Content in Virtual and Augmented Reality
Technical Disclosure Commons Defensive Publications Series October 02, 2017 Geo-Located Content in Virtual and Augmented Reality Thomas Anglaret Follow this and additional works at: http://www.tdcommons.org/dpubs_series
More information