FlexAR: A Tangible Augmented Reality Experience for Teaching Anatomy

Similar documents
Body Cursor: Supporting Sports Training with the Out-of-Body Sence

DESIGN STYLE FOR BUILDING INTERIOR 3D OBJECTS USING MARKER BASED AUGMENTED REALITY

ICOS: Interactive Clothing System

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

Human Autonomous Vehicles Interactions: An Interdisciplinary Approach

BoBoiBoy Interactive Holographic Action Card Game Application

P15083: Virtual Visualization for Anatomy Teaching, Training and Surgery Simulation Applications. Gate Review

Social Rules for Going to School on a Robot

HCITools: Strategies and Best Practices for Designing, Evaluating and Sharing Technical HCI Toolkits

Published in: Proceedings of the Workshop on What to Study in HCI at CHI 2015 Conference on Human Factors in Computing Systems

synchrolight: Three-dimensional Pointing System for Remote Video Communication

VISUALIZING CONTINUITY BETWEEN 2D AND 3D GRAPHIC REPRESENTATIONS

Xdigit: An Arithmetic Kinect Game to Enhance Math Learning Experiences

CONTENT RICH INTERACTIVE, AND IMMERSIVE EXPERIENCES, IN ADVERTISING, MARKETING, AND EDUCATION

Drumtastic: Haptic Guidance for Polyrhythmic Drumming Practice

ISCW 2001 Tutorial. An Introduction to Augmented Reality

Learning Based Interface Modeling using Augmented Reality

Toward an Augmented Reality System for Violin Learning Support

COLLABORATION WITH TANGIBLE AUGMENTED REALITY INTERFACES.

Capacitive Face Cushion for Smartphone-Based Virtual Reality Headsets

The Mixed Reality Book: A New Multimedia Reading Experience

Enhanced Virtual Transparency in Handheld AR: Digital Magnifying Glass

Image Interpretation System for Informed Consent to Patients by Use of a Skeletal Tracking

Designing for End-User Programming through Voice: Developing Study Methodology

Portfolio. Swaroop Kumar Pal swarooppal.wordpress.com github.com/swarooppal1088

Enhancing Shipboard Maintenance with Augmented Reality

Pop-up Java: An Augmented Reality Mobile Game to Teach Java. Richard Myers. TSYS School of Computer Science, Columbus State University, USA

Interactions and Applications for See- Through interfaces: Industrial application examples

AR Tamagotchi : Animate Everything Around Us

Virtual Reality Calendar Tour Guide

Open Research Online The Open University s repository of research publications and other research outputs

Admin. Today: Designing for Virtual Reality VR and 3D interfaces Interaction design for VR Prototyping for VR

Beyond: collapsible tools and gestures for computational design

Geo-Located Content in Virtual and Augmented Reality

Virtual Reality as Innovative Approach to the Interior Designing

Remote Shoulder-to-shoulder Communication Enhancing Co-located Sensation

Towards a Google Glass Based Head Control Communication System for People with Disabilities. James Gips, Muhan Zhang, Deirdre Anderson

mixed reality mixed reality & (tactile and) tangible interaction (tactile and) tangible interaction class housekeeping about me

Visual & Virtual Configure-Price-Quote (CPQ) Report. June 2017, Version Novus CPQ Consulting, Inc. All Rights Reserved

REPORT ON THE CURRENT STATE OF FOR DESIGN. XL: Experiments in Landscape and Urbanism

Virtual gasoline engine based on augment reality for mechanical engineering education

Figure 1. The game was developed to be played on a large multi-touch tablet and multiple smartphones.

Tilak Maharashtra Vidyapeeth, Pune. BFA- Game Design

Beyond Actuated Tangibles: Introducing Robots to Interactive Tabletops

Advances In Natural And Applied Sciences 2018 April; 12(4): pages DOI: /anas

Challenges for Establishing a Latin American Community in HCI/UX

ExTouch: Spatially-aware embodied manipulation of actuated objects mediated by augmented reality

Augmented Reality With Engineering Graphics

Applying virtual reality technology to architectural design teaching

Implementation of Augmented Reality System for Smartphone Advertisements

Development of an Augmented Reality Aided CNC Training Scenario

HUMAN COMPUTER INTERFACE

Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces

A Quality Watch Android Based Application for Monitoring Robotic Arm Statistics Using Augmented Reality

LANEY COLLEGE COURSE OUTLINE

Short Course on Computational Illumination

The Holographic Human for surgical navigation using Microsoft HoloLens

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

Job Description. Commitment: Must be available to work full-time hours, M-F for weeks beginning Summer of 2018.

Augmented reality as an aid for the use of machine tools

iwindow Concept of an intelligent window for machine tools using augmented reality

Interior Design using Augmented Reality Environment

Virtual and Augmented Reality for Cabin Crew Training: Practical Applications

ALOIS TOMÁNEK FORMS OF PUPPETS

Open Archive TOULOUSE Archive Ouverte (OATAO)

Integrated Driving Aware System in the Real-World: Sensing, Computing and Feedback

City University of Hong Kong. Course Syllabus. offered by School of Creative Media with effect from Semester A 2018 /19

Culturally Sensitive Design for Privacy: A case study of the Arabian Gulf

National 5 Graphic Communication Assignment Assessment task

3D VR Puzzle for Anatomy Education

AUGMENTED REALITY AS AN AID FOR THE USE OF MACHINE TOOLS

Multimodal Interaction Concepts for Mobile Augmented Reality Applications

Interior Design with Augmented Reality

R (2) Controlling System Application with hands by identifying movements through Camera

Department of Computer Science and Engineering The Chinese University of Hong Kong. Year Final Year Project

MSMS Software for VR Simulations of Neural Prostheses and Patient Training and Rehabilitation

Published in: Proceedings of the 8th International Conference on Tangible, Embedded and Embodied Interaction

Shared Presence and Collaboration Using a Co-Located Humanoid Robot

CS 354R: Computer Game Technology

Avatar gesture library details

Project Plan Augmented Reality Mechanic Training

Simulation of Water Inundation Using Virtual Reality Tools for Disaster Study: Opportunity and Challenges

An Environment For Long-Term Engagement with Personal Genomic Data

Physical Affordances of Check-in Stations for Museum Exhibits

CompuScholar, Inc. Alignment to Utah Game Development Fundamentals 2 Standards

Real-time AR Edutainment System Using Sensor Based Motion Recognition

Arcaid: Addressing Situation Awareness and Simulator Sickness in a Virtual Reality Pac-Man Game

Squishy Circuits as a Tangible Interface

University of Huddersfield Repository

Exhibition Strategy of Digital 3D Data of Object in Archives using Digitally Mediated Technologies for High User Experience

Knowledge Acquisition and Representation in Facility Management

New interface approaches for telemedicine

Omni-Directional Catadioptric Acquisition System

COMMUNITY COLLEGE OF CITY UNIVERSITY CITY UNIVERSITY OF HONG KONG

Learning Media Based on Augmented Reality Applied on the Lesson of Electrical Network Protection System

Virtual Reality Based Scalable Framework for Travel Planning and Training

Support Notes (Issue 1) September Play and Learn. Certificate in Digital Applications (DA204) Game Making

E-Learning in Virtual- und Augmented Reality. Hannes Kaufmann

Moving Web 3d Content into GearVR

Transcription:

FlexAR: A Tangible Augmented Reality Experience for Teaching Anatomy Michael Saenz Texas A&M University 401 Joe Routt Boulevard College Station, TX 77843 msaenz015@gmail.com Kelly Maset Texas A&M University 401 Joe Routt Boulevard College Station, TX 77843 kellymaset@gmail.com Joshua Strunk Texas A&M University 401 Joe Routt Boulevard College Station, TX 77843 joshua.p.strunk@gmail.com Paste the appropriate copyright/license statement here. ACM now supports three different publication options: ACM copyright: ACM holds the copyright on the work. This is the historical approach. License: The author(s) retain copyright, but ACM receives an exclusive publication license. Open Access: The author(s) wish to pay for the work to be open access. The additional fee must be paid to ACM. This text field is large enough to hold the appropriate release statement assuming it is single-spaced in Verdana 7 point font. Please do not change the size of this text box. Every submission will be assigned their own unique DOI string to be included here. Abstract In this paper we present FlexAR, a tangible augmented reality (TAR)[5] application for anatomy education. Traditionally, anatomy has been taught primarily in two dimensions, particularly for those in non-medical fields such as artists and physical education practitioners. Medical students often gain hands-on experience through cadaver dissection[8]. However, with dissection becoming less practical, the need to find alternative solutions for teaching anatomy, in three dimensions, is becoming more pressing. We propose FlexAR as a teaching tool for providing students both the written information found in printed study materials and the ability to examine structures in 3D previously available only through dissection. Users of our prototype manipulate a tangible skeletal model of a human arm affixed with augmented reality (AR) targets. An AR-enabled device records this interaction through a camera and projects a digital 3D model consisting of the bones and major muscles of the arm over the physical model onscreen. Users can examine both gross anatomy as well as muscle flexion and extension. The user is also able to interact through a graphical user interface (GUI) to highlight and display information for individual muscles. From initial feedback we received, we believe that once expanded FlexAR could be effective as a standalone or supplementary tool for both group and individual learning of anatomy.

Author Keywords Augmented Reality; Tangible User Interface; education; human anatomy ACM Classification Keywords H.5.m. Information interfaces and presentation (e.g., HCI): Miscellaneous. See: http://www.acm.org/about/class/1998/ Optional section to be included in your final version, but strongly encouraged. Introduction A significant drawback of learning anatomy through traditional reading material is the difficulty students have piecing together disparate 2D images into knowledge of 3D structures. In the past, this has been solved through the use of cadavers for dissection, which gives students hands-on experience with anatomical structures. Recently, however, this has become less practical for a number of reasons. For one, this method of instruction does not afford students room to make mistakes or repeat procedures. Furthermore, as rising ethical concerns limit the availability of cadavers for dissection, the costs of acquiring them rise. In light of these issues, we present FlexAR, an AR application that combines a tangible interface and GUI to teach anatomy. FlexAR combines the written information available from traditional reading materials with the spatial learning one would acquire from anatomical dissection. Users have the ability to study anatomy at their preferred pace, freely exploring gross anatomy or selecting specific structures for closer examination. Related Work Prior work relating to the use of AR for anatomy education is detailed by Juanes [3]. This paper introduces a tool for augmenting 2D images from a book with static 3D models on mobile devices. This allows the user to view the structure of particular body parts in 3D space without the need for physical models. However, we believe that having a dynamic model will improve the user s ability to understand spatial relationships and the effect of movement on different systems. For our research we focus on demonstrating the flexion and extension of various muscle groups as the result of movement of the arm using an articulated tangible as a controller. Another related application which uses a TAR interface is ARnatomy. ARnatomy aims to uses a tangible user interface (TUI) by using models of dog bones to control the display of information on a mobile device such as a smartphone or tablet[4]. Though this application does include dynamic text, there is little interaction between the user and the tangibles themselves; the tangible controls only the location of the text onscreen and there is no interaction between the user and mobile device. Thus, this application is useful primarily as a tool for memorizing written information. In FlexAR we fuse interaction with a GUI and TUI, allowing the user to manipulate a physical model to drive the animation of a 3D digital overlay and highlight and display the information of individual muscles. System Description FlexAR consists of a camera device and a tangible skeletal arm which is used to drive the interaction. The TAR interface provides an ideal system for the user to explore the model in 3D space. Traditional instruction

relies on materials such as books, diagrams, and standalone physical anatomical models; our current prototype combines written information with a tactile model to serve as a self-contained learning module. match the range of motion possible with the physical model. Once the assets were completed, they were exported to Unity for integration with the application. The primary interaction for users using our FlexAR prototype is done by manipulating our tangible skeletal arm. As the user manipulates the tangible a computer model of the arm with muscles and other anatomical features tracks to the tangible. In addition of tracking the tangible the muscles update to reflect proper extension and flexion. In addition to exploring the movement and structure of muscles the user is able to, through a GUI, select individual muscles. When a muscle becomes selected its model changes textures to become highlighted. In addition to highlighting the muscle, textual information, similar to that which could be found in a text book, is also presented. Design Implementation The augmented reality system was implemented using a number of programs and development tools. The 3D overlay was created in Autodesk Maya[1] before being imported into the multiplatform game engine Unity[7] for integration with our application. The assets for the 3D overlay were developed in Maya using our physical arm model and Gray s Anatomy [2] as reference. To enhance immersion, the physical and digital models had to align as closely as possible in appearance and be anatomically correct. After the skeleton was modeled, each muscle was modeled and textured separately in order to allow them to be selected individually. The movement and deformations were created using a combination of a simple rig and blendshapes, which were set up in such a way as to Figure 1. Front, side, and back views of the final digital assets in their anatomically neutral starting position. We used Vuforia, a mobile AR library implemented by QUALCOMM Incorporated[6], to hanle the capturing and tracking of our augmented reality targets. Vuforia relies on camera feed to track image targets, projecting a 3D overlay relative to the position of detected targets onscreen. For FlexAR, we used 4 targets: 1 to determine the basic position of the arm and the others to control the rotation of the shoulder, elbow, and wrist joints of the 3D model.

devices. For this iteration we focused on usability, interaction, and user experience. Figure 2. High-level system overview of the Vuforia SDK Unity Extension. Prototype Design In our initial primary observations, we found that users had difficulty selecting individual muscles in 3D space. A common suggestion was to introduce direct interaction with a GUI rather than using 3D overlay selection. Rather than having to select individual muscles, users could instead display information using tabs labeled with the names of the muscles. We found that this implementation provided a more intuitive experience for the user. Preliminary Observations Protocol To test FlexAR, we allowed users to interact with our prototype in a lab environment. We gathered nine university-level participants from a variety of backgrounds including art, dance, and computer engineering. We then divided them into three groups of equal size, to test the three different builds of FlexAR. The devices FlexAR was built for are: a desktop with a webcam, an Android tablet, and Epson Moverio BT-200 glasses. Each group interacted with only one of these Users were brought to the lab one at a time. After a brief questionnaire and instruction period, each was asked to explore the application with minimal direction. In the first phase, we observed the interactions involving only the tangible. Users in the desktop and wearable groups were allowed to manipulate the tangible themselves, while those testing the application on the tablet held the device while a secondary user manipulated the tangible. Next, users were instructed to interact with the GUI. Users were then given the opportunity to freely interact with the prototype. Once they finished interacting with the prototype, they were given a short post-quiz and asked to give feedback. Discussion The feedback we received was generally positive. Users listed a number of areas in which they believed FlexAR would be useful. Several students compared it favorably to using anatomy textbooks to train medical personnel. A few mentioned its usefulness as a reference in the process of creating joint systems for 3D animation, and one user who specialized in sculpting expressed interest in being able to view static structures from multiple angles. Users were also asked to list the type of device they felt was best suited to this application. The general consensus was that the tablet was too large and difficult to handle at the same time as the tangible. What we found interesting, however, was the feedback regarding the desktop and the smart glasses. Many users stated that the desktop would be ideal for learning as a group or class while the glasses would be

most useful in individual learning settings or where mobility was the most important factor. Using the desktop, one person - such as an instructor or group leader - could interact with the application in front of the camera while the others observed the screen. This would be most useful for guided learning. In contrast, the glasses would work best for those wishing to study independently or during individual assignments. Consequently, we plan to focus on these two devices in our continuing research. Future Work Motivated by the positive initial feedback we received, we are currently furthering our research and expanding our prototype to include not only the arm but also other key regions such as the torso. A larger tangible would likely enhance the user s sense of immersion into the application and bring us closer to our goal of being able to replace cadaver dissection as an effective method of teaching anatomy spatially. Additionally, we plan to move to the Vuforia SDK for Digital Eyewear, a software library designed specifically to take advantage of smart glass technology, once it is released to the public. Conclusion FlexAR is a prototype tool for teaching anatomy through the use of augmented reality. We believe that it contributes to education by giving users both written and 3D visual information about anatomy without the need for dissection or traditional study materials. In its current state it is useful as a tool for studying the muscles of the arm, but when expanded shows promise as an application for teaching the anatomy of multiple complete body systems, both for individuals and large groups across a wide range of disciplines. Figures 3-5. From left to right: users of a desktop with webcam, Android tablet, and Epson Moverio BT-200 glasses.

Acknowledgements We thank Dr. Jinsil Hwaryoung Seo for sharing her research with us and helping us procure the supplies needed for our research. We also thank Erica Malone for providing us with anatomical study tools and guidance during the development of the 3D model. References [1] Autodesk Maya. Available from http://www.autodesk.com/products/maya/overview/ [2] H. Gray, T. P. Pick, and R. Howden, "Myology," in Gray's Anatomy. Philadelphia, PA: Running Press, 1974, ch. 4, sec. 7, pp. 442-455. [3] J. Juanes, D. Hernández, P. Ruisoto, E. García, G. Villarrubia, and A. Prats. 2014. Augmented reality techniques, using mobile devices, for learning human anatomy. In Proceedings of the Second International Conference on Technological Ecosystems for Enhancing Multiculturality (TEEM '14). ACM, New York, NY, USA, 7-11. [4] J. Seo, J. Storey, J. Chavez, D. Reyna, J. Suh, and M. Pine. 2014. ARnatomy: tangible AR app for learning gross anatomy. In ACM SIGGRAPH 2014 Posters (SIGGRAPH '14). ACM, New York, NY, USA, Article 25, 1 page. [5] M.Billinghurst, H.Kato, and I.Poupyrev. Tangible augmented reality. In Proc. of SIGGRAPH Asia '08 (Courses), pages 1--10, 2008. [6] Qualcomm Vuforia. Available from https://www.qualcomm.com/products/vuforia/ [7] Unity. Available from http://unity3d.com/ [8] Winkelmann, A. 2007. Anatomical dissection as a teaching method in medical school: a review of the evidence. Med Education vol. 41(1), 15--22.