3D Data Navigation via Natural User Interfaces
|
|
- Roy Thornton
- 5 years ago
- Views:
Transcription
1 3D Data Navigation via Natural User Interfaces Francisco R. Ortega PhD Candidate and GAANN Fellow Co-Advisors: Dr. Rishe and Dr. Barreto Committee Members: Dr. Raju, Dr. Clarke and Dr. Zeng GAANN Fellowship Supervisor: Dr. Milani
2 The most profound technologies are those that disappear The computer for the 21st century Mark Weiser, 1999 Father of ubiquitous computing
3 WIMP Windows Icons Menu Pointers Paradigm Most successful input paradigm
4 Bill Buxton Three State Model Stylus Example
5 Bill Buxton Three State Model Mouse Example
6 Fitt s Law
7 Post-WIMP The set of components that will be required for the correct interaction in this new paradigm: Multi Touch Vision Tracking Speech Recognition Pen Based Interaction Virtual Devices Etc.
8 NUI and Other Devices
9 Software Visual Studio 2010 C++ Ogre3D OpenGL Windows 7 Additional libraries
10 Motivation Find methods that allow users to interact with technology seamlessly Add building blocks to Post-WIMP Contribute a easy adaptable framework to software without full gesture support Help domain specific problems via MT Neural Sciences (e.g., MRI)
11 Example: 3D-2D-3D
12 Domain Specific Example MRI 3D and MRI data slices
13 BACKGROUND Understanding Touch Multi-Touch Interaction Gesture Recognition Virtual Devices Design Guidelines Current Open Systems / Frameworks
14 Understanding Touch Interactions Oblique finger orientation w/o sensors Rotations and Translations Combined or separate? One or two hands One hand better for simpler tasks (rotations) Two hands better for separable type of tasks (anchor gestures)
15 Multi-Touch Techniques Hancock et el, 2007 One two three touches Algorithm
16 Gesture Recognition Methods Hidden Markov Models Sezgin and Davis, Neural Networks Pittman, Featured-Based Classifiers Rubine, Template Matching Kara and Stahovic, Geometric Recognizers Wobbrock and Wilson, 2007 Etc.
17 Gesture Recognizer Example The $1 and the $N algorithms are examples of a combination of : Geometric Recognizers Template Matching
18 The $1 Algorithm (Wobbrock and Wilson, 2007) Resample the points in the path (N=64) Find indicative angle between the centroid and gesture s first point Rotate gesture to Indicative angle = 0 Scale gesture to reference square Calculate the distance between: Candidate gesture to each stored template
19 The $N Algorithm (extension to $1) Anthony and Wobbrock, 2010
20 Virtual Devices New set of devices that either: Emulate Physical Devices Find new ways for interaction that are not possible with current physical devices Examples Virtual Keyboard Virtual Sphere
21 Virtual Keyboard
22 Virtual Sphere Chen et el Simulates a 3D Trackball X axis: Left-Right Y Axis: Top-Down Z Axis: Circular Fashion *
23 Design Guidelines I (Bowman et el. 2005) Use existing manipulation techniques unless an application will benefit greatly from creating new ones. Swipe Gesture Perform task analysis when selecting a 3D interaction technique. How precise was the Rotation? Match the interaction to the device. Consider the specific device properties
24 Design Guidelines II (Bowman et el. 2005) Reduce wasted motion. Amplify the user hand motion Reducethe number of degrees of freedom (DOF) whenever possible. Constraint 3D interaction when possible There is no single best 3D manipulation technique
25 Set-up Methods / Implementation Main Display 3M 22 M2256PW Monitor Windows 7 MT Drivers and Library Recognition and Navigation Algorithms
26 Camera Setup
27 Gestures [GestureWorks] Anchor Rotate Two Hand (5) Divide 4 Finger Rotate 5 Finger Hold
28 Basic Gestures
29 Gesture Action: Gesture Rotation
30 Beyond Basic
31
32 Point of Clouds
33 How does it work? Receive Trace Collection from Operating System Generic Touch Events: Up, Move, Down events Detect Trace Event Uses a MAP to keep collections of traces Tap, Double Tap and Drag Delegate responsibility to Trace Manager Trace Manager is aware of different recognizers Allow event and poll techniques
34 How does it work? (cont.) Decision Manager Collects possible gestures from recognizers Using a priority criteria, selects Gesture Action Manager Selects action with correct values Name of Gestures Callback function (or raised event) Transformation Matrix
35 Overview of the Framework System Diagram A system level view of our proposed framework Class Diagram The current classes that are part of our framework MT FSM Multi-Touch Finite State Machine
36 System Diagram
37 Evaluation Pilot Program (3-5 subjects) Correct bias produced by Instruments Experimental procedures Environmental Factors **
38 Main Experiment Main Experiment (30 Subjects) Repeated Measure design Compares Our approach to an existing Approach Will measure: Execution time: measures gesture actions to accomplish task e.g. : User will navigate within a 3D maze Accuracy of the movement: measures steps to finish What was the path taken to accomplish task? Likert scale: measures usability of the user
39 Preliminary Results
40 Preliminary Results Support for other devices for comparison Wiimote 3D Mouse Gamepad
41
42
43
44 Questions For references, please me to Thank you
Direct Manipulation. and Instrumental Interaction. CS Direct Manipulation
Direct Manipulation and Instrumental Interaction 1 Review: Interaction vs. Interface What s the difference between user interaction and user interface? Interface refers to what the system presents to the
More informationGestureCommander: Continuous Touch-based Gesture Prediction
GestureCommander: Continuous Touch-based Gesture Prediction George Lucchese george lucchese@tamu.edu Jimmy Ho jimmyho@tamu.edu Tracy Hammond hammond@cs.tamu.edu Martin Field martin.field@gmail.com Ricardo
More informationBuilding a bimanual gesture based 3D user interface for Blender
Modeling by Hand Building a bimanual gesture based 3D user interface for Blender Tatu Harviainen Helsinki University of Technology Telecommunications Software and Multimedia Laboratory Content 1. Background
More informationVICs: A Modular Vision-Based HCI Framework
VICs: A Modular Vision-Based HCI Framework The Visual Interaction Cues Project Guangqi Ye, Jason Corso Darius Burschka, & Greg Hager CIRL, 1 Today, I ll be presenting work that is part of an ongoing project
More informationInterface Design V: Beyond the Desktop
Interface Design V: Beyond the Desktop Rob Procter Further Reading Dix et al., chapter 4, p. 153-161 and chapter 15. Norman, The Invisible Computer, MIT Press, 1998, chapters 4 and 15. 11/25/01 CS4: HCI
More informationClassifying 3D Input Devices
IMGD 5100: Immersive HCI Classifying 3D Input Devices Robert W. Lindeman Associate Professor Department of Computer Science Worcester Polytechnic Institute gogo@wpi.edu Motivation The mouse and keyboard
More informationE90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright
E90 Project Proposal 6 December 2006 Paul Azunre Thomas Murray David Wright Table of Contents Abstract 3 Introduction..4 Technical Discussion...4 Tracking Input..4 Haptic Feedack.6 Project Implementation....7
More informationClassifying 3D Input Devices
IMGD 5100: Immersive HCI Classifying 3D Input Devices Robert W. Lindeman Associate Professor Department of Computer Science Worcester Polytechnic Institute gogo@wpi.edu But First Who are you? Name Interests
More informationCricut Design Space App for ipad User Manual
Cricut Design Space App for ipad User Manual Cricut Explore design-and-cut system From inspiration to creation in just a few taps! Cricut Design Space App for ipad 1. ipad Setup A. Setting up the app B.
More informationIMGD 4000 Technical Game Development II Interaction and Immersion
IMGD 4000 Technical Game Development II Interaction and Immersion Robert W. Lindeman Associate Professor Human Interaction in Virtual Environments (HIVE) Lab Department of Computer Science Worcester Polytechnic
More informationVisual Interpretation of Hand Gestures as a Practical Interface Modality
Visual Interpretation of Hand Gestures as a Practical Interface Modality Frederik C. M. Kjeldsen Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate
More informationAdvancements in Gesture Recognition Technology
IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 4, Issue 4, Ver. I (Jul-Aug. 2014), PP 01-07 e-issn: 2319 4200, p-issn No. : 2319 4197 Advancements in Gesture Recognition Technology 1 Poluka
More informationR (2) Controlling System Application with hands by identifying movements through Camera
R (2) N (5) Oral (3) Total (10) Dated Sign Assignment Group: C Problem Definition: Controlling System Application with hands by identifying movements through Camera Prerequisite: 1. Web Cam Connectivity
More informationMarkerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces
Markerless 3D Gesture-based Interaction for Handheld Augmented Reality Interfaces Huidong Bai The HIT Lab NZ, University of Canterbury, Christchurch, 8041 New Zealand huidong.bai@pg.canterbury.ac.nz Lei
More informationInteracting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)
Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception
More informationEnabling Cursor Control Using on Pinch Gesture Recognition
Enabling Cursor Control Using on Pinch Gesture Recognition Benjamin Baldus Debra Lauterbach Juan Lizarraga October 5, 2007 Abstract In this project we expect to develop a machine-user interface based on
More information3D Modelling Is Not For WIMPs Part II: Stylus/Mouse Clicks
3D Modelling Is Not For WIMPs Part II: Stylus/Mouse Clicks David Gauldie 1, Mark Wright 2, Ann Marie Shillito 3 1,3 Edinburgh College of Art 79 Grassmarket, Edinburgh EH1 2HJ d.gauldie@eca.ac.uk, a.m.shillito@eca.ac.uk
More informationHuman Computer Interaction Lecture 04 [ Paradigms ]
Human Computer Interaction Lecture 04 [ Paradigms ] Imran Ihsan Assistant Professor www.imranihsan.com imranihsan.com HCIS1404 - Paradigms 1 why study paradigms Concerns how can an interactive system be
More informationUUIs Ubiquitous User Interfaces
UUIs Ubiquitous User Interfaces Alexander Nelson April 16th, 2018 University of Arkansas - Department of Computer Science and Computer Engineering The Problem As more and more computation is woven into
More informationA Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang, Dong-jun Seo, and Dong-seok Jung,
IJCSNS International Journal of Computer Science and Network Security, VOL.11 No.9, September 2011 55 A Study on the control Method of 3-Dimensional Space Application using KINECT System Jong-wook Kang,
More informationDirect Manipulation. and Instrumental Interaction. Direct Manipulation 1
Direct Manipulation and Instrumental Interaction Direct Manipulation 1 Direct Manipulation Direct manipulation is when a virtual representation of an object is manipulated in a similar way to a real world
More informationUniversity of California, Santa Barbara. CS189 Fall 17 Capstone. VR Telemedicine. Product Requirement Documentation
University of California, Santa Barbara CS189 Fall 17 Capstone VR Telemedicine Product Requirement Documentation Jinfa Zhu Kenneth Chan Shouzhi Wan Xiaohe He Yuanqi Li Supervised by Ole Eichhorn Helen
More informationVirtual Grasping Using a Data Glove
Virtual Grasping Using a Data Glove By: Rachel Smith Supervised By: Dr. Kay Robbins 3/25/2005 University of Texas at San Antonio Motivation Navigation in 3D worlds is awkward using traditional mouse Direct
More informationDirect Manipulation. and Instrumental Interaction. Direct Manipulation
Direct Manipulation and Instrumental Interaction Direct Manipulation 1 Direct Manipulation Direct manipulation is when a virtual representation of an object is manipulated in a similar way to a real world
More informationDRAFT: SPARSH UI: A MULTI-TOUCH FRAMEWORK FOR COLLABORATION AND MODULAR GESTURE RECOGNITION. Desirée Velázquez NSF REU Intern
Proceedings of the World Conference on Innovative VR 2009 WINVR09 July 12-16, 2008, Brussels, Belgium WINVR09-740 DRAFT: SPARSH UI: A MULTI-TOUCH FRAMEWORK FOR COLLABORATION AND MODULAR GESTURE RECOGNITION
More informationDesign a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison
e-issn 2455 1392 Volume 2 Issue 10, October 2016 pp. 34 41 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com Design a Model and Algorithm for multi Way Gesture Recognition using Motion and
More informationWhat was the first gestural interface?
stanford hci group / cs247 Human-Computer Interaction Design Studio What was the first gestural interface? 15 January 2013 http://cs247.stanford.edu Theremin Myron Krueger 1 Myron Krueger There were things
More informationSmartCanvas: A Gesture-Driven Intelligent Drawing Desk System
SmartCanvas: A Gesture-Driven Intelligent Drawing Desk System Zhenyao Mo +1 213 740 4250 zmo@graphics.usc.edu J. P. Lewis +1 213 740 9619 zilla@computer.org Ulrich Neumann +1 213 740 0877 uneumann@usc.edu
More informationOutline. Paradigms for interaction. Introduction. Chapter 5 : Paradigms. Introduction Paradigms for interaction (15)
Outline 01076568 Human Computer Interaction Chapter 5 : Paradigms Introduction Paradigms for interaction (15) ดร.ชมพ น ท จ นจาคาม [kjchompo@gmail.com] สาขาว ชาว ศวกรรมคอมพ วเตอร คณะว ศวกรรมศาสตร สถาบ นเทคโนโลย
More informationGART: The Gesture and Activity Recognition Toolkit
GART: The Gesture and Activity Recognition Toolkit Kent Lyons, Helene Brashear, Tracy Westeyn, Jung Soo Kim, and Thad Starner College of Computing and GVU Center Georgia Institute of Technology Atlanta,
More informationCATIA V5 Workbook Release V5-6R2013
CATIA V5 Workbook Release V5-6R2013 Richard Cozzens SDC PUBLICATIONS Better Textbooks. Lower Prices. www.sdcpublications.com Powered by TCPDF (www.tcpdf.org) Visit the following websites to learn more
More informationRobust Hand Gesture Recognition for Robotic Hand Control
Robust Hand Gesture Recognition for Robotic Hand Control Ankit Chaudhary Robust Hand Gesture Recognition for Robotic Hand Control 123 Ankit Chaudhary Department of Computer Science Northwest Missouri State
More informationA Gestural Interaction Design Model for Multi-touch Displays
Songyang Lao laosongyang@ vip.sina.com A Gestural Interaction Design Model for Multi-touch Displays Xiangan Heng xianganh@ hotmail ABSTRACT Media platforms and devices that allow an input from a user s
More informationVirtual Touch Human Computer Interaction at a Distance
International Journal of Computer Science and Telecommunications [Volume 4, Issue 5, May 2013] 18 ISSN 2047-3338 Virtual Touch Human Computer Interaction at a Distance Prasanna Dhisale, Puja Firodiya,
More information6 Ubiquitous User Interfaces
6 Ubiquitous User Interfaces Viktoria Pammer-Schindler May 3, 2016 Ubiquitous User Interfaces 1 Days and Topics March 1 March 8 March 15 April 12 April 26 (10-13) April 28 (9-14) May 3 May 10 Administrative
More informationStereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.
Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.
More informationTouch Interfaces. Jeff Avery
Touch Interfaces Jeff Avery Touch Interfaces In this course, we have mostly discussed the development of web interfaces, with the assumption that the standard input devices (e.g., mouse, keyboards) are
More informationCSE 165: 3D User Interaction. Lecture #11: Travel
CSE 165: 3D User Interaction Lecture #11: Travel 2 Announcements Homework 3 is on-line, due next Friday Media Teaching Lab has Merge VR viewers to borrow for cell phone based VR http://acms.ucsd.edu/students/medialab/equipment
More informationA Method for Temporal Hand Gesture Recognition
A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University Jacksonville, AL 36265 (256) 782-5103 newj@ksl.jsu.edu ABSTRACT Ongoing efforts at
More informationInterior Design with Augmented Reality
Interior Design with Augmented Reality Ananda Poudel and Omar Al-Azzam Department of Computer Science and Information Technology Saint Cloud State University Saint Cloud, MN, 56301 {apoudel, oalazzam}@stcloudstate.edu
More informationProjection Based HCI (Human Computer Interface) System using Image Processing
GRD Journals- Global Research and Development Journal for Volume 1 Issue 5 April 2016 ISSN: 2455-5703 Projection Based HCI (Human Computer Interface) System using Image Processing Pankaj Dhome Sagar Dhakane
More informationToolkit For Gesture Classification Through Acoustic Sensing
Toolkit For Gesture Classification Through Acoustic Sensing Pedro Soldado pedromgsoldado@ist.utl.pt Instituto Superior Técnico, Lisboa, Portugal October 2015 Abstract The interaction with touch displays
More informationA SURVEY ON GESTURE RECOGNITION TECHNOLOGY
A SURVEY ON GESTURE RECOGNITION TECHNOLOGY Deeba Kazim 1, Mohd Faisal 2 1 MCA Student, Integral University, Lucknow (India) 2 Assistant Professor, Integral University, Lucknow (india) ABSTRACT Gesture
More informationBuilding a gesture based information display
Chair for Com puter Aided Medical Procedures & cam par.in.tum.de Building a gesture based information display Diplomarbeit Kickoff Presentation by Nikolas Dörfler Feb 01, 2008 Chair for Computer Aided
More informationRecent Progress on Wearable Augmented Interaction at AIST
Recent Progress on Wearable Augmented Interaction at AIST Takeshi Kurata 12 1 Human Interface Technology Lab University of Washington 2 AIST, Japan kurata@ieee.org Weavy The goal of the Weavy project team
More informationUser s handbook Last updated in December 2017
User s handbook Last updated in December 2017 Contents Contents... 2 System info and options... 3 Mindesk VR-CAD interface basics... 4 Controller map... 5 Global functions... 6 Tool palette... 7 VR Design
More informationCS 247 Project 2. Part 1. Reflecting On Our Target Users. Jorge Cueto Edric Kyauk Dylan Moore Victoria Wee
1 CS 247 Project 2 Jorge Cueto Edric Kyauk Dylan Moore Victoria Wee Part 1 Reflecting On Our Target Users Our project presented our team with the task of redesigning the Snapchat interface for runners,
More informationUniversity of California, Santa Barbara. CS189 Fall 17 Capstone. VR Telemedicine. Product Requirement Documentation
University of California, Santa Barbara CS189 Fall 17 Capstone VR Telemedicine Product Requirement Documentation Jinfa Zhu Kenneth Chan Shouzhi Wan Xiaohe He Yuanqi Li Supervised by Ole Eichhorn Helen
More informationSaphira Robot Control Architecture
Saphira Robot Control Architecture Saphira Version 8.1.0 Kurt Konolige SRI International April, 2002 Copyright 2002 Kurt Konolige SRI International, Menlo Park, California 1 Saphira and Aria System Overview
More informationVirtual Environment Interaction Based on Gesture Recognition and Hand Cursor
Virtual Environment Interaction Based on Gesture Recognition and Hand Cursor Chan-Su Lee Kwang-Man Oh Chan-Jong Park VR Center, ETRI 161 Kajong-Dong, Yusong-Gu Taejon, 305-350, KOREA +82-42-860-{5319,
More informationUsing Hands and Feet to Navigate and Manipulate Spatial Data
Using Hands and Feet to Navigate and Manipulate Spatial Data Johannes Schöning Institute for Geoinformatics University of Münster Weseler Str. 253 48151 Münster, Germany j.schoening@uni-muenster.de Florian
More informationFLORIDA INTERNATIONAL UNIVERSITY. Miami, Florida 3D NAVIGATION WITH SIX DEGREES-OF-FREEDOM USING A MULTI-TOUCH DISPLAY
FLORIDA INTERNATIONAL UNIVERSITY Miami, Florida 3D NAVIGATION WITH SIX DEGREES-OF-FREEDOM USING A MULTI-TOUCH DISPLAY A dissertation submitted in partial fulfillment of the requirements for the degree
More information3D Interaction Techniques
3D Interaction Techniques Hannes Interactive Media Systems Group (IMS) Institute of Software Technology and Interactive Systems Based on material by Chris Shaw, derived from Doug Bowman s work Why 3D Interaction?
More information3D User Interfaces. Using the Kinect and Beyond. John Murray. John Murray
Using the Kinect and Beyond // Center for Games and Playable Media // http://games.soe.ucsc.edu John Murray John Murray Expressive Title Here (Arial) Intelligence Studio Introduction to Interfaces User
More informationThe 8 th International Scientific Conference elearning and software for Education Bucharest, April 26-27, / X
The 8 th International Scientific Conference elearning and software for Education Bucharest, April 26-27, 2012 10.5682/2066-026X-12-103 DEVELOPMENT OF A NATURAL USER INTERFACE FOR INTUITIVE PRESENTATIONS
More informationThe Control of Avatar Motion Using Hand Gesture
The Control of Avatar Motion Using Hand Gesture ChanSu Lee, SangWon Ghyme, ChanJong Park Human Computing Dept. VR Team Electronics and Telecommunications Research Institute 305-350, 161 Kajang-dong, Yusong-gu,
More informationHand Data Glove: A Wearable Real-Time Device for Human- Computer Interaction
Hand Data Glove: A Wearable Real-Time Device for Human- Computer Interaction Piyush Kumar 1, Jyoti Verma 2 and Shitala Prasad 3 1 Department of Information Technology, Indian Institute of Information Technology,
More informationDouble-side Multi-touch Input for Mobile Devices
Double-side Multi-touch Input for Mobile Devices Double side multi-touch input enables more possible manipulation methods. Erh-li (Early) Shen Jane Yung-jen Hsu National Taiwan University National Taiwan
More informationIntroduction to HCI. CS4HC3 / SE4HC3/ SE6DO3 Fall Instructor: Kevin Browne
Introduction to HCI CS4HC3 / SE4HC3/ SE6DO3 Fall 2011 Instructor: Kevin Browne brownek@mcmaster.ca Slide content is based heavily on Chapter 1 of the textbook: Designing the User Interface: Strategies
More informationVEWL: A Framework for Building a Windowing Interface in a Virtual Environment Daniel Larimer and Doug A. Bowman Dept. of Computer Science, Virginia Tech, 660 McBryde, Blacksburg, VA dlarimer@vt.edu, bowman@vt.edu
More informationInput devices and interaction. Ruth Aylett
Input devices and interaction Ruth Aylett Contents Tracking What is available Devices Gloves, 6 DOF mouse, WiiMote Why is it important? Interaction is basic to VEs We defined them as interactive in real-time
More informationResearch Seminar. Stefano CARRINO fr.ch
Research Seminar Stefano CARRINO stefano.carrino@hefr.ch http://aramis.project.eia- fr.ch 26.03.2010 - based interaction Characterization Recognition Typical approach Design challenges, advantages, drawbacks
More informationINTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT
INTERACTION AND SOCIAL ISSUES IN A HUMAN-CENTERED REACTIVE ENVIRONMENT TAYSHENG JENG, CHIA-HSUN LEE, CHI CHEN, YU-PIN MA Department of Architecture, National Cheng Kung University No. 1, University Road,
More informationFrictioned Micromotion Input for Touch Sensitive Devices
Technical Disclosure Commons Defensive Publications Series May 18, 2015 Frictioned Micromotion Input for Touch Sensitive Devices Samuel Huang Follow this and additional works at: http://www.tdcommons.org/dpubs_series
More informationSketchpad Ivan Sutherland (1962)
Sketchpad Ivan Sutherland (1962) 7 Viewable on Click here https://www.youtube.com/watch?v=yb3saviitti 8 Sketchpad: Direct Manipulation Direct manipulation features: Visibility of objects Incremental action
More informationEasy Input For Gear VR Documentation. Table of Contents
Easy Input For Gear VR Documentation Table of Contents Setup Prerequisites Fresh Scene from Scratch In Editor Keyboard/Mouse Mappings Using Model from Oculus SDK Components Easy Input Helper Pointers Standard
More informationService Robots in an Intelligent House
Service Robots in an Intelligent House Jesus Savage Bio-Robotics Laboratory biorobotics.fi-p.unam.mx School of Engineering Autonomous National University of Mexico UNAM 2017 OUTLINE Introduction A System
More informationKINECT HANDS-FREE. Rituj Beniwal. Department of Electrical Engineering Indian Institute of Technology, Kanpur. Pranjal Giri
KINECT HANDS-FREE Rituj Beniwal Pranjal Giri Agrim Bari Raman Pratap Singh Akash Jain Department of Aerospace Engineering Indian Institute of Technology, Kanpur Atharva Mulmuley Department of Chemical
More informationJob Description. Commitment: Must be available to work full-time hours, M-F for weeks beginning Summer of 2018.
Research Intern Director of Research We are seeking a summer intern to support the team to develop prototype 3D sensing systems based on state-of-the-art sensing technologies along with computer vision
More informationTangible User Interfaces
Tangible User Interfaces Seminar Vernetzte Systeme Prof. Friedemann Mattern Von: Patrick Frigg Betreuer: Michael Rohs Outline Introduction ToolStone Motivation Design Interaction Techniques Taxonomy for
More informationMicrosoft Scrolling Strip Prototype: Technical Description
Microsoft Scrolling Strip Prototype: Technical Description Primary features implemented in prototype Ken Hinckley 7/24/00 We have done at least some preliminary usability testing on all of the features
More informationRingEdit: A Control Point Based Editing Approach in Sketch Recognition Systems
RingEdit: A Control Point Based Editing Approach in Sketch Recognition Systems Yuxiang Zhu, Joshua Johnston, and Tracy Hammond Department of Computer Science and Engineering Texas A&M University College
More informationCHAPTER 1. INTRODUCTION 16
1 Introduction The author s original intention, a couple of years ago, was to develop a kind of an intuitive, dataglove-based interface for Computer-Aided Design (CAD) applications. The idea was to interact
More informationSense. 3D scanning application for Intel RealSense 3D Cameras. Capture your world in 3D. User Guide. Original Instructions
Sense 3D scanning application for Intel RealSense 3D Cameras Capture your world in 3D User Guide Original Instructions TABLE OF CONTENTS 1 INTRODUCTION.... 3 COPYRIGHT.... 3 2 SENSE SOFTWARE SETUP....
More informationA Novel Human Computer Interaction Paradigm for Volume Visualization in Projection-Based. Environments
Virtual Environments 1 A Novel Human Computer Interaction Paradigm for Volume Visualization in Projection-Based Virtual Environments Changming He, Andrew Lewis, and Jun Jo Griffith University, School of
More informationMURDOCH RESEARCH REPOSITORY
MURDOCH RESEARCH REPOSITORY This is the author s final version of the work, as accepted for publication following peer review but without the publisher s layout or pagination. The definitive version is
More informationPERFORMANCE IN A HAPTIC ENVIRONMENT ABSTRACT
PERFORMANCE IN A HAPTIC ENVIRONMENT Michael V. Doran,William Owen, and Brian Holbert University of South Alabama School of Computer and Information Sciences Mobile, Alabama 36688 (334) 460-6390 doran@cis.usouthal.edu,
More informationOwner s Guide. DB-303 Version 1.0 Copyright Pulse Code, Inc. 2009, All Rights Reserved
Owner s Guide DB-303 Version 1.0 www.pulsecodeinc.com/db-303 Copyright Pulse Code, Inc. 2009, All Rights Reserved INTRODUCTION Thank you for purchasing the DB-303 Digital Bass Line. The DB-303 is a bass
More informationAuditory System For a Mobile Robot
Auditory System For a Mobile Robot PhD Thesis Jean-Marc Valin Department of Electrical Engineering and Computer Engineering Université de Sherbrooke, Québec, Canada Jean-Marc.Valin@USherbrooke.ca Motivations
More informationUsability Evaluation of Multi- Touch-Displays for TMA Controller Working Positions
Sesar Innovation Days 2014 Usability Evaluation of Multi- Touch-Displays for TMA Controller Working Positions DLR German Aerospace Center, DFS German Air Navigation Services Maria Uebbing-Rumke, DLR Hejar
More informationCOMPARATIVE STUDY AND ANALYSIS FOR GESTURE RECOGNITION METHODOLOGIES
http:// COMPARATIVE STUDY AND ANALYSIS FOR GESTURE RECOGNITION METHODOLOGIES Rafiqul Z. Khan 1, Noor A. Ibraheem 2 1 Department of Computer Science, A.M.U. Aligarh, India 2 Department of Computer Science,
More informationAbstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction
Creating a Collaborative Multi Touch Computer Aided Design Program Cole Anagnost, Thomas Niedzielski, Desirée Velázquez, Prasad Ramanahally, Stephen Gilbert Iowa State University { someguy tomn deveri
More informationBeyond Actuated Tangibles: Introducing Robots to Interactive Tabletops
Beyond Actuated Tangibles: Introducing Robots to Interactive Tabletops Sowmya Somanath Department of Computer Science, University of Calgary, Canada. ssomanat@ucalgary.ca Ehud Sharlin Department of Computer
More informationHeads up interaction: glasgow university multimodal research. Eve Hoggan
Heads up interaction: glasgow university multimodal research Eve Hoggan www.tactons.org multimodal interaction Multimodal Interaction Group Key area of work is Multimodality A more human way to work Not
More informationMobile Interaction with the Real World
Andreas Zimmermann, Niels Henze, Xavier Righetti and Enrico Rukzio (Eds.) Mobile Interaction with the Real World Workshop in conjunction with MobileHCI 2009 BIS-Verlag der Carl von Ossietzky Universität
More informationInteraction Techniques using Head Mounted Displays and Handheld Devices for Outdoor Augmented Reality
Interaction Techniques using Head Mounted Displays and Handheld Devices for Outdoor Augmented Reality by Rahul Budhiraja A thesis submitted in partial fulfillment of the requirements for the Degree of
More informationIMMIView. A multi-user solution for design review in real-time. Real Time Image Processing manuscript No. (will be inserted by the editor)
Real Time Image Processing manuscript No. (will be inserted by the editor) Ricardo Jota Bruno Araújo Luís Bruno João M. Pereira Joaquim A. Jorge IMMIView A multi-user solution for design review in real-time
More informationThe 21 st Century Wireless Classroom Network for AP Calculus
The 21 st Century Wireless Classroom Network for AP Calculus In this exploratory hands-on workshop, we will be solving Calculus problems with the HP Prime Graphing Calculator and the HP Wireless Classroom
More informationEnhancedTable: Supporting a Small Meeting in Ubiquitous and Augmented Environment
EnhancedTable: Supporting a Small Meeting in Ubiquitous and Augmented Environment Hideki Koike 1, Shin ichiro Nagashima 1, Yasuto Nakanishi 2, and Yoichi Sato 3 1 Graduate School of Information Systems,
More informationChapter 1 - Introduction
1 "We all agree that your theory is crazy, but is it crazy enough?" Niels Bohr (1885-1962) Chapter 1 - Introduction Augmented reality (AR) is the registration of projected computer-generated images over
More informationŞahinSim: A Flight Simulator for End-Game Simulations
ŞahinSim: A Flight Simulator for End-Game Simulations Özer Özaydın, D. Turgay Altılar Department of Computer Science ITU Informatics Institute Maslak, Istanbul, 34457, Turkey ozaydinoz@itu.edu.tr altilar@cs.itu.edu.tr
More informationTouch & Gesture. HCID 520 User Interface Software & Technology
Touch & Gesture HCID 520 User Interface Software & Technology Natural User Interfaces What was the first gestural interface? Myron Krueger There were things I resented about computers. Myron Krueger
More informationAn Implementation Review of Occlusion-Based Interaction in Augmented Reality Environment
An Implementation Review of Occlusion-Based Interaction in Augmented Reality Environment Mohamad Shahrul Shahidan, Nazrita Ibrahim, Mohd Hazli Mohamed Zabil, Azlan Yusof College of Information Technology,
More informationVR4D: An Immersive and Collaborative Experience to Improve the Interior Design Process
VR4D: An Immersive and Collaborative Experience to Improve the Interior Design Process Amine Chellali, Frederic Jourdan, Cédric Dumas To cite this version: Amine Chellali, Frederic Jourdan, Cédric Dumas.
More informationRealtime 3D Computer Graphics Virtual Reality
Realtime 3D Computer Graphics Virtual Reality Virtual Reality Input Devices Special input devices are required for interaction,navigation and motion tracking (e.g., for depth cue calculation): 1 WIMP:
More informationGUI and Gestures. CS334 Fall Daniel G. Aliaga Department of Computer Science Purdue University
GUI and Gestures CS334 Fall 2013 Daniel G. Aliaga Department of Computer Science Purdue University User Interfaces Human Computer Interaction Graphical User Interfaces History 2D interfaces VR/AR Interfaces
More informationAndroid User manual. Intel Education Lab Camera by Intellisense CONTENTS
Intel Education Lab Camera by Intellisense Android User manual CONTENTS Introduction General Information Common Features Time Lapse Kinematics Motion Cam Microscope Universal Logger Pathfinder Graph Challenge
More informationGuidelines for choosing VR Devices from Interaction Techniques
Guidelines for choosing VR Devices from Interaction Techniques Jaime Ramírez Computer Science School Technical University of Madrid Campus de Montegancedo. Boadilla del Monte. Madrid Spain http://decoroso.ls.fi.upm.es
More informationInteraction Techniques for Immersive Virtual Environments: Design, Evaluation, and Application
Interaction Techniques for Immersive Virtual Environments: Design, Evaluation, and Application Doug A. Bowman Graphics, Visualization, and Usability Center College of Computing Georgia Institute of Technology
More information3D Projected Imagery Freespace Control Unit (PIFCU): Design, Implementation, and Analysis
3D Projected Imagery Freespace Control Unit (PIFCU): Design, Implementation, and Analysis Holly Tina Ferguson Department of Computer Science & Engineering MI 2012-2013 Advisor: Dr. Stephen Turner 11/11/2013
More information