Multi-Modal User Interaction

Similar documents
Perceptual Interfaces. Matthew Turk s (UCSB) and George G. Robertson s (Microsoft Research) slides on perceptual p interfaces

Short Course on Computational Illumination

Introduction to Haptics

Virtual Reality Calendar Tour Guide

User Interface Agents

Advancements in Gesture Recognition Technology

HUMAN-COMPUTER INTERACTION: OVERVIEW ON STATE OF THE ART TECHNOLOGY

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright

Geo-Located Content in Virtual and Augmented Reality

Multimodal Research at CPK, Aalborg

CONTENT RICH INTERACTIVE, AND IMMERSIVE EXPERIENCES, IN ADVERTISING, MARKETING, AND EDUCATION

Interactive Tables. ~Avishek Anand Supervised by: Michael Kipp Chair: Vitaly Friedman

Augmented Reality Lecture notes 01 1

AR Glossary. Terms. AR Glossary 1

A SURVEY OF MOBILE APPLICATION USING AUGMENTED REALITY

MIRACLE: Mixed Reality Applications for City-based Leisure and Experience. Mark Billinghurst HIT Lab NZ October 2009

LCC 3710 Principles of Interaction Design. Readings. Sound in Interfaces. Speech Interfaces. Speech Applications. Motivation for Speech Interfaces

Realtime 3D Computer Graphics Virtual Reality

Mario Romero 2014/11/05. Multimodal Interaction and Interfaces Mixed Reality

Sketching Interface. Larry Rudolph April 24, Pervasive Computing MIT SMA 5508 Spring 2006 Larry Rudolph

Sketching Interface. Motivation

ISO JTC 1 SC 24 WG9 G E R A R D J. K I M K O R E A U N I V E R S I T Y

GLOSSARY for National Core Arts: Media Arts STANDARDS

Ubiquitous Computing Summer Episode 16: HCI. Hannes Frey and Peter Sturm University of Trier. Hannes Frey and Peter Sturm, University of Trier 1

Collaborating with a Mobile Robot: An Augmented Reality Multimodal Interface

ENHANCED HUMAN-AGENT INTERACTION: AUGMENTING INTERACTION MODELS WITH EMBODIED AGENTS BY SERAFIN BENTO. MASTER OF SCIENCE in INFORMATION SYSTEMS

Portfolio. Swaroop Kumar Pal swarooppal.wordpress.com github.com/swarooppal1088

6 Ubiquitous User Interfaces

Human Factors. We take a closer look at the human factors that affect how people interact with computers and software:

Mixed and Augmented Reality Reference Model as of January 2014

Human Robot Interaction (HRI)

What was the first gestural interface?

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

Virtual Environments. Ruth Aylett

Product Requirements Document

GUIBDSS Gestural User Interface Based Digital Sixth Sense The wearable computer

Classifying 3D Input Devices

3D Interaction Techniques

RV - AULA 05 - PSI3502/2018. User Experience, Human Computer Interaction and UI

Interactive Simulation: UCF EIN5255. VR Software. Audio Output. Page 4-1

Determining Optimal Player Position, Distance, and Scale from a Point of Interest on a Terrain

HUMAN-COMPUTER INTERACTION: OVERVIEW ON STATE OF THE ART

Introduction to HCI. CS4HC3 / SE4HC3/ SE6DO3 Fall Instructor: Kevin Browne

ISCW 2001 Tutorial. An Introduction to Augmented Reality

Immersion in Multimodal Gaming

Touch & Gesture. HCID 520 User Interface Software & Technology

Heads up interaction: glasgow university multimodal research. Eve Hoggan

Human Robot Dialogue Interaction. Barry Lumpkin

Classifying 3D Input Devices

Interface Design V: Beyond the Desktop

ARMY RDT&E BUDGET ITEM JUSTIFICATION (R2 Exhibit)

Multi-sensory Tracking of Elders in Outdoor Environments on Ambient Assisted Living

Multi-Modal User Interaction. Lecture 3: Eye Tracking and Applications

Naturalness in the Design of Computer Hardware - The Forgotten Interface?

Multi-modal Human-Computer Interaction. Attila Fazekas.

Enhancing Shipboard Maintenance with Augmented Reality

Interior Design with Augmented Reality

VIRTUAL REALITY Introduction. Emil M. Petriu SITE, University of Ottawa

Lecture 2: Sensors. Zheng-Hua Tan

UUIs Ubiquitous User Interfaces

Definitions and Application Areas

Charting Past, Present, and Future Research in Ubiquitous Computing

Advances in Human!!!!! Computer Interaction

Multi-modal Human-computer Interaction

EMPOWERING THE CONNECTED FIELD FORCE WORKER WITH ADVANCED ANALYTICS MATTHEW SHORT ACCENTURE LABS

Mission Space. Value-based use of augmented reality in support of critical contextual environments

CHAPTER 1. INTRODUCTION 16

CS 315 Intro to Human Computer Interaction (HCI)

Mixed / Augmented Reality in Action

The Mixed Reality Book: A New Multimedia Reading Experience

ELG 5121/CSI 7631 Fall Projects Overview. Projects List

revolutionizing Subhead Can Be Placed Here healthcare Anders Gronstedt, Ph.D., President, Gronstedt Group September 22, 2017

Augmented & Virtual Reality. Grand Computers Club May 18, 2016

Mobile and Pervasive Game Technologies. Joel Ross ICS 62 05/19/2011

CS415 Human Computer Interaction

interactive laboratory

Week-1 [8/29, 31, 9/2]: Introduction, Discussion of Lab Platforms (Jetson)

mixed reality mixed reality & (tactile and) tangible interaction (tactile and) tangible interaction class housekeeping about me

Virtual Reality in Neuro- Rehabilitation and Beyond

The 8 th International Scientific Conference elearning and software for Education Bucharest, April 26-27, / X

User Interface Software Projects

AR Tamagotchi : Animate Everything Around Us

Augmented Reality 3D Pop-up Book: An Educational Research Study

Augmented and Virtual Reality

Revolution of the Information Viewport

CSC 2524, Fall 2017 AR/VR Interaction Interface

User Interaction and Perception from the Correlation of Dynamic Visual Responses Melinda Piper

CS415 Human Computer Interaction

IMGD 4000 Technical Game Development II Interaction and Immersion

Perception, Meaning and Transmodal Design

Central Valley School District (Computer Literacy 7) Curriculum Map (Grade 7) Week 1

Unit 23. QCF Level 3 Extended Certificate Unit 23 Human Computer Interaction

Do-It-Yourself Object Identification Using Augmented Reality for Visually Impaired People

SIXTH SENSE TECHNOLOGY A STEP AHEAD

Technology offer. Aerial obstacle detection software for the visually impaired

Inventory of a Potential Mobile Augmented Reality Genre for Learning

HAPTIC USER INTERFACES Final lecture

D S R G. Alina Mashko, GUI universal and global design. Department of vehicle technology. Faculty of Transportation Sciences

AAU SUMMER SCHOOL PROGRAMMING SOCIAL ROBOTS FOR HUMAN INTERACTION LECTURE 10 MULTIMODAL HUMAN-ROBOT INTERACTION

Recent Progress on Augmented-Reality Interaction in AIST

Transcription:

Multi-Modal User Interaction Lecture 4: Multiple Modalities Zheng-Hua Tan Department of Electronic Systems Aalborg University, Denmark zt@es.aau.dk MMUI, IV, Zheng-Hua Tan 1 Outline Multimodal interface Various modalities and their combination Perceptual user interface MMUI, IV, Zheng-Hua Tan 2 1

Multimodal system characteristics Recognition of simultaneous or alternative individual modes Type and size of gesture vocabulary speech vocabulary (and grammar) Type of signal fusion Type of platform and applications MMUI, IV, Zheng-Hua Tan 3 Challenges Development of cognitive theories to guide multimodal system design Development of effective natural language processing Dialogue processing Error-handling techniques Function robustly and adaptively Support for collaborative multiperson use http://www.geekstir.com/project-natal-milo-xbox-360 MMUI, IV, Zheng-Hua Tan 4 2

What is multimodal interface Humans perceive the world through senses. Touch, Smell, Sight, Hearing, and Taste A mode = Communication through one sense Computers process information through modes Keyboard, Microphone, Camera etc. Multimodal Interfaces try to combine several different modes of communicating: Speech, gesture, sketch Provide user with multiple modalities (communication skills) Multiple styles of interaction Simultaneous or not Fine-grained distinctions: Visual: Graphics, Text, Simulation Auditory: Speech, Non-verbal sounds (Skantze, 2010) MMUI, IV, Zheng-Hua Tan 5 Multimedia vs multimodal Multimedia more than one mode of communication is output to the user E.g. a sound clip attached to a presentation. Media channels: Text, graphics, animation, video: all visual media Multimodal computer processes more than one mode of communication. E.g. the combined input of speech and touch in new mobile phones Sensory modalities: Visual, auditory, tactile, (Skantze, 2010) MMUI, IV, Zheng-Hua Tan 6 3

Potential Output Modalities Visual: Visualization 3D GUIs Virtual/Augmented Reality Auditory: Speech Embodied Conversational Sound Haptics (tactile) Force feedback Low freq. bass Pain Taste? Scent? (Skantze, 2010) MMUI, IV, Zheng-Hua Tan 7 Possible input modalities Speech or other sounds Head movements (facial expression, gaze) Pointing, pen, touch Body movement/gestures multimodal interaction Motion controller (accelerometer) Tangibles Positioning Brain? Biomodalities? (sweat, pulse, respiration) (Skantze, 2010) MMUI, IV, Zheng-Hua Tan 8 4

Outline Multimodal interface Various modalities and their combination Perceptual user interface MMUI, IV, Zheng-Hua Tan 9 Multimodal speech and pen-gesture applications Interpret speech and pen-based gestureal input in a robust manner Bolt s Put That There concept MMUI, IV, Zheng-Hua Tan 10 5

IBM s human-centric word processor Architectural flow of signal and language processing Delete this word <points to word>. MMUI, IV, Zheng-Hua Tan 11 Boeing s speech and gesture system Give me that <points to an object>. Fly forward, or a flying gesture. MMUI, IV, Zheng-Hua Tan 12 6

Microsoft MIPAD: A multimodal interaction prototype (Huang et al. 2001) MMUI, IV, Zheng-Hua Tan 13 Microsoft MIPAD: A multimodal interaction prototype (Huang et al. 2001) MMUI, IV, Zheng-Hua Tan 14 7

Microsoft MIPAD: A multimodal interaction prototype (Huang et al. 2001) MMUI, IV, Zheng-Hua Tan 15 Typical info flow in a multimodal architecture (Oviatt et al. 2000) MMUI, IV, Zheng-Hua Tan 16 8

Interactive surface Scratch input Touch table MMUI, IV, Zheng-Hua Tan 17 Scratch input (C. Harrison, 2008) MMUI, IV, Zheng-Hua Tan 18 9

Touch table MMUI, IV, Zheng-Hua Tan 19 Reactable Fiducial markers used in reactivision MMUI, IV, Zheng-Hua Tan 20 10

Camera Phone Based Motion Sensing (Wang et al. 2006) MMUI, IV, Zheng-Hua Tan 21 Egocentric interaction Exploits the spatial relation between user and device and uses changes in this relation as input commands. (T. Luel and F. Mazzone, 2009) (M.H. Justesen, et al. 2010) MMUI, IV, Zheng-Hua Tan 22 11

3D sensing a) an image with persons and information overlay Elderly care, survelience b) detected foreground and information (Andersen, et al. 2010) 23 Finding information Google it! Layar First Mobile Augmented Reality Browser The world is the interface! 24 12

Layar First Mobile Augmented Reality Browser submit new layers, manage their layers and accounts interfaces http://layar.com/ : founded in June 2009 in Amsterdam 25 Layar cont. Superposition of multiple layers: Reality, Design Layout, Point of Interest (POI) Layar ar browses ressources rces on the server to display the POI. Layar uses the HTTP GET request (Requests a representation of the specified resource. A simple action of retrieval.) Layar allows the creation of layers by developers. No license costs involved creating a layer. 26 13

Layar cont. 500 layers developed by from individuals to small enterprises to large companies; 2000 layers in development. Over 1 million active end-users. Applications for marketing. Support all Android devices and the iphone 3GS. A Symbian version is in development. Need internet connection, camera, GPS and compass. 27 The world is my interface Mobile devices can be used to interact with the "Internet of Things. Sensors in smartphones to revolutionize the UI: microphones cameras motion sensors proximity sensors, and location sensors. Many application examples http://www.lucidproject.org/ 28 14

Brain-computer interface (BCI) (Schalk 2004) MMUI, IV, Zheng-Hua Tan 29 Outline Multimodal interface Various modalities and their combination Perceptual user interface MMUI, IV, Zheng-Hua Tan 30 15

The media equation Nass and Reeves s initial intuitions: What seems most obvious is that media are tools, pieces of hardware, not players in social life. Like all other tools, it seems that media simply help people accomplish tasks, learn new information, or entertain themselves. People don t have social relationships with tools. MMUI, IV, Zheng-Hua Tan 31 The media equation Their experiments subsequently convinced them that these intuitions were wrong, and that people do not predominately view media as tools. People tend to equate media and real life the media equation: Media = real life Individuals interactions with computers, television, and new media are fundamentally social and natural, just like interactions in real life. To bypass the media equation requires effort and is difficult to sustain. MMUI, IV, Zheng-Hua Tan 32 16

Perceptual user interface Highly interactive, multimodal interfaces modeled after natural human-to-human interaction, with the goal of enabling people to interact with technology in a similar fashion to how they interact with each other and with the physical world. (Matthew Turk) MMUI, IV, Zheng-Hua Tan 33 Perceptual user interface Vision based interfaces Gesture recognition Full body tracking Head tracking Eye-gaze tracking Audio based interfaces Interaction between man and machine should be based on the very same concepts as that between humans, i.e., it should be intuitive, multi-modal and based on emotion. - Reeves and Nass (1996), The Media Equation. MMUI, IV, Zheng-Hua Tan 34 17

Summary Multimodal interface Various modalities and their combination Perceptual user interface MMUI, IV, Zheng-Hua Tan 35 18