Applying Vision to Intelligent Human-Computer Interaction

Size: px
Start display at page:

Download "Applying Vision to Intelligent Human-Computer Interaction"

Transcription

1 Applying Vision to Intelligent Human-Computer Interaction Guangqi Ye Department of Computer Science The Johns Hopkins University Baltimore, MD October 21,

2 Vision for Natural HCI Advantages Affordable, non-intrusive, rich info. Crucial in multimodal interface Speech/gesture system Vision-Based HCI: 3D interface, natural HandVu and ARToolkit by M. Turk, et. al 2

3 Motivation Haptics + Vision Remove constant contact limit. Gestures for vision-based HCI Intuitive with representation power Applications: 3D VE, tele-op., surgical Addressed problems Visual data collection Analysis, model and recognition 3

4 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment 4

5 Vision + Haptics 3D Registration via visual tracking Remove limitation of constant contact Different passive objects to generate various sensation 5

6 Vision: Hand Segmentation Model background: color histograms Foreground detection: histogram matching Skin modeling: Gaussian model on Hue 6

7 Vision: Fingertip Tracking Fingertip detection: model-based Tracking: prediction (Kalman) + local detection 7

8 Haptics Module 3-D registration: Interaction simulation Examples: planes, buttons 8

9 Experimetal Results System: Pentimum III PC, 12fps 9

10 Vision + Haptics: Video 10

11 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment Conclusions 11

12 Visual Modeling of Gestures: General Framework Gesture generation Gesture recognition 12

13 Related Research in Modeling Gestures for HCI 13

14 Targeted Problems Analysis: mostly tracking-based Our approach: using localized parser Model: single modality (static/dynamic) Our model: coherent multimodal framework Recognition: Limited vocabulary/users Our contributions: large-scale experiment 14

15 Visual Interaction Cues(VICs) Paradigm Site-centered interaction Example: cell phone buttons 15

16 VICs State Mode Extend interaction functionality 3D gestures 16

17 VICs Principle: Sited Interaction Component mapping 17

18 Localized Parsers Low-level Parsers Motion, shape Learning-Based Modeling Neural Networks, HMMs 18

19 System Architecture 19

20 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment Conclusions 20

21 4D-Touchpad System Geometric calib. Homography-based Chromatic calib. Affine model for appearance transform 21

22 System Calibration Example 22

23 Hand Detection Foreground segmentation Image difference Modeling skin color Thresholding in YUV space Training: 16 users, 98% accuracy Hand region detection Merge skin pixels in segmented foreground 23

24 Hand Detection Example 24

25 Integrated into Existing Interface Shape parser + state-based gesture modeling 25

26 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment Conclusions 26

27 Efficient Motion Capture of 3D Gesture Capturing shape and motion in local space Appearance feature volume Region-based stereo matching Motion: differencing appearance 27

28 Appearance Feature Example 28

29 Posture Modeling Using 3D Feature Model 1: 3-layer neural networks Input: raw feature NN: 20 hidden nodes Posture Pick Push Press-Left Press-Right Stop Grab Drop Silence Training 99.97% % % % % % % 99.98% Testing 99.18% 99.93% 99.89% 99.96% % 99.82% 99.82% 98.56% 29

30 Posture Modeling Using 3D Feature Model 2: histogram-based ML Input: vector quantization, 96 clusters Posture Pick Push Press-Left Press-Right Stop Grab Drop Silence Training 96.95% 96.98% % 99.07% 99.80% 98.28% % 98.90% Testing 97.50% % 94.83% 98.15% % 95.00% 98.85% 98.68% 30

31 Dynamic Gesture Modeling Hidden Markov Models Input: VQ, 96 symbols Extension: modeling stop state p(s T ) Gesture Standard Training Standard Testing Extended Training Extended Testing Twist Twist-Anti Flip Negative Overall

32 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment Conclusions 32

33 Model Multimodal Gestures Low-level gesture as Gesture Words 3 classes: static, dynamic, parameterized High-level gesture Sequence of GWords Bigram model to capture constraints 33

34 Example Model 34

35 Learning and Inference Learning the bigram: maximum likelihood Inference: greedy-choice for online Choose path with maximum p(v t v t-1 )p(s t v t ) 35

36 Outline Vision/Haptics system Modular framework for VBHCI 4DT platform Novel scheme for hand motion capture Modeling composite gestures Human factors experiment Conclusions 36

37 Human Factors Experiment Gesture vocabulary: 14 gesture words Multi-Modal: posture, parameterized and dynamic gestures 9 possible gesture sentences Data collecting 16 volunteers, including 7 female 5 training and 3 testing sequences Gesture cuing: video + text 37

38 Example Video Cuing 38

39 Modeling Parameterized Gesture Three Gestures: moving, rotate, resize Region tracking on segmented image Pyramid SSD tracker: X =R(Θ)X + T Template: 150 x 150 Evaluation Average residual error: 5.5/6.0/6.7 pixels 39

40 Composite Gesture Modeling Result Gesture Pushing Twisting Twisting-Anti Dropping Flipping Moving Rotating Stopping Resizing Total Sequences Ratio 97.14% % 96.42% 96.55% 96.89% 94.29% 92.59% % 96.67% 96.47% 40

41 User Feedback on Gesturebased Interface Gesture vocabulary Easy to learn: 100% agree Fatigue compared to GUI with mouse 50%: comparable, 38%: more tired, 12% less Overall convenience compared to GUI with mouse 44%: more comfortable 44%: comparable 12%: more awkward 41

42 Contributions Vision+Haptics: novel multimodal interface VICs/4DT: a new framework for VBHCI and data collection Efficient motion capture for gesture analysis Heterogeneous gestures modeling Large-scale gesture experiments 42

43 Acknowledgement Dr. G. Hager Dr. D. Burschka, J. Corso, A. Okamura Dr. J. Eisner, R. Etienne-cummings, I. Shafran CIRL Lab X. Dai, L. Lu, S. Lee, M. Dewan, N. Howie, H. Lin, S. Seshanami Haptics Explorative Lab J. Abott, P. Marayong 43

44 Thanks 44

VICs: A Modular Vision-Based HCI Framework

VICs: A Modular Vision-Based HCI Framework VICs: A Modular Vision-Based HCI Framework The Visual Interaction Cues Project Guangqi Ye, Jason Corso Darius Burschka, & Greg Hager CIRL, 1 Today, I ll be presenting work that is part of an ongoing project

More information

Research Seminar. Stefano CARRINO fr.ch

Research Seminar. Stefano CARRINO  fr.ch Research Seminar Stefano CARRINO stefano.carrino@hefr.ch http://aramis.project.eia- fr.ch 26.03.2010 - based interaction Characterization Recognition Typical approach Design challenges, advantages, drawbacks

More information

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005. Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.

More information

VisHap: Augmented Reality Combining Haptics and Vision

VisHap: Augmented Reality Combining Haptics and Vision VisHap: Augmented Reality Combining Haptics and Vision Guangqi Ye 1, Jason J. Corso 1, Gregory D. Hager 1, Allison M. Okamura 1,2 Departments of 1 Computer Science and 2 Mechanical Engineering The Johns

More information

Visual Interpretation of Hand Gestures as a Practical Interface Modality

Visual Interpretation of Hand Gestures as a Practical Interface Modality Visual Interpretation of Hand Gestures as a Practical Interface Modality Frederik C. M. Kjeldsen Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Graduate

More information

LOOKING AHEAD: UE4 VR Roadmap. Nick Whiting Technical Director VR / AR

LOOKING AHEAD: UE4 VR Roadmap. Nick Whiting Technical Director VR / AR LOOKING AHEAD: UE4 VR Roadmap Nick Whiting Technical Director VR / AR HEADLINE AND IMAGE LAYOUT RECENT DEVELOPMENTS RECENT DEVELOPMENTS At Epic, we drive our engine development by creating content. We

More information

VICs: A Modular Vision-Based HCI Framework

VICs: A Modular Vision-Based HCI Framework Proceedings of ICVS 2003, pp. 257-267 April 2003, Graz, Austria VICs: A Modular Vision-Based HCI Framework Guangqi Ye, Jason Corso, Darius Burschka, and Gregory D. Hager The Johns Hopkins University Computational

More information

Computer Vision. Howie Choset Introduction to Robotics

Computer Vision. Howie Choset   Introduction to Robotics Computer Vision Howie Choset http://www.cs.cmu.edu.edu/~choset Introduction to Robotics http://generalrobotics.org What is vision? What is computer vision? Edge Detection Edge Detection Interest points

More information

Perceptual Interfaces. Matthew Turk s (UCSB) and George G. Robertson s (Microsoft Research) slides on perceptual p interfaces

Perceptual Interfaces. Matthew Turk s (UCSB) and George G. Robertson s (Microsoft Research) slides on perceptual p interfaces Perceptual Interfaces Adapted from Matthew Turk s (UCSB) and George G. Robertson s (Microsoft Research) slides on perceptual p interfaces Outline Why Perceptual Interfaces? Multimodal interfaces Vision

More information

Advanced Techniques for Mobile Robotics Location-Based Activity Recognition

Advanced Techniques for Mobile Robotics Location-Based Activity Recognition Advanced Techniques for Mobile Robotics Location-Based Activity Recognition Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz Activity Recognition Based on L. Liao, D. J. Patterson, D. Fox,

More information

A Kinect-based 3D hand-gesture interface for 3D databases

A Kinect-based 3D hand-gesture interface for 3D databases A Kinect-based 3D hand-gesture interface for 3D databases Abstract. The use of natural interfaces improves significantly aspects related to human-computer interaction and consequently the productivity

More information

Neural Networks The New Moore s Law

Neural Networks The New Moore s Law Neural Networks The New Moore s Law Chris Rowen, PhD, FIEEE CEO Cognite Ventures December 216 Outline Moore s Law Revisited: Efficiency Drives Productivity Embedded Neural Network Product Segments Efficiency

More information

Virtual Touch Human Computer Interaction at a Distance

Virtual Touch Human Computer Interaction at a Distance International Journal of Computer Science and Telecommunications [Volume 4, Issue 5, May 2013] 18 ISSN 2047-3338 Virtual Touch Human Computer Interaction at a Distance Prasanna Dhisale, Puja Firodiya,

More information

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison

Design a Model and Algorithm for multi Way Gesture Recognition using Motion and Image Comparison e-issn 2455 1392 Volume 2 Issue 10, October 2016 pp. 34 41 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com Design a Model and Algorithm for multi Way Gesture Recognition using Motion and

More information

Today. CS 395T Visual Recognition. Course content. Administration. Expectations. Paper reviews

Today. CS 395T Visual Recognition. Course content. Administration. Expectations. Paper reviews Today CS 395T Visual Recognition Course logistics Overview Volunteers, prep for next week Thursday, January 18 Administration Class: Tues / Thurs 12:30-2 PM Instructor: Kristen Grauman grauman at cs.utexas.edu

More information

Advanced Man-Machine Interaction

Advanced Man-Machine Interaction Signals and Communication Technology Advanced Man-Machine Interaction Fundamentals and Implementation Bearbeitet von Karl-Friedrich Kraiss 1. Auflage 2006. Buch. XIX, 461 S. ISBN 978 3 540 30618 4 Format

More information

REAL TIME GESTURE RECOGNITION SYSTEM FOR ADAS CHEE YING XUAN A REPORT SUBMITTED TO. Universiti Tunku Abdul Rahman

REAL TIME GESTURE RECOGNITION SYSTEM FOR ADAS CHEE YING XUAN A REPORT SUBMITTED TO. Universiti Tunku Abdul Rahman REAL TIME GESTURE RECOGNITION SYSTEM FOR ADAS BY CHEE YING XUAN A REPORT SUBMITTED TO Universiti Tunku Abdul Rahman in partial fulfilment of the requirements for the degree of BACHELOR OF INFORMATION SYSTEMS

More information

COMPARATIVE STUDY AND ANALYSIS FOR GESTURE RECOGNITION METHODOLOGIES

COMPARATIVE STUDY AND ANALYSIS FOR GESTURE RECOGNITION METHODOLOGIES http:// COMPARATIVE STUDY AND ANALYSIS FOR GESTURE RECOGNITION METHODOLOGIES Rafiqul Z. Khan 1, Noor A. Ibraheem 2 1 Department of Computer Science, A.M.U. Aligarh, India 2 Department of Computer Science,

More information

Virtual Grasping Using a Data Glove

Virtual Grasping Using a Data Glove Virtual Grasping Using a Data Glove By: Rachel Smith Supervised By: Dr. Kay Robbins 3/25/2005 University of Texas at San Antonio Motivation Navigation in 3D worlds is awkward using traditional mouse Direct

More information

Service Robots in an Intelligent House

Service Robots in an Intelligent House Service Robots in an Intelligent House Jesus Savage Bio-Robotics Laboratory biorobotics.fi-p.unam.mx School of Engineering Autonomous National University of Mexico UNAM 2017 OUTLINE Introduction A System

More information

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3

More information

Different Hand Gesture Recognition Techniques Using Perceptron Network

Different Hand Gesture Recognition Techniques Using Perceptron Network Different Hand Gesture Recognition Techniques Using Perceptron Network Nidhi Chauhan Department of Computer Science & Engg. Suresh Gyan Vihar University, Jaipur(Raj.) Email: nidhi99.chauhan@gmail.com Abstract

More information

3D-Position Estimation for Hand Gesture Interface Using a Single Camera

3D-Position Estimation for Hand Gesture Interface Using a Single Camera 3D-Position Estimation for Hand Gesture Interface Using a Single Camera Seung-Hwan Choi, Ji-Hyeong Han, and Jong-Hwan Kim Department of Electrical Engineering, KAIST, Gusung-Dong, Yusung-Gu, Daejeon, Republic

More information

Short Course on Computational Illumination

Short Course on Computational Illumination Short Course on Computational Illumination University of Tampere August 9/10, 2012 Matthew Turk Computer Science Department and Media Arts and Technology Program University of California, Santa Barbara

More information

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception

More information

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY

A SURVEY ON GESTURE RECOGNITION TECHNOLOGY A SURVEY ON GESTURE RECOGNITION TECHNOLOGY Deeba Kazim 1, Mohd Faisal 2 1 MCA Student, Integral University, Lucknow (India) 2 Assistant Professor, Integral University, Lucknow (india) ABSTRACT Gesture

More information

Available theses in robotics (November 2017) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin

Available theses in robotics (November 2017) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Available theses in robotics (November 2017) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Ergonomic positioning of bulky objects Thesis 1 Robot acts as a 3rd hand for workpiece positioning: Muscular

More information

Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane

Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane Journal of Communication and Computer 13 (2016) 329-337 doi:10.17265/1548-7709/2016.07.002 D DAVID PUBLISHING Development of a Finger Mounted Type Haptic Device Using a Plane Approximated to Tangent Plane

More information

Vision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab

Vision-based User-interfaces for Pervasive Computing. CHI 2003 Tutorial Notes. Trevor Darrell Vision Interface Group MIT AI Lab Vision-based User-interfaces for Pervasive Computing Tutorial Notes Vision Interface Group MIT AI Lab Table of contents Biographical sketch..ii Agenda..iii Objectives.. iv Abstract..v Introduction....1

More information

Maneesh Dewan. Prepared on: April 11, 2007

Maneesh Dewan. Prepared on: April 11, 2007 Maneesh Dewan maneesh@cs.jhu.edu www.cs.jhu.edu/~maneesh 307, E. University Parkway, 3400 N. Charles Street, Baltimore, MD 21218. NEB B28, Baltimore, MD 21218. Phone: (410) 900 8804 (C) Phone: (410) 516

More information

Single Chip for Imaging, Color Segmentation, Histogramming and Pattern Matching

Single Chip for Imaging, Color Segmentation, Histogramming and Pattern Matching Paper Title: Single Chip for Imaging, Color Segmentation, Histogramming and Pattern Matching Authors: Ralph Etienne-Cummings 1,2, Philippe Pouliquen 1,2, M. Anthony Lewis 1 Affiliation: 1 Iguana Robotics,

More information

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi

An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi An Evaluation of Automatic License Plate Recognition Vikas Kotagyale, Prof.S.D.Joshi Department of E&TC Engineering,PVPIT,Bavdhan,Pune ABSTRACT: In the last decades vehicle license plate recognition systems

More information

Sketching Interface. Larry Rudolph April 24, Pervasive Computing MIT SMA 5508 Spring 2006 Larry Rudolph

Sketching Interface. Larry Rudolph April 24, Pervasive Computing MIT SMA 5508 Spring 2006 Larry Rudolph Sketching Interface Larry April 24, 2006 1 Motivation Natural Interface touch screens + more Mass-market of h/w devices available Still lack of s/w & applications for it Similar and different from speech

More information

Sketching Interface. Motivation

Sketching Interface. Motivation Sketching Interface Larry Rudolph April 5, 2007 1 1 Natural Interface Motivation touch screens + more Mass-market of h/w devices available Still lack of s/w & applications for it Similar and different

More information

Data Quality Monitoring of the CMS Pixel Detector

Data Quality Monitoring of the CMS Pixel Detector Data Quality Monitoring of the CMS Pixel Detector 1 * Purdue University Department of Physics, 525 Northwestern Ave, West Lafayette, IN 47906 USA E-mail: petra.merkel@cern.ch We present the CMS Pixel Data

More information

Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin

Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Ergonomic positioning of bulky objects Thesis 1 Robot acts as a 3rd hand for workpiece positioning: Muscular fatigue

More information

Automated hand recognition as a human-computer interface

Automated hand recognition as a human-computer interface Automated hand recognition as a human-computer interface Sergii Shelpuk SoftServe, Inc. sergii.shelpuk@gmail.com Abstract This paper investigates applying Machine Learning to the problem of turning a regular

More information

Robotic System Simulation and Modeling Stefan Jörg Robotic and Mechatronic Center

Robotic System Simulation and Modeling Stefan Jörg Robotic and Mechatronic Center Robotic System Simulation and ing Stefan Jörg Robotic and Mechatronic Center Outline Introduction The SAFROS Robotic System Simulator Robotic System ing Conclusions Folie 2 DLR s Mirosurge: A versatile

More information

Virtual Environment Interaction Based on Gesture Recognition and Hand Cursor

Virtual Environment Interaction Based on Gesture Recognition and Hand Cursor Virtual Environment Interaction Based on Gesture Recognition and Hand Cursor Chan-Su Lee Kwang-Man Oh Chan-Jong Park VR Center, ETRI 161 Kajong-Dong, Yusong-Gu Taejon, 305-350, KOREA +82-42-860-{5319,

More information

OBJECTIVE OF THE BOOK ORGANIZATION OF THE BOOK

OBJECTIVE OF THE BOOK ORGANIZATION OF THE BOOK xv Preface Advancement in technology leads to wide spread use of mounting cameras to capture video imagery. Such surveillance cameras are predominant in commercial institutions through recording the cameras

More information

Hand & Upper Body Based Hybrid Gesture Recognition

Hand & Upper Body Based Hybrid Gesture Recognition Hand & Upper Body Based Hybrid Gesture Prerna Sharma #1, Naman Sharma *2 # Research Scholor, G. B. P. U. A. & T. Pantnagar, India * Ideal Institue of Technology, Ghaziabad, India Abstract Communication

More information

Driver Assistance for "Keeping Hands on the Wheel and Eyes on the Road"

Driver Assistance for Keeping Hands on the Wheel and Eyes on the Road ICVES 2009 Driver Assistance for "Keeping Hands on the Wheel and Eyes on the Road" Cuong Tran and Mohan Manubhai Trivedi Laboratory for Intelligent and Safe Automobiles (LISA) University of California

More information

Hand Gesture Recognition System Using Camera

Hand Gesture Recognition System Using Camera Hand Gesture Recognition System Using Camera Viraj Shinde, Tushar Bacchav, Jitendra Pawar, Mangesh Sanap B.E computer engineering,navsahyadri Education Society sgroup of Institutions,pune. Abstract - In

More information

DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING SEMINAR REPORT ON GESTURE RECOGNITION SUBMITTED BY PRAKRUTHI.V ( )

DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING SEMINAR REPORT ON GESTURE RECOGNITION SUBMITTED BY PRAKRUTHI.V ( ) DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING PONDICHERRY ENGINEERING COLLEGE SEMINAR REPORT ON GESTURE RECOGNITION SUBMITTED BY PRAKRUTHI.V (283175132) PRATHIBHA ANNAPURNA.P (283175135) SARANYA.S (283175174)

More information

Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL

Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL Design and Development of a Marker-based Augmented Reality System using OpenCV and OpenGL Yap Hwa Jentl, Zahari Taha 2, Eng Tat Hong", Chew Jouh Yeong" Centre for Product Design and Manufacturing (CPDM).

More information

Activity monitoring and summarization for an intelligent meeting room

Activity monitoring and summarization for an intelligent meeting room IEEE Workshop on Human Motion, Austin, Texas, December 2000 Activity monitoring and summarization for an intelligent meeting room Ivana Mikic, Kohsia Huang, Mohan Trivedi Computer Vision and Robotics Research

More information

Interaction via motion observation

Interaction via motion observation Interaction via motion observation M A Foyle 1 and R J McCrindle 2 School of Systems Engineering, University of Reading, Reading, UK mfoyle@iee.org, r.j.mccrindle@reading.ac.uk www.sse.reading.ac.uk ABSTRACT

More information

CS415 Human Computer Interaction

CS415 Human Computer Interaction CS415 Human Computer Interaction Lecture 10 Advanced HCI Universal Design & Intro to Cognitive Models October 30, 2016 Sam Siewert Summary of Thoughts on ITS Collective Wisdom of Our Classes (2015, 2016)

More information

Haptic Virtual Fixtures for Robot-Assisted Manipulation

Haptic Virtual Fixtures for Robot-Assisted Manipulation Haptic Virtual Fixtures for Robot-Assisted Manipulation Jake J. Abbott, Panadda Marayong, and Allison M. Okamura Department of Mechanical Engineering, The Johns Hopkins University {jake.abbott, pmarayong,

More information

신경망기반자동번역기술. Konkuk University Computational Intelligence Lab. 김강일

신경망기반자동번역기술. Konkuk University Computational Intelligence Lab.  김강일 신경망기반자동번역기술 Konkuk University Computational Intelligence Lab. http://ci.konkuk.ac.kr kikim01@kunkuk.ac.kr 김강일 Index Issues in AI and Deep Learning Overview of Machine Translation Advanced Techniques in

More information

Coded Exposure HDR Light-Field Video Recording

Coded Exposure HDR Light-Field Video Recording Coded Exposure HDR Light-Field Video Recording David C. Schedl, Clemens Birklbauer, and Oliver Bimber* Johannes Kepler University Linz *firstname.lastname@jku.at Exposure Sequence long exposed short HDR

More information

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition Hetal R. Thaker Atmiya Institute of Technology & science, Kalawad Road, Rajkot Gujarat, India C. K. Kumbharana,

More information

MRT: Mixed-Reality Tabletop

MRT: Mixed-Reality Tabletop MRT: Mixed-Reality Tabletop Students: Dan Bekins, Jonathan Deutsch, Matthew Garrett, Scott Yost PIs: Daniel Aliaga, Dongyan Xu August 2004 Goals Create a common locus for virtual interaction without having

More information

Real-time human control of robots for robot skill synthesis (and a bit

Real-time human control of robots for robot skill synthesis (and a bit Real-time human control of robots for robot skill synthesis (and a bit about imitation) Erhan Oztop JST/ICORP, ATR/CNS, JAPAN 1/31 IMITATION IN ARTIFICIAL SYSTEMS (1) Robotic systems that are able to imitate

More information

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Komal Hasija 1, Rajani Mehta 2 Abstract Recognition is a very effective area of research in regard of security with the involvement

More information

Virtual Experiments as a Tool for Active Engagement

Virtual Experiments as a Tool for Active Engagement Virtual Experiments as a Tool for Active Engagement Lei Bao Stephen Stonebraker Gyoungho Lee Physics Education Research Group Department of Physics The Ohio State University Context Cues and Knowledge

More information

Vision Based Hand Gesture Recognition

Vision Based Hand Gesture Recognition Vision Based Hand Gesture Recognition Yanmin Zhu 1, Zhibo Yang 2, Bo Yuan 3 Intelligent Computing Lab, Division of Informatics Graduate School at Shenzhen, Tsinghua University Shenzhen 518055, P. R. China

More information

Activities at SC 24 WG 9: An Overview

Activities at SC 24 WG 9: An Overview Activities at SC 24 WG 9: An Overview G E R A R D J. K I M, C O N V E N E R I S O J T C 1 S C 2 4 W G 9 Mixed and Augmented Reality (MAR) ISO SC 24 and MAR ISO-IEC JTC 1 SC 24 Have developed standards

More information

Recognition Of Vehicle Number Plate Using MATLAB

Recognition Of Vehicle Number Plate Using MATLAB Recognition Of Vehicle Number Plate Using MATLAB Mr. Ami Kumar Parida 1, SH Mayuri 2,Pallabi Nayk 3,Nidhi Bharti 4 1Asst. Professor, Gandhi Institute Of Engineering and Technology, Gunupur 234Under Graduate,

More information

CSSE463: Image Recognition Day 2

CSSE463: Image Recognition Day 2 CSSE463: Image Recognition Day 2 Roll call Announcements: Moodle has drop box for Lab 1 Next class: lots more Matlab how-to (bring your laptop) Questions? Today: Color and color features Do questions 1-2

More information

CHAPTER 1. INTRODUCTION 16

CHAPTER 1. INTRODUCTION 16 1 Introduction The author s original intention, a couple of years ago, was to develop a kind of an intuitive, dataglove-based interface for Computer-Aided Design (CAD) applications. The idea was to interact

More information

Modern Control Theoretic Approach for Gait and Behavior Recognition. Charles J. Cohen, Ph.D. Session 1A 05-BRIMS-023

Modern Control Theoretic Approach for Gait and Behavior Recognition. Charles J. Cohen, Ph.D. Session 1A 05-BRIMS-023 Modern Control Theoretic Approach for Gait and Behavior Recognition Charles J. Cohen, Ph.D. ccohen@cybernet.com Session 1A 05-BRIMS-023 Outline Introduction - Behaviors as Connected Gestures Gesture Recognition

More information

Hand Segmentation for Hand Gesture Recognition

Hand Segmentation for Hand Gesture Recognition Hand Segmentation for Hand Gesture Recognition Sonal Singhai Computer Science department Medicaps Institute of Technology and Management, Indore, MP, India Dr. C.S. Satsangi Head of Department, information

More information

P15083: Virtual Visualization for Anatomy Teaching, Training and Surgery Simulation Applications. Gate Review

P15083: Virtual Visualization for Anatomy Teaching, Training and Surgery Simulation Applications. Gate Review P15083: Virtual Visualization for Anatomy Teaching, Training and Surgery Simulation Applications Gate Review Agenda review of starting objectives customer requirements, engineering requirements 50% goal,

More information

Overview of Challenges in the Development of Autonomous Mobile Robots. August 23, 2011

Overview of Challenges in the Development of Autonomous Mobile Robots. August 23, 2011 Overview of Challenges in the Development of Autonomous Mobile Robots August 23, 2011 What is in a Robot? Sensors Effectors and actuators (i.e., mechanical) Used for locomotion and manipulation Controllers

More information

ISO JTC 1 SC 24 WG9 G E R A R D J. K I M K O R E A U N I V E R S I T Y

ISO JTC 1 SC 24 WG9 G E R A R D J. K I M K O R E A U N I V E R S I T Y New Work Item Proposal: A Standard Reference Model for Generic MAR Systems ISO JTC 1 SC 24 WG9 G E R A R D J. K I M K O R E A U N I V E R S I T Y What is a Reference Model? A reference model (for a given

More information

Improved SIFT Matching for Image Pairs with a Scale Difference

Improved SIFT Matching for Image Pairs with a Scale Difference Improved SIFT Matching for Image Pairs with a Scale Difference Y. Bastanlar, A. Temizel and Y. Yardımcı Informatics Institute, Middle East Technical University, Ankara, 06531, Turkey Published in IET Electronics,

More information

Robust Hand Gesture Recognition for Robotic Hand Control

Robust Hand Gesture Recognition for Robotic Hand Control Robust Hand Gesture Recognition for Robotic Hand Control Ankit Chaudhary Robust Hand Gesture Recognition for Robotic Hand Control 123 Ankit Chaudhary Department of Computer Science Northwest Missouri State

More information

Multi-Modal User Interaction

Multi-Modal User Interaction Multi-Modal User Interaction Lecture 4: Multiple Modalities Zheng-Hua Tan Department of Electronic Systems Aalborg University, Denmark zt@es.aau.dk MMUI, IV, Zheng-Hua Tan 1 Outline Multimodal interface

More information

Masters of Engineering in Electrical Engineering Course Syllabi ( ) City University of New York--College of Staten Island

Masters of Engineering in Electrical Engineering Course Syllabi ( ) City University of New York--College of Staten Island City University of New York--College of Staten Island Masters of Engineering in Electrical Engineering Course Syllabi (2017-2018) Required Core Courses ELE 600/ MTH 6XX Probability Theory and Stochastic

More information

SIMULATION VOICE RECOGNITION SYSTEM FOR CONTROLING ROBOTIC APPLICATIONS

SIMULATION VOICE RECOGNITION SYSTEM FOR CONTROLING ROBOTIC APPLICATIONS SIMULATION VOICE RECOGNITION SYSTEM FOR CONTROLING ROBOTIC APPLICATIONS 1 WAHYU KUSUMA R., 2 PRINCE BRAVE GUHYAPATI V 1 Computer Laboratory Staff., Department of Information Systems, Gunadarma University,

More information

Gesture Recognition with Real World Environment using Kinect: A Review

Gesture Recognition with Real World Environment using Kinect: A Review Gesture Recognition with Real World Environment using Kinect: A Review Prakash S. Sawai 1, Prof. V. K. Shandilya 2 P.G. Student, Department of Computer Science & Engineering, Sipna COET, Amravati, Maharashtra,

More information

Classification for Motion Game Based on EEG Sensing

Classification for Motion Game Based on EEG Sensing Classification for Motion Game Based on EEG Sensing Ran WEI 1,3,4, Xing-Hua ZHANG 1,4, Xin DANG 2,3,4,a and Guo-Hui LI 3 1 School of Electronics and Information Engineering, Tianjin Polytechnic University,

More information

Research of an Algorithm on Face Detection

Research of an Algorithm on Face Detection , pp.217-222 http://dx.doi.org/10.14257/astl.2016.141.47 Research of an Algorithm on Face Detection Gong Liheng, Yang Jingjing, Zhang Xiao School of Information Science and Engineering, Hebei North University,

More information

Vision Review: Image Processing. Course web page:

Vision Review: Image Processing. Course web page: Vision Review: Image Processing Course web page: www.cis.udel.edu/~cer/arv September 7, Announcements Homework and paper presentation guidelines are up on web page Readings for next Tuesday: Chapters 6,.,

More information

GESTURE RECOGNITION WITH 3D CNNS

GESTURE RECOGNITION WITH 3D CNNS April 4-7, 2016 Silicon Valley GESTURE RECOGNITION WITH 3D CNNS Pavlo Molchanov Xiaodong Yang Shalini Gupta Kihwan Kim Stephen Tyree Jan Kautz 4/6/2016 Motivation AGENDA Problem statement Selecting the

More information

FP7 ICT Call 6: Cognitive Systems and Robotics

FP7 ICT Call 6: Cognitive Systems and Robotics FP7 ICT Call 6: Cognitive Systems and Robotics Information day Luxembourg, January 14, 2010 Libor Král, Head of Unit Unit E5 - Cognitive Systems, Interaction, Robotics DG Information Society and Media

More information

Enabling Cursor Control Using on Pinch Gesture Recognition

Enabling Cursor Control Using on Pinch Gesture Recognition Enabling Cursor Control Using on Pinch Gesture Recognition Benjamin Baldus Debra Lauterbach Juan Lizarraga October 5, 2007 Abstract In this project we expect to develop a machine-user interface based on

More information

Definitions and Application Areas

Definitions and Application Areas Definitions and Application Areas Ambient intelligence: technology and design Fulvio Corno Politecnico di Torino, 2013/2014 http://praxis.cs.usyd.edu.au/~peterris Summary Definition(s) Application areas

More information

Heads up interaction: glasgow university multimodal research. Eve Hoggan

Heads up interaction: glasgow university multimodal research. Eve Hoggan Heads up interaction: glasgow university multimodal research Eve Hoggan www.tactons.org multimodal interaction Multimodal Interaction Group Key area of work is Multimodality A more human way to work Not

More information

What was the first gestural interface?

What was the first gestural interface? stanford hci group / cs247 Human-Computer Interaction Design Studio What was the first gestural interface? 15 January 2013 http://cs247.stanford.edu Theremin Myron Krueger 1 Myron Krueger There were things

More information

Pattern Adaptive and Finger Image-guided Keypad Interface for In-vehicle Information Systems

Pattern Adaptive and Finger Image-guided Keypad Interface for In-vehicle Information Systems INTERNATIONAL JOURNAL ON SMART SENSING AND INTELLIGENT SYSTEMS, VOL. 1, NO. 3, SEPETEMBER 28 Pattern Adaptive and Finger Image-guided Keypad Interface for In-vehicle Information Systems Sang-Ho Kim, Kosuke

More information

Discriminative Training for Automatic Speech Recognition

Discriminative Training for Automatic Speech Recognition Discriminative Training for Automatic Speech Recognition 22 nd April 2013 Advanced Signal Processing Seminar Article Heigold, G.; Ney, H.; Schluter, R.; Wiesler, S. Signal Processing Magazine, IEEE, vol.29,

More information

Abstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction

Abstract. Keywords: Multi Touch, Collaboration, Gestures, Accelerometer, Virtual Prototyping. 1. Introduction Creating a Collaborative Multi Touch Computer Aided Design Program Cole Anagnost, Thomas Niedzielski, Desirée Velázquez, Prasad Ramanahally, Stephen Gilbert Iowa State University { someguy tomn deveri

More information

INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction

INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction INTAIRACT: Joint Hand Gesture and Fingertip Classification for Touchless Interaction Xavier Suau 1,MarcelAlcoverro 2, Adolfo Lopez-Mendez 3, Javier Ruiz-Hidalgo 2,andJosepCasas 3 1 Universitat Politécnica

More information

Navigation of PowerPoint Using Hand Gestures

Navigation of PowerPoint Using Hand Gestures Navigation of PowerPoint Using Hand Gestures Dnyanada R Jadhav 1, L. M. R. J Lobo 2 1 M.E Department of Computer Science & Engineering, Walchand Institute of technology, Solapur, India 2 Associate Professor

More information

PRACTICAL IMAGE AND VIDEO PROCESSING USING MATLAB

PRACTICAL IMAGE AND VIDEO PROCESSING USING MATLAB PRACTICAL IMAGE AND VIDEO PROCESSING USING MATLAB OGE MARQUES Florida Atlantic University *IEEE IEEE PRESS WWILEY A JOHN WILEY & SONS, INC., PUBLICATION CONTENTS LIST OF FIGURES LIST OF TABLES FOREWORD

More information

Enhanced Waveform Interpolative Coding at 4 kbps

Enhanced Waveform Interpolative Coding at 4 kbps Enhanced Waveform Interpolative Coding at 4 kbps Oded Gottesman, and Allen Gersho Signal Compression Lab. University of California, Santa Barbara E-mail: [oded, gersho]@scl.ece.ucsb.edu Signal Compression

More information

Gesture recognition based on arm tracking for human-robot interaction

Gesture recognition based on arm tracking for human-robot interaction Gesture recognition based on arm tracking for human-robot interaction Markos Sigalas, Haris Baltzakis and Panos Trahanias Institute of Computer Science Foundation for Research and Technology - Hellas (FORTH)

More information

Computer and Machine Vision

Computer and Machine Vision Computer and Machine Vision Lecture Week 7 Part-2 (Exam #1 Review) February 26, 2014 Sam Siewert Outline of Week 7 Basic Convolution Transform Speed-Up Concepts for Computer Vision Hough Linear Transform

More information

The Hand Gesture Recognition System Using Depth Camera

The Hand Gesture Recognition System Using Depth Camera The Hand Gesture Recognition System Using Depth Camera Ahn,Yang-Keun VR/AR Research Center Korea Electronics Technology Institute Seoul, Republic of Korea e-mail: ykahn@keti.re.kr Park,Young-Choong VR/AR

More information

RoboCup. Presented by Shane Murphy April 24, 2003

RoboCup. Presented by Shane Murphy April 24, 2003 RoboCup Presented by Shane Murphy April 24, 2003 RoboCup: : Today and Tomorrow What we have learned Authors Minoru Asada (Osaka University, Japan), Hiroaki Kitano (Sony CS Labs, Japan), Itsuki Noda (Electrotechnical(

More information

MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES

MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES MATLAB DIGITAL IMAGE/SIGNAL PROCESSING TITLES -2018 S.NO PROJECT CODE 1 ITIMP01 2 ITIMP02 3 ITIMP03 4 ITIMP04 5 ITIMP05 6 ITIMP06 7 ITIMP07 8 ITIMP08 9 ITIMP09 `10 ITIMP10 11 ITIMP11 12 ITIMP12 13 ITIMP13

More information

Multi-Modal User Interaction. Lecture 3: Eye Tracking and Applications

Multi-Modal User Interaction. Lecture 3: Eye Tracking and Applications Multi-Modal User Interaction Lecture 3: Eye Tracking and Applications Zheng-Hua Tan Department of Electronic Systems Aalborg University, Denmark zt@es.aau.dk 1 Part I: Eye tracking Eye tracking Tobii eye

More information

Human-Computer Intelligent Interaction: A Survey

Human-Computer Intelligent Interaction: A Survey Human-Computer Intelligent Interaction: A Survey Michael Lew 1, Erwin M. Bakker 1, Nicu Sebe 2, and Thomas S. Huang 3 1 LIACS Media Lab, Leiden University, The Netherlands 2 ISIS Group, University of Amsterdam,

More information

SIGVerse - A Simulation Platform for Human-Robot Interaction Jeffrey Too Chuan TAN and Tetsunari INAMURA National Institute of Informatics, Japan The

SIGVerse - A Simulation Platform for Human-Robot Interaction Jeffrey Too Chuan TAN and Tetsunari INAMURA National Institute of Informatics, Japan The SIGVerse - A Simulation Platform for Human-Robot Interaction Jeffrey Too Chuan TAN and Tetsunari INAMURA National Institute of Informatics, Japan The 29 th Annual Conference of The Robotics Society of

More information

Perception and Perspective in Robotics

Perception and Perspective in Robotics Perception and Perspective in Robotics Paul Fitzpatrick MIT CSAIL USA experimentation helps perception Rachel: We have got to find out if [ugly naked guy]'s alive. Monica: How are we going to do that?

More information

Development of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane

Development of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane Development of A Finger Mounted Type Haptic Device Using A Plane Approximated to Tangent Plane Makoto Yoda Department of Information System Science Graduate School of Engineering Soka University, Soka

More information

Vision-Based Speaker Detection Using Bayesian Networks

Vision-Based Speaker Detection Using Bayesian Networks Appears in Computer Vision and Pattern Recognition (CVPR 99), Ft. Collins, CO, June, 1999. Vision-Based Speaker Detection Using Bayesian Networks James M. Rehg Cambridge Research Lab Compaq Computer Corp.

More information

A Real Time Static & Dynamic Hand Gesture Recognition System

A Real Time Static & Dynamic Hand Gesture Recognition System International Journal of Engineering Inventions e-issn: 2278-7461, p-issn: 2319-6491 Volume 4, Issue 12 [Aug. 2015] PP: 93-98 A Real Time Static & Dynamic Hand Gesture Recognition System N. Subhash Chandra

More information