EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON

Size: px
Start display at page:

Download "EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON"

Transcription

1 EXPERIMENTAL BILATERAL CONTROL TELEMANIPULATION USING A VIRTUAL EXOSKELETON Josep Amat 1, Alícia Casals 2, Manel Frigola 2, Enric Martín 2 1Robotics Institute. (IRI) UPC / CSIC Llorens Artigas 4-6, 2a pt Barcelona, SPAIN 2Dep. Automatic Control and Computer Engineering Pau Gargallo n. 5, Barcelona Universitat Politècnica de Catalunya (UPC) Abstract. The introduction of less invasive interfaces in control usually carry with them new drawbacks such as less perception, less dexterity, etc. This work tackles the experimentation of new means of perceptual feedback in teleoperation, when the operator guides the task by means of a visual exoskeleton, and consequently without any mechanical device that can be provided with haptic devices. The work describes the evaluation of augmented images and sound feedback as alternative means for the bilateral control. 1. Introduction When working on teleoperated tasks, the control of the slave arms can be performed by means of different elements. Such elements can be control devices, such as joysticks, or gesture based systems using an adequate exoskeleton. The advantage of using a mechanical exoskeleton is that it can incorporate sensorial feedback, forces and torques, in a more natural or immersive way [1], compared to the use of other devices, such as joystick type [2]. Nevertheless, such mechanical structures become heavy and burden. With the aim of making the teleoperation tasks easier through the use of exoskeletons, avoiding at the same time the drawbacks of wearing mechanical elements, which are highly invasive, the possibility of using a virtual exoskeleton has been studied. This new interface is based on a stereoscopic vision system that detects the operator arms and computes their spatial position [3, 4]. The first results obtained [5] were positive in what refers to the operator s mobility and easy of movements. But, the system introduces as main drawbacks; first, movement delays, when the operator moves excessively quickly; second, a lower precision in the slave arms positioning, and third, and more important, the lack of force feedback. This kind of feedback helps significantly, in many cases, in the development of a task, providing more safety and efficiency. Limitations of the system dynamics, when the operator moves too quickly for the control, do not appear when the operator is covered by a mechanical exoskeleton, due to the physical or ergonomic constraints, that restrict the operator s movements. Nevertheless, these control limitations, produced by this more comfortable and agile system, do not annoy the user after a short training period. Poor precision of the controlled movements, due to the measurement errors produced by the 3D vision system that detects and tracks the operator s movements, can be easily corrected by the operator, by closing the loop through the visual feedback during the execution of the teleoperated task. The lack of force feedback during the teleoperation process has motivated the study of alternative sensorial feedback. We have experimented different systems based on the acoustic perception of the efforts carried out, as well as on the visualisation, over the screen, of the synthetic images, from which the effort exerted over the environment can be interpreted. This superposition does not reduce the visual capabilities. This work focuses on the evaluation of the improvement in efficiency through the use of these new means of sensorial feedback. The visual or acoustic feedback perception is evaluated in relation to the different kind of teleoperated tasks.

2 2. System description Two stereo cameras focusing at the operator and a movement tracking system act as a virtual exoskeleton that provide the teleoperation control unit with the orders produced through the operator s movements. Fig. 1 shows the structure of the system, a schema containing the system main modules. The measure of forces and torques applied to the robot end-effector, as well as the camera that visualises the working area provide the additional information about the development of the teleoperated task. These data is used to generate the sensorial feedback to the operator. The sensed data is converted to sound as well as to synthetic images to be merged with the visualised image on the operator screen, in such a way that the user can interpret the remote touch, forces, etc, both visually and acoustically. The whole process consists of two parts, first, sensor data processing for extracting the required process information, and second, their transformation to different perceptive information, easier to interpret by the user through different perception means, in this case, sound and augmented images as depicted below. Fig. 1 System overview 2. 1 Robot guidance from gestures Through the detection and tracking of the operator s arms, a vision system interprets some simple gestures for robot guidance. 3D perception of the operator s arms is based on the adjustment of the moving images, obtained from two or more cameras, with a rough model of a person, of multicylindrical type. Fig. 2 shows the two camera views of the scene with the superimposed multicylindrical model adjusted to the body. Fig. 2. Adjustment of the model to the operator arms

3 The gestures based guidance system operates based on the information provided by a model of the human body, which is specifically designed to recognise a human body shape. Several singular points are detected from some extracted features of a sequence of images. A first classification of these points is made of the most significant body parts, the head and arms. The search of singular points and their classification to identify these relevant parts is also based on the prediction and tracking of the trajectories of each part. The tracking algorithm is suitable to compensate the possible punctual lost detection of the recognition system, provided they last a time period short enough. The recognition system can provide new position references of the body parts when they appear clearly in the image for the first time or after a short occlusion period. 2.2 Image generation The purpose of the image generation module is to provide the user with visual information from the force and torque data, so as to increase the perception feedback and thus creating a more immersive environment. The goal of the added visual information is twofold: first, to indicate the magnitude and direction of the forces and torques applied over the robot end-effector, and second, to overdraw a synthetic image over the areass of the image where the visibility is poor due to the task itself (smoke, turbidness, occlusions...). The visual indication of the applied forces and torques can be displayed in two different ways according to the user s preferences, either adding arrows and reference axis to the images or changing the colour of the area of the image corresponding to the point of application of forces. In this case, the hue value of the corresponding area is rotated an angle proportional to the measured force. The information to be drawn in the images is built by a 3D graphic engine (OpenGL based) that maintains a threedimensional model of the robot and the scene. So the virtual images are used to enhance video data, the work can be classified as augmented reality, according to Milgram taxonomy [6] of real world and virtual reality interaction. On the other hand, when a given area of the image loses clarity, making difficult the remote teleoperation, a synthetic image, corresponding to the model of the missed region of the image, is superposed in that position. With the aim of improving efficiency without losing the available image information, a system for mixing the video and the synthetic images has been developed. The system does not operate in a fixed form, switching from the real image to synthetic data, but changing in a progressive way, a gradual mixing that superposes the synthetic image over the real one operating according to two criteria: At local level in the image, to make the synthetic image grow progressively over the zones with low visibility. At model level, solidifying the modelled object bodies, in accordance with the level of visibility, from wire-frame type minimally invasive models up to solid models. The achievement of this progressive image superposition has required the use of a local visibility measure function, based on the local gradient histogram. The objective is to progressively superpose the synthetic video over the real images acquired from the scene in real time, from the measurement of C (x, y), the local clarity, or image visibility function. Consequently, the progressive appearance of the synthetic image over the real one is not binary, but progressive, as a function of the clarity or medium transparency. This gradual change, thus affects both, the spatial image, that is, the boundaries from the real to the mixed area move progressively as a function of the clarity or medium transparency, as well as the weight with which the synthetic image appears over the real one in the mixed area [7]. The adequate placement of the superposed synthetic data over the real image, the registration process, is affected by the system delays that can produce a mismatching between both images, so we use computer vision techniques for robot tool localisation and segmentation in the images, thus enabling the dynamic correction of the positioning error. Figure 3 shows the process followed to create images displaying the real scene with force indications taken from the sensors placed on the robot end-effector. First, the system obtains the image of the task scene from a calibrated camera. Additionally, the robot controller gives the position and orientation of the robot tool and also the force and torque measured in the robot wrist. From the contours of the original image, and with the camera calibration data and robot localisation, the robot tool is segmented. In parallel, the graphics based engine creates a foreground image with the graphical force indications. These 'on image indication' can be done in several ways: arrows showing the torque values, arrows showing force direction and magnitude, modification of the hue of the pixels on the virtual image. The teleoperator

4 control module will choose which one is needed or desired in every moment through visualisation options parameter. Some augmented images are shown in the results section. Video In ROBOT ARM Contour Extraction Tool Segmentation Tool Area Scene Video H ( F, T ) Coloured Video TCP ROBOT TOOL Arrows Video Signal Arrow Generation Operator Tracking Sound Generation Virtual Exosqueleton Teleoperation Data F,T TCP TELEOPERATION CONTROL Visualisation Options Task Control Fig. 3. The image generation system 2.3 Sound generation The aim of this sound generation module is to render the significant remote audio signals into new sounds, as realistic as possible, enabling the user to relate their characteristics to the effort carried out. In order to construct an effective augmented sound, two auditory effects are considered: 3D sound effect and efforts magnification. Interaural time difference is a common effect used to create the feeling of hearing the sound coming from a particular direction. It usually involves the control of the amount of delay between the stereo sound channels (see [8] for other possibilities). This effect is used here to provide the user with the auditory perception of the point of contact between the tool tip and the object surface. For instance, when a milling operation erases the surface of an object, a milling stereo sound is generated in the operator headphones. The required data is acquired from a piezoelectric sensor placed on the robot end-effector. The virtual position of the sound source is determined considering the viewing direction, the estimated point of contact and the tool center point. The user can modify or switch off this sound effect changing the auditory options of the teleoperation program interface.

5 Power Audio In Sound Acquisition Hall Sensor Power Spectrum Milling Sound Detection Actuator Current F,T ROBOT ARM F E I A Frequency and Amplitude Modulator F + T Sound Spectrum Synthesiser Left Channel Right Channel t Delay Force/ 3D Sound Effect Force Sensor TCP ROBOT TOOL PiezoElectric Sensor to Headphones Operator Tracking Video Out Image Generation Video In TELEOPERATION CONTROL Acoustic Options Task Control Virtual Exosqueleton Teleoperation Data Fig. 4. The audio generation system To produce the force effort magnification, the system isolates the characteristic sound of the machining process from other sounds such as those generated by the mill actuator or by the gears. Once the milling sound is detected, the system replaces the original sound by a more friendly synthetic sound. This synthetic sound slightly resembles the original milling sound but is less noisy and is shifted to lower frequencies. To make the sound more natural, the harmonics of the synthetic sound are modified by means of the main milling frequency (f E ). Then, with the aim of increasing the user's perception of the mill efforts, the synthetic sound is shifted in frequency, being the magnitude of the shift the average of the mill actuator current (I A ). The synthetic sound is modulated in amplitude as well. The modulation factor chosen is the maximum between the magnitude of the force vector F and the magnitude of torque vector T (multiplied by the length of the tool). 3. Results The experimental system has been tested by several users in order to evaluate qualitatively the improvement of teleoperation efficiency, as well as to evaluate whether the effectiveness of the auditory and visual indication improves with the use of this new means of feedback, the bilateral control. Figure 5 shows a snapshot of the robot and its environment in the working scenario. The task performed consisted in milling a hole in a plastic part, deeply enough as to make possible to insert a peg of know dimensions in it.

6 Fig. 5. The milling teleoperated task Fig. 6 and 7 show the two ways of representing the generated visual information of forces and torques. While in fig. 6 the representation is symbolic, with the arrows representation, the torques around each axis and the magnitude and orientation of the force vector, in fig. 7, the chroma changes according to the magnitude of the measured force. The majority of users consider the augmented sound generated by the system very valuable. The synthetic sound generated, and shown in figure 8c, is evaluated as clear, with no noise, and meaningful by the majority of the users. This sound is synthesised from the spectrum of the real sound, as described above. The additional help of the 3D sound effect is considered in general correct, but it annoys some of the users, due to the fact that it can be perceived as infering a false movement of the tool. Some other users had understood the behaviour of the sound effect and they use it efficiently to generate the circular trajectory around the hole. Fig. 6. Torque (left) and force (right) representation, adding simbolic information in the images.

7 Fig. 7. Force representation changing the chroma of the image according to the measured force (from left to right, the force detected in the tool tip increases). 200 Power Spectrum (the mill is rotating free) D C B A a) [Hz] D C B Power Spectrum (milling) A E 50 b) [Hz] 200 Synthetic sound generated c) [Hz] Fig. 8. Results of the audio generation. a) Power spectrum of the mill actuator sound rotating in the free air and the main characteristic frequencies (A,B,C,D). b) Power spectrum of the mill actuator sound and the main milling frequency (E). c) The generated synthetic milling sound. 4. Conclusions The continuous introduction of new techniques for the execution of teleoperation tasks requires the experimentation of new perceptual feedback means. Having developed a vision based virtual exoskeleton there appears new operating situations since the operator does not wear any physical device, and therefore haptics is not possible. The system described proposes and evaluates, as additional information to the remote video images, new feedback means, consisting of augmented images together with sound generation. The video images are merged with visual representation of the forces applied on the robot tool, while a synthetic sound is generated from the real working environment sound. The subjective perception of the different means of information feedback does not enable us to conclude taking a decision for a best option, therefore, different options for feedback representation is

8 left to the user according to their preferences. Further experimentation and the improvement from their conclusions of the feedback data is necessary. The extracted sensorial data can be used not only for the teleoperation bilateral control in the sense of generating data perception to the user, but also to introduce new correction strategies to the robot arm. This aspect opens new possibilities towards more advanced teleoperated systems able to receive higher level commands. 5. References [1] Ohya J., Miyasato T., Nakatsu R. Virtual Reality Technologies for Multimedia Communications. In Mixed Reality Merging Real and Virtual Worlds. Ed. Y. Ohta, H. Takamura. Springer-Verlag, NY, [2] M. Bergamasco. Force replication to the human operator: The development of arm and hand exoskeleton as haptic devices. In The Seventh Int. Symp. on Robotics Research, Germany, pp , 1995 [3] Azarbayejani A. and Pentland A. Real-time self-calibrating stereo person tracking using 3-D shape estimation from blobs features. In Proceedings IEEE Conference on Computer Vision and Pattern Recognition, Vienna, [4] Amat J., Casals C., Frigola F., Pages J.. Possibilities of Man-Machine interaction through the perception of human gestures. In journal Contributions to Science 1(2): , Institut d Estudis Catalans, Barcelona, [5] J. Amat, M. Frigola, A. Casals. Virtual Exoskeleton for Telemanipulation. Lecture Notes in Control and Information Sciences. Experimental Robotics VII. Eds. D. Rus, S. Singh, Springer, pp.21-31, [6] P. Milgram, Herman Colquhoun Jr. A Taxonomy of Real and Virtual World Display Integration, Mixed Reality - Merging Real and Virtual Worlds, Ohmsha (Tokyo)-Springer Verlag (Berlin), pp. 5-30, ISBN (Springer Verlag) [7] A. Casals, J. Fernandez and J. Amat, Augmented reality to assist teleoperation working with reduced visual conditions. IEEE Int. Conference on Robotics and Automation, Washington-USA, 2002 [8] A. Mouchtaris, P. Reveliotis and C. Kyriakakis. Non-minimum phase inverse filter methods for immersive audio rendering. Acoustics, Speech, and Signal Processing, Proceedings., 1999 IEEE International Conference on, Volume: 6, Page(s): vol.6, 1999.

HUMAN Robot Cooperation Techniques in Surgery

HUMAN Robot Cooperation Techniques in Surgery HUMAN Robot Cooperation Techniques in Surgery Alícia Casals Institute for Bioengineering of Catalonia (IBEC), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain alicia.casals@upc.edu Keywords:

More information

Chapter 2 Introduction to Haptics 2.1 Definition of Haptics

Chapter 2 Introduction to Haptics 2.1 Definition of Haptics Chapter 2 Introduction to Haptics 2.1 Definition of Haptics The word haptic originates from the Greek verb hapto to touch and therefore refers to the ability to touch and manipulate objects. The haptic

More information

Haptic Tele-Assembly over the Internet

Haptic Tele-Assembly over the Internet Haptic Tele-Assembly over the Internet Sandra Hirche, Bartlomiej Stanczyk, and Martin Buss Institute of Automatic Control Engineering, Technische Universität München D-829 München, Germany, http : //www.lsr.ei.tum.de

More information

2. Introduction to Computer Haptics

2. Introduction to Computer Haptics 2. Introduction to Computer Haptics Seungmoon Choi, Ph.D. Assistant Professor Dept. of Computer Science and Engineering POSTECH Outline Basics of Force-Feedback Haptic Interfaces Introduction to Computer

More information

VIRTUAL REALITY Introduction. Emil M. Petriu SITE, University of Ottawa

VIRTUAL REALITY Introduction. Emil M. Petriu SITE, University of Ottawa VIRTUAL REALITY Introduction Emil M. Petriu SITE, University of Ottawa Natural and Virtual Reality Virtual Reality Interactive Virtual Reality Virtualized Reality Augmented Reality HUMAN PERCEPTION OF

More information

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES

MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL REALITY TECHNOLOGIES INTERNATIONAL CONFERENCE ON ENGINEERING AND PRODUCT DESIGN EDUCATION 4 & 5 SEPTEMBER 2008, UNIVERSITAT POLITECNICA DE CATALUNYA, BARCELONA, SPAIN MECHANICAL DESIGN LEARNING ENVIRONMENTS BASED ON VIRTUAL

More information

Using Simulation to Design Control Strategies for Robotic No-Scar Surgery

Using Simulation to Design Control Strategies for Robotic No-Scar Surgery Using Simulation to Design Control Strategies for Robotic No-Scar Surgery Antonio DE DONNO 1, Florent NAGEOTTE, Philippe ZANNE, Laurent GOFFIN and Michel de MATHELIN LSIIT, University of Strasbourg/CNRS,

More information

The Haptic Impendance Control through Virtual Environment Force Compensation

The Haptic Impendance Control through Virtual Environment Force Compensation The Haptic Impendance Control through Virtual Environment Force Compensation OCTAVIAN MELINTE Robotics and Mechatronics Department Institute of Solid Mechanicsof the Romanian Academy ROMANIA octavian.melinte@yahoo.com

More information

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real... v preface Motivation Augmented reality (AR) research aims to develop technologies that allow the real-time fusion of computer-generated digital content with the real world. Unlike virtual reality (VR)

More information

HeroX - Untethered VR Training in Sync'ed Physical Spaces

HeroX - Untethered VR Training in Sync'ed Physical Spaces Page 1 of 6 HeroX - Untethered VR Training in Sync'ed Physical Spaces Above and Beyond - Integrating Robotics In previous research work I experimented with multiple robots remotely controlled by people

More information

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright

E90 Project Proposal. 6 December 2006 Paul Azunre Thomas Murray David Wright E90 Project Proposal 6 December 2006 Paul Azunre Thomas Murray David Wright Table of Contents Abstract 3 Introduction..4 Technical Discussion...4 Tracking Input..4 Haptic Feedack.6 Project Implementation....7

More information

Advanced Man-Machine Interaction

Advanced Man-Machine Interaction Signals and Communication Technology Advanced Man-Machine Interaction Fundamentals and Implementation Bearbeitet von Karl-Friedrich Kraiss 1. Auflage 2006. Buch. XIX, 461 S. ISBN 978 3 540 30618 4 Format

More information

Virtual Environments. Ruth Aylett

Virtual Environments. Ruth Aylett Virtual Environments Ruth Aylett Aims of the course 1. To demonstrate a critical understanding of modern VE systems, evaluating the strengths and weaknesses of the current VR technologies 2. To be able

More information

A Very High Level Interface to Teleoperate a Robot via Web including Augmented Reality

A Very High Level Interface to Teleoperate a Robot via Web including Augmented Reality A Very High Level Interface to Teleoperate a Robot via Web including Augmented Reality R. Marín, P. J. Sanz and J. S. Sánchez Abstract The system consists of a multirobot architecture that gives access

More information

The Effect of Haptic Feedback on Basic Social Interaction within Shared Virtual Environments

The Effect of Haptic Feedback on Basic Social Interaction within Shared Virtual Environments The Effect of Haptic Feedback on Basic Social Interaction within Shared Virtual Environments Elias Giannopoulos 1, Victor Eslava 2, María Oyarzabal 2, Teresa Hierro 2, Laura González 2, Manuel Ferre 2,

More information

VICs: A Modular Vision-Based HCI Framework

VICs: A Modular Vision-Based HCI Framework VICs: A Modular Vision-Based HCI Framework The Visual Interaction Cues Project Guangqi Ye, Jason Corso Darius Burschka, & Greg Hager CIRL, 1 Today, I ll be presenting work that is part of an ongoing project

More information

Haptics CS327A

Haptics CS327A Haptics CS327A - 217 hap tic adjective relating to the sense of touch or to the perception and manipulation of objects using the senses of touch and proprioception 1 2 Slave Master 3 Courtesy of Walischmiller

More information

Perception. Introduction to HRI Simmons & Nourbakhsh Spring 2015

Perception. Introduction to HRI Simmons & Nourbakhsh Spring 2015 Perception Introduction to HRI Simmons & Nourbakhsh Spring 2015 Perception my goals What is the state of the art boundary? Where might we be in 5-10 years? The Perceptual Pipeline The classical approach:

More information

iwindow Concept of an intelligent window for machine tools using augmented reality

iwindow Concept of an intelligent window for machine tools using augmented reality iwindow Concept of an intelligent window for machine tools using augmented reality Sommer, P.; Atmosudiro, A.; Schlechtendahl, J.; Lechler, A.; Verl, A. Institute for Control Engineering of Machine Tools

More information

Robot Sensors Introduction to Robotics Lecture Handout September 20, H. Harry Asada Massachusetts Institute of Technology

Robot Sensors Introduction to Robotics Lecture Handout September 20, H. Harry Asada Massachusetts Institute of Technology Robot Sensors 2.12 Introduction to Robotics Lecture Handout September 20, 2004 H. Harry Asada Massachusetts Institute of Technology Touch Sensor CCD Camera Vision System Ultrasonic Sensor Photo removed

More information

Comparison of Haptic and Non-Speech Audio Feedback

Comparison of Haptic and Non-Speech Audio Feedback Comparison of Haptic and Non-Speech Audio Feedback Cagatay Goncu 1 and Kim Marriott 1 Monash University, Mebourne, Australia, cagatay.goncu@monash.edu, kim.marriott@monash.edu Abstract. We report a usability

More information

PHYSICAL ROBOTS PROGRAMMING BY IMITATION USING VIRTUAL ROBOT PROTOTYPES

PHYSICAL ROBOTS PROGRAMMING BY IMITATION USING VIRTUAL ROBOT PROTOTYPES Bulletin of the Transilvania University of Braşov Series I: Engineering Sciences Vol. 6 (55) No. 2-2013 PHYSICAL ROBOTS PROGRAMMING BY IMITATION USING VIRTUAL ROBOT PROTOTYPES A. FRATU 1 M. FRATU 2 Abstract:

More information

HAPTIC GUIDANCE BASED ON HARMONIC FUNCTIONS FOR THE EXECUTION OF TELEOPERATED ASSEMBLY TASKS. Carlos Vázquez Jan Rosell,1

HAPTIC GUIDANCE BASED ON HARMONIC FUNCTIONS FOR THE EXECUTION OF TELEOPERATED ASSEMBLY TASKS. Carlos Vázquez Jan Rosell,1 Preprints of IAD' 2007: IFAC WORKSHOP ON INTELLIGENT ASSEMBLY AND DISASSEMBLY May 23-25 2007, Alicante, Spain HAPTIC GUIDANCE BASED ON HARMONIC FUNCTIONS FOR THE EXECUTION OF TELEOPERATED ASSEMBLY TASKS

More information

Bibliography. Conclusion

Bibliography. Conclusion the almost identical time measured in the real and the virtual execution, and the fact that the real execution with indirect vision to be slower than the manipulation on the simulated environment. The

More information

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine)

Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Interacting within Virtual Worlds (based on talks by Greg Welch and Mark Mine) Presentation Working in a virtual world Interaction principles Interaction examples Why VR in the First Place? Direct perception

More information

Telecommunication and remote-controlled

Telecommunication and remote-controlled Spatial Interfaces Editors: Frank Steinicke and Wolfgang Stuerzlinger Telexistence: Enabling Humans to Be Virtually Ubiquitous Susumu Tachi The University of Tokyo Telecommunication and remote-controlled

More information

VIEW: Visual Interactive Effective Worlds Lorentz Center International Center for workshops in the Sciences June Dr.

VIEW: Visual Interactive Effective Worlds Lorentz Center International Center for workshops in the Sciences June Dr. Virtual Reality & Presence VIEW: Visual Interactive Effective Worlds Lorentz Center International Center for workshops in the Sciences 25-27 June 2007 Dr. Frederic Vexo Virtual Reality & Presence Outline:

More information

THE BEATING EQUALIZER AND ITS APPLICATION TO THE SYNTHESIS AND MODIFICATION OF PIANO TONES

THE BEATING EQUALIZER AND ITS APPLICATION TO THE SYNTHESIS AND MODIFICATION OF PIANO TONES J. Rauhala, The beating equalizer and its application to the synthesis and modification of piano tones, in Proceedings of the 1th International Conference on Digital Audio Effects, Bordeaux, France, 27,

More information

JEPPIAAR ENGINEERING COLLEGE

JEPPIAAR ENGINEERING COLLEGE JEPPIAAR ENGINEERING COLLEGE Jeppiaar Nagar, Rajiv Gandhi Salai 600 119 DEPARTMENT OFMECHANICAL ENGINEERING QUESTION BANK VII SEMESTER ME6010 ROBOTICS Regulation 013 JEPPIAAR ENGINEERING COLLEGE Jeppiaar

More information

Ubiquitous Computing Summer Episode 16: HCI. Hannes Frey and Peter Sturm University of Trier. Hannes Frey and Peter Sturm, University of Trier 1

Ubiquitous Computing Summer Episode 16: HCI. Hannes Frey and Peter Sturm University of Trier. Hannes Frey and Peter Sturm, University of Trier 1 Episode 16: HCI Hannes Frey and Peter Sturm University of Trier University of Trier 1 Shrinking User Interface Small devices Narrow user interface Only few pixels graphical output No keyboard Mobility

More information

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Mel Spectrum Analysis of Speech Recognition using Single Microphone International Journal of Engineering Research in Electronics and Communication Mel Spectrum Analysis of Speech Recognition using Single Microphone [1] Lakshmi S.A, [2] Cholavendan M [1] PG Scholar, Sree

More information

Interactive Virtual Environments

Interactive Virtual Environments Interactive Virtual Environments Introduction Emil M. Petriu, Dr. Eng., FIEEE Professor, School of Information Technology and Engineering University of Ottawa, Ottawa, ON, Canada http://www.site.uottawa.ca/~petriu

More information

Differences in Fitts Law Task Performance Based on Environment Scaling

Differences in Fitts Law Task Performance Based on Environment Scaling Differences in Fitts Law Task Performance Based on Environment Scaling Gregory S. Lee and Bhavani Thuraisingham Department of Computer Science University of Texas at Dallas 800 West Campbell Road Richardson,

More information

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES International Journal of Advanced Research in Engineering and Technology (IJARET) Volume 9, Issue 3, May - June 2018, pp. 177 185, Article ID: IJARET_09_03_023 Available online at http://www.iaeme.com/ijaret/issues.asp?jtype=ijaret&vtype=9&itype=3

More information

Toward an Augmented Reality System for Violin Learning Support

Toward an Augmented Reality System for Violin Learning Support Toward an Augmented Reality System for Violin Learning Support Hiroyuki Shiino, François de Sorbier, and Hideo Saito Graduate School of Science and Technology, Keio University, Yokohama, Japan {shiino,fdesorbi,saito}@hvrl.ics.keio.ac.jp

More information

SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4

SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4 SOPA version 2 Revised July 7 2014 SOPA project September 21, 2014 Contents 1 Introduction 2 2 Basic concept 3 3 Capturing spatial audio 4 4 Sphere around your head 5 5 Reproduction 7 5.1 Binaural reproduction......................

More information

ROBOT VISION. Dr.M.Madhavi, MED, MVSREC

ROBOT VISION. Dr.M.Madhavi, MED, MVSREC ROBOT VISION Dr.M.Madhavi, MED, MVSREC Robotic vision may be defined as the process of acquiring and extracting information from images of 3-D world. Robotic vision is primarily targeted at manipulation

More information

International Journal of Innovative Research in Engineering Science and Technology APRIL 2018 ISSN X

International Journal of Innovative Research in Engineering Science and Technology APRIL 2018 ISSN X HIGH DYNAMIC RANGE OF MULTISPECTRAL ACQUISITION USING SPATIAL IMAGES 1 M.Kavitha, M.Tech., 2 N.Kannan, M.E., and 3 S.Dharanya, M.E., 1 Assistant Professor/ CSE, Dhirajlal Gandhi College of Technology,

More information

VR based HCI Techniques & Application. November 29, 2002

VR based HCI Techniques & Application. November 29, 2002 VR based HCI Techniques & Application November 29, 2002 stefan.seipel@hci.uu.se What is Virtual Reality? Coates (1992): Virtual Reality is electronic simulations of environments experienced via head mounted

More information

Intelligent Systems, Control and Automation: Science and Engineering

Intelligent Systems, Control and Automation: Science and Engineering Intelligent Systems, Control and Automation: Science and Engineering Volume 64 Series Editor S. G. Tzafestas For further volumes: http://www.springer.com/series/6259 Matjaž Mihelj Janez Podobnik Haptics

More information

6.869 Advances in Computer Vision Spring 2010, A. Torralba

6.869 Advances in Computer Vision Spring 2010, A. Torralba 6.869 Advances in Computer Vision Spring 2010, A. Torralba Due date: Wednesday, Feb 17, 2010 Problem set 1 You need to submit a report with brief descriptions of what you did. The most important part is

More information

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface

Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface Evaluation of Visuo-haptic Feedback in a 3D Touch Panel Interface Xu Zhao Saitama University 255 Shimo-Okubo, Sakura-ku, Saitama City, Japan sheldonzhaox@is.ics.saitamau.ac.jp Takehiro Niikura The University

More information

Speech/Music Change Point Detection using Sonogram and AANN

Speech/Music Change Point Detection using Sonogram and AANN International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 6, Number 1 (2016), pp. 45-49 International Research Publications House http://www. irphouse.com Speech/Music Change

More information

LASER ASSISTED COMBINED TELEOPERATION AND AUTONOMOUS CONTROL

LASER ASSISTED COMBINED TELEOPERATION AND AUTONOMOUS CONTROL ANS EPRRSD - 13 th Robotics & remote Systems for Hazardous Environments 11 th Emergency Preparedness & Response Knoxville, TN, August 7-10, 2011, on CD-ROM, American Nuclear Society, LaGrange Park, IL

More information

Booklet of teaching units

Booklet of teaching units International Master Program in Mechatronic Systems for Rehabilitation Booklet of teaching units Third semester (M2 S1) Master Sciences de l Ingénieur Université Pierre et Marie Curie Paris 6 Boite 164,

More information

Robust Haptic Teleoperation of a Mobile Manipulation Platform

Robust Haptic Teleoperation of a Mobile Manipulation Platform Robust Haptic Teleoperation of a Mobile Manipulation Platform Jaeheung Park and Oussama Khatib Stanford AI Laboratory Stanford University http://robotics.stanford.edu Abstract. This paper presents a new

More information

International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, ISSN

International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18,   ISSN International Journal of Computer Engineering and Applications, Volume XII, Issue IV, April 18, www.ijcea.com ISSN 2321-3469 AUGMENTED REALITY FOR HELPING THE SPECIALLY ABLED PERSONS ABSTRACT Saniya Zahoor

More information

CAPACITIES FOR TECHNOLOGY TRANSFER

CAPACITIES FOR TECHNOLOGY TRANSFER CAPACITIES FOR TECHNOLOGY TRANSFER The Institut de Robòtica i Informàtica Industrial (IRI) is a Joint University Research Institute of the Spanish Council for Scientific Research (CSIC) and the Technical

More information

Surgical robot simulation with BBZ console

Surgical robot simulation with BBZ console Review Article on Thoracic Surgery Surgical robot simulation with BBZ console Francesco Bovo 1, Giacomo De Rossi 2, Francesco Visentin 2,3 1 BBZ srl, Verona, Italy; 2 Department of Computer Science, Università

More information

LCC 3710 Principles of Interaction Design. Readings. Sound in Interfaces. Speech Interfaces. Speech Applications. Motivation for Speech Interfaces

LCC 3710 Principles of Interaction Design. Readings. Sound in Interfaces. Speech Interfaces. Speech Applications. Motivation for Speech Interfaces LCC 3710 Principles of Interaction Design Class agenda: - Readings - Speech, Sonification, Music Readings Hermann, T., Hunt, A. (2005). "An Introduction to Interactive Sonification" in IEEE Multimedia,

More information

Exploring Surround Haptics Displays

Exploring Surround Haptics Displays Exploring Surround Haptics Displays Ali Israr Disney Research 4615 Forbes Ave. Suite 420, Pittsburgh, PA 15213 USA israr@disneyresearch.com Ivan Poupyrev Disney Research 4615 Forbes Ave. Suite 420, Pittsburgh,

More information

FUNDAMENTALS ROBOT TECHNOLOGY. An Introduction to Industrial Robots, T eleoperators and Robot Vehicles. D J Todd. Kogan Page

FUNDAMENTALS ROBOT TECHNOLOGY. An Introduction to Industrial Robots, T eleoperators and Robot Vehicles. D J Todd. Kogan Page FUNDAMENTALS of ROBOT TECHNOLOGY An Introduction to Industrial Robots, T eleoperators and Robot Vehicles D J Todd &\ Kogan Page First published in 1986 by Kogan Page Ltd 120 Pentonville Road, London Nl

More information

CONTROLLING METHODS AND CHALLENGES OF ROBOTIC ARM

CONTROLLING METHODS AND CHALLENGES OF ROBOTIC ARM CONTROLLING METHODS AND CHALLENGES OF ROBOTIC ARM Aniket D. Kulkarni *1, Dr.Sayyad Ajij D. *2 *1(Student of E&C Department, MIT Aurangabad, India) *2(HOD of E&C department, MIT Aurangabad, India) aniket2212@gmail.com*1,

More information

Passive Bilateral Teleoperation

Passive Bilateral Teleoperation Passive Bilateral Teleoperation Project: Reconfigurable Control of Robotic Systems Over Networks Márton Lırinc Dept. Of Electrical Engineering Sapientia University Overview What is bilateral teleoperation?

More information

Università di Roma La Sapienza. Medical Robotics. A Teleoperation System for Research in MIRS. Marilena Vendittelli

Università di Roma La Sapienza. Medical Robotics. A Teleoperation System for Research in MIRS. Marilena Vendittelli Università di Roma La Sapienza Medical Robotics A Teleoperation System for Research in MIRS Marilena Vendittelli the DLR teleoperation system slave three versatile robots MIRO light-weight: weight < 10

More information

Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances

Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances Spatial Interfaces and Interactive 3D Environments for Immersive Musical Performances Florent Berthaut and Martin Hachet Figure 1: A musician plays the Drile instrument while being immersed in front of

More information

Medical Robotics. Part II: SURGICAL ROBOTICS

Medical Robotics. Part II: SURGICAL ROBOTICS 5 Medical Robotics Part II: SURGICAL ROBOTICS In the last decade, surgery and robotics have reached a maturity that has allowed them to be safely assimilated to create a new kind of operating room. This

More information

MEAM 520. Haptic Rendering and Teleoperation

MEAM 520. Haptic Rendering and Teleoperation MEAM 520 Haptic Rendering and Teleoperation Katherine J. Kuchenbecker, Ph.D. General Robotics, Automation, Sensing, and Perception Lab (GRASP) MEAM Department, SEAS, University of Pennsylvania Lecture

More information

Kinect Interface for UC-win/Road: Application to Tele-operation of Small Robots

Kinect Interface for UC-win/Road: Application to Tele-operation of Small Robots Kinect Interface for UC-win/Road: Application to Tele-operation of Small Robots Hafid NINISS Forum8 - Robot Development Team Abstract: The purpose of this work is to develop a man-machine interface for

More information

ISMCR2004. Abstract. 2. The mechanism of the master-slave arm of Telesar II. 1. Introduction. D21-Page 1

ISMCR2004. Abstract. 2. The mechanism of the master-slave arm of Telesar II. 1. Introduction. D21-Page 1 Development of Multi-D.O.F. Master-Slave Arm with Bilateral Impedance Control for Telexistence Riichiro Tadakuma, Kiyohiro Sogen, Hiroyuki Kajimoto, Naoki Kawakami, and Susumu Tachi 7-3-1 Hongo, Bunkyo-ku,

More information

PERCEPTUAL EFFECTS IN ALIGNING VIRTUAL AND REAL OBJECTS IN AUGMENTED REALITY DISPLAYS

PERCEPTUAL EFFECTS IN ALIGNING VIRTUAL AND REAL OBJECTS IN AUGMENTED REALITY DISPLAYS 41 st Annual Meeting of Human Factors and Ergonomics Society, Albuquerque, New Mexico. Sept. 1997. PERCEPTUAL EFFECTS IN ALIGNING VIRTUAL AND REAL OBJECTS IN AUGMENTED REALITY DISPLAYS Paul Milgram and

More information

AR 2 kanoid: Augmented Reality ARkanoid

AR 2 kanoid: Augmented Reality ARkanoid AR 2 kanoid: Augmented Reality ARkanoid B. Smith and R. Gosine C-CORE and Memorial University of Newfoundland Abstract AR 2 kanoid, Augmented Reality ARkanoid, is an augmented reality version of the popular

More information

Computer Assisted Medical Interventions

Computer Assisted Medical Interventions Outline Computer Assisted Medical Interventions Force control, collaborative manipulation and telemanipulation Bernard BAYLE Joint course University of Strasbourg, University of Houston, Telecom Paris

More information

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005.

Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays. Habib Abi-Rached Thursday 17 February 2005. Stereo-based Hand Gesture Tracking and Recognition in Immersive Stereoscopic Displays Habib Abi-Rached Thursday 17 February 2005. Objective Mission: Facilitate communication: Bandwidth. Intuitiveness.

More information

2.1 Dual-Arm Humanoid Robot A dual-arm humanoid robot is actuated by rubbertuators, which are McKibben pneumatic artiæcial muscles as shown in Figure

2.1 Dual-Arm Humanoid Robot A dual-arm humanoid robot is actuated by rubbertuators, which are McKibben pneumatic artiæcial muscles as shown in Figure Integrating Visual Feedback and Force Feedback in 3-D Collision Avoidance for a Dual-Arm Humanoid Robot S. Charoenseang, A. Srikaew, D. M. Wilkes, and K. Kawamura Center for Intelligent Systems Vanderbilt

More information

A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems

A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems F. Steinicke, G. Bruder, H. Frenz 289 A Multimodal Locomotion User Interface for Immersive Geospatial Information Systems Frank Steinicke 1, Gerd Bruder 1, Harald Frenz 2 1 Institute of Computer Science,

More information

are in front of some cameras and have some influence on the system because of their attitude. Since the interactor is really made aware of the impact

are in front of some cameras and have some influence on the system because of their attitude. Since the interactor is really made aware of the impact Immersive Communication Damien Douxchamps, David Ergo, Beno^ t Macq, Xavier Marichal, Alok Nandi, Toshiyuki Umeda, Xavier Wielemans alterface Λ c/o Laboratoire de Télécommunications et Télédétection Université

More information

Augmented and Virtual Reality

Augmented and Virtual Reality CS-3120 Human-Computer Interaction Augmented and Virtual Reality Mikko Kytö 7.11.2017 From Real to Virtual [1] Milgram, P., & Kishino, F. (1994). A taxonomy of mixed reality visual displays. IEICE TRANSACTIONS

More information

Visual Search using Principal Component Analysis

Visual Search using Principal Component Analysis Visual Search using Principal Component Analysis Project Report Umesh Rajashekar EE381K - Multidimensional Digital Signal Processing FALL 2000 The University of Texas at Austin Abstract The development

More information

Gesture Recognition with Real World Environment using Kinect: A Review

Gesture Recognition with Real World Environment using Kinect: A Review Gesture Recognition with Real World Environment using Kinect: A Review Prakash S. Sawai 1, Prof. V. K. Shandilya 2 P.G. Student, Department of Computer Science & Engineering, Sipna COET, Amravati, Maharashtra,

More information

SMart wearable Robotic Teleoperated surgery

SMart wearable Robotic Teleoperated surgery SMart wearable Robotic Teleoperated surgery This project has received funding from the European Union s Horizon 2020 research and innovation programme under grant agreement No 732515 Context Minimally

More information

Haptic presentation of 3D objects in virtual reality for the visually disabled

Haptic presentation of 3D objects in virtual reality for the visually disabled Haptic presentation of 3D objects in virtual reality for the visually disabled M Moranski, A Materka Institute of Electronics, Technical University of Lodz, Wolczanska 211/215, Lodz, POLAND marcin.moranski@p.lodz.pl,

More information

Texture recognition using force sensitive resistors

Texture recognition using force sensitive resistors Texture recognition using force sensitive resistors SAYED, Muhammad, DIAZ GARCIA,, Jose Carlos and ALBOUL, Lyuba Available from Sheffield Hallam University Research

More information

Haplug: A Haptic Plug for Dynamic VR Interactions

Haplug: A Haptic Plug for Dynamic VR Interactions Haplug: A Haptic Plug for Dynamic VR Interactions Nobuhisa Hanamitsu *, Ali Israr Disney Research, USA nobuhisa.hanamitsu@disneyresearch.com Abstract. We demonstrate applications of a new actuator, the

More information

Novel machine interface for scaled telesurgery

Novel machine interface for scaled telesurgery Novel machine interface for scaled telesurgery S. Clanton, D. Wang, Y. Matsuoka, D. Shelton, G. Stetten SPIE Medical Imaging, vol. 5367, pp. 697-704. San Diego, Feb. 2004. A Novel Machine Interface for

More information

Interactive Simulation: UCF EIN5255. VR Software. Audio Output. Page 4-1

Interactive Simulation: UCF EIN5255. VR Software. Audio Output. Page 4-1 VR Software Class 4 Dr. Nabil Rami http://www.simulationfirst.com/ein5255/ Audio Output Can be divided into two elements: Audio Generation Audio Presentation Page 4-1 Audio Generation A variety of audio

More information

MEAM 520. Haptic Rendering and Teleoperation

MEAM 520. Haptic Rendering and Teleoperation MEAM 520 Haptic Rendering and Teleoperation Katherine J. Kuchenbecker, Ph.D. General Robotics, Automation, Sensing, and Perception Lab (GRASP) MEAM Department, SEAS, University of Pennsylvania Lecture

More information

Sound source localization and its use in multimedia applications

Sound source localization and its use in multimedia applications Notes for lecture/ Zack Settel, McGill University Sound source localization and its use in multimedia applications Introduction With the arrival of real-time binaural or "3D" digital audio processing,

More information

Computing for Engineers in Python

Computing for Engineers in Python Computing for Engineers in Python Lecture 10: Signal (Image) Processing Autumn 2011-12 Some slides incorporated from Benny Chor s course 1 Lecture 9: Highlights Sorting, searching and time complexity Preprocessing

More information

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB

Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Analysis of Various Methodology of Hand Gesture Recognition System using MATLAB Komal Hasija 1, Rajani Mehta 2 Abstract Recognition is a very effective area of research in regard of security with the involvement

More information

Perception. Read: AIMA Chapter 24 & Chapter HW#8 due today. Vision

Perception. Read: AIMA Chapter 24 & Chapter HW#8 due today. Vision 11-25-2013 Perception Vision Read: AIMA Chapter 24 & Chapter 25.3 HW#8 due today visual aural haptic & tactile vestibular (balance: equilibrium, acceleration, and orientation wrt gravity) olfactory taste

More information

English PRO-642. Advanced Features: On-Screen Display

English PRO-642. Advanced Features: On-Screen Display English PRO-642 Advanced Features: On-Screen Display 1 Adjusting the Camera Settings The joystick has a middle button that you click to open the OSD menu. This button is also used to select an option that

More information

Teleoperation Based on the Hidden Robot Concept

Teleoperation Based on the Hidden Robot Concept IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART A: SYSTEMS AND HUMANS, VOL. 31, NO. 1, JANUARY 2001 1 Teleoperation Based on the Hidden Robot Concept Abderrahmane Kheddar Abstract Overlaying classical

More information

8.2 IMAGE PROCESSING VERSUS IMAGE ANALYSIS Image processing: The collection of routines and

8.2 IMAGE PROCESSING VERSUS IMAGE ANALYSIS Image processing: The collection of routines and 8.1 INTRODUCTION In this chapter, we will study and discuss some fundamental techniques for image processing and image analysis, with a few examples of routines developed for certain purposes. 8.2 IMAGE

More information

A FRAMEWORK FOR TELEPRESENT GAME-PLAY IN LARGE VIRTUAL ENVIRONMENTS

A FRAMEWORK FOR TELEPRESENT GAME-PLAY IN LARGE VIRTUAL ENVIRONMENTS A FRAMEWORK FOR TELEPRESENT GAME-PLAY IN LARGE VIRTUAL ENVIRONMENTS Patrick Rößler, Frederik Beutler, and Uwe D. Hanebeck Intelligent Sensor-Actuator-Systems Laboratory Institute of Computer Science and

More information

Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction

Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction Shopping Together: A Remote Co-shopping System Utilizing Spatial Gesture Interaction Minghao Cai 1(B), Soh Masuko 2, and Jiro Tanaka 1 1 Waseda University, Kitakyushu, Japan mhcai@toki.waseda.jp, jiro@aoni.waseda.jp

More information

Perception in Immersive Virtual Reality Environments ROB ALLISON DEPT. OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE YORK UNIVERSITY, TORONTO

Perception in Immersive Virtual Reality Environments ROB ALLISON DEPT. OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE YORK UNIVERSITY, TORONTO Perception in Immersive Virtual Reality Environments ROB ALLISON DEPT. OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE YORK UNIVERSITY, TORONTO Overview Basic concepts and ideas of virtual environments

More information

Visuo-Haptic Interface for Teleoperation of Mobile Robot Exploration Tasks

Visuo-Haptic Interface for Teleoperation of Mobile Robot Exploration Tasks Visuo-Haptic Interface for Teleoperation of Mobile Robot Exploration Tasks Nikos C. Mitsou, Spyros V. Velanas and Costas S. Tzafestas Abstract With the spread of low-cost haptic devices, haptic interfaces

More information

HAND-SHAPED INTERFACE FOR INTUITIVE HUMAN- ROBOT COMMUNICATION THROUGH HAPTIC MEDIA

HAND-SHAPED INTERFACE FOR INTUITIVE HUMAN- ROBOT COMMUNICATION THROUGH HAPTIC MEDIA HAND-SHAPED INTERFACE FOR INTUITIVE HUMAN- ROBOT COMMUNICATION THROUGH HAPTIC MEDIA RIKU HIKIJI AND SHUJI HASHIMOTO Department of Applied Physics, School of Science and Engineering, Waseda University 3-4-1

More information

Application Areas of AI Artificial intelligence is divided into different branches which are mentioned below:

Application Areas of AI   Artificial intelligence is divided into different branches which are mentioned below: Week 2 - o Expert Systems o Natural Language Processing (NLP) o Computer Vision o Speech Recognition And Generation o Robotics o Neural Network o Virtual Reality APPLICATION AREAS OF ARTIFICIAL INTELLIGENCE

More information

Jane Li. Assistant Professor Mechanical Engineering Department, Robotic Engineering Program Worcester Polytechnic Institute

Jane Li. Assistant Professor Mechanical Engineering Department, Robotic Engineering Program Worcester Polytechnic Institute Jane Li Assistant Professor Mechanical Engineering Department, Robotic Engineering Program Worcester Polytechnic Institute Use an example to explain what is admittance control? You may refer to exoskeleton

More information

Privacy-Protected Camera for the Sensing Web

Privacy-Protected Camera for the Sensing Web Privacy-Protected Camera for the Sensing Web Ikuhisa Mitsugami 1, Masayuki Mukunoki 2, Yasutomo Kawanishi 2, Hironori Hattori 2, and Michihiko Minoh 2 1 Osaka University, 8-1, Mihogaoka, Ibaraki, Osaka

More information

Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin

Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Available theses in robotics (March 2018) Prof. Paolo Rocco Prof. Andrea Maria Zanchettin Ergonomic positioning of bulky objects Thesis 1 Robot acts as a 3rd hand for workpiece positioning: Muscular fatigue

More information

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods 19 An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods T.Arunachalam* Post Graduate Student, P.G. Dept. of Computer Science, Govt Arts College, Melur - 625 106 Email-Arunac682@gmail.com

More information

Control a 2-Axis Servomechanism by Gesture Recognition using a Generic WebCam

Control a 2-Axis Servomechanism by Gesture Recognition using a Generic WebCam Tavares, J. M. R. S.; Ferreira, R. & Freitas, F. / Control a 2-Axis Servomechanism by Gesture Recognition using a Generic WebCam, pp. 039-040, International Journal of Advanced Robotic Systems, Volume

More information

Haptic Virtual Fixtures for Robot-Assisted Manipulation

Haptic Virtual Fixtures for Robot-Assisted Manipulation Haptic Virtual Fixtures for Robot-Assisted Manipulation Jake J. Abbott, Panadda Marayong, and Allison M. Okamura Department of Mechanical Engineering, The Johns Hopkins University {jake.abbott, pmarayong,

More information

Continuous Rotation Control of Robotic Arm using Slip Rings for Mars Rover

Continuous Rotation Control of Robotic Arm using Slip Rings for Mars Rover International Conference on Mechanical, Industrial and Materials Engineering 2017 (ICMIME2017) 28-30 December, 2017, RUET, Rajshahi, Bangladesh. Paper ID: AM-270 Continuous Rotation Control of Robotic

More information

Performance Issues in Collaborative Haptic Training

Performance Issues in Collaborative Haptic Training 27 IEEE International Conference on Robotics and Automation Roma, Italy, 1-14 April 27 FrA4.4 Performance Issues in Collaborative Haptic Training Behzad Khademian and Keyvan Hashtrudi-Zaad Abstract This

More information

A NOVEL CONTROL SYSTEM FOR ROBOTIC DEVICES

A NOVEL CONTROL SYSTEM FOR ROBOTIC DEVICES A NOVEL CONTROL SYSTEM FOR ROBOTIC DEVICES THAIR A. SALIH, OMAR IBRAHIM YEHEA COMPUTER DEPT. TECHNICAL COLLEGE/ MOSUL EMAIL: ENG_OMAR87@YAHOO.COM, THAIRALI59@YAHOO.COM ABSTRACT It is difficult to find

More information

VISUAL REQUIREMENTS ON AUGMENTED VIRTUAL REALITY SYSTEM

VISUAL REQUIREMENTS ON AUGMENTED VIRTUAL REALITY SYSTEM Annals of the University of Petroşani, Mechanical Engineering, 8 (2006), 73-78 73 VISUAL REQUIREMENTS ON AUGMENTED VIRTUAL REALITY SYSTEM JOZEF NOVÁK-MARCINČIN 1, PETER BRÁZDA 2 Abstract: Paper describes

More information