Auditory System For a Mobile Robot

Size: px
Start display at page:

Download "Auditory System For a Mobile Robot"

Transcription

1 Auditory System For a Mobile Robot PhD Thesis Jean-Marc Valin Department of Electrical Engineering and Computer Engineering Université de Sherbrooke, Québec, Canada Jean-Marc.Valin@USherbrooke.ca

2 Motivations Robots need information about their environment in order to be intelligent Artificial vision has been popular for a long time, but artificial audition is new Robust audition is essential for humanrobot interaction (cocktail party effect)

3 Approaches To Artificial Audition Single microphone Two microphones (binaural audition) Human-robot interaction Unreliable Imitate human auditory system Limited localisation and separation Microphone array audition More information available Simpler processing

4 Objectives Localise and track simultaneous moving sound sources Separate sound sources Perform automatic speech recognition Remain within robotics constraints complexity, algorithmic delay robustness to noise and reverberation weight/space/adaptability moving sources, moving robot

5 Experimental Setup Eight microphones on the Spartacus robot Two configurations Noisy conditions Two environments Reverberation time Lab (E1) 350 ms Hall (E2) 1 s cube (C1) shell(c2)

6 Sound Source Localisation

7 Approaches to Sound Source Localisation Binaural Microphone array Interaural phase difference (delay) Interaural intensity difference Estimation through TDOAs Subspace methods (MUSIC) Direct search (steered beamformer) Post-processing Kalman filtering Particle filtering

8 Steered Beamformer Delay-and-sum beamformer Maximise output energy Frequency domain computation

9 Spectral Weighting Normal cross-correlation peaks are very wide PHAse Transform (PHAT) has narrow peaks Apply weighting Weight according to noise and reverberation Models the precedence effect Sensitivity is decreased after a loud sound

10 Direction Search Finding directions with highest energy Fixed number of sources Q=4 Lookup-and-sum algorithm 25 times less complex

11 Post-Processing: Particle Filtering Need to track sources over time Steered beamformer output is noisy Representing pdf as particles One set of (1000) particles per source State=[position, speed]

12 Particle Filtering Steps 1) Prediction 2) Instantaneous probabilities estimation As a function of steered beamformer energy

13 Particle Filtering Steps (cont.) 3) Source-observation assignment Need to know which observation is related to which tracked source Compute : Probability that q is a false alarm : Probability that q is source j : Probability that q is a new source

14 Particle Filtering Steps (cont.) 4) Particle weights update Merging past and present information Taking into account source-observation assignment 5) Addition or removal of sources 6) Estimation of source positions Weighted mean of the particle positions 7) Resampling

15 Localisation Results (E1) Detection accuracy over distance Localisation accuracy

16 Tracking Results Two sources crossing with C2 Video E1 E2

17 Tracking Results (cont.) Four moving sources with C2 E1 E2

18 Sound Source Separation & Speech Recognition

19 Overview of Sound Source Separation Frequency domain processing Simple, low complexity Linear source separation Non-linear post-filter Tracking information Microphones X n k, l Sources S m k, l Geometric Y m k, l Postsource filter separation Separated Sources S m k, l

20 Geometric Source Separation Frequency domain: Constrained optimization Minimize correlation of the outputs: Subject to geometric constraint: Modifications to original GSS algorithm Instantaneous computation of correlations Regularisation

21 Multi-Source Post-Filter

22 Interference Estimation Source separation leaks Incomplete adaptation Inaccuracy in localization Reverberation/diffraction Imperfect microphones Estimation from other separated sources

23 Reverberation Estimation Exponential decay model Example: 500 Hz frequency bin

24 Results (SNR) Three speakers C2 (shell), E1 (lab) 15 12,5 10 SNR (db) 7,5 5 Source 1 2,5 Source 2 0 Source 3-2,5-5 -7,5 Input Delayandsum GSS GSS + singlesource GSS + multisource

25 Speech Recognition Accuracy (Nuance) Proposed post-filter reduces errors by 50% Reverberation removal helps in E2 only No significant difference between C1 and C2 E2, C2, 3 speakers 90% Digit recognition 85% 80% 3 speakers: 83% 75% Right 70% 2 speakers: 90% Front 65% microphone Word correct (%) separated Left 60% 55% 50% GSS only Post-filter (no dereverb.) Proposed system

26 Man vs. Machine How does a human compare? 90% Word correct (%) 85% 80% 75% 70% 65% 60% 55% 50% Listener 1 Is it fair? Yes and no! Listener 2 Listener 3 Listener 4 Listener 5 Proposed system

27 Real-Time Application Video from AAAI conference

28 Speech Recognition With Missing Feature Theory Speech is transformed into features (~12) Not all features are reliable MFT = ignore unreliable features Compute missing feature mask Use the mask to compute probabilities

29 Missing Feature Mask Interference: unreliable Stationary noise: reliable black: reliable white: unreliable

30 Results (MFT) Japanese isolated word recognition (SIG2 robot, CTK) 3 simultaneous sources 200-word vocabulary 30, 60, 90 degrees separation Word correct (%) Right Front Left GSS GSS+postfilter GSS+postfilter+MFT

31 Summary of the System

32 Conclusion What have we achieved? Localisation and tracking of sound sources Separation of multiple sources Robust basis for human-robot interaction What are the main innovations? Frequency-domain steered beamformer Particle filtering source-observation assignment Separation post-filtering for multiple sources and reverberation Integration with missing feature theory

33 Where From Here? Future work Complete dialogue system Echo cancellation for the robot's own voice Use human-inspired techniques Environmental sound recognition Embedded implementation Other applications Video-conference: automatically follow speaker with a camera Automatic transcription

34 Questions? Comments?

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research

Improving Meetings with Microphone Array Algorithms. Ivan Tashev Microsoft Research Improving Meetings with Microphone Array Algorithms Ivan Tashev Microsoft Research Why microphone arrays? They ensure better sound quality: less noises and reverberation Provide speaker position using

More information

Indoor Sound Localization

Indoor Sound Localization MIN-Fakultät Fachbereich Informatik Indoor Sound Localization Fares Abawi Universität Hamburg Fakultät für Mathematik, Informatik und Naturwissenschaften Fachbereich Informatik Technische Aspekte Multimodaler

More information

Embedded Auditory System for Small Mobile Robots

Embedded Auditory System for Small Mobile Robots Embedded Auditory System for Small Mobile Robots Simon Brière, Jean-Marc Valin, François Michaud, Dominic Létourneau Abstract Auditory capabilities would allow small robots interacting with people to act

More information

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer

Michael Brandstein Darren Ward (Eds.) Microphone Arrays. Signal Processing Techniques and Applications. With 149 Figures. Springer Michael Brandstein Darren Ward (Eds.) Microphone Arrays Signal Processing Techniques and Applications With 149 Figures Springer Contents Part I. Speech Enhancement 1 Constant Directivity Beamforming Darren

More information

The psychoacoustics of reverberation

The psychoacoustics of reverberation The psychoacoustics of reverberation Steven van de Par Steven.van.de.Par@uni-oldenburg.de July 19, 2016 Thanks to Julian Grosse and Andreas Häußler 2016 AES International Conference on Sound Field Control

More information

Automotive three-microphone voice activity detector and noise-canceller

Automotive three-microphone voice activity detector and noise-canceller Res. Lett. Inf. Math. Sci., 005, Vol. 7, pp 47-55 47 Available online at http://iims.massey.ac.nz/research/letters/ Automotive three-microphone voice activity detector and noise-canceller Z. QI and T.J.MOIR

More information

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B.

Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya 2, B. Yamuna 2, H. Divya 2, B. Shiva Kumar 2, B. www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 4 Issue 4 April 2015, Page No. 11143-11147 Speech Enhancement Using Beamforming Dr. G. Ramesh Babu 1, D. Lavanya

More information

Simultaneous Recognition of Speech Commands by a Robot using a Small Microphone Array

Simultaneous Recognition of Speech Commands by a Robot using a Small Microphone Array 2012 2nd International Conference on Computer Design and Engineering (ICCDE 2012) IPCSIT vol. 49 (2012) (2012) IACSIT Press, Singapore DOI: 10.7763/IPCSIT.2012.V49.14 Simultaneous Recognition of Speech

More information

Lateralisation of multiple sound sources by the auditory system

Lateralisation of multiple sound sources by the auditory system Modeling of Binaural Discrimination of multiple Sound Sources: A Contribution to the Development of a Cocktail-Party-Processor 4 H.SLATKY (Lehrstuhl für allgemeine Elektrotechnik und Akustik, Ruhr-Universität

More information

Robust Speech Recognition Group Carnegie Mellon University. Telephone: Fax:

Robust Speech Recognition Group Carnegie Mellon University. Telephone: Fax: Robust Automatic Speech Recognition In the 21 st Century Richard Stern (with Alex Acero, Yu-Hsiang Chiu, Evandro Gouvêa, Chanwoo Kim, Kshitiz Kumar, Amir Moghimi, Pedro Moreno, Hyung-Min Park, Bhiksha

More information

A MICROPHONE ARRAY INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE

A MICROPHONE ARRAY INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE A MICROPHONE ARRA INTERFACE FOR REAL-TIME INTERACTIVE MUSIC PERFORMANCE Daniele Salvati AVIRES lab Dep. of Mathematics and Computer Science, University of Udine, Italy daniele.salvati@uniud.it Sergio Canazza

More information

Speaker Localization in Noisy Environments Using Steered Response Voice Power

Speaker Localization in Noisy Environments Using Steered Response Voice Power 112 IEEE Transactions on Consumer Electronics, Vol. 61, No. 1, February 2015 Speaker Localization in Noisy Environments Using Steered Response Voice Power Hyeontaek Lim, In-Chul Yoo, Youngkyu Cho, and

More information

Robust Low-Resource Sound Localization in Correlated Noise

Robust Low-Resource Sound Localization in Correlated Noise INTERSPEECH 2014 Robust Low-Resource Sound Localization in Correlated Noise Lorin Netsch, Jacek Stachurski Texas Instruments, Inc. netsch@ti.com, jacek@ti.com Abstract In this paper we address the problem

More information

EE1.el3 (EEE1023): Electronics III. Acoustics lecture 20 Sound localisation. Dr Philip Jackson.

EE1.el3 (EEE1023): Electronics III. Acoustics lecture 20 Sound localisation. Dr Philip Jackson. EE1.el3 (EEE1023): Electronics III Acoustics lecture 20 Sound localisation Dr Philip Jackson www.ee.surrey.ac.uk/teaching/courses/ee1.el3 Sound localisation Objectives: calculate frequency response of

More information

Automatic Text-Independent. Speaker. Recognition Approaches Using Binaural Inputs

Automatic Text-Independent. Speaker. Recognition Approaches Using Binaural Inputs Automatic Text-Independent Speaker Recognition Approaches Using Binaural Inputs Karim Youssef, Sylvain Argentieri and Jean-Luc Zarader 1 Outline Automatic speaker recognition: introduction Designed systems

More information

/07/$ IEEE 111

/07/$ IEEE 111 DESIGN AND IMPLEMENTATION OF A ROBOT AUDITION SYSTEM FOR AUTOMATIC SPEECH RECOGNITION OF SIMULTANEOUS SPEECH Shun ichi Yamamoto, Kazuhiro Nakadai, Mikio Nakano, Hiroshi Tsujino, Jean-Marc Valin, Kazunori

More information

Students: Avihay Barazany Royi Levy Supervisor: Kuti Avargel In Association with: Zoran, Haifa

Students: Avihay Barazany Royi Levy Supervisor: Kuti Avargel In Association with: Zoran, Haifa Students: Avihay Barazany Royi Levy Supervisor: Kuti Avargel In Association with: Zoran, Haifa Spring 2008 Introduction Problem Formulation Possible Solutions Proposed Algorithm Experimental Results Conclusions

More information

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment International Journal of Electronics Engineering Research. ISSN 975-645 Volume 9, Number 4 (27) pp. 545-556 Research India Publications http://www.ripublication.com Study Of Sound Source Localization Using

More information

Binaural Hearing. Reading: Yost Ch. 12

Binaural Hearing. Reading: Yost Ch. 12 Binaural Hearing Reading: Yost Ch. 12 Binaural Advantages Sounds in our environment are usually complex, and occur either simultaneously or close together in time. Studies have shown that the ability to

More information

Figure 1. SIG ACAM 100 and OptiNav BeamformX at InterNoise 2015.

Figure 1. SIG ACAM 100 and OptiNav BeamformX at InterNoise 2015. SIG ACAM 100 with OptiNav BeamformX Signal Interface Group s (SIG) ACAM 100 is a microphone array for locating and analyzing sound sources in real time. Combined with OptiNav s BeamformX software, it makes

More information

Estimation of Reverberation Time from Binaural Signals Without Using Controlled Excitation

Estimation of Reverberation Time from Binaural Signals Without Using Controlled Excitation Estimation of Reverberation Time from Binaural Signals Without Using Controlled Excitation Sampo Vesa Master s Thesis presentation on 22nd of September, 24 21st September 24 HUT / Laboratory of Acoustics

More information

Microphone Array Design and Beamforming

Microphone Array Design and Beamforming Microphone Array Design and Beamforming Heinrich Löllmann Multimedia Communications and Signal Processing heinrich.loellmann@fau.de with contributions from Vladi Tourbabin and Hendrik Barfuss EUSIPCO Tutorial

More information

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES ROOM AND CONCERT HALL ACOUSTICS The perception of sound by human listeners in a listening space, such as a room or a concert hall is a complicated function of the type of source sound (speech, oration,

More information

Audio data fuzzy fusion for source localization

Audio data fuzzy fusion for source localization International Neural Network Society 13-16 September, 2013, Halkidiki, Greece Audio data fuzzy fusion for source localization M. Malcangi Università degli Studi di Milano Department of Computer Science

More information

SGN Audio and Speech Processing

SGN Audio and Speech Processing Introduction 1 Course goals Introduction 2 SGN 14006 Audio and Speech Processing Lectures, Fall 2014 Anssi Klapuri Tampere University of Technology! Learn basics of audio signal processing Basic operations

More information

Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues

Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues Effects of Reverberation on Pitch, Onset/Offset, and Binaural Cues DeLiang Wang Perception & Neurodynamics Lab The Ohio State University Outline of presentation Introduction Human performance Reverberation

More information

Eyes n Ears: A System for Attentive Teleconferencing

Eyes n Ears: A System for Attentive Teleconferencing Eyes n Ears: A System for Attentive Teleconferencing B. Kapralos 1,3, M. Jenkin 1,3, E. Milios 2,3 and J. Tsotsos 1,3 1 Department of Computer Science, York University, North York, Canada M3J 1P3 2 Department

More information

Evaluating Real-time Audio Localization Algorithms for Artificial Audition in Robotics

Evaluating Real-time Audio Localization Algorithms for Artificial Audition in Robotics Evaluating Real-time Audio Localization Algorithms for Artificial Audition in Robotics Anthony Badali, Jean-Marc Valin,François Michaud, and Parham Aarabi University of Toronto Dept. of Electrical & Computer

More information

Mei Wu Acoustics. By Mei Wu and James Black

Mei Wu Acoustics. By Mei Wu and James Black Experts in acoustics, noise and vibration Effects of Physical Environment on Speech Intelligibility in Teleconferencing (This article was published at Sound and Video Contractors website www.svconline.com

More information

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm International OPEN ACCESS Journal Of Modern Engineering Research (IJMER) Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm A.T. Rajamanickam, N.P.Subiramaniyam, A.Balamurugan*,

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Architectural Acoustics Session 1pAAa: Advanced Analysis of Room Acoustics:

More information

NOISE ESTIMATION IN A SINGLE CHANNEL

NOISE ESTIMATION IN A SINGLE CHANNEL SPEECH ENHANCEMENT FOR CROSS-TALK INTERFERENCE by Levent M. Arslan and John H.L. Hansen Robust Speech Processing Laboratory Department of Electrical Engineering Box 99 Duke University Durham, North Carolina

More information

Stefan Launer, Lyon, January 2011 Phonak AG, Stäfa, CH

Stefan Launer, Lyon, January 2011 Phonak AG, Stäfa, CH State of art and Challenges in Improving Speech Intelligibility in Hearing Impaired People Stefan Launer, Lyon, January 2011 Phonak AG, Stäfa, CH Content Phonak Stefan Launer, Speech in Noise Workshop,

More information

Leak Energy Based Missing Feature Mask Generation for ICA and GSS and Its Evaluation with Simultaneous Speech Recognition

Leak Energy Based Missing Feature Mask Generation for ICA and GSS and Its Evaluation with Simultaneous Speech Recognition Leak Energy Based Missing Feature Mask Generation for ICA and GSS and Its Evaluation with Simultaneous Speech Recognition Shun ichi Yamamoto, Ryu Takeda, Kazuhiro Nakadai, Mikio Nakano, Hiroshi Tsujino,

More information

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas

Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor. Presented by Amir Kiperwas Emanuël A. P. Habets, Jacob Benesty, and Patrick A. Naylor Presented by Amir Kiperwas 1 M-element microphone array One desired source One undesired source Ambient noise field Signals: Broadband Mutually

More information

Recent Advances in Acoustic Signal Extraction and Dereverberation

Recent Advances in Acoustic Signal Extraction and Dereverberation Recent Advances in Acoustic Signal Extraction and Dereverberation Emanuël Habets Erlangen Colloquium 2016 Scenario Spatial Filtering Estimated Desired Signal Undesired sound components: Sensor noise Competing

More information

Joint Position-Pitch Decomposition for Multi-Speaker Tracking

Joint Position-Pitch Decomposition for Multi-Speaker Tracking Joint Position-Pitch Decomposition for Multi-Speaker Tracking SPSC Laboratory, TU Graz 1 Contents: 1. Microphone Arrays SPSC circular array Beamforming 2. Source Localization Direction of Arrival (DoA)

More information

Sound Source Localization using HRTF database

Sound Source Localization using HRTF database ICCAS June -, KINTEX, Gyeonggi-Do, Korea Sound Source Localization using HRTF database Sungmok Hwang*, Youngjin Park and Younsik Park * Center for Noise and Vibration Control, Dept. of Mech. Eng., KAIST,

More information

Meeting Corpora Hardware Overview & ASR Accuracies

Meeting Corpora Hardware Overview & ASR Accuracies Meeting Corpora Hardware Overview & ASR Accuracies George Jose (153070011) Guide : Dr. Preeti Rao Indian Institute of Technology, Bombay 22 July, 2016 1/18 Outline 1 AMI Meeting Corpora 2 3 2/18 AMI Meeting

More information

ROBUST echo cancellation requires a method for adjusting

ROBUST echo cancellation requires a method for adjusting 1030 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 3, MARCH 2007 On Adjusting the Learning Rate in Frequency Domain Echo Cancellation With Double-Talk Jean-Marc Valin, Member,

More information

Automatic Speech Recognition Improved by Two-Layered Audio-Visual Integration For Robot Audition

Automatic Speech Recognition Improved by Two-Layered Audio-Visual Integration For Robot Audition 9th IEEE-RAS International Conference on Humanoid Robots December 7-, 29 Paris, France Automatic Speech Recognition Improved by Two-Layered Audio-Visual Integration For Robot Audition Takami Yoshida, Kazuhiro

More information

Monaural and Binaural Speech Separation

Monaural and Binaural Speech Separation Monaural and Binaural Speech Separation DeLiang Wang Perception & Neurodynamics Lab The Ohio State University Outline of presentation Introduction CASA approach to sound separation Ideal binary mask as

More information

TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones and Source Counting

TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones and Source Counting TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones Source Counting Ali Pourmohammad, Member, IACSIT Seyed Mohammad Ahadi Abstract In outdoor cases, TDOA-based methods

More information

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation

Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation Dual Transfer Function GSC and Application to Joint Noise Reduction and Acoustic Echo Cancellation Gal Reuven Under supervision of Sharon Gannot 1 and Israel Cohen 2 1 School of Engineering, Bar-Ilan University,

More information

Factors Governing the Intelligibility of Speech Sounds

Factors Governing the Intelligibility of Speech Sounds HSR Journal Club JASA, vol(19) No(1), Jan 1947 Factors Governing the Intelligibility of Speech Sounds N. R. French and J. C. Steinberg 1. Introduction Goal: Determine a quantitative relationship between

More information

Self Localization Using A Modulated Acoustic Chirp

Self Localization Using A Modulated Acoustic Chirp Self Localization Using A Modulated Acoustic Chirp Brian P. Flanagan The MITRE Corporation, 7515 Colshire Dr., McLean, VA 2212, USA; bflan@mitre.org ABSTRACT This paper describes a robust self localization

More information

Optimal Adaptive Filtering Technique for Tamil Speech Enhancement

Optimal Adaptive Filtering Technique for Tamil Speech Enhancement Optimal Adaptive Filtering Technique for Tamil Speech Enhancement Vimala.C Project Fellow, Department of Computer Science Avinashilingam Institute for Home Science and Higher Education and Women Coimbatore,

More information

SGN Audio and Speech Processing

SGN Audio and Speech Processing SGN 14006 Audio and Speech Processing Introduction 1 Course goals Introduction 2! Learn basics of audio signal processing Basic operations and their underlying ideas and principles Give basic skills although

More information

In air acoustic vector sensors for capturing and processing of speech signals

In air acoustic vector sensors for capturing and processing of speech signals University of Wollongong Research Online University of Wollongong Thesis Collection University of Wollongong Thesis Collections 2011 In air acoustic vector sensors for capturing and processing of speech

More information

Different Approaches of Spectral Subtraction Method for Speech Enhancement

Different Approaches of Spectral Subtraction Method for Speech Enhancement ISSN 2249 5460 Available online at www.internationalejournals.com International ejournals International Journal of Mathematical Sciences, Technology and Humanities 95 (2013 1056 1062 Different Approaches

More information

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks

Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks Distance Estimation and Localization of Sound Sources in Reverberant Conditions using Deep Neural Networks Mariam Yiwere 1 and Eun Joo Rhee 2 1 Department of Computer Engineering, Hanbat National University,

More information

Advances in Direction-of-Arrival Estimation

Advances in Direction-of-Arrival Estimation Advances in Direction-of-Arrival Estimation Sathish Chandran Editor ARTECH HOUSE BOSTON LONDON artechhouse.com Contents Preface xvii Acknowledgments xix Overview CHAPTER 1 Antenna Arrays for Direction-of-Arrival

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Architectural Acoustics Session 2aAAa: Adapting, Enhancing, and Fictionalizing

More information

EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION

EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION University of Kentucky UKnowledge University of Kentucky Master's Theses Graduate School 2007 EXPERIMENTAL EVALUATION OF MODIFIED PHASE TRANSFORM FOR SOUND SOURCE DETECTION Anand Ramamurthy University

More information

PERFORMANCE COMPARISON BETWEEN STEREAUSIS AND INCOHERENT WIDEBAND MUSIC FOR LOCALIZATION OF GROUND VEHICLES ABSTRACT

PERFORMANCE COMPARISON BETWEEN STEREAUSIS AND INCOHERENT WIDEBAND MUSIC FOR LOCALIZATION OF GROUND VEHICLES ABSTRACT Approved for public release; distribution is unlimited. PERFORMANCE COMPARISON BETWEEN STEREAUSIS AND INCOHERENT WIDEBAND MUSIC FOR LOCALIZATION OF GROUND VEHICLES September 1999 Tien Pham U.S. Army Research

More information

Improvement in Listening Capability for Humanoid Robot HRP-2

Improvement in Listening Capability for Humanoid Robot HRP-2 2010 IEEE International Conference on Robotics and Automation Anchorage Convention District May 3-8, 2010, Anchorage, Alaska, USA Improvement in Listening Capability for Humanoid Robot HRP-2 Toru Takahashi,

More information

All-Neural Multi-Channel Speech Enhancement

All-Neural Multi-Channel Speech Enhancement Interspeech 2018 2-6 September 2018, Hyderabad All-Neural Multi-Channel Speech Enhancement Zhong-Qiu Wang 1, DeLiang Wang 1,2 1 Department of Computer Science and Engineering, The Ohio State University,

More information

VQ Source Models: Perceptual & Phase Issues

VQ Source Models: Perceptual & Phase Issues VQ Source Models: Perceptual & Phase Issues Dan Ellis & Ron Weiss Laboratory for Recognition and Organization of Speech and Audio Dept. Electrical Eng., Columbia Univ., NY USA {dpwe,ronw}@ee.columbia.edu

More information

Onset Detection Revisited

Onset Detection Revisited simon.dixon@ofai.at Austrian Research Institute for Artificial Intelligence Vienna, Austria 9th International Conference on Digital Audio Effects Outline Background and Motivation 1 Background and Motivation

More information

Audio Imputation Using the Non-negative Hidden Markov Model

Audio Imputation Using the Non-negative Hidden Markov Model Audio Imputation Using the Non-negative Hidden Markov Model Jinyu Han 1,, Gautham J. Mysore 2, and Bryan Pardo 1 1 EECS Department, Northwestern University 2 Advanced Technology Labs, Adobe Systems Inc.

More information

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION Lin Wang 1,2, Heping Ding 2 and Fuliang Yin 1 1 School of Electronic and Information Engineering, Dalian

More information

Bandwidth Extension for Speech Enhancement

Bandwidth Extension for Speech Enhancement Bandwidth Extension for Speech Enhancement F. Mustiere, M. Bouchard, M. Bolic University of Ottawa Tuesday, May 4 th 2010 CCECE 2010: Signal and Multimedia Processing 1 2 3 4 Current Topic 1 2 3 4 Context

More information

Convention Paper Presented at the 131st Convention 2011 October New York, USA

Convention Paper Presented at the 131st Convention 2011 October New York, USA Audio Engineering Society Convention Paper Presented at the 131st Convention 211 October 2 23 New York, USA This paper was peer-reviewed as a complete manuscript for presentation at this Convention. Additional

More information

Voice Activity Detection

Voice Activity Detection Voice Activity Detection Speech Processing Tom Bäckström Aalto University October 2015 Introduction Voice activity detection (VAD) (or speech activity detection, or speech detection) refers to a class

More information

Multiple Sound Sources Localization Using Energetic Analysis Method

Multiple Sound Sources Localization Using Energetic Analysis Method VOL.3, NO.4, DECEMBER 1 Multiple Sound Sources Localization Using Energetic Analysis Method Hasan Khaddour, Jiří Schimmel Department of Telecommunications FEEC, Brno University of Technology Purkyňova

More information

BIOLOGICALLY INSPIRED BINAURAL ANALOGUE SIGNAL PROCESSING

BIOLOGICALLY INSPIRED BINAURAL ANALOGUE SIGNAL PROCESSING Brain Inspired Cognitive Systems August 29 September 1, 2004 University of Stirling, Scotland, UK BIOLOGICALLY INSPIRED BINAURAL ANALOGUE SIGNAL PROCESSING Natasha Chia and Steve Collins University of

More information

arxiv: v1 [cs.sd] 4 Dec 2018

arxiv: v1 [cs.sd] 4 Dec 2018 LOCALIZATION AND TRACKING OF AN ACOUSTIC SOURCE USING A DIAGONAL UNLOADING BEAMFORMING AND A KALMAN FILTER Daniele Salvati, Carlo Drioli, Gian Luca Foresti Department of Mathematics, Computer Science and

More information

Integrated Vision and Sound Localization

Integrated Vision and Sound Localization Integrated Vision and Sound Localization Parham Aarabi Safwat Zaky Department of Electrical and Computer Engineering University of Toronto 10 Kings College Road, Toronto, Ontario, Canada, M5S 3G4 parham@stanford.edu

More information

Binaural segregation in multisource reverberant environments

Binaural segregation in multisource reverberant environments Binaural segregation in multisource reverberant environments Nicoleta Roman a Department of Computer Science and Engineering, The Ohio State University, Columbus, Ohio 43210 Soundararajan Srinivasan b

More information

Microphone Array project in MSR: approach and results

Microphone Array project in MSR: approach and results Microphone Array project in MSR: approach and results Ivan Tashev Microsoft Research June 2004 Agenda Microphone Array project Beamformer design algorithm Implementation and hardware designs Demo Motivation

More information

Speech Quality Assessment for Wideband Communication Scenarios

Speech Quality Assessment for Wideband Communication Scenarios Speech Quality Assessment for Wideband Communication Scenarios H. W. Gierlich, S. Völl, F. Kettler (HEAD acoustics GmbH) P. Jax (IND, RWTH Aachen) Workshop on Wideband Speech Quality in Terminals and Networks

More information

GSM Interference Cancellation For Forensic Audio

GSM Interference Cancellation For Forensic Audio Application Report BACK April 2001 GSM Interference Cancellation For Forensic Audio Philip Harrison and Dr Boaz Rafaely (supervisor) Institute of Sound and Vibration Research (ISVR) University of Southampton,

More information

Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self

Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Paul Fitzpatrick and Artur M. Arsenio CSAIL, MIT Modal and amodal features Modal and amodal features (following

More information

SOUND SOURCE LOCATION METHOD

SOUND SOURCE LOCATION METHOD SOUND SOURCE LOCATION METHOD Michal Mandlik 1, Vladimír Brázda 2 Summary: This paper deals with received acoustic signals on microphone array. In this paper the localization system based on a speaker speech

More information

Local Relative Transfer Function for Sound Source Localization

Local Relative Transfer Function for Sound Source Localization Local Relative Transfer Function for Sound Source Localization Xiaofei Li 1, Radu Horaud 1, Laurent Girin 1,2, Sharon Gannot 3 1 INRIA Grenoble Rhône-Alpes. {firstname.lastname@inria.fr} 2 GIPSA-Lab &

More information

POSSIBLY the most noticeable difference when performing

POSSIBLY the most noticeable difference when performing IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 15, NO. 7, SEPTEMBER 2007 2011 Acoustic Beamforming for Speaker Diarization of Meetings Xavier Anguera, Associate Member, IEEE, Chuck Wooters,

More information

3 RD GENERATION BE HEARD AND HEAR, LOUD AND CLEAR

3 RD GENERATION BE HEARD AND HEAR, LOUD AND CLEAR 3 RD GENERATION BE HEARD AND HEAR, LOUD AND CLEAR The ultimate voice and communications solution, MaxxVoice is a suite of state-of-the-art technologies created by Waves Audio, recipient of a 2011 Technical

More information

Sound Processing Technologies for Realistic Sensations in Teleworking

Sound Processing Technologies for Realistic Sensations in Teleworking Sound Processing Technologies for Realistic Sensations in Teleworking Takashi Yazu Makoto Morito In an office environment we usually acquire a large amount of information without any particular effort

More information

Acoustic Beamforming for Speaker Diarization of Meetings

Acoustic Beamforming for Speaker Diarization of Meetings JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY 2007 1 Acoustic Beamforming for Speaker Diarization of Meetings Xavier Anguera, Member, IEEE, Chuck Wooters, Member, IEEE, Javier Hernando, Member,

More information

An analysis of blind signal separation for real time application

An analysis of blind signal separation for real time application University of Wollongong Research Online University of Wollongong Thesis Collection 1954-2016 University of Wollongong Thesis Collections 2006 An analysis of blind signal separation for real time application

More information

Localization of underwater moving sound source based on time delay estimation using hydrophone array

Localization of underwater moving sound source based on time delay estimation using hydrophone array Journal of Physics: Conference Series PAPER OPEN ACCESS Localization of underwater moving sound source based on time delay estimation using hydrophone array To cite this article: S. A. Rahman et al 2016

More information

Robust Speech Recognition Based on Binaural Auditory Processing

Robust Speech Recognition Based on Binaural Auditory Processing INTERSPEECH 2017 August 20 24, 2017, Stockholm, Sweden Robust Speech Recognition Based on Binaural Auditory Processing Anjali Menon 1, Chanwoo Kim 2, Richard M. Stern 1 1 Department of Electrical and Computer

More information

Revision 1.1 May Front End DSP Audio Technologies for In-Car Applications ROADMAP 2016

Revision 1.1 May Front End DSP Audio Technologies for In-Car Applications ROADMAP 2016 Revision 1.1 May 2016 Front End DSP Audio Technologies for In-Car Applications ROADMAP 2016 PAGE 2 EXISTING PRODUCTS 1. Hands-free communication enhancement: Voice Communication Package (VCP-7) generation

More information

Speech Enhancement Techniques using Wiener Filter and Subspace Filter

Speech Enhancement Techniques using Wiener Filter and Subspace Filter IJSTE - International Journal of Science Technology & Engineering Volume 3 Issue 05 November 2016 ISSN (online): 2349-784X Speech Enhancement Techniques using Wiener Filter and Subspace Filter Ankeeta

More information

Calibration of Microphone Arrays for Improved Speech Recognition

Calibration of Microphone Arrays for Improved Speech Recognition MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Calibration of Microphone Arrays for Improved Speech Recognition Michael L. Seltzer, Bhiksha Raj TR-2001-43 December 2001 Abstract We present

More information

Sound Source Localization in Median Plane using Artificial Ear

Sound Source Localization in Median Plane using Artificial Ear International Conference on Control, Automation and Systems 28 Oct. 14-17, 28 in COEX, Seoul, Korea Sound Source Localization in Median Plane using Artificial Ear Sangmoon Lee 1, Sungmok Hwang 2, Youngjin

More information

COM 12 C 288 E October 2011 English only Original: English

COM 12 C 288 E October 2011 English only Original: English Question(s): 9/12 Source: Title: INTERNATIONAL TELECOMMUNICATION UNION TELECOMMUNICATION STANDARDIZATION SECTOR STUDY PERIOD 2009-2012 Audience STUDY GROUP 12 CONTRIBUTION 288 P.ONRA Contribution Additional

More information

Speech Enhancement Based On Noise Reduction

Speech Enhancement Based On Noise Reduction Speech Enhancement Based On Noise Reduction Kundan Kumar Singh Electrical Engineering Department University Of Rochester ksingh11@z.rochester.edu ABSTRACT This paper addresses the problem of signal distortion

More information

Speech Enhancement: Reduction of Additive Noise in the Digital Processing of Speech

Speech Enhancement: Reduction of Additive Noise in the Digital Processing of Speech Speech Enhancement: Reduction of Additive Noise in the Digital Processing of Speech Project Proposal Avner Halevy Department of Mathematics University of Maryland, College Park ahalevy at math.umd.edu

More information

Assessment of General Applicability of Ego Noise Estimation

Assessment of General Applicability of Ego Noise Estimation 211 IEEE International Conference on Robotics and Automation Shanghai International Conference Center May 9-13, 211, Shanghai, China Assessment of General Applicability of Ego Estimation Applications to

More information

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming

Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming Speech and Audio Processing Recognition and Audio Effects Part 3: Beamforming Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Engineering

More information

Separation and Recognition of multiple sound source using Pulsed Neuron Model

Separation and Recognition of multiple sound source using Pulsed Neuron Model Separation and Recognition of multiple sound source using Pulsed Neuron Model Kaname Iwasa, Hideaki Inoue, Mauricio Kugler, Susumu Kuroyanagi, Akira Iwata Nagoya Institute of Technology, Gokiso-cho, Showa-ku,

More information

A HYPOTHESIS TESTING APPROACH FOR REAL-TIME MULTICHANNEL SPEECH SEPARATION USING TIME-FREQUENCY MASKS. Ryan M. Corey and Andrew C.

A HYPOTHESIS TESTING APPROACH FOR REAL-TIME MULTICHANNEL SPEECH SEPARATION USING TIME-FREQUENCY MASKS. Ryan M. Corey and Andrew C. 6 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 3 6, 6, SALERNO, ITALY A HYPOTHESIS TESTING APPROACH FOR REAL-TIME MULTICHANNEL SPEECH SEPARATION USING TIME-FREQUENCY MASKS

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 12 Speech Signal Processing 14/03/25 http://www.ee.unlv.edu/~b1morris/ee482/

More information

Introduction to Audio Watermarking Schemes

Introduction to Audio Watermarking Schemes Introduction to Audio Watermarking Schemes N. Lazic and P. Aarabi, Communication over an Acoustic Channel Using Data Hiding Techniques, IEEE Transactions on Multimedia, Vol. 8, No. 5, October 2006 Multimedia

More information

Airo Interantional Research Journal September, 2013 Volume II, ISSN:

Airo Interantional Research Journal September, 2013 Volume II, ISSN: Airo Interantional Research Journal September, 2013 Volume II, ISSN: 2320-3714 Name of author- Navin Kumar Research scholar Department of Electronics BR Ambedkar Bihar University Muzaffarpur ABSTRACT Direction

More information

Robust Voice Activity Detection Based on Discrete Wavelet. Transform

Robust Voice Activity Detection Based on Discrete Wavelet. Transform Robust Voice Activity Detection Based on Discrete Wavelet Transform Kun-Ching Wang Department of Information Technology & Communication Shin Chien University kunching@mail.kh.usc.edu.tw Abstract This paper

More information

A Simple Two-Microphone Array Devoted to Speech Enhancement and Source Tracking

A Simple Two-Microphone Array Devoted to Speech Enhancement and Source Tracking A Simple Two-Microphone Array Devoted to Speech Enhancement and Source Tracking A. Álvarez, P. Gómez, R. Martínez and, V. Nieto Departamento de Arquitectura y Tecnología de Sistemas Informáticos Universidad

More information

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter

Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter 1 Gupteswar Sahu, 2 D. Arun Kumar, 3 M. Bala Krishna and 4 Jami Venkata Suman Assistant Professor, Department of ECE,

More information

From Binaural Technology to Virtual Reality

From Binaural Technology to Virtual Reality From Binaural Technology to Virtual Reality Jens Blauert, D-Bochum Prominent Prominent Features of of Binaural Binaural Hearing Hearing - Localization Formation of positions of the auditory events (azimuth,

More information