Acoustics Research Institute

Similar documents
Computational Perception. Sound localization 2

Binaural Hearing. Reading: Yost Ch. 12

Listening with Headphones

A Virtual Audio Environment for Testing Dummy- Head HRTFs modeling Real Life Situations

Binaural hearing. Prof. Dan Tollin on the Hearing Throne, Oldenburg Hearing Garden

Auditory Localization

Spatial Audio Reproduction: Towards Individualized Binaural Sound

The role of fine structure in bilateral cochlear implantation

Upper hemisphere sound localization using head-related transfer functions in the median plane and interaural differences

Sound Source Localization using HRTF database

EXPLORATION OF A BIOLOGICALLY INSPIRED MODEL FOR SOUND SOURCE LOCALIZATION IN 3D SPACE

URBANA-CHAMPAIGN. CS 498PS Audio Computing Lab. 3D and Virtual Sound. Paris Smaragdis. paris.cs.illinois.

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics

PAPER Enhanced Vertical Perception through Head-Related Impulse Response Customization Based on Pinna Response Tuning in the Median Plane

Convention Paper Presented at the 139th Convention 2015 October 29 November 1 New York, USA

From Binaural Technology to Virtual Reality

EE1.el3 (EEE1023): Electronics III. Acoustics lecture 20 Sound localisation. Dr Philip Jackson.

3D sound image control by individualized parametric head-related transfer functions

NEAR-FIELD VIRTUAL AUDIO DISPLAYS

c 2014 Michael Friedman

Sound source localization and its use in multimedia applications

Ivan Tashev Microsoft Research

ORIENTATION IN SIMPLE VIRTUAL AUDITORY SPACE CREATED WITH MEASURED HRTF

Convention Paper 9712 Presented at the 142 nd Convention 2017 May 20 23, Berlin, Germany

Audio Engineering Society. Convention Paper. Presented at the 124th Convention 2008 May Amsterdam, The Netherlands

Introduction. 1.1 Surround sound

Enhancing 3D Audio Using Blind Bandwidth Extension

Audio Engineering Society. Convention Paper. Presented at the 131st Convention 2011 October New York, NY, USA

III. Publication III. c 2005 Toni Hirvonen.

Study on method of estimating direct arrival using monaural modulation sp. Author(s)Ando, Masaru; Morikawa, Daisuke; Uno

BINAURAL RECORDING SYSTEM AND SOUND MAP OF MALAGA

Creating three dimensions in virtual auditory displays *

Extracting the frequencies of the pinna spectral notches in measured head related impulse responses

Computational Perception /785

HRTF adaptation and pattern learning

Tara J. Martin Boston University Hearing Research Center, 677 Beacon Street, Boston, Massachusetts 02215

Binaural Hearing- Human Ability of Sound Source Localization

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

University of Huddersfield Repository

Intensity Discrimination and Binaural Interaction

Convention Paper Presented at the 125th Convention 2008 October 2 5 San Francisco, CA, USA

Proceedings of Meetings on Acoustics

A triangulation method for determining the perceptual center of the head for auditory stimuli

PERSONALIZED HEAD RELATED TRANSFER FUNCTION MEASUREMENT AND VERIFICATION THROUGH SOUND LOCALIZATION RESOLUTION

Personalization of head-related transfer functions in the median plane based on the anthropometry of the listener s pinnae a)

3D Sound Simulation over Headphones

HRIR Customization in the Median Plane via Principal Components Analysis

Convention Paper 9870 Presented at the 143 rd Convention 2017 October 18 21, New York, NY, USA

Accurate sound reproduction from two loudspeakers in a living room

Auditory Distance Perception. Yan-Chen Lu & Martin Cooke

Reproduction of Surround Sound in Headphones

Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA. Why Ambisonics Does Work

Sound localization Sound localization in audio-based games for visually impaired children

THE DEVELOPMENT OF A DESIGN TOOL FOR 5-SPEAKER SURROUND SOUND DECODERS

A learning, biologically-inspired sound localization model

Circumaural transducer arrays for binaural synthesis

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 A MODEL OF THE HEAD-RELATED TRANSFER FUNCTION BASED ON SPECTRAL CUES

Virtual Reality Presentation of Loudspeaker Stereo Recordings

The Human Auditory System

TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones and Source Counting

Externalization in binaural synthesis: effects of recording environment and measurement procedure

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.

WAVELET-BASED SPECTRAL SMOOTHING FOR HEAD-RELATED TRANSFER FUNCTION FILTER DESIGN

3D Audio Systems through Stereo Loudspeakers

Localization of the Speaker in a Real and Virtual Reverberant Room. Abstract

Convention Paper Presented at the 144 th Convention 2018 May 23 26, Milan, Italy

University of Huddersfield Repository

Aalborg Universitet. Audibility of time switching in dynamic binaural synthesis Hoffmann, Pablo Francisco F.; Møller, Henrik

3D audio overview : from 2.0 to N.M (?)

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

Virtual Audio Systems

A binaural auditory model and applications to spatial sound evaluation

Factors Affecting Auditory Localization and Situational Awareness in the Urban Battlefield

The psychoacoustics of reverberation

Exploiting envelope fluctuations to achieve robust extraction and intelligent integration of binaural cues

Spatial audio is a field that

Monaural and Binaural Speech Separation

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

Dataset of head-related transfer functions measured with a circular loudspeaker array

Psychoacoustics of 3D Sound Recording: Research and Practice

Minimum Audible Movement Angles for Discriminating Upward from Downward Trajectories of Smooth Virtual Source Motion within a Sagittal Plane

THE TEMPORAL and spectral structure of a sound signal

CRANIAL TRANSITIONS FOR SOPRANO SAXOPHONE AND ELECTRONIC PROCESSING. Jonas Braasch

ANALYZING NOTCH PATTERNS OF HEAD RELATED TRANSFER FUNCTIONS IN CIPIC AND SYMARE DATABASES. M. Shahnawaz, L. Bianchi, A. Sarti, S.

Envelopment and Small Room Acoustics

Final Exam Study Guide: Introduction to Computer Music Course Staff April 24, 2015

Proceedings of Meetings on Acoustics

Sound Source Localization in Median Plane using Artificial Ear

SPAT. Binaural Encoding Tool. Multiformat Room Acoustic Simulation & Localization Processor. Flux All rights reserved

[ V. Ralph Algazi and Richard O. Duda ] [ Exploiting head motion for immersive communication]

BIOLOGICALLY INSPIRED BINAURAL ANALOGUE SIGNAL PROCESSING

A classification-based cocktail-party processor

MANY emerging applications require the ability to render

IEEE/ACM TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 26, NO. 7, JULY

Assessing the contribution of binaural cues for apparent source width perception via a functional model

On distance dependence of pinna spectral patterns in head-related transfer functions

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.

INVESTIGATING BINAURAL LOCALISATION ABILITIES FOR PROPOSING A STANDARDISED TESTING ENVIRONMENT FOR BINAURAL SYSTEMS

Virtual Acoustic Space as Assistive Technology

Transcription:

Austrian Academy of Sciences Acoustics Research Institute Spatial SpatialHearing: Hearing: Single SingleSound SoundSource Sourcein infree FreeField Field Piotr PiotrMajdak Majdak&&Bernhard BernhardLaback Laback http://www.kfs.oeaw.ac.at http://www.kfs.oeaw.ac.at piotr@majdak.com piotr@majdak.com Institut für Schallforschung, Österreichische Akademie der Wissenschaften; A-1040 Wien, Wohllebengasse 12-14 Tel: +43 1/51581-2511 Fax: +43 1/51581-2530 email: piotr@majdak.com

Spatial Hearing Object identification (outside the field-of-view) Multiple sound sources: Analysis of the auditory scene Focusing on a target (coctail-party effect) Enclosed auditory spaces: Multiple echos of the sound (reverberation) Complex perceptual effects: Sound perception (directivity, distance) Room perception Sound-room interaction (Jot, 1999)

Spatial Hearing: Simple Case Single sound source in free field: No room (no reflections) y c a r u Acc No precedence effect No distance (Direction only) Pre Elevation Azimuth n o i s ci

Localization of Sound Sources Signals from two receivers available: Aligned in the horizontal plane Asymmetries in the geometry of the individual receivers Direction-dependent spectral change of incoming sound Williams (2002)

Sound Localization in the Horizontal Plane Interaural level differences (ILDs) AL Interaural time differences (ITDs) Left: Right: AR ITD 0 ILD: AL AR ITD

Interaural Level Differences ILD (db) Broadband ILD of dummy head: Azimuth angle (deg)

Interaural Level Differences Frequency dependent: ILD (sphere model) Wave length larger than head diameter: Little effect Wave length smaller than head diameter: Large attenuation due to shadowing effects Azim ut h ang le (de g) Fr eq ue nc y Diffraction (H z) Wave length in the range of the head diameter: Blauert (1974)

Interaural Level Differences Perceptual threshold (just noticable diff., JND): In the order of 1 db (Hall, 1964 and others) Depends on the sound level (Herschkowitz & Durlach, 1969) ILD JND (db) Herschkowitz & Durlach (1969) Sensation level (db)

Interaural Level Differences Perceptual threshold: In the order of 1 db (Hall, 1964 and others) Depends on the sound level (Herschkowitz & Durlach, 1969) (Francart and Wouters, 2007) ILD JND (db) Small dependency on frequency Francart and Wouters (2007) and frequency shift between the two ears Frequency shift

Interaural Level Differences Perceptual threshold: In the order of 1 db (Hall, 1964 and others) Depends on the sound level (Herschkowitz & Durlach, 1969) (Francart and Wouters, 2007) Depends on the lateral sound position (Bernstein, 2004) ILD JND (db) Small dependency on frequency Francart and Wouters (2007) and frequency shift between the two ears Frequency shift

Sound Localization in the Horizontal Plane Interaural level differences (ILDs) Left: AL Interaural time differences (ITDs) Right: AR ITD 0 ITD

Interaural Time Differences ITD= r sin c (Woodworth & Schlosberg, 1962) Begault (2001)

Interaural Time Differences ITD= r sin c r re : (Woodworth & Schlosberg, 1962) r e =0.51 x 1 0.18 x3 0.032 (Algazi et al., 2001) Algazi et al. (2001)

Interaural Time Differences Estimated ITD (ms) Azimuth (deg) Physical range: +/- 800 µs Time (ms) Azimuth (deg)

Interaural Time Differences Perceptual threshold: Best conditions: In the order of 10 µs ITD (µs) Zwislocki & Feldman, 1956 Frequency (Hz)

Interaural Time Differences Perceptual threshold: Best conditions: In the order of 10 µs Pure tones: depends on frequency Ambiguity of the ongoing ITD in the phase Percent Correct (%) Refractory time of the neurons Henning (1974) ITD (µs)

Interaural Time Differences Perceptual threshold in complex signals: Pure-tone ITD ITD in the fine structure ITDENV 0 ITDFS 0 Percent Correct (%) Henning (1974) ITD (µs)

Interaural Time Differences Perceptual threshold in complex signals: Pure-tone ITD ITD in the fine structure Modulations: ITD in the envelope ITDENV 0 ITDFS 0 Percent Correct (%) Henning (1974) ITD (µs)

Interaural Time Differences Lateralization based on ITD: Low-frequency (pure-tone) ITD: strong cue Sidedness High-frequency (envelope) ITD: weaker cue Bernstein (2001) ITD (µs)

Localization Cues for the Horizontal Plane ILD (broadband)? ITD (broadband)? Envelope ITD (high frequencies)? Spectral cues? Interaural spectral differences? Monaural cues? Duplex theory (Rayleigh 1907 & others) Low frequency range: ITDs High frequency range: ILDs Does the duplex theory still hold?

Duplex Theory Revisited Macpherson & Middlebrooks (2002) ILD weight: 0.52 (broadband); 0.24 (low-pass); 0.82 (high-pass) ITD weight: 0.82 (broadband); 0.88 (low-pass); 0.24 (high-pass) Envelope ITD weight (broadband): dep. on onset and modulation Interaural spectral difference weight: Same as broadband ILD Monaural (near-ear) spectrum weight: 0.03 (broadband); 0.03 (high-pass)

Localization Cues for the Horizontal Plane High-frequency ILD Low-frequency ITD Onset-ITD and ITD in the ongoing modulation Spectral information not relevant (neither monaural nor binaural) Valid for the lateral dimension in horizontal plane only! (Rayleigh 1876)

Cone of Confusion ITD-based front-back ambiguity: Modeled Measured Whigthman & Kistler (1997) Can be resolved with the help of spectral cues

Cone of Confusion Left panel: Right panel: Scrambled spectrum Results: More front-back confusions Response Angle (Deg) Flat spectrum Larger elevation error Whigthman & Kistler (1997) Target Angle (Deg)

Perceptually-Relevant Coordinate System Geodetic coordinate system: Azimuth & Elevation Horizontal-polar coordinate system: Lateral angle: binaural disperity Polar angle: spectral cues Spectral cue Elevation Azimuth Binaural disperity

Spectral Cues Head-related transfer functions (HRTFs) Describe the filtering effect of the head, torso, pinna Depend on the position of the sound source Time-domain: Head-related impulse responses (HRIRs) Williams (2002)

HRTF Measurement Open ear canal Wightman & Kistler (1996) Wightman & Kistler (1996

HRTF Measurement Closed ear canal: simpler (less variability) Probe microphone MØLLER et al (1995) Electret capsule

HRTF Measurement System identification for many positions

HRTFs Polar angle (deg) Polar angle (deg) In the median (mid-sagittal) plane Frequency (khz) Frequency (khz)

Directional Transfer Functions (DTFs) Model for HRTFs: H f =C f D f N log H i f i=1 Polar angle (deg) Polar angle (deg) Consider directional cues only 1 C f = N Frequency (khz) Frequency (khz)

Phase Spectrum? Model for the HRTFs: H f = H ap f H min f H ap f =e i f H min f = H f e i All-pass filter (delay ITD) ap min f Minimum phase system Kulkarni et al. (1999)

Phase Spectrum Kistler & Wightman (1992) Perceptually not relevant Wightman & Kistler (1992) Kulkarni et al. (1999) Macpherson & Middlebrooks (2002) Hartmann et al. (2010) Target-Response Correlation Left/Right Front/Back Up/Down Original phase Min.phase+delay

Signal Synthesis for Virtual Acoustics Real Virtual Apply ITD (if DTFs modeled by min. phase+delay) Present the binaural signals via headphones Response Angle (Deg) Filter signal with corresponding pair of DTFs Whigthman & Kistler (1997) Target Angle (Deg)

Signal Synthesis for Virtual Acoustics Middlebrooks (1999) So: Generic HRTFs for all?

Subject-Dependency of HRTFs HRTFs depend on the anthropometry HRTFs of 27 subjects Sorted by the first notch The same position Algazi et al (2001) Algazi et al (2001)

Subject-Dependency of HRTFs Localization with others' ears? Less externalization More front-back confusions Effect depends on subject compatibility: differences in anthropometry differences in HRTFs Middlebrooks (1999)

Plasticity in Sound Localization Ability to recalibrate the auditory system Pre-test Post-test Hofman et al (1998) Hofman et al (1998)

Plasticity in Sound Localization: Supervised Training (Walder, Laback, Majdak, 2010) Training (days) Pre-tests Post-tests

Further Factors Affecting Sound Localization in Free Field Head movements: Help to resolve front-back confusions (Perret & Noble, 1997) Blauert (1974)

Further Factors Affecting Sound Localization in Free Field Head movements: Help to resolve front-back confusions (Perret & Noble, 1997) Vision: Visual feedback providing consistent information about the environment helps (Majdak et al. 2010)

Further Factors Affecting Sound Localization in Free Field Head movements: Help to resolve front-back confusions (Perret & Noble, 1997) Vision: Visual feedback providing consistent information about the environment helps (Majdak et al. 2010) Experience: Training on localization using own HRTFs helps (Majdak et al. 2010)

Summary Sound localization (single source, no room): Lateral positions: Relevant cues: Binaural and broadband Spectral shape negligible Easily derived from anthropometry Vertical positions (also front vs. back): Relevant cues: Monaural spectral shape Complex relation to the anthropometry Individualized HRTFs required Generic HRTFs: problem not solved yet Long-term recalibration to modified HRTFs possible A way towards an optimized generic HRTF?

Summary Sound localization (single source, no room): Lateral positions: Relevant cues: Binaural and broadband Spectral shape negligible Easily derived from anthropometry Vertical positions (also front vs. back): Relevant cues: Monaural spectral shape Complex relation to the anthropometry Individualized HRTFs required Generic HRTFs: problem not solved yet Long-term recalibration to modified HRTFs possible A way towards an optimized generic HRTF?

Summary Sound localization (single source, no room): Lateral positions: Relevant cues: Binaural and broadband Spectral shape negligible Easily derived from anthropometry Vertical positions (also front vs. back): Relevant cues: Monaural spectral shape Complex relation to the anthropometry Individualized HRTFs required Generic HRTFs: problem not solved yet Long-term recalibration to modified HRTFs possible A way towards an optimized generic HRTF?