ORIENTATION IN SIMPLE VIRTUAL AUDITORY SPACE CREATED WITH MEASURED HRTF

Similar documents
Virtual Acoustic Space as Assistive Technology

A Virtual Audio Environment for Testing Dummy- Head HRTFs modeling Real Life Situations

Spatial Audio Reproduction: Towards Individualized Binaural Sound

HRIR Customization in the Median Plane via Principal Components Analysis

Acoustics Research Institute

BINAURAL RECORDING SYSTEM AND SOUND MAP OF MALAGA

Sound Source Localization using HRTF database

Enhancing 3D Audio Using Blind Bandwidth Extension

Convention Paper 9870 Presented at the 143 rd Convention 2017 October 18 21, New York, NY, USA

Auditory Localization

Study on method of estimating direct arrival using monaural modulation sp. Author(s)Ando, Masaru; Morikawa, Daisuke; Uno

Audio Engineering Society. Convention Paper. Presented at the 131st Convention 2011 October New York, NY, USA

PERSONALIZED HEAD RELATED TRANSFER FUNCTION MEASUREMENT AND VERIFICATION THROUGH SOUND LOCALIZATION RESOLUTION

Proceedings of Meetings on Acoustics

Upper hemisphere sound localization using head-related transfer functions in the median plane and interaural differences

Proceedings of Meetings on Acoustics

Validation of lateral fraction results in room acoustic measurements

Convention Paper 9712 Presented at the 142 nd Convention 2017 May 20 23, Berlin, Germany

Convention Paper Presented at the 139th Convention 2015 October 29 November 1 New York, USA

Externalization in binaural synthesis: effects of recording environment and measurement procedure

Aalborg Universitet. Binaural Technique Hammershøi, Dorte; Møller, Henrik. Published in: Communication Acoustics. Publication date: 2005

Computational Perception. Sound localization 2

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 A MODEL OF THE HEAD-RELATED TRANSFER FUNCTION BASED ON SPECTRAL CUES

PAPER Enhanced Vertical Perception through Head-Related Impulse Response Customization Based on Pinna Response Tuning in the Median Plane

Computational Perception /785

3D sound image control by individualized parametric head-related transfer functions

Proceedings of Meetings on Acoustics

Improved Head Related Transfer Function Generation and Testing for Acoustic Virtual Reality Development

Convention e-brief 433

Binaural Hearing. Reading: Yost Ch. 12

III. Publication III. c 2005 Toni Hirvonen.

WAVELET-BASED SPECTRAL SMOOTHING FOR HEAD-RELATED TRANSFER FUNCTION FILTER DESIGN

A triangulation method for determining the perceptual center of the head for auditory stimuli

ANALYZING NOTCH PATTERNS OF HEAD RELATED TRANSFER FUNCTIONS IN CIPIC AND SYMARE DATABASES. M. Shahnawaz, L. Bianchi, A. Sarti, S.

INVESTIGATING BINAURAL LOCALISATION ABILITIES FOR PROPOSING A STANDARDISED TESTING ENVIRONMENT FOR BINAURAL SYSTEMS

Improving room acoustics at low frequencies with multiple loudspeakers and time based room correction

Sound localization Sound localization in audio-based games for visually impaired children

URBANA-CHAMPAIGN. CS 498PS Audio Computing Lab. 3D and Virtual Sound. Paris Smaragdis. paris.cs.illinois.

Analysis of Frontal Localization in Double Layered Loudspeaker Array System

Psychoacoustic Cues in Room Size Perception

VIRTUAL ACOUSTICS: OPPORTUNITIES AND LIMITS OF SPATIAL SOUND REPRODUCTION

Testing of Objective Audio Quality Assessment Models on Archive Recordings Artifacts

THE USE OF VOLUME VELOCITY SOURCE IN TRANSFER MEASUREMENTS

Modeling Head-Related Transfer Functions Based on Pinna Anthropometry

Virtual Reality Presentation of Loudspeaker Stereo Recordings

HRTF measurement on KEMAR manikin

THE MATLAB IMPLEMENTATION OF BINAURAL PROCESSING MODEL SIMULATING LATERAL POSITION OF TONES WITH INTERAURAL TIME DIFFERENCES

Introduction. 1.1 Surround sound

Measuring impulse responses containing complete spatial information ABSTRACT

Spatial Audio & The Vestibular System!

Spatial audio is a field that

SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4

Ivan Tashev Microsoft Research

HRTF adaptation and pattern learning

A binaural auditory model and applications to spatial sound evaluation

Comparison of binaural microphones for externalization of sounds

3D Sound System with Horizontally Arranged Loudspeakers

Listening with Headphones

SIMULATION OF SMALL HEAD-MOVEMENTS ON A VIRTUAL AUDIO DISPLAY USING HEADPHONE PLAYBACK AND HRTF SYNTHESIS. György Wersényi

Acoustical Active Noise Control

3D Sound Simulation over Headphones

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 VIRTUAL AUDIO REPRODUCED IN A HEADREST

Sound source localization and its use in multimedia applications

Application Note: Headphone Electroacoustic Measurements

Finding the Prototype for Stereo Loudspeakers

Proceedings of Meetings on Acoustics

THE INTERACTION BETWEEN HEAD-TRACKER LATENCY, SOURCE DURATION, AND RESPONSE TIME IN THE LOCALIZATION OF VIRTUAL SOUND SOURCES

EFFECTS OF PHYSICAL CONFIGURATIONS ON ANC HEADPHONE PERFORMANCE

On distance dependence of pinna spectral patterns in head-related transfer functions

MEASURING DIRECTIVITIES OF NATURAL SOUND SOURCES WITH A SPHERICAL MICROPHONE ARRAY

TDE-ILD-HRTF-Based 2D Whole-Plane Sound Source Localization Using Only Two Microphones and Source Counting

3D Audio Systems through Stereo Loudspeakers

Creating three dimensions in virtual auditory displays *

The analysis of multi-channel sound reproduction algorithms using HRTF data

PERSONAL 3D AUDIO SYSTEM WITH LOUDSPEAKERS

Live multi-track audio recording

Personalization of head-related transfer functions in the median plane based on the anthropometry of the listener s pinnae a)

The effect of 3D audio and other audio techniques on virtual reality experience

IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES. Q. Meng, D. Sen, S. Wang and L. Hayes

METHOD OF ESTIMATING DIRECTION OF ARRIVAL OF SOUND SOURCE FOR MONAURAL HEARING BASED ON TEMPORAL MODULATION PERCEPTION

Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA. Why Ambisonics Does Work

APPLICATION OF THE HEAD RELATED TRANSFER FUNCTIONS IN ROOM ACOUSTICS DESIGN USING BEAMFORMING

Method of acoustical estimation of an auditorium

Proceedings of Meetings on Acoustics

Robotic Spatial Sound Localization and Its 3-D Sound Human Interface

Binaural hearing. Prof. Dan Tollin on the Hearing Throne, Oldenburg Hearing Garden

Selection of Microphones for Diffusion Measurement Method

Convention Paper Presented at the 130th Convention 2011 May London, UK

EFFECT OF ARTIFICIAL MOUTH SIZE ON SPEECH TRANSMISSION INDEX. Ken Stewart and Densil Cabrera

PSYCHOACOUSTIC EVALUATION OF DIFFERENT METHODS FOR CREATING INDIVIDUALIZED, HEADPHONE-PRESENTED VAS FROM B-FORMAT RIRS

Tara J. Martin Boston University Hearing Research Center, 677 Beacon Street, Boston, Massachusetts 02215

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Loudness & Temporal resolution

Two-channel Separation of Speech Using Direction-of-arrival Estimation And Sinusoids Plus Transients Modeling

24. TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November Alexander Lindau*, Stefan Weinzierl*

Intensity Discrimination and Binaural Interaction

29th TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November 2016

Evaluation of a new stereophonic reproduction method with moving sweet spot using a binaural localization model

Influence of artificial mouth s directivity in determining Speech Transmission Index

Interpolation of Head-Related Transfer Functions

Transcription:

ORIENTATION IN SIMPLE VIRTUAL AUDITORY SPACE CREATED WITH MEASURED HRTF F. Rund, D. Štorek, O. Glaser, M. Barda Faculty of Electrical Engineering Czech Technical University in Prague, Prague, Czech Republic Abstract Using own measured HRTF for each user is considered as the best way to obtain a highquality perception of sound source virtual position. Without proper equipment the measuring process is very time consuming, so we decided to use the virtual auditory space with small resolution only at first. In this work, we got the HRTF in 15 positions for 7 different subjects with using MLS signal. Then a series of virtually positioned sound was created using MATLAB algorithm. Necessity of equalization the room, loudspeaker and headphones for later listening is also discussed. Finally, we made a test sequence fixed to verify orientation in virtual acoustic space. Each subject had to listen to the sequence and mark positions in questionnaire, where he considers the sound source is located. Unknown to subject each sequence was doubled for estimation whether the subject is sure about sound source position or guesses. Wideband noise bursts modulated by low-frequency sine were chosen as stimulus. 1 Introduction Creating the three-dimensional sound for entertainment, commercial, and scientific systems by using HRTF is well-known nowadays. There have been many researches due to reach the best quality in virtual sound source positioning. HRTF is a complex function, which captures the spectral changes of signal that occurs when a sound wave propagates from sound source to the listener s outer ear. The listener evaluates the spectral content and latency of both signals from left and right ear, and according to this estimates the location of sound source. Head-Related Transfer Function depends on frequency, azimuth, elevation, range and it also significantly varies from person-to-person [2]. We can write it down as H = f(φ, θ, ω, r, subject). HRTF can be transferred through inverse Fourier Transformation into time domain, where it is represented as HRIR (Head Related Impulse Response), which represents the impulse response of path between sound source and entrance to the left (right) ear. If we know the HRIR for both channels, we can create a stereo signal from monaural source as implied for left channel by Eq. (1). It is necessary to keep both channels separate, therefore headphones have to be used. y L ( t) HRIR ( t) x( t) (1) HRTF is subjectively depended, so in general everyone must have own set of HRTF from required directions. There are two basic ways how to get personal HRTF. At first, we can measure a whole set. A disadvantage is that we need a net of measured points with sufficient density, so it requires a lot of time spending with measuring process. However, this approach provides the best results in final perception of sound. Other way is to create a mathematical model according to simplemeasured anthropometrical parameters. Nowadays, models are constantly improved with good results, but in terms of sound source perception quality is worse than measured set. In this work we got HRTF in 15 positions (5 azimuth and 3 elevation levels) for 7 different subjects. Measurement points were selected only in frontal area with 45 degrees step in horizontal and median plane from the center of view (θ = 0, φ = 0 ), as shown on Fig. 1. As the first step, the positions are enough for basic resolution in virtual auditory space, because we can combine main directions of right / left and up / down. L

Figure 1: Measured points only for frontal area 2 Measurement of HRTF set In order to get a proper HRIR set, we arranged a simple measuring set, which consists of tiltable loudspeaker mounted on extensible stand, swivel chair with calibrated pointer, two small microphones, sound card, amplifier, and measuring software. For particular types see Table 1. In best way all measurements of HRTS should be done in anechoic chamber, but we considered available baffled studio sufficient for our experiment. Each subject was seated on the swivel chair, and both microphones were attached by plaster on the beginning of ear canal. For reducing the influence of canal cavity we used medical earplugs. At the beginning we asked the subject not to move and look straight ahead. After measuring process subject turns for another 45 degrees. This was done for 3 elevations of loudspeaker. EASERA software using MLS measuring signal [3] was used for obtaining HRIR. The data was stored as stereo wav files and then processed in MATLAB 2006a version. It was necessary to check microphone position permanently and keep the system balanced, because even small position change caused incorrect microphone gain, what makes well perceptible offset in final virtual sound positioning. Measuring time for one subject was app. 40 minutes. Figure 2: Measuring of HRTF a) whole measuring set b) detail on microphone attachment. Table 1: MEASURING EQUIPMENT USED DURING EXPERIMENT Microphones: 2x Sennheiser MKE 2 Gold Loudspeaker: wideband TVM ARZ 6608 Amplifier: Brüel & Kjær 2706 Sound card: Fireface 400 Meas. software: EASERA Processing: Matlab 2006a

3 Equalization of measured HRTF When we obtain HRIR, it includes even room response for measuring MLS signal, so HRIR is actually multiplied by strongly attenuated but still distinct reflections. We designed the in-room measurement for first reflection coming from the floor. It is necessary to eliminate all samples of HRIR after first reflection comes. That causes explicit HRIR distortion, but this variant is more accurate than the one with reflections included. There are different times of the reflection arrival, but we were operating only with average. It finally led to app. 8.5 ms duration of HRIR (from beginning) what is about 820 samples in using 96 khz sampling frequency as shown on Eq. (2). reflected _ sound _ path HRIR _ length sampling _ frequency (2) sound _ velocity Another adjustment is then needed to compensate influence of loudspeaker, which distorts flat spectra of measuring MLS signal. The same compensation is also needed for headphones transfer function compensation, because this characteristic distorts HRTF too. In general, we have to compensate all elements presented during measurement and binaural listening between sound source and listener. All these adjustments were made in frequency domain by dividing Fourier transformation of measured HRIR with appropriate inverse transfer function. HRIR_L = ifft(fft(hrir_l)./fft(headphones_ir_l)./fft(loudspeaker_l)); HRIR_R = ifft(fft(hrir_r)./fft(headphones_ir_r)./fft(loudspeaker_r)); Last step in creating virtually positioned sound is convolution between input signal and both appropriate adjusted HRIRs for left and right channel. After it both monaural signals are put together into one stereo wav file. signal_l = conv(sum,hrir_l); signal_r = conv(sum,hrir_r); signal = [signal_r ; signal_l]'; wavwrite(signal1, 44100, 16, ['signal_binaural.wav']) Whole procedure of creating virtually positioned sound [out] from monaural source [x(t)] is depicted on Fig. 3. Figure 3: Scheme of HRTF equalization and generating stimuli

4 Stimuli creating A wideband stimulus is considered as the most suitable for the best results in sound source position perception, because every frequency band is involved by different principle of localization cues [1]. Using a wideband stimulus we combine all localization mechanisms, so the final perception is supposed to be more specified. In [6] is used a White Gaussian Noise modulated (WGN) by sin(40hz). Modulation causes constant amount of leading edges, which also improves sound source localization. Stimulus used in this experiment is depicted on Fig. 4. Figure 4: Modulated noise stimulus in time domain After filtration with HRTF resp. convolution with HRIR the output signal spectra (both channels) is uniquely shaped according to appropriate HRTF for desired direction. Spectral notches and peaks can be seen on characteristic frequency bands. Filtration and shaping of narrowband spectra of WGN stimulus for one channel and one direction is shown on Fig 5. Figure 5: Changing spectral parameters of noise-like stimulus using HRTF 5 Subjective test of sound source location perception For final testing of perception a sequence of positioned WGN wav files was created. The sound source was virtually positioned (ideally) into the same locations, where all 15 HRIRs were measured. We tried to verify whether compensation of headphones and loudspeaker is really needed and how it actually affects the final perception. All sequenses were made in 3 variants: measured HRTF, compensated loudspeaker, and compensated headphones and loudspeaker. We used AKG K 55 headphones for this test. The Virtual Auditory Space we used can be represented as shown on Fig. 6. Every subject firstly listened to the tutorial sequence, which went through all 15 points in order: A1, A2, A3, B1, B2, and so on. We considered this important, because informal test shown that first contact with virtually positioned sound can make the subject confused. On this tutorial sequence the subject was allowed to set the volume to feel comfortable.

The final orientation test consists of 30 virtually positioned sounds. Every sound was introduced by 1 khz tone non-positioned signalization beep and three times repeated. After that the subject had 6 seconds to fill a gap on questionnaire with number of a sample. Each following sample in sequence had to differ at least in one step in elevation and one in azimuth for bigger subjective sound source movement. After first 15 samples, when each position occurred, the whole sequence was repeated without subject s knowledge. Comparing results in both same sequences tells more whether the subject guesses or is sure about virtual sound source position. Figure 5: Virtual auditory space scheme rear view Results of this task were not as good as we predicted. We thought about almost 100% accuracy because of quite big distances between measuring points, but it was only 10 46%. In Table 2 we can see factual information. During the tests we took notice of very strong sensitivity on microphone gain offset. It is necessary to take care of gain balance in the beginning of measurement in C position both microphones must be symmetric fixed, because even app. 3-4 mm deviation causes as many as 20 error in localization in azimuth plane. All subjects were able to distinguish side of incoming sound, but results in median plane were not so precise. Columns azim. and elev. in Table 2 show RMSE for every subject for both planes. Values are related to step of 45. In sequence 2 and 3 (with compensations) an externalization effect [4] is perceptible when we hear the virtual sound source out of the head, so it makes the source more real even the stimulus is only wideband noise. Subjects DS and MB, who were also authors, show better results although they didn t know the sequence order in advance. We think it is important to get used to virtual positioning and sound character first for improving orientation in virtual acoustic space. Table 2: RESULTS OF ORIENTATION IN VIRTUAL AUDITORY SPACE Sequence 1 Sequence 2 Sequence 3 Subject azim. elev. correct azim. elev. correct azim. elev. correct SM 0,66 1,15 4 0,73 1,11 6 0,75 1,06 6 MB 0,71 0,88 10 0,58 1,05 7 0,71 1,24 4 PS 0,98 0,79 6 0,63 1,18 3 1,12 0,89 5 TS 0,86 1,11 5 0,93 1,13 3 1,05 0,91 8 BK 0,88 1,03 6 0,73 1,24 6 1,62 1,22 4 DS 0,48 0,84 10 0,48 0,58 14 0,58 0,73 11 average 0,76 0,97 6,8 0,68 1,05 6,5 0,97 1,01 6,3

6 Evaluation of microphone in-ear position influence Microphone technical documentation declares spherical directional characteristic. We wanted to verify, if or how measured HRTF is changed according to microphone position [2]. The measurement was done for position C2, which means the sound was situated directly in front of the manikin we used. For real application on live subjects we used position (a), because this one is most comfortable in light of microphone attachment on subject s body. Other two positions were suggested with focus on keeping microphone centre on the same place. All three variants are shown on Fig. 7. Figure 7: Three different types of microphone attachment The transfer function does not vary significantly and keeps all trends same up to 10 khz for all three microphone positions. Above this frequency, variants (a) and (b) show almost the same behavior, but (c) shows over 10 db higher notch on approximately 16 khz. This frequency area is important especially for elevation cues. Three position-dependent HRTFs are shown on Fig. 8. Figure 8: HRTF measured for three types of microphone attachment There is also a question of non-easy connection between actual shape of Head Related Transfer Function and perception of sound adjusted by it. Every frequency band is involved by various principles and interaction of sound and subject s body, resp. torso, head and pinna, and also has a unique role in localization for different directions. For our purposes, small deviations in HRTF behavior were neglected.

7 Results Head Related Transfer Function was measured for 15 locations on 7 subjects. HRTF is very sensitive for any gain offsets, so the system configuration has to be permanently checked. After that test sequence of noisy-like stimulus using unique HRTF set for each subject was made in order to verify orientation in Virtual Auditory Space. Differences in perception for three types of equalization were tested (none, loudspeaker, loudspeaker + headphones). Compensation brings an externalization effect, which moves the sound source perception out of subject s head, so it makes the source more real. It also improves the resolution in virtual space, but in out experiment it was not proved because of unwanted gain offset. After that verification of microphone position influence was done. There were as a matter of fact the same trends of HRTF behavior up to 10 khz, but above this level higher frequencies were slightly attenuated for one of microphone position. Other two variants were almost identical, so we finally neglected the microphone position influence. Final results of this experiment were not sufficient, because we expected much more precise orientation with certain position determination. Now we want to extend this experiment by more precise and dense measuring of HRTF with using a head-tracking system, because possibility to make head movements during localization which shift the sound source position improves subject s estimation [4]. Also possibility of learning-to-hear virtual positioned sound, as mentioned in section 5, is desirable to verify. This research is aimed to develop interfaces of assistive technologies (image sonification, virtual navigation ) for visually impaired. Acknowledgements The project "Orientation in Simple Virtual Auditory Space Created with Measured HRTF" was supported by the Grant Agency of the Czech Technical University in Prague, grant No. SGS10/082/OHK3/1T/13. References [1] Wenzel, E. M., Arruda, M., Kistler, D. J., Wightman, F. L., Localization Using Nonidividualized Head-Related Transfer Functions, J. Acoust. Soc. Am., vol. 94, pp. 111-123 (July 1993) [2] Algazi, R., Aveando, C., Thomson, D., Dependence of Subject and Measurement Position in Binaural Signal Acquisition, J. Audio. Eng. Soc., vol. 47, no 11, pp. 937-947, Nov. 1999 [3] Kadlec, F., Zpracování akustického signálu, ČVUT FEL, Praha 2002 [4] Wersényi, G., Localization a HRTF-based Minimum-Audiable-Angle Listening test for GUIB applications, in Electronic Journal «Technical Acoustics», 2007 [5] Susnik, R., Sodnik, J., Tomazic, S., Measurements of Auditory Navigation in Virtual Acoustic Space, Perceptual Interfaces and Reality Laboratory, UMIACS, University of Lubljana, Slovenia, 2004 [6] Susnik, R., Sodnik, J., Tomazic, S., Sound Source Choice in HRTF Acoustic Imaging, University of Lubljana, Slovenia, 2003 Ing. František Rund, Ph.D. Department of Radioelectronics, FEE, CTU in Prague, Technická 2, 166 27 Praha 6, Czech Republic email:xrund@fel.cvut.cz Ing. Dominik Štorek Department of Radioelectronics, FEE, CTU in Prague, Technická 2, 166 27 Praha 6, Czech Republic, email: storedom@fel.cvut.cz