Proceedings of Meetings on Acoustics

Similar documents
ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

Ivan Tashev Microsoft Research

SPHERICAL MICROPHONE ARRAY BASED IMMERSIVE AUDIO SCENE RENDERING. Adam M. O Donovan, Dmitry N. Zotkin, Ramani Duraiswami

29th TONMEISTERTAGUNG VDT INTERNATIONAL CONVENTION, November 2016

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics

INVESTIGATING BINAURAL LOCALISATION ABILITIES FOR PROPOSING A STANDARDISED TESTING ENVIRONMENT FOR BINAURAL SYSTEMS

From Binaural Technology to Virtual Reality

A Toolkit for Customizing the ambix Ambisonics-to- Binaural Renderer

Measuring impulse responses containing complete spatial information ABSTRACT

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics

Ambisonics Directional Room Impulse Response as a New SOFA Convention

Abstract. 1. Introduction and Motivation. 3. Methods. 2. Related Work Omni Directional Stereo Imaging

From acoustic simulation to virtual auditory displays

Validation of lateral fraction results in room acoustic measurements

Spatialization and Timbre for Effective Auditory Graphing

Proceedings of Meetings on Acoustics

SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4

6-channel recording/reproduction system for 3-dimensional auralization of sound fields

Listening with Headphones

Visualization of Compact Microphone Array Room Impulse Responses

Sound Source Localization using HRTF database

Psychoacoustic Cues in Room Size Perception

Introduction. 1.1 Surround sound

Recent Advances in Acoustic Signal Extraction and Dereverberation

The psychoacoustics of reverberation

Binaural Hearing. Reading: Yost Ch. 12

REAL TIME WALKTHROUGH AURALIZATION - THE FIRST YEAR

Robotic Spatial Sound Localization and Its 3-D Sound Human Interface

HRIR Customization in the Median Plane via Principal Components Analysis

Multiple Sound Sources Localization Using Energetic Analysis Method

Institute for Media Technology Electronic Media Technology (ELMT)

Direction-Dependent Physical Modeling of Musical Instruments

Auditory System For a Mobile Robot

AN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES

Sound source localization and its use in multimedia applications

Waves Nx VIRTUAL REALITY AUDIO

Advanced techniques for the determination of sound spatialization in Italian Opera Theatres

RIR Estimation for Synthetic Data Acquisition

Acoustics Research Institute

Spatialisation accuracy of a Virtual Performance System

MANY emerging applications require the ability to render

Binaural room impulse response database acquired from a variable acoustics classroom

Room impulse response measurement with a spherical microphone array, application to room and building acoustics

Accurate sound reproduction from two loudspeakers in a living room

Soundfield Navigation using an Array of Higher-Order Ambisonics Microphones

Proceedings of Meetings on Acoustics

Study Of Sound Source Localization Using Music Method In Real Acoustic Environment

Blind source separation and directional audio synthesis for binaural auralization of multiple sound sources using microphone array recordings

HRTF adaptation and pattern learning

On distance dependence of pinna spectral patterns in head-related transfer functions

Book Chapters. Refereed Journal Publications J11

Proceedings of Meetings on Acoustics

Circumaural transducer arrays for binaural synthesis

DECORRELATION TECHNIQUES FOR THE RENDERING OF APPARENT SOUND SOURCE WIDTH IN 3D AUDIO DISPLAYS. Guillaume Potard, Ian Burnett

Spatial Audio & The Vestibular System!

SGN Audio and Speech Processing

Auditory Localization

SGN Audio and Speech Processing

CAPTURING ANCIENT THEATERS SOUND SIGNATURE USING BEAMFORMING

New acoustical techniques for measuring spatial properties in concert halls

Audio Engineering Society Convention Paper 5449

Monaural and Binaural Speech Separation

A Virtual Audio Environment for Testing Dummy- Head HRTFs modeling Real Life Situations

MPEG-4 Structured Audio Systems

3D AUDIO AR/VR CAPTURE AND REPRODUCTION SETUP FOR AURALIZATION OF SOUNDSCAPES

Enhancing 3D Audio Using Blind Bandwidth Extension

Wave Field Analysis Using Virtual Circular Microphone Arrays

Audio Engineering Society. Convention Paper. Presented at the 131st Convention 2011 October New York, NY, USA

Improving room acoustics at low frequencies with multiple loudspeakers and time based room correction

Room Impulse Response Modeling in the Sub-2kHz Band using 3-D Rectangular Digital Waveguide Mesh

Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA. Why Ambisonics Does Work

IMPLEMENTATION AND APPLICATION OF A BINAURAL HEARING MODEL TO THE OBJECTIVE EVALUATION OF SPATIAL IMPRESSION

Proceedings of Meetings on Acoustics

Microphone Array Design and Beamforming

TARGET SPEECH EXTRACTION IN COCKTAIL PARTY BY COMBINING BEAMFORMING AND BLIND SOURCE SEPARATION

Proceedings of Meetings on Acoustics

SIA Software Company, Inc.

The Human Auditory System

Sound rendering in Interactive Multimodal Systems. Federico Avanzini

Proceedings of Meetings on Acoustics

A Database of Anechoic Microphone Array Measurements of Musical Instruments

Speech Intelligibility Enhancement using Microphone Array via Intra-Vehicular Beamforming

VIRTUAL ACOUSTICS: OPPORTUNITIES AND LIMITS OF SPATIAL SOUND REPRODUCTION

Binaural Sound Source Localization Based on Steered Beamformer with Spherical Scatterer

Automatic Evaluation of Hindustani Learner s SARGAM Practice

APPLICATION OF THE HEAD RELATED TRANSFER FUNCTIONS IN ROOM ACOUSTICS DESIGN USING BEAMFORMING

University of Huddersfield Repository

Potential and Limits of a High-Density Hemispherical Array of Loudspeakers for Spatial Hearing and Auralization Research

396 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 2, FEBRUARY 2011

Subband Analysis of Time Delay Estimation in STFT Domain

Extracting the frequencies of the pinna spectral notches in measured head related impulse responses

Spatial Audio Reproduction: Towards Individualized Binaural Sound

Time-Lapse Panoramas for the Egyptian Heritage

MEASURING DIRECTIVITIES OF NATURAL SOUND SOURCES WITH A SPHERICAL MICROPHONE ARRAY

Validation of a Virtual Sound Environment System for Testing Hearing Aids

EFFECT OF STIMULUS SPEED ERROR ON MEASURED ROOM ACOUSTIC PARAMETERS

Conference Paper Presented at the Conference on Audio for Virtual and Augmented Reality 2018 August 20 22, Redmond, WA, USA

Reducing comb filtering on different musical instruments using time delay estimation

Transcription:

Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Architectural Acoustics Session 1pAAa: Advanced Analysis of Room Acoustics: Looking Beyond ISO 3382 II 1pAAa5. Spectro-temporal analysis, visualization and identification of listening-space measurements Adam E. O'Donovan, Dmitry N. Zotkin and Ramani Duraiswami* *Corresponding author's address: Institute for Advanced Computer Studies, University of Maryland College Park, A. V. Williams Building, College Park, Maryland 20740, ramani@umiacs.umd.edu Measurement of a listening space might be done to characterize it in a gross way, or to identify some deficiencies in the space, which can then be corrected. Alternately, these measurements might be performed to create inputs to an auralization software, which might seek to recreate a virtual listening experience. We propose that it might be possible to perform audio-visual measurements of a listening space that allow the entire response of the listening space to be understood, visualized and recreated. The goal is to understand the response completely in terms of its components: along direction, frequency, early and late characteristics, and finally at the level of the structural elements of the scattering surfaces in the space. Measurements with spherical arrays of cameras and microphones provide measurements that allow the response to be decomposed in the desired fashion. Post-processing software that allows the measurements to be analyzed instantly following the measurement will also be described, and demonstrated. Published by the Acoustical Society of America through the American Institute of Physics 2013 Acoustical Society of America [DOI: 10.1121/1.4800008] Received 22 Jan 2013; published 2 Jun 2013 Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 1

ROOM AND CONCERT HALL ACOUSTICS The perception of sound by human listeners in a listening space, such as a room or a concert hall is a complicated function of the type of source sound (speech, oration, concert, music and type), of the intended outcome (comprehension, comfort, enjoyment, localization), of the room (geometry, reflecting surfaces/absorbers, size) and the listener (HRTF, hearing acuity). Quite obviously, reducing this tremendous variability to a single standard is an extremely hard task. The ISO 3382 standard [1] specifies several performance space acoustic parameters (Part 1) and room acoustics parameters (Part 2) for this. These measurements include varying quantities (decay times, clarity, etc.) extracted in varying frequency bands, for both single microphone and binaural measurements. The binaural measurements attempt to account for the effects of the human head, torso and pinnae on the measurements, but do not account for inter-personal variations. To go beyond this standard, one needs to understand the impulse response for every source-receiver position combination of interest. In the case of binaural listening, this needs to be done in a way that particular individual head-related transfer functions can be employed. MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES To obtain a finer understanding of the listening characteristics, individual impulse responses must be measured for each direction and their structure understood from the geometry and materials used to construct the space. In particular for the early stage, the interaction of the source sound with each scattering surface, in each frequency band can be easily characterized using the VisiSonics AudioVisual Panoramic Camera. This device provides several microphones (64 or 128), and several high definition cameras (5 or 15), which provide an ability to record sound, and to co-register the sound. The device allows 24 bit recordings, at 3 user-selectable gain levels, and 48 khz sampling. All the cameras and microphones are mounted on a head-sized sphere (0.2 m diameter). The use of an advanced digital microphone architecture (Zotkin et al 2013) allows the audio measurements from all cameras and microphones be sent over a combination of IEEE1394 (video) and USB 2 (audio), or over USB 3 (audio and video). A NVIDIA GPU card equipped laptop allows both real time analysis of a scene, and of sophisticated on-scene computations. A panoramic video image is computed from the multiple video cameras by using a GPU based stitching algorithm. An overlaid audio image is computed by employing several thousand plane-wave or spherical beamformers on a grid of directions on the sphere (O Donovan et al, 2007). The magnitude of these beamformers in particular frequency bands are mapped to false color values and overlaid via alpha-blending as a transparent texture over the panoramic video. The raw audio and video data are streamed to disk, and can be used for more careful late analysis. Real-time analysis possible via the video stream includes identification of prominent reflections, and the geometric sources of their origin. A Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 2

sample screenshot with a speaking user in a room is shown below. SPATIAL, TEMPORAL AND SPECTRAL ANALYSES Various off-line analyses can be performed via the tool. These include the computation of directional impulse responses, the analysis of the reverberation in temporal windows, the analysis in spectral windows, the online computation of spectrograms with various weightings and frequency scales, and others. Some preliminary analysis results can be seen in following video posted online. The interface for performing spectro-temporal spatial analysis of the acoustic scene is shown above. Window 1 shows the time domain plot of the signal recorded. A sliding red box allows a particular time interval of interest to be selected. Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 3

Window 2 shows a spectrogram of the selected time region. The spectrogram is a useful tool for finding structure in the recorded audio. It can be used to select a frequency and time sub region to generate an acoustic image. In the following images, screenshots are shown from a recording made in a reverberant chamber. The image below shows the first frame of the acoustic image. It was generated by computing the acoustic images for all frequency bins within the user selected frequency range of a 256 sample frame of data. For each successive image the sample frame is advanced by 10 samples and the corresponding image is generated. This sequence can then be played to explore the dynamics of the instantaneous spatial distribution of the incident sound field. Notice the thin red line in the spectrogram. This line indicates the extent of the data used to generate the image in the current frame. The image above shows when the impulse onset has occurred; the spatial distribution of the acoustic field is very localized to the driver creating the signal, and a bulls-eye pattern is seen. Because the analysis frame time can be selected accurately, we can advance the frames, until the first reflection is distinctly seen. In this case the image above shows that it is on the leftmost wall of the reverberant chamber. The corresponding bulls-eye pattern of the reflection is seen clearly in the image below, as is the temporal pattern of the Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 4

corresponding reflection. In the image above, the frame has been advanced to show the second distinct reflection occurring on the center wall. Advancing 100 ms (the image above) shows just how diffuse the field has already gotten in this reverberant chamber. CONCLUSIONS AND ONGOING WORK The images above just indicate a small selection of analyses possible. Users of the device have used it to quickly identify problematic structures in rooms. It is also possible to compute integrated quantities over time and frequency. Further, using known sources we can compute impulse responses. In a version of the tool under development, we can compute impulse responses corresponding to any direction. Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 5

REFERENCES ISO-3382, Acoustics - Measurement of room acoustic parameters -- Part 1: Performance spaces (2009) ISO-3382, Acoustics - Measurement of room acoustic parameters -- Part 2: Reverberation time in ordinary rooms (2008) A. O Donovan, R. Duraiswami, Jan Neumann (2007). Microphone Arrays as Generalized Cameras for Integrated Audio Visual Processing. IEEE CVPR. A. O'Donovan, D. N. Zotkin and R. Duraiswami (2008). A spherical microphone array based system for immersive audio scene rendering. Proc. 14th International Conference on Auditory Displays. Paris, France. A. O'Donovan, R. Duraiswami and D.N. Zotkin. (2010). Automatic matched filter recovery via the audio camera. Proc. IEEE ICASSP, (pp. 2826-2829). Dallas, TX. A. O Donovan, R. Duraiswami and D.N. Zotkin. (2008). Imaging Concert Hall Acoustics using Audio and Visual Cameras. IEEE ICASSP, (pp. 5284-5287). Las Vegas, NV. A. O'Donovan, R. Duraiswami and Nail A. Gumerov. (2007). Real Time Capture of Audio Images and Their Use with Video. IEEE WASPAA, (pp. 1-8). Proceedings of Meetings on Acoustics, Vol. 19, 015021 (2013) Page 6