SOPA version 2. Revised July SOPA project. September 21, Introduction 2. 2 Basic concept 3. 3 Capturing spatial audio 4

Similar documents
SOPA version 3. SOPA project. July 22, Principle Introduction Direction of propagation Speed of propagation...

Sound source localization and its use in multimedia applications

Physics 115 Lecture 13. Fourier Analysis February 22, 2018

Sound Processing Technologies for Realistic Sensations in Teleworking

Auditory Localization

Analysis of Frontal Localization in Double Layered Loudspeaker Array System

Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA. Why Ambisonics Does Work

Predicting localization accuracy for stereophonic downmixes in Wave Field Synthesis

TE 302 DISCRETE SIGNALS AND SYSTEMS. Chapter 1: INTRODUCTION

Binaural Hearing. Reading: Yost Ch. 12

Introduction. 1.1 Surround sound

University of Huddersfield Repository

Timbral Distortion in Inverse FFT Synthesis

Spatial Audio Transmission Technology for Multi-point Mobile Voice Chat

Proceedings of Meetings on Acoustics

Lecture 3 Complex Exponential Signals

Final Exam Study Guide: Introduction to Computer Music Course Staff April 24, 2015

Pulse Code Modulation (PCM)

Multi-Loudspeaker Reproduction: Surround Sound

Laboratory Assignment 5 Amplitude Modulation

The Discrete Fourier Transform. Claudia Feregrino-Uribe, Alicia Morales-Reyes Original material: Dr. René Cumplido

ELEC3242 Communications Engineering Laboratory Amplitude Modulation (AM)

B360 Ambisonics Encoder. User Guide

PERSONAL 3D AUDIO SYSTEM WITH LOUDSPEAKERS

Speech Compression. Application Scenarios

Surround: The Current Technological Situation. David Griesinger Lexicon 3 Oak Park Bedford, MA

Spatial Audio & The Vestibular System!

Lecture 7 Frequency Modulation

Signal Characteristics

Computational Perception /785

Introduction to signals and systems

ROOM AND CONCERT HALL ACOUSTICS MEASUREMENTS USING ARRAYS OF CAMERAS AND MICROPHONES

The psychoacoustics of reverberation

Spatial Audio Reproduction: Towards Individualized Binaural Sound

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 VIRTUAL AUDIO REPRODUCED IN A HEADREST

Spatial audio is a field that

CMPT 368: Lecture 4 Amplitude Modulation (AM) Synthesis

Measuring impulse responses containing complete spatial information ABSTRACT

Virtual Acoustic Space as Assistive Technology

Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau

Sound Source Localization using HRTF database

Interference in stimuli employed to assess masking by substitution. Bernt Christian Skottun. Ullevaalsalleen 4C Oslo. Norway

The quality of the transmission signal The characteristics of the transmission medium. Some type of transmission medium is required for transmission:

Enhancing 3D Audio Using Blind Bandwidth Extension

Dr. Cahit Karakuş ANALOG SİNYALLER

8.2 IMAGE PROCESSING VERSUS IMAGE ANALYSIS Image processing: The collection of routines and

Outline. Context. Aim of our projects. Framework

Modulation. Digital Data Transmission. COMP476 Networked Computer Systems. Analog and Digital Signals. Analog and Digital Examples.

Spectrum Analysis: The FFT Display

Proceedings of Meetings on Acoustics

A spatial squeezing approach to ambisonic audio compression

PHYC 500: Introduction to LabView. Exercise 9 (v 1.1) Spectral content of waveforms. M.P. Hasselbeck, University of New Mexico

Laboratory Assignment 4. Fourier Sound Synthesis

5.1 Graphing Sine and Cosine Functions.notebook. Chapter 5: Trigonometric Functions and Graphs

Study on method of estimating direct arrival using monaural modulation sp. Author(s)Ando, Masaru; Morikawa, Daisuke; Uno

Advanced Audiovisual Processing Expected Background

Three-dimensional sound field simulation using the immersive auditory display system Sound Cask for stage acoustics

WHY BOTHER WITH STEREO?

Structure of Speech. Physical acoustics Time-domain representation Frequency domain representation Sound shaping

8A. ANALYSIS OF COMPLEX SOUNDS. Amplitude, loudness, and decibels

Fundamentals of Music Technology

Digital Signal Processing. VO Embedded Systems Engineering Armin Wasicek WS 2009/10

Spectrum. Additive Synthesis. Additive Synthesis Caveat. Music 270a: Modulation

Direction-Dependent Physical Modeling of Musical Instruments

Audio Engineering Society Convention Paper Presented at the 110th Convention 2001 May Amsterdam, The Netherlands

Reducing comb filtering on different musical instruments using time delay estimation

6-channel recording/reproduction system for 3-dimensional auralization of sound fields

Signals A Preliminary Discussion EE442 Analog & Digital Communication Systems Lecture 2

Sound Waves and Beats

The analysis of multi-channel sound reproduction algorithms using HRTF data

Lecture 2: SIGNALS. 1 st semester By: Elham Sunbu

Acoustic Communication System Using Mobile Terminal Microphones

Notes on Fourier transforms

Improving room acoustics at low frequencies with multiple loudspeakers and time based room correction

Virtual Sound Source Positioning and Mixing in 5.1 Implementation on the Real-Time System Genesis

Since the advent of the sine wave oscillator

Blind source separation and directional audio synthesis for binaural auralization of multiple sound sources using microphone array recordings

Chapter 4. Digital Audio Representation CS 3570

YEDITEPE UNIVERSITY ENGINEERING FACULTY COMMUNICATION SYSTEMS LABORATORY EE 354 COMMUNICATION SYSTEMS

Interpolation Error in Waveform Table Lookup

Audio /Video Signal Processing. Lecture 1, Organisation, A/D conversion, Sampling Gerald Schuller, TU Ilmenau

EC 6501 DIGITAL COMMUNICATION UNIT - II PART A

A Database of Anechoic Microphone Array Measurements of Musical Instruments

Influence of artificial mouth s directivity in determining Speech Transmission Index

German Jordanian University Department of Communication Engineering Digital Communication Systems Lab. CME 313-Lab

Wave Field Analysis Using Virtual Circular Microphone Arrays

Master MVA Analyse des signaux Audiofréquences Audio Signal Analysis, Indexing and Transformation

ORIENTATION IN SIMPLE VIRTUAL AUDITORY SPACE CREATED WITH MEASURED HRTF

PART II Practical problems in the spectral analysis of speech signals

Proceedings of Meetings on Acoustics

Laboratory Assignment 2 Signal Sampling, Manipulation, and Playback

Linux Audio Conference 2009

Spatialization and Timbre for Effective Auditory Graphing

Developing a Versatile Audio Synthesizer TJHSST Senior Research Project Computer Systems Lab

396 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 19, NO. 2, FEBRUARY 2011

Trigonometric functions and sound

Robotic Spatial Sound Localization and Its 3-D Sound Human Interface

Music 270a: Modulation

Modulation. Digital Data Transmission. COMP476 Networked Computer Systems. Sine Waves vs. Square Waves. Fourier Series. Modulation

University of Huddersfield Repository

Transcription:

SOPA version 2 Revised July 7 2014 SOPA project September 21, 2014 Contents 1 Introduction 2 2 Basic concept 3 3 Capturing spatial audio 4 4 Sphere around your head 5 5 Reproduction 7 5.1 Binaural reproduction...................... 7 5.2 Database configuration...................... 8 5.3 Panoramic sound field system.................. 8 6 SOPA version 2 file 13 1

Fig. 1: Yaw, pitch and roll 1 Introduction In this document, what is new in SOPA version 2 will be introduced. Since directions only in the horizontal plane were dealt with in SOPA version 1, elevation of the sound could not be given to the listener. Although the listeners could control the panning, they could not control tilt. In SOPA version 2, spatial information not only in the horizontal but also on the vertical dimension are encoded and transmitted. By encoding spatial audio as SOPA version 2, therefore, elevation of the sound can be reproduced and the listeners can interactively control tilt. In this document, panning means the rotation in the horizontal plane or the rotation around the Y-axis and it corresponds to yaw of the listner s head. Tilt is the rotation around the X-axis and it corresponds to pitch of the listener s head. As can be seen in Fig. 1, roll is the rotation around the Z-axis. By using the SOPA format, the panoramic sounds can be presented to the listeners not only through headphones but also by a multi-channel loudspeaker system. How to reproduce the panoramic sounds by a multi-channel loudspeaker system will be described, too. 2

Fig. 2: Capturing a monaural signal What is captured at the observation point can be regarded as the sum of sine waves of the frequencies between f 0 Hz and f N 1 Hz inclusive. Each sine wave can be completely specified by its amplitude and phase response. 2 Basic concept It is known that any given waveform can be subdivided to one or more sinusoidal signals. In the case of Fig. 2, the monaural signal captured at the observation point can be thought to be a mixture of sine waves of the frequencies between f 0 Hz and f N 1 Hz inclusive. Here, it is apparent that the number of the components is N no matter how many sound sources exist around the observation point. Each sine wave can be completely specified by its amplitude and phase response. It means that any monaural signal can be reproduced by using its spectral information. However, the monaural signal does not contain any spatial information. What are necessary to transmit the spatial audio are not only the amplitude and phase response of the signal but also the directional information of the sine waves. As long as it is propagating, the sine wave has its direction of propagation. The basic concept of the SOPA technology is that the spatial audio information can be reproduced by using the spectral information and the directional information of each frequency component. When there is more than one microphone, the direction of propagation can be calculated by a simple trigonometric function since the speed of the sound wave in the air is known. On the 2-dimensional plane, the minimum number of microphones needed to determine the direction is 3 and in the 3

Fig. 3: Top view of a 4-channel miniature head simulator 3-dimensional space, at least 4 microphones are needed. 3 Capturing spatial audio Three-dimensional spatial audio information can be captured with a 4-channel miniature head simulator whose photograph is shown in Fig. 3. The 4-channel miniature head simulator consists of 4 omnidirectional microphones. In our model shown in Fig. 3, these microphones are placed at the vertices of a tetrahedron. Each side of the tetrahedron is 30 mm long. One of the microphones is expediently defined as a reference microphone and the others are defined as comparison microphones. Since the speed of the sound wave in the air is known, from the phase difference between signals captured by the reference and comparison microphones, the direction of propagation can be calculated for each frequency bin in the temporal window. Here, we define the direction from which the sine wave is propagating as sound image direction. Needless to say, the sound image direction is not necessarily corresponds to the sound source direction. In the SOPA version 2 format, the sound image directions are encoded as 8-bit unsigned integers and can be transmitted online along with the audio data stream of the reference signal (signal captured by the reference microphone). The number of the PCM data contained in each temporal window (frame) has to be 2 n (n is an integer), that is 1,024, 2,048, 4,096, for instance. How the sound image directions are encoded as 8-bit binary data is given in the next section. 4

Fig. 5: Perspective view of the imaginary sphere Fig. 4: Top view of the horizontal plane 4 Sphere around your head In the SOPA format, directional information is represented by an 8-bit unsigned integer. It means that 256 numbers are assignable to different directions. In SOPA version 1, however, only 73 numbers are used. Numbers from 1 to 72 are assigned to directions in the horizontal plane as can be seen in Fig. 4 and the number 0 is used as the marker that indicate the beginning of the frames [1]. Fig. 5 shows an imaginary 3-dimensional sphere assumed in SOPA version 2. The center of the sphere is the reference point where the listener s head is assumed to be. The sphere is divided latitudinally into 13 levels as shown in Fig. 6 and Table 1. The latitudes of 90 and -90 degrees represent the right above (top direction) and right below (bottom direction), respectively. Each latitude is segmented longitudinally. The numbers of the segments are presented in Table 1. The total number of the segments is 254 (1 + 8 + 16 + 24 + 30 + 32 + 32 + 32 + 30 + 24 + 16 + 8 + 1). In other words, the surface of the imaginary sphere is divided into 254 sectors in SOPA version 5

Table 1: Longitudinal segments Fig. 6: Latitudes of the sphere Latitudes Number of Longitudinal directions (degree) segments 90 1 Top 75 8 Every 1/4 rad. 60 16 Every 1/8 rad. 45 24 Every 1/12 rad. 30 30 Every 1/15 rad. 15 32 Every 1/16 rad. 0 32 Every 1/16 rad. -15 32 Every 1/16 rad. -30 30 Every 1/15 rad. -45 24 Every 1/12 rad. -60 16 Every 1/8 rad. -75 8 Every 1/4 rad. -90 1 Bottom Front Back Fig. 7: Front and back views of the sphere 2. In the SOPA version 2 format, integers between 0 and 253 inclusive are assigned, one by one, to the sectors. The numbers 0 and 253 are assigned to the top (latitude of 90 degrees) and bottom (latitude of -90 degrees) directions, respectively. The sectors with their numbers are shown in Fig. 7 and Fig. 8. Now, the sound image direction can be represented by an 8-bit binary. The sector with the number x directly faces the sector with the number 253 - x. For instance, the sectors with the numbers 4 and 249 directly face one another. Since the sectors on the surface of the sphere represent the relative directions from the reference point, the sectors with the numbers x 6

Top Bottom Fig. 8: Top and bottom views of the sphere Integers between 0 and 253 inclusive are assigned, one by one, to the sectors on the surface of the sphere. and 253 - x have the opposite directions to each other. Increasing the latitude corresponds to the elevation of the sound. By using SOPA version 2, not only the longitudinal directions but also the elevation of the sounds can be dealt with. It allows the creators to express not only yaw but also pitch and roll of the listener s head. 5 Reproduction Since the SOPA format conveys a stream of monaural audio data (16-bit PCM) and the sound image direction (8-bit binary) for each frequency bin of each temporal window, what is needed to reproduce SOPA data as the panoramic sound is to make each frequency component sound as if it is propagating from its sound image direction. Panoramic sounds can be generated from the SOPA data and presented to a listener or listeners either through stereo headphones or by a multi-channel loudspeaker system. 5.1 Binaural reproduction To generate the binaural signals out of the SOPA data, the HRTF (Head Related Transfer Function) database is used. By decoding the SOPA data, the sound image direction can be extracted for each frequency bin. To reproduce the signal as the binaural signal, the application has to read the HRTF data of the corresponding direction from the HRTF databases and compose the 7

Temporal HRTF. For more information about the Temporal HRTF, see the reference[1]. The binaural signals can be generated by superimposing the left and right Temporal HRTFs to the reference signal in each temporal window. After adding the signals to those of the previous frame by the overlap-add method, the binaural signals can be generated and the generated signals are reproduced through headphones [2]. 5.2 Database configuration In our applications, 2 database files are used. They are the binary files named hrtf3d512.bin and phase3d512.bin. The former contains the amplitude spectra and the latter contains the phase spectra of the HRTF. Each of them consists of a binary data stream of 16-bit signed integers based on HRIR data recorded at a sampling rate of 44,100 Hz. Fig. 9 shows the data stream in hrtf3d512.bin. As can be seen in the figure, the file contains 254 subsets of the data, each of which corresponds to one of 254 directions. The numbers on the horizontal axis correspond to the numbers assigned to the sectors on the surface of the imaginary sphere shown in the Fig. 5. Each subset contains 512 data, each of which corresponds to a frequency bin. Fig. 10 shows one of these subsets. The subset is the HRTF amplitude spectrum for a particular angle represented in a linear scale and multiplied by 2,048 to fit in the 16-bit scale. Since it is based on the data sampled at 44,100 Hz, the horizontal axis represents the frequencies between 0 and 44,100 Hz. Phase3d512.bin also consists of 254 subsets of data. Each subset contains 512 signed integers. The phase values range from π to π (in units of radians). To fit them into the 16-bit linear scale, the values are multiplied by 10,000. Therefore, an angle of π rad. is represented by 31,415 instead of 3.1415. Fig. 11 shows a single subset, where 256 on the X-axis corresponds to the Nyquist frequency, which is 22,050 Hz. 5.3 Panoramic sound field system By using the HRTF database, binaural signals can be generated so that the panoramic sounds can be presented to a listener. It is not convenient, however, when there is more than one listener. To let all the listeners experience the panoramic sounds at the same time, sufficient number of headphones have to be prepared. There is another option, however, to generate the panoramic 8

Fig. 9: Data stream in an HRTF database Fig. 10: Single subset of an HRTF database Fig. 11: Single subset of a phase database 9

Fig. 12: A prototype of the panoramic sound field system As a prototype, a 4-channel panoramic sound field system was produced. Loudspeakers are placed at the vertices of a square in the horizontal plane. sound out of the SOPA data. It is named Panoramic Sound Field System. It uses a multi-channel loudspeaker system instead of stereo headphones. A panoramic sound field system consists of a SOPA decoder software and a multi-channel loudspeaker system. A prototype of the system is shown in Fig. 12. Although this prototype consists of 4 loudspeakers, the number of the loudspeakers can be arranged if necessary. The SOPA decoder for the panoramic sound field system simply adjust proportion distributed to the loudspeakers for each frequency bin according to the sound image direction. The closer the sound image direction is to the direction of the loudspeaker, the higher is the proportion distributed to the loudspeaker (Fig. 13). If the sound image direction of a component lies in the middle of two loudspeakers as shown in the left illustration of Fig. 13, the component is equally distributed to these loudspeakers. A phantom source of the component is generated at the middle of them by summing localization. Fig. 14 shows the flow of processing. A SOPA file conveys a monaural audio data stream (reference signal) and the directional data that contains the sound image directions. The spectrum of the reference signal is extracted for each temporal window by applying a fast Fourier transform to the reference signal. Only the amplitude is modified for each frequency bin according 10

Fig. 13: Sound image direction and signal distribution Proportion distributed to loudspeakers is determined for each frequency bin according to the sound image direction of the corresponding frequency. The size of the arrows schematically represents proportion distributed to the loudspeaker. to the sound image direction of the corresponding frequency. After modifying the amplitude, signals of the time domain are generated by applying an inverse Fourier transform to the modified spectrum. The HRTF database is not used in the panoramic sound field system. The modification of the amplitude and the inverse Fourier transform are carried out for each of 4 loudspeaker channels. The sound image direction is a relative direction and can be altered when the listener s viewing axis changes. In most of the conventional surround sound systems, however, panning of the sounds cannot be altered programmably. In the panoramic sound field system, proportion distributed to the loudspeakers can be determined immediately before the data are reproduced. This makes it possible for the listener to control panning of sounds during its reproduction. As can be seen in Fig 14, by biasing the directional data, panning of the sounds can be modified quite easily. In the case of the prototype shown in Fig. 12, loudspeakers are placed at the vertices of a square in the horizontal plane. As long as the loudspeakers are 2-dimensionally arranged like this, spatial information only in the horizontal plane can be dealt with. It can be expanded, however, to 3-dimensional by arranging the loudspeakers in 3-dimensionally. In many of the loudspeaker array technologies such as the wave field synthesis [3] and the Boundary Surface Control [4], the acoustic wave front is artificially synthesized by using some tens or hundreds of loudspeakers. They are based on Huygens Principle. Since in these methods, a sound field 11

Fig. 14: Flow of processing The flow of processing in the panoramic sound field system is shown. By modifying the directional data, panning of the sounds can be controlled interactively. is synthesized by controlling sound pressure at many different points on the given plane, relatively wide listening area can be provided. Not like these methods, in the panoramic sound field system, spatial audio information at a single point is reproduced instead of synthesizing a sound field. In the panoramic sound field system, therefore, only a limited listening area can be provided. The accuracy of the spatial information depends on the listening spot. It may be optimum at the reference point from where each loudspeaker is at the same distance. Even its limited listening area is taken into consideration, the panoramic sound field system may be attractive for its user-friendliness. It requires far less microphones and loudspeakers than the wave field synthesis does especially when dealing with the 3-dimensional space. 12

Fig. 15: Header and the beginning of the data stream in a SOPA file What are modified in the version 2 are indicated by the yellow and orange colors. For details about the file format, see the reference [1]. 6 SOPA version 2 file The file structure of SOPA version 2 is mostly the same as that of the previous version described in the reference [1] except for 2 modifications. One of the modifications is the version number specified in the header part. As can be seen in Fig. 15, there is 2.0.0.0 instead of 1.0.0.0 for the SOPA file version in the header (area with yellow color in Fig. 15). The other modification is made to the marker in the data stream. As previously mentioned, the surface of the imaginary sphere is divided into 254 segments and each number between 0 and 253 is assigned to one of the segments. The numbers 254 and 255 are not assigned to any directions. In SOPA version 2, the number 255 is used as the marker that indicate the beginning of the frames. Since the frame size is not specified in the header of the sopa file, we have to insert markers in the data stream to let the program know the frame size. In the SOPA file, the sound image directions for the frequency bins are represented by 8-bit unsigned integers. In the case of the f 0 (dc) component, however, the sound image direction has no importance. The addresses corresponding to the f 0 component (area shown in orange background color in Fig. 15) can contain any integer, and we put 255s there. FF in hex, 255 in decimal can be seen in Fig. 15. Since it is not used for any other 13

frequencies, 255s appear only at addresses corresponding to f 0 and can be used as markers that indicate the beginning of the frames. By counting the number of addresses between the first 255 and the next 255, the program automatically extracts the size of the frame. As you can easily guess, the size of a SOPA file is identical to a stereo WAV file. This is far smaller than the data needed in the wave field synthesis that requires tens or hundreds of microphones and loudspeakers. Although it provides a wider listening area, the wave field synthesis costs not only the hardware resources but also the network resources to transmit the number of audio data to synthesize the wave front. In spite of its very small size, the SOPA file can convey enough spatial information to generate the panoramic sounds. The SOPA technology will considerably save not only computational but also network resources. References [1] doc, https://staff.aist.go.jp/ashihara-k/documents/doc.pdf [2] Kousuke Taki, Shogo Kiryu and Kaoru Ashihara, Capturing spatial audio information by using a miniature head simulator. Proceedings of the 21st International Congress on Acoustics, 3aED7, 2013 [3] Günter Theile, Wave Field Synthesis - A Promising Spatial Audio Rendering Concept, Proc. of the 7th Int. Conference of digital audio effects, 125-132, 2004 [4] Ise, S., Development of an immersive auditory display Sound Cask for transferring musical skill in a remote environment, Proceedings of ASJ 2014 Autumn Meeting, 1287-1290, 2014 14