Journal of the Acoustical Society of America 88
|
|
- Melvyn Lawson
- 6 years ago
- Views:
Transcription
1 The following article appeared in Journal of the Acoustical Society of America 88: and may be found at Copyright (1990) Acoustical Society of America. This article may be downloaded for personal use only. Any other use requires prior permission of the author and the Acoustical Society of America.
2 Analytical expressions for the tonotopic sensory scale Hartmut Traunm011er Institutionen fsr lingvistik, Stockholms universitet, S Stockholm, Sweden (Received 16 August 1989; accepted for publication 20 February 1990) Accuracy and simplicity of analytical expressions for the relations between frequency and critical bandwidth as well as critical-band rate (in Bark) are assessed for the purpose of applications in speech perception research and in speech technology. The equivalent rectangular bandwidth (ERB) is seen as a measure of frequency resolution, while the classical critical-band rate is considered a measure of tonotopic position. For the conversion of frequency to critical-band rate, and vice versa, the inversible formula z = [26.81/( / f) ] is proposed. Within the frequency range of the perceptually essential vowel formants ( khz), it agrees to within _ Bark with the Bark scale, originally published in the form of a table. PACS numbers: Cq, Ar, Fe INTRODUCTION Two processes are generally assumed to contribute to auditory frequency resolution. First, the hearing system is capable of performing an "oscillographic" analysis of the set of neural signals originating in the cochlea. This process is limited to frequencies that can be resolved in the pattern of neural responses. While single neurons are not likely to fire more frequently than 500 times per second even at high stimulus intensities, frequencies between 0.5 and 1.5 khz can still be handled in the temporal domain, albeit less efficiently, on the basis of the signals from a large number of neurons. The capability and limitations of a frequency analysis in the temporal domain are demonstrated vividly by cochlear implant patients whose sole auditory input is an undifferentiated electrical stimulation of the auditory nerve. The second process covers the whole auditory frequency range. Any sound entering a normal functioning cochlea is subjecto a spectral analysis, resulting in a frequency-toplace transformation. The cochlea can be regarded as a bank of filters whose outputs are ordered tonotopically, with the filters closest to the base responding maximally to the highest frequencies. The tonotopic order is known to be maintained in the structure of the neural network at higher levels in the hearing system. The "notch-noise method" has often been used in investigations of auditory frequency selectivity. It involves the determination of the detection threshold for a sinusoid, centered in a spectral notch of a noise, as a function of the width of the notch. On the basis of results obtained with this method, auditory frequency selecivity can be described in terms of the equivalent rectangular bandwidth (ERB) as a function of center frequency (Moore and Glasberg, 1983). Since the two processes mentioned above both contribute to the detection of the sinusoid, the ERB, or ERB rate should not be taken as a measure of the tonotopic scale as such. A quantity related to the ERB, though not identical with it, is the classical critical bandwidth (CB) (Zwicker et al., 1957). Measurement of the CB typically involves loudness summation experiments. Different summation rules have been found to hold for auditory stimuli, depending on whether their frequency components are separated by more or less than the CB. The CB and the ERB have been found to be proportional and equivalent for center frequencies above 500 Hz. For lower frequencies, there is a discrepancy, as shown in Fig. 1. In this range, the ERB decreases with decreasing center frequency, while the CB remains close to constant. The discrepancy can be explained by the reasona- ble assumption that the analysis within the temporal domain is irrelevanto loudnessummation as long as loudness variations are not audible as such, while it contributes substantially to frequency resolution for f< 500 Hz. Consequently, the CB should not be taken as a measure of frequency resolution, but CB rate may be taken as a measure of the tonotopic sensory scale. In the familiar CB-rate scale (see Fig. 2), the CB has been chosen to serve as a natural unit of the tonotopic sensory scale. Standard values for the relation between frequen- O. I ;.0 IO ß I I I I I I S S; z,.o l <f) FIG. 1. Equivalent rectangular bandwidth, according to the formula B = f2 q_ f , given by Moore and Glasberg ( 1983 ) (curve), and critical bandwidth, according to Zwicker's ( 1961 ) table (marks), as a function of frequency. 97 J. Acoust. Soc. Am. 88 (1), July /90/ Acoustical Society of America 97
3 24 Frequency (khz> O.S S.O 10 I I! I! I I I. ANALYTICAL EXPRESSIONS A. Expressions for critical-band rate In rough approximation, the relation betweenf and z is linear for f< 500 Hz (z =f/100) and logarithmic for higher frequencies. Figure 3 (a) shows the error functions of two logarithmic approximations to the CB scale. One of these, Eq. (1), has been suggested by Zwicker and Terhardt (1980). It gives values that agree with the tabulated ones to within _ 0.25 Bark in the range 0.6 <f< 7.2 khz. The other approximation, Eq. (2), satisfies our stricter standards of no more than _ 0.05-Bark deviation at the cost of a reduction in the range of validity, to 1.0 < f< 3.6 khz: z = 14.2 log(f/1000) 4-8.7, ( 1 ) z = ln(f) (2) lg(f> FIG. 2. Critical-band rate z as a function offrequencyf The plus sign ( + ) represents data from Zwicker ( 1961 ). The curve corresponds to Eq. (6). cyfand CB rate z have been proposed by Zwicker ( 1961 ) in the form of a table. The CB-rate scale has been applied extensively in research on psychoacoustics and speech perception. For most of these applications, it would be more convenient to have the relation between z and fspecified in the form of an equation instead of a table. Several equations that approximate the tabulated values have also been published (Tjomov, 1971; Schroeder, 1977; Zwicker and Terhardt, 1980; Traunm011er, 1983). In the following, the error functions of these equations will be compared. Recent studies of speech sound suggest that the tonotopic distances (CB-rate differences) between prominent peaks in their spectra are fundamental to the perception of their phonetic quality. More specifically, it has been suggested that the spectral peaks shaped by the formants and the fundamental have the same relative tonotopic locations in linguistically identical vowels uttered by speakers different in age and sex (Traunm011er, 1983, 1988; Syrdal and Gopal, 1986). While differences in speaker size appear to be reflected in a tonotopic translation of the spectral peaks, differences in vocal effort appear to be reflected in a linear tonotopic compression/expansion (Traunm011er, 1988). In order to test these hypotheses, both in theory and by means of speech synthesis, a convenient and accurate method of conversion from frequency to CB rate, and vice versa, is needed. Our requirements include that the function have a simple inverse and that it be accurate preferably to within _ 0.05 Bark in the range of essential vowel formant frequencies of men, women, and children. This rigorous claim for accuracy prevents the introduction of any avoidable error in addition to that inherent in the table (Zwicker, 1961 ). However, it should be noticed that the absolute width of the critical band, and its definition, is irrelevan to the applications we have in mind, as long as the obtained scales remain proportional. 98 J. Acoust. Soc. Am., Vol. 88, No. 1, July 1990 In these and in all the following equations, frequencyf is to be expressed in Hz and CB rate z in CB units (Bark). A mathematical function that is linear at one extreme and logarithmic at the other extreme, the sinus-hyperbolicus function, has been used by Tjomov ( 1971 ), Eq. (3), and by Schroeder (1977), Eq. (4), to calculate CB rate. The error functions of both equations are shown in Fig. 3(b). f = 600 sinh (z/6.7) 4-20, ( 3 ) z = 6.7 ln{[ (f-- 20)/600] + ([ (f- 20)/600] 2 + 1) /2} (inverse), f = 650 sinh (z?7), (4) z = 7 ln((f/650) + [ (f/650) ] /2) (inverse). As compared with the tabulated values, Tjomov's equation (3) is accurate to within to Bark for f< 4.5 khz and Schroeder's equation (4) to within 4-03 Bark for f< 4.0 khz. These equations are accurate enough for some applications in which frequency components above 4 khz may be neglected, as they are in some systems of telephonic communication. Approximations covering the whole auditory frequency range can be achieved in various ways by appropriate combinations of mathematical functions. For the most part, however, this yields equations that lack a simple inverse. The most accurate of the equations given by Zwicker and Terhardt (1980), z = 13 atn( f) atn(f/7500) 2, (5) is of this kind. It agrees with the table to within to Bark over the whole range of auditory perception [see Fig. 3 (c) ]. The waviness of the error function tells us, however, that there is room for improvement. The equation also clearly falls short of our standards. If, e.g., we want to compare the tonotopic distances between two pairs of spectral peaks, we might obtain an error of up to 0.9 Bark. An approximation that has a simple inverse and meets our standards is achieved by considering z to be related to log(f) by a logistic function, also known as "growth curve." Such an approximation, Eq. (6), has been proposed by Traunm011er (1983). Its error function is shown in Fig. 3(d): Hartmut Traunmdller: Tonotopic sensory scale 98
4 A. Fregue'ncy f (khz) 0.0 O.S I I I I l.o I I I I I½ I I'''''1'... I... I ee e 3 o _ 4 CB-rote z (LobIe vo, lue) (a) 3 v. N 2 I - A. ß 0 -_ - -.! _ O v-2- N ' i _. ß o. -.3! - --e... I...,I,,,,,I... I,,,,,I CB-rote z (Loble v lue) (b) a, I ß /, I I! I I I 0.0 O.S I I'''''1'''''1'... I'... I 3 2 ß 0 2,,, i,,,,, I,,,, I... I,,,,, I o z, CB-rale z (tab!e v lue) (c).3 ß 2 f - ß 0... c - o v -,2 N o _.... I,,,,, I,,,,, I,,,, I,,,,, I CB- cte z (tc le vclue) o FIG. 3. ( a)-(d) Error functions of various approximations of the CB-rate scale. The error is defined as the difference between the calculated value and that in Zwicker's ( 1961 ) table. It is plotted in steps of 0.5 Bark for each frequency value in that table. (a) Logarithmic approximations: curve with marks, Eq. ( 1 ) [given by Zwicker and Terhardt (1980)]; curve without marks, Eq. (2). (b) Sinus-hyperbolicus approximations: lower curve, Eq. (3) [given by Tjomov ( 1971 )]; upper curve, Eq. (4) [given by Schroeder (1977)]. (c) An overall approximation, Eq. ( 5 ), given by Zwicker and Terhardt (1980).(d) A logistic "growth-curve" approximation: lower curve with error scale at the left, Eq. (6) [given by Traunmiiller (1983)]; upper curve, shown vertically displaced, with error scale at the right, Eq. (6) with corrections (7) and (8). z = [26.81f/( f) ] , (6) f= 1960(z )/( z) (inverse). The values obtained with Eq. (6) deviate from the tabulated ones by less than +_ 0.05 Bark for 0.2 <f< 6.7 khz. At the low-frequency end of the scale, the deviation from the table (Zwicker, 1961 ) sums up to Bark for f= 0 Hz ( Bark for f= 20 Hz). At least in part, this deviation is due to biased rounding of the bandwidth values in Zwicker's table. For frequencies below 400 Hz, the standard width of the critical band was set uniformly equal to 100 Hz. This appears to have been done in order to obtain the mnemonically simple relation z = f? 100. The original bandwidth data (Zwicker et al., 1957 ) indicate B 90 Hz for the lower frequencies in that range. The values listed in the table for f< 100 Hz are particularly questionable because they can hardly be said to be based on any reliable experimental evi- dence. Equation (6) may representhe tonotopic scale well enough down to the lowest frequencies for which it can be determined experimentally. The deviation at the high-frequency end of the scale remains unaccounted for. Calculating z with Eq. (6), close agreement with the table can be achieved over the whole auditory frequency range by added corrections, bending the error function straight at both ends of the scale, in the following way: for calculated z < 2.0 Bark: z'=z+o. 15(2--z), (7) for calculated z > 20 Bark: z' = z q (z ). (8) Since this is an easily inverted procedure, the calculation off for a given zis not a problem. The error function obtained with these corrections is also shown in Fig. 3 (d). The values calculated in this way agree with the table for f> 100 Hz to within -F 0.05 Bark. Correction (7), however, simulates also the above-mentioned bias at low frequencies. 99 J. Acoust. $oc. Am., Vol. 88, No. 1, July 1990 Hartmut Traunm Jller: Tonotopic sensory scale 99
5 F e u, ency f (khz) 0.0 O.E; A..O 8.0 for critical bands centered at z obtained by Eq. (6) without corrections. The values calculated by Eq. (10) agree with Zwicker's table to within + 6% for 0.27 <f<5.8 khz. Within that range, the error function is similar to that obtained by Eq. (9). The error functions of both equations are shown in Fig. 4. ß 0 ' v rn 2 ACKNOWLEDGMENT The preparation of this paper has been supported by a grant from HSFR, the Swedish Council for Research in the Humanities and Social Sciences. u.i I,, I,,,, I,,,,, I,,,, I 0 G , CB-rote z (Loble value) FIG. 4. Error functions for critical bandwidth calculated with Eq. (9) (curve with marks) and Eq. (10) (curve without marks), as compared with Zwicker's ( 1961 ) table values (see also Fig. 1 ). B. Expressions for critical bandwidth Zwicker and Terhardt (1980) proposed the equation B ( f2) 0'69 (9) to calculate critical bandwidth B as a function of center fre- quencyf While Eq. (9) is very accurate, it cannot easily be integrated to obtain CB rate. The authors' equation for CB rate (5)'is not compatible with Eq. (9). Proceeding from Eq. (6), critical bandwidths B can be calculated as B = 52548/(z z ) (lo) Moore, B.C. J., and Glasberg, B. R. (1983). "Suggested formulae for calculating auditory-filter bandwidths and excitation patterns," J. Acoust. Soc. Am. 74, Schroeder, M. R. (1977). "Recognition of complex acoustic signals," in Life Sciences Research Report 5 (Dahlem Konferenzen), edited by T. H. Bullock (Abakon Verlag, Berlin), pp Syrdal, A. K., and Gopal, H. S. (1986). "A perceptual model of vowel recognition based on the auditory representation of American English vowels," J. Acoust. Soc. Am. 79, Tjomov, V. L. (1971). "A model to describe the results of psychoacoustical experiments on steady-state stimuli," in Analiz Rechevykh $ignalov Chelovekom, edited by G. V. Gershuni (Nauka, Leningrad), pp Traunmiiller, H. (1983). "On vowels: Perception of spectral features, related aspects of production and sociophonetic dimensions," Ph.D. thesis, University of Stockholm. Traunmiiller, H. (1988). "Paralinguistic variation and invariance in the characteristic frequencies of vowels," Phonetica 45, Zwicker, E. ( 1961). "Subdivision of the audible frequency range into critical bands (Frequenzgruppen)," J. Accoust. Soc. Am. 33, 248. Zwicker, E., Flottorp, G., and Stevens, S.S. (1957). "Critical bandwidth in loudness summation," J. Acoust. Soc. Am. 29, Zwicker, E., and Terhardt, E. (1980). "Analytical expressions for criticalband rate and critical bandwidth as a function of frequency," J. Acoust. Soc. Am. 68, J. Acoust. Soc. Am., Vol. 88, No. 1, July 1990 Hartmut Traunm(Jller: Tonotopic sensory scale 100
Psycho-acoustics (Sound characteristics, Masking, and Loudness)
Psycho-acoustics (Sound characteristics, Masking, and Loudness) Tai-Shih Chi ( 冀泰石 ) Department of Communication Engineering National Chiao Tung University Mar. 20, 2008 Pure tones Mathematics of the pure
More informationAuditory modelling for speech processing in the perceptual domain
ANZIAM J. 45 (E) ppc964 C980, 2004 C964 Auditory modelling for speech processing in the perceptual domain L. Lin E. Ambikairajah W. H. Holmes (Received 8 August 2003; revised 28 January 2004) Abstract
More informationPerception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.
Perception of pitch BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb 2008. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence Erlbaum,
More informationPerception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.
Perception of pitch BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb 2009. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence
More informationHCS 7367 Speech Perception
HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Power spectrum model of masking Assumptions: Only frequencies within the passband of the auditory filter contribute to masking. Detection is based
More informationPerception of pitch. Importance of pitch: 2. mother hemp horse. scold. Definitions. Why is pitch important? AUDL4007: 11 Feb A. Faulkner.
Perception of pitch AUDL4007: 11 Feb 2010. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence Erlbaum, 2005 Chapter 7 1 Definitions
More information19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007
19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 MODELING SPECTRAL AND TEMPORAL MASKING IN THE HUMAN AUDITORY SYSTEM PACS: 43.66.Ba, 43.66.Dc Dau, Torsten; Jepsen, Morten L.; Ewert,
More informationAuditory filters at low frequencies: ERB and filter shape
Auditory filters at low frequencies: ERB and filter shape Spring - 2007 Acoustics - 07gr1061 Carlos Jurado David Robledano Spring 2007 AALBORG UNIVERSITY 2 Preface The report contains all relevant information
More informationYou know about adding up waves, e.g. from two loudspeakers. AUDL 4007 Auditory Perception. Week 2½. Mathematical prelude: Adding up levels
AUDL 47 Auditory Perception You know about adding up waves, e.g. from two loudspeakers Week 2½ Mathematical prelude: Adding up levels 2 But how do you get the total rms from the rms values of two signals
More information2920 J. Acoust. Soc. Am. 102 (5), Pt. 1, November /97/102(5)/2920/5/$ Acoustical Society of America 2920
Detection and discrimination of frequency glides as a function of direction, duration, frequency span, and center frequency John P. Madden and Kevin M. Fire Department of Communication Sciences and Disorders,
More informationarxiv: v1 [eess.as] 30 Dec 2017
LOGARITHMI FREQUEY SALIG AD OSISTET FREQUEY OVERAGE FOR THE SELETIO OF AUDITORY FILTERAK ETER FREQUEIES Shoufeng Lin arxiv:8.75v [eess.as] 3 Dec 27 Department of Electrical and omputer Engineering, urtin
More informationIII. Publication III. c 2005 Toni Hirvonen.
III Publication III Hirvonen, T., Segregation of Two Simultaneously Arriving Narrowband Noise Signals as a Function of Spatial and Frequency Separation, in Proceedings of th International Conference on
More informationA Pole Zero Filter Cascade Provides Good Fits to Human Masking Data and to Basilar Membrane and Neural Data
A Pole Zero Filter Cascade Provides Good Fits to Human Masking Data and to Basilar Membrane and Neural Data Richard F. Lyon Google, Inc. Abstract. A cascade of two-pole two-zero filters with level-dependent
More informationAN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES
Proceedings of the COST G-6 Conference on Digital Audio Effects (DAFX-), Verona, Italy, December 7-9,2 AN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES Tapio Lokki Telecommunications
More informationAuditory Based Feature Vectors for Speech Recognition Systems
Auditory Based Feature Vectors for Speech Recognition Systems Dr. Waleed H. Abdulla Electrical & Computer Engineering Department The University of Auckland, New Zealand [w.abdulla@auckland.ac.nz] 1 Outlines
More informationTHE PERCEPTION OF ALL-PASS COMPONENTS IN TRANSFER FUNCTIONS
PACS Reference: 43.66.Pn THE PERCEPTION OF ALL-PASS COMPONENTS IN TRANSFER FUNCTIONS Pauli Minnaar; Jan Plogsties; Søren Krarup Olesen; Flemming Christensen; Henrik Møller Department of Acoustics Aalborg
More informationEffect of filter spacing and correct tonotopic representation on melody recognition: Implications for cochlear implants
Effect of filter spacing and correct tonotopic representation on melody recognition: Implications for cochlear implants Kalyan S. Kasturi and Philipos C. Loizou Dept. of Electrical Engineering The University
More informationTone-in-noise detection: Observed discrepancies in spectral integration. Nicolas Le Goff a) Technische Universiteit Eindhoven, P.O.
Tone-in-noise detection: Observed discrepancies in spectral integration Nicolas Le Goff a) Technische Universiteit Eindhoven, P.O. Box 513, NL-5600 MB Eindhoven, The Netherlands Armin Kohlrausch b) and
More informationAcoustics, signals & systems for audiology. Week 9. Basic Psychoacoustic Phenomena: Temporal resolution
Acoustics, signals & systems for audiology Week 9 Basic Psychoacoustic Phenomena: Temporal resolution Modulating a sinusoid carrier at 1 khz (fine structure) x modulator at 100 Hz (envelope) = amplitudemodulated
More informationCOM325 Computer Speech and Hearing
COM325 Computer Speech and Hearing Part III : Theories and Models of Pitch Perception Dr. Guy Brown Room 145 Regent Court Department of Computer Science University of Sheffield Email: g.brown@dcs.shef.ac.uk
More informationAcoustics, signals & systems for audiology. Week 4. Signals through Systems
Acoustics, signals & systems for audiology Week 4 Signals through Systems Crucial ideas Any signal can be constructed as a sum of sine waves In a linear time-invariant (LTI) system, the response to a sinusoid
More informationPerceptual Speech Enhancement Using Multi_band Spectral Attenuation Filter
Perceptual Speech Enhancement Using Multi_band Spectral Attenuation Filter Sana Alaya, Novlène Zoghlami and Zied Lachiri Signal, Image and Information Technology Laboratory National Engineering School
More informationBlock diagram of proposed general approach to automatic reduction of speech wave to lowinformation-rate signals.
XIV. SPEECH COMMUNICATION Prof. M. Halle G. W. Hughes J. M. Heinz Prof. K. N. Stevens Jane B. Arnold C. I. Malme Dr. T. T. Sandel P. T. Brady F. Poza C. G. Bell O. Fujimura G. Rosen A. AUTOMATIC RESOLUTION
More informationCitation for published version (APA): Lijzenga, J. (1997). Discrimination of simplified vowel spectra Groningen: s.n.
University of Groningen Discrimination of simplified vowel spectra Lijzenga, Johannes IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please
More informationStructure of Speech. Physical acoustics Time-domain representation Frequency domain representation Sound shaping
Structure of Speech Physical acoustics Time-domain representation Frequency domain representation Sound shaping Speech acoustics Source-Filter Theory Speech Source characteristics Speech Filter characteristics
More informationComparison of Spectral Analysis Methods for Automatic Speech Recognition
INTERSPEECH 2013 Comparison of Spectral Analysis Methods for Automatic Speech Recognition Venkata Neelima Parinam, Chandra Vootkuri, Stephen A. Zahorian Department of Electrical and Computer Engineering
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Loudness & Temporal resolution
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Loudness & Temporal resolution Absolute thresholds & Loudness Name some ways these concepts are crucial to audiologists Sivian & White (1933) JASA
More informationAUDL GS08/GAV1 Auditory Perception. Envelope and temporal fine structure (TFS)
AUDL GS08/GAV1 Auditory Perception Envelope and temporal fine structure (TFS) Envelope and TFS arise from a method of decomposing waveforms The classic decomposition of waveforms Spectral analysis... Decomposes
More informationRASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991
RASTA-PLP SPEECH ANALYSIS Hynek Hermansky Nelson Morgan y Aruna Bayya Phil Kohn y TR-91-069 December 1991 Abstract Most speech parameter estimation techniques are easily inuenced by the frequency response
More informationComplex Sounds. Reading: Yost Ch. 4
Complex Sounds Reading: Yost Ch. 4 Natural Sounds Most sounds in our everyday lives are not simple sinusoidal sounds, but are complex sounds, consisting of a sum of many sinusoids. The amplitude and frequency
More informationResults of Egan and Hake using a single sinusoidal masker [reprinted with permission from J. Acoust. Soc. Am. 22, 622 (1950)].
XVI. SIGNAL DETECTION BY HUMAN OBSERVERS Prof. J. A. Swets Prof. D. M. Green Linda E. Branneman P. D. Donahue Susan T. Sewall A. MASKING WITH TWO CONTINUOUS TONES One of the earliest studies in the modern
More informationDETERMINATION OF EQUAL-LOUDNESS RELATIONS AT HIGH FREQUENCIES
DETERMINATION OF EQUAL-LOUDNESS RELATIONS AT HIGH FREQUENCIES Rhona Hellman 1, Hisashi Takeshima 2, Yo^iti Suzuki 3, Kenji Ozawa 4, and Toshio Sone 5 1 Department of Psychology and Institute for Hearing,
More informationTemporal resolution AUDL Domain of temporal resolution. Fine structure and envelope. Modulating a sinusoid. Fine structure and envelope
Modulating a sinusoid can also work this backwards! Temporal resolution AUDL 4007 carrier (fine structure) x modulator (envelope) = amplitudemodulated wave 1 2 Domain of temporal resolution Fine structure
More informationIntroduction to cochlear implants Philipos C. Loizou Figure Captions
http://www.utdallas.edu/~loizou/cimplants/tutorial/ Introduction to cochlear implants Philipos C. Loizou Figure Captions Figure 1. The top panel shows the time waveform of a 30-msec segment of the vowel
More informationBinaural Hearing. Reading: Yost Ch. 12
Binaural Hearing Reading: Yost Ch. 12 Binaural Advantages Sounds in our environment are usually complex, and occur either simultaneously or close together in time. Studies have shown that the ability to
More informationSynthesis Algorithms and Validation
Chapter 5 Synthesis Algorithms and Validation An essential step in the study of pathological voices is re-synthesis; clear and immediate evidence of the success and accuracy of modeling efforts is provided
More informationQuantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation
Quantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation Peter J. Murphy and Olatunji O. Akande, Department of Electronic and Computer Engineering University
More informationSignals, Sound, and Sensation
Signals, Sound, and Sensation William M. Hartmann Department of Physics and Astronomy Michigan State University East Lansing, Michigan Л1Р Contents Preface xv Chapter 1: Pure Tones 1 Mathematics of the
More informationMultichannel level alignment, part I: Signals and methods
Suokuisma, Zacharov & Bech AES 5th Convention - San Francisco Multichannel level alignment, part I: Signals and methods Pekka Suokuisma Nokia Research Center, Speech and Audio Systems Laboratory, Tampere,
More informationWARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS
NORDIC ACOUSTICAL MEETING 12-14 JUNE 1996 HELSINKI WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS Helsinki University of Technology Laboratory of Acoustics and Audio
More informationTHE USE OF ARTIFICIAL NEURAL NETWORKS IN THE ESTIMATION OF THE PERCEPTION OF SOUND BY THE HUMAN AUDITORY SYSTEM
INTERNATIONAL JOURNAL ON SMART SENSING AND INTELLIGENT SYSTEMS VOL. 8, NO. 3, SEPTEMBER 2015 THE USE OF ARTIFICIAL NEURAL NETWORKS IN THE ESTIMATION OF THE PERCEPTION OF SOUND BY THE HUMAN AUDITORY SYSTEM
More informationEE482: Digital Signal Processing Applications
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 12 Speech Signal Processing 14/03/25 http://www.ee.unlv.edu/~b1morris/ee482/
More informationTNS Journal Club: Efficient coding of natural sounds, Lewicki, Nature Neurosceince, 2002
TNS Journal Club: Efficient coding of natural sounds, Lewicki, Nature Neurosceince, 2002 Rich Turner (turner@gatsby.ucl.ac.uk) Gatsby Unit, 18/02/2005 Introduction The filters of the auditory system have
More informationUsing the Gammachirp Filter for Auditory Analysis of Speech
Using the Gammachirp Filter for Auditory Analysis of Speech 18.327: Wavelets and Filterbanks Alex Park malex@sls.lcs.mit.edu May 14, 2003 Abstract Modern automatic speech recognition (ASR) systems typically
More informationHearing and Deafness 2. Ear as a frequency analyzer. Chris Darwin
Hearing and Deafness 2. Ear as a analyzer Chris Darwin Frequency: -Hz Sine Wave. Spectrum Amplitude against -..5 Time (s) Waveform Amplitude against time amp Hz Frequency: 5-Hz Sine Wave. Spectrum Amplitude
More informationTesting of Objective Audio Quality Assessment Models on Archive Recordings Artifacts
POSTER 25, PRAGUE MAY 4 Testing of Objective Audio Quality Assessment Models on Archive Recordings Artifacts Bc. Martin Zalabák Department of Radioelectronics, Czech Technical University in Prague, Technická
More informationSignals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend
Signals & Systems for Speech & Hearing Week 6 Bandpass filters & filterbanks Practical spectral analysis Most analogue signals of interest are not easily mathematically specified so applying a Fourier
More informationMachine recognition of speech trained on data from New Jersey Labs
Machine recognition of speech trained on data from New Jersey Labs Frequency response (peak around 5 Hz) Impulse response (effective length around 200 ms) 41 RASTA filter 10 attenuation [db] 40 1 10 modulation
More informationFinal Exam Study Guide: Introduction to Computer Music Course Staff April 24, 2015
Final Exam Study Guide: 15-322 Introduction to Computer Music Course Staff April 24, 2015 This document is intended to help you identify and master the main concepts of 15-322, which is also what we intend
More informationANALYSIS AND EVALUATION OF IRREGULARITY IN PITCH VIBRATO FOR STRING-INSTRUMENT TONES
Abstract ANALYSIS AND EVALUATION OF IRREGULARITY IN PITCH VIBRATO FOR STRING-INSTRUMENT TONES William L. Martens Faculty of Architecture, Design and Planning University of Sydney, Sydney NSW 2006, Australia
More informationHARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS
HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS Sean Enderby and Zlatko Baracskai Department of Digital Media Technology Birmingham City University Birmingham, UK ABSTRACT In this paper several
More informationSpeech/Music Change Point Detection using Sonogram and AANN
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 6, Number 1 (2016), pp. 45-49 International Research Publications House http://www. irphouse.com Speech/Music Change
More informationspeech signal S(n). This involves a transformation of S(n) into another signal or a set of signals
16 3. SPEECH ANALYSIS 3.1 INTRODUCTION TO SPEECH ANALYSIS Many speech processing [22] applications exploits speech production and perception to accomplish speech analysis. By speech analysis we extract
More informationEnhancement of Speech Signal by Adaptation of Scales and Thresholds of Bionic Wavelet Transform Coefficients
ISSN (Print) : 232 3765 An ISO 3297: 27 Certified Organization Vol. 3, Special Issue 3, April 214 Paiyanoor-63 14, Tamil Nadu, India Enhancement of Speech Signal by Adaptation of Scales and Thresholds
More informationPhase and Feedback in the Nonlinear Brain. Malcolm Slaney (IBM and Stanford) Hiroko Shiraiwa-Terasawa (Stanford) Regaip Sen (Stanford)
Phase and Feedback in the Nonlinear Brain Malcolm Slaney (IBM and Stanford) Hiroko Shiraiwa-Terasawa (Stanford) Regaip Sen (Stanford) Auditory processing pre-cosyne workshop March 23, 2004 Simplistic Models
More informationREPORT ITU-R BS Short-term loudness metering. Foreword
Rep. ITU-R BS.2103-1 1 REPORT ITU-R BS.2103-1 Short-term loudness metering (Question ITU-R 2/6) (2007-2008) Foreword This Report is in two parts. The first part discusses the need for different types of
More informationAn introduction to physics of Sound
An introduction to physics of Sound Outlines Acoustics and psycho-acoustics Sound? Wave and waves types Cycle Basic parameters of sound wave period Amplitude Wavelength Frequency Outlines Phase Types of
More informationBASIC ELECTRONICS PROF. T.S. NATARAJAN DEPT OF PHYSICS IIT MADRAS
BASIC ELECTRONICS PROF. T.S. NATARAJAN DEPT OF PHYSICS IIT MADRAS LECTURE-13 Basic Characteristic of an Amplifier Simple Transistor Model, Common Emitter Amplifier Hello everybody! Today in our series
More informationSpectro-Temporal Methods in Primary Auditory Cortex David Klein Didier Depireux Jonathan Simon Shihab Shamma
Spectro-Temporal Methods in Primary Auditory Cortex David Klein Didier Depireux Jonathan Simon Shihab Shamma & Department of Electrical Engineering Supported in part by a MURI grant from the Office of
More informationFREQUENCY WARPED ALL-POLE MODELING OF VOWEL SPECTRA: DEPENDENCE ON VOICE AND VOWEL QUALITY. Pushkar Patwardhan and Preeti Rao
Proceedings of Workshop on Spoken Language Processing January 9-11, 23, T.I.F.R., Mumbai, India. FREQUENCY WARPED ALL-POLE MODELING OF VOWEL SPECTRA: DEPENDENCE ON VOICE AND VOWEL QUALITY Pushkar Patwardhan
More informationContent-based Processing for Masking Minimization in Multi-track Recordings
Content-based Processing for Masking Minimization in Multi-track Recordings Sebastian Vega Lopez Department of Information and Communication Technologies Universitat Pompeu Fabra A thesis submitted for
More informationUniversity of Washington Department of Electrical Engineering Computer Speech Processing EE516 Winter 2005
University of Washington Department of Electrical Engineering Computer Speech Processing EE516 Winter 2005 Lecture 5 Slides Jan 26 th, 2005 Outline of Today s Lecture Announcements Filter-bank analysis
More informationSOUND QUALITY EVALUATION OF FAN NOISE BASED ON HEARING-RELATED PARAMETERS SUMMARY INTRODUCTION
SOUND QUALITY EVALUATION OF FAN NOISE BASED ON HEARING-RELATED PARAMETERS Roland SOTTEK, Klaus GENUIT HEAD acoustics GmbH, Ebertstr. 30a 52134 Herzogenrath, GERMANY SUMMARY Sound quality evaluation of
More informationNOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or
NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or other reproductions of copyrighted material. Any copying
More informationI. INTRODUCTION J. Acoust. Soc. Am. 110 (3), Pt. 1, Sep /2001/110(3)/1628/13/$ Acoustical Society of America
On the upper cutoff frequency of the auditory critical-band envelope detectors in the context of speech perception a) Oded Ghitza Media Signal Processing Research, Agere Systems, Murray Hill, New Jersey
More informationThe source-filter model of speech production"
24.915/24.963! Linguistic Phonetics! The source-filter model of speech production" Glottal airflow Output from lips 400 200 0.1 0.2 0.3 Time (in secs) 30 20 10 0 0 1000 2000 3000 Frequency (Hz) Source
More informationEffect of bandwidth extension to telephone speech recognition in cochlear implant users
Effect of bandwidth extension to telephone speech recognition in cochlear implant users Chuping Liu Department of Electrical Engineering, University of Southern California, Los Angeles, California 90089
More informationPerceptive Speech Filters for Speech Signal Noise Reduction
International Journal of Computer Applications (975 8887) Volume 55 - No. *, October 22 Perceptive Speech Filters for Speech Signal Noise Reduction E.S. Kasthuri and A.P. James School of Computer Science
More informationSpeech, Hearing and Language: work in progress. Volume 12
Speech, Hearing and Language: work in progress Volume 12 2 Construction of a rotary vibrator and its application in human tactile communication Abbas HAYDARI and Stuart ROSEN Department of Phonetics and
More informationINTRODUCTION TO ACOUSTIC PHONETICS 2 Hilary Term, week 6 22 February 2006
1. Resonators and Filters INTRODUCTION TO ACOUSTIC PHONETICS 2 Hilary Term, week 6 22 February 2006 Different vibrating objects are tuned to specific frequencies; these frequencies at which a particular
More informationHi-Fi voice: observations on the distribution of energy in the singing voice spectrum above 5 khz
Hi-Fi voice: observations on the distribution of energy in the singing voice spectrum above 5 khz S. O Ternström Kungliga Tekniska Högskolan, Dept. of Speech, Music & Hearing, Lindstedtsvägen 24, SE-100
More informationA Silicon Model of an Auditory Neural Representation of Spectral Shape
A Silicon Model of an Auditory Neural Representation of Spectral Shape John Lazzaro 1 California Institute of Technology Pasadena, California, USA Abstract The paper describes an analog integrated circuit
More informationAudible Aliasing Distortion in Digital Audio Synthesis
56 J. SCHIMMEL, AUDIBLE ALIASING DISTORTION IN DIGITAL AUDIO SYNTHESIS Audible Aliasing Distortion in Digital Audio Synthesis Jiri SCHIMMEL Dept. of Telecommunications, Faculty of Electrical Engineering
More informationNeural Processing of Amplitude-Modulated Sounds: Joris, Schreiner and Rees, Physiol. Rev. 2004
Neural Processing of Amplitude-Modulated Sounds: Joris, Schreiner and Rees, Physiol. Rev. 2004 Richard Turner (turner@gatsby.ucl.ac.uk) Gatsby Computational Neuroscience Unit, 02/03/2006 As neuroscientists
More informationPerformance study of Text-independent Speaker identification system using MFCC & IMFCC for Telephone and Microphone Speeches
Performance study of Text-independent Speaker identification system using & I for Telephone and Microphone Speeches Ruchi Chaudhary, National Technical Research Organization Abstract: A state-of-the-art
More informationSpectral and temporal processing in the human auditory system
Spectral and temporal processing in the human auditory system To r s t e n Da u 1, Mo rt e n L. Jepsen 1, a n d St e p h a n D. Ew e r t 2 1Centre for Applied Hearing Research, Ørsted DTU, Technical University
More informationFFT 1 /n octave analysis wavelet
06/16 For most acoustic examinations, a simple sound level analysis is insufficient, as not only the overall sound pressure level, but also the frequency-dependent distribution of the level has a significant
More informationVOICE QUALITY SYNTHESIS WITH THE BANDWIDTH ENHANCED SINUSOIDAL MODEL
VOICE QUALITY SYNTHESIS WITH THE BANDWIDTH ENHANCED SINUSOIDAL MODEL Narsimh Kamath Vishweshwara Rao Preeti Rao NIT Karnataka EE Dept, IIT-Bombay EE Dept, IIT-Bombay narsimh@gmail.com vishu@ee.iitb.ac.in
More informationHuman Auditory Periphery (HAP)
Human Auditory Periphery (HAP) Ray Meddis Department of Human Sciences, University of Essex Colchester, CO4 3SQ, UK. rmeddis@essex.ac.uk A demonstrator for a human auditory modelling approach. 23/11/2003
More informationINTRODUCTION. Address and author to whom correspondence should be addressed. Electronic mail:
Detection of time- and bandlimited increments and decrements in a random-level noise Michael G. Heinz Speech and Hearing Sciences Program, Division of Health Sciences and Technology, Massachusetts Institute
More informationLinguistic Phonetics. Spectral Analysis
24.963 Linguistic Phonetics Spectral Analysis 4 4 Frequency (Hz) 1 Reading for next week: Liljencrants & Lindblom 1972. Assignment: Lip-rounding assignment, due 1/15. 2 Spectral analysis techniques There
More informationABSTRACT. Title of Document: SPECTROTEMPORAL MODULATION LISTENERS. Professor, Dr.Shihab Shamma, Department of. Electrical Engineering
ABSTRACT Title of Document: SPECTROTEMPORAL MODULATION SENSITIVITY IN HEARING-IMPAIRED LISTENERS Golbarg Mehraei, Master of Science, 29 Directed By: Professor, Dr.Shihab Shamma, Department of Electrical
More informationTime-Frequency Distributions for Automatic Speech Recognition
196 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 9, NO. 3, MARCH 2001 Time-Frequency Distributions for Automatic Speech Recognition Alexandros Potamianos, Member, IEEE, and Petros Maragos, Fellow,
More informationLoudspeaker Distortion Measurement and Perception Part 2: Irregular distortion caused by defects
Loudspeaker Distortion Measurement and Perception Part 2: Irregular distortion caused by defects Wolfgang Klippel, Klippel GmbH, wklippel@klippel.de Robert Werner, Klippel GmbH, r.werner@klippel.de ABSTRACT
More informationAUDL Final exam page 1/7 Please answer all of the following questions.
AUDL 11 28 Final exam page 1/7 Please answer all of the following questions. 1) Consider 8 harmonics of a sawtooth wave which has a fundamental period of 1 ms and a fundamental component with a level of
More informationAspiration Noise during Phonation: Synthesis, Analysis, and Pitch-Scale Modification. Daryush Mehta
Aspiration Noise during Phonation: Synthesis, Analysis, and Pitch-Scale Modification Daryush Mehta SHBT 03 Research Advisor: Thomas F. Quatieri Speech and Hearing Biosciences and Technology 1 Summary Studied
More informationINFLUENCE OF FREQUENCY DISTRIBUTION ON INTENSITY FLUCTUATIONS OF NOISE
INFLUENCE OF FREQUENCY DISTRIBUTION ON INTENSITY FLUCTUATIONS OF NOISE Pierre HANNA SCRIME - LaBRI Université de Bordeaux 1 F-33405 Talence Cedex, France hanna@labriu-bordeauxfr Myriam DESAINTE-CATHERINE
More informationPerceived Pitch of Synthesized Voice with Alternate Cycles
Journal of Voice Vol. 16, No. 4, pp. 443 459 2002 The Voice Foundation Perceived Pitch of Synthesized Voice with Alternate Cycles Xuejing Sun and Yi Xu Department of Communication Sciences and Disorders,
More informationDistortion products and the perceived pitch of harmonic complex tones
Distortion products and the perceived pitch of harmonic complex tones D. Pressnitzer and R.D. Patterson Centre for the Neural Basis of Hearing, Dept. of Physiology, Downing street, Cambridge CB2 3EG, U.K.
More informationQuarterly Progress and Status Report. A note on the vocal tract wall impedance
Dept. for Speech, Music and Hearing Quarterly Progress and Status Report A note on the vocal tract wall impedance Fant, G. and Nord, L. and Branderud, P. journal: STL-QPSR volume: 17 number: 4 year: 1976
More informationImagine the cochlea unrolled
2 2 1 1 1 1 1 Cochlea & Auditory Nerve: obligatory stages of auditory processing Think of the auditory periphery as a processor of signals 2 2 1 1 1 1 1 Imagine the cochlea unrolled Basilar membrane motion
More informationBark and ERB Bilinear Transforms
Bark and ERB Bilinear Transforms Julius O. Smith III Center for Computer Research in Music and Acoustics (CCRMA), Stanford University Stanford, CA 9435 USA Jonathan S. Abel Human Factors Research Division
More informationModelling the sensation of fluctuation strength
Product Sound Quality and Multimodal Interaction: Paper ICA016-113 Modelling the sensation of fluctuation strength Alejandro Osses Vecchi (a), Rodrigo García León (a), Armin Kohlrausch (a,b) (a) Human-Technology
More informationScienceDirect. Unsupervised Speech Segregation Using Pitch Information and Time Frequency Masking
Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 46 (2015 ) 122 126 International Conference on Information and Communication Technologies (ICICT 2014) Unsupervised Speech
More informationAUDL 4007 Auditory Perception. Week 1. The cochlea & auditory nerve: Obligatory stages of auditory processing
AUDL 4007 Auditory Perception Week 1 The cochlea & auditory nerve: Obligatory stages of auditory processing 1 Think of the ear as a collection of systems, transforming sounds to be sent to the brain 25
More informationPrinciples of Musical Acoustics
William M. Hartmann Principles of Musical Acoustics ^Spr inger Contents 1 Sound, Music, and Science 1 1.1 The Source 2 1.2 Transmission 3 1.3 Receiver 3 2 Vibrations 1 9 2.1 Mass and Spring 9 2.1.1 Definitions
More informationSpeech Synthesis; Pitch Detection and Vocoders
Speech Synthesis; Pitch Detection and Vocoders Tai-Shih Chi ( 冀泰石 ) Department of Communication Engineering National Chiao Tung University May. 29, 2008 Speech Synthesis Basic components of the text-to-speech
More informationPattern Recognition. Part 6: Bandwidth Extension. Gerhard Schmidt
Pattern Recognition Part 6: Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Institute of Electrical and Information Engineering Digital Signal Processing and System Theory
More informationDERIVATION OF TRAPS IN AUDITORY DOMAIN
DERIVATION OF TRAPS IN AUDITORY DOMAIN Petr Motlíček, Doctoral Degree Programme (4) Dept. of Computer Graphics and Multimedia, FIT, BUT E-mail: motlicek@fit.vutbr.cz Supervised by: Dr. Jan Černocký, Prof.
More informationLinguistics 401 LECTURE #2. BASIC ACOUSTIC CONCEPTS (A review)
Linguistics 401 LECTURE #2 BASIC ACOUSTIC CONCEPTS (A review) Unit of wave: CYCLE one complete wave (=one complete crest and trough) The number of cycles per second: FREQUENCY cycles per second (cps) =
More information