The EarSpring Model for the Loudness Response in Unimpaired Human Hearing

Similar documents
Chapter 12. Preview. Objectives The Production of Sound Waves Frequency of Sound Waves The Doppler Effect. Section 1 Sound Waves

Sound is the human ear s perceived effect of pressure changes in the ambient air. Sound can be modeled as a function of time.

Hearing and Deafness 2. Ear as a frequency analyzer. Chris Darwin

Perception of pitch. Importance of pitch: 2. mother hemp horse. scold. Definitions. Why is pitch important? AUDL4007: 11 Feb A. Faulkner.

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.

Perception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.

AUDL GS08/GAV1 Signals, systems, acoustics and the ear. Loudness & Temporal resolution

Dynamic Vibration Absorber

Chapter 3. Meeting 3, Psychoacoustics, Hearing, and Reflections

AUDL 4007 Auditory Perception. Week 1. The cochlea & auditory nerve: Obligatory stages of auditory processing

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

APPENDIX MATHEMATICS OF DISTORTION PRODUCT OTOACOUSTIC EMISSION GENERATION: A TUTORIAL

Imagine the cochlea unrolled

Psycho-acoustics (Sound characteristics, Masking, and Loudness)

Week 1. Signals & Systems for Speech & Hearing. Sound is a SIGNAL 3. You may find this course demanding! How to get through it:

Phase and Feedback in the Nonlinear Brain. Malcolm Slaney (IBM and Stanford) Hiroko Shiraiwa-Terasawa (Stanford) Regaip Sen (Stanford)

Physics 101. Lecture 21 Doppler Effect Loudness Human Hearing Interference of Sound Waves Reflection & Refraction of Sound

describe sound as the transmission of energy via longitudinal pressure waves;

Acoustics, signals & systems for audiology. Week 4. Signals through Systems

SOUND 1 -- ACOUSTICS 1

Principles of Musical Acoustics

Signals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend

Preliminary study of the vibration displacement measurement by using strain gauge

Results of Egan and Hake using a single sinusoidal masker [reprinted with permission from J. Acoust. Soc. Am. 22, 622 (1950)].

Distortion products and the perceived pitch of harmonic complex tones

SOUND. Second, the energy is transferred from the source in the form of a longitudinal sound wave.

Musical Acoustics, C. Bertulani. Musical Acoustics. Lecture 14 Timbre / Tone quality II

8.3 Basic Parameters for Audio

Detection of external stimuli Response to the stimuli Transmission of the response to the brain

CHAPTER 6 INTRODUCTION TO SYSTEM IDENTIFICATION

of bamboo. notes. in the D4. learning to. amplitudes and. pipe. The the.5% to. each. individual. 2% range.

Standard Octaves and Sound Pressure. The superposition of several independent sound sources produces multifrequency noise: i=1

Chapter 10: Compensation of Power Transmission Systems

ENGINEERING STAFF REPORT. The JBL Model L40 Loudspeaker System. Mark R. Gander, Design Engineer

HCS 7367 Speech Perception

A102 Signals and Systems for Hearing and Speech: Final exam answers

What is Sound? Part II

Digitally controlled Active Noise Reduction with integrated Speech Communication

NCERT solution for Sound

PHYS 130 Exam #2 3/26/2009

Lecture 18 Stability of Feedback Control Systems

AUDL Final exam page 1/7 Please answer all of the following questions.

THE MATLAB IMPLEMENTATION OF BINAURAL PROCESSING MODEL SIMULATING LATERAL POSITION OF TONES WITH INTERAURAL TIME DIFFERENCES

Chapter 2 A Silicon Model of Auditory-Nerve Response

Copyright 2009 Pearson Education, Inc.

George Mason University Signals and Systems I Spring 2016

Lesson 3 Measurement of sound

Week I AUDL Signals & Systems for Speech & Hearing. Sound is a SIGNAL. You may find this course demanding! How to get through it: What is sound?

Intext Exercise 1 Question 1: How does the sound produced by a vibrating object in a medium reach your ear?

Force versus Frequency Figure 1.

FM AND BESSEL ZEROS TUTORIAL QUESTIONS using the WAVE ANALYSER without a WAVE ANALYSER...137

Whole geometry Finite-Difference modeling of the violin

Section 1: Sound. Sound and Light Section 1

A mechanical wave is a disturbance which propagates through a medium with little or no net displacement of the particles of the medium.

An introduction to physics of Sound

SOUND QUALITY EVALUATION OF FAN NOISE BASED ON HEARING-RELATED PARAMETERS SUMMARY INTRODUCTION

A study of Vibration Analysis for Gearbox Casing Using Finite Element Analysis

CHAPTER 12 SOUND. Sound: Sound is a form of energy which produces a sensation of hearing in our ears.

COM325 Computer Speech and Hearing

Chapter 15 Supplement HPS. Harmonic Motion

MUS 302 ENGINEERING SECTION

Sound. Production of Sound

Noise from Pulsating Supercavities Prepared by:

Chapter 7. Waves and Sound

Analysis on Acoustic Attenuation by Periodic Array Structure EH KWEE DOE 1, WIN PA PA MYO 2

The psychoacoustics of reverberation

Tuesday, March 22nd, 9:15 11:00

Knowledge Integration Module 2 Fall 2016

Lab 9 AC FILTERS AND RESONANCE

Experiment 1 LRC Transients

Name Date Class _. Holt Science Spectrum

Lecture 7: Superposition and Fourier Theorem

Experiment 2: Transients and Oscillations in RLC Circuits

Signals, Sound, and Sensation

HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS

Physics I Notes: Chapter 13 Sound

Lecture PowerPoints. Chapter 12 Physics: Principles with Applications, 6 th edition Giancoli

Psychology in Your Life

Back to. Communication Products Group. Technical Notes. Adjustment and Performance of Variable Equalizers

Chapter 4 PSY 100 Dr. Rick Grieve Western Kentucky University

SOUND & MUSIC. Sound & Music 1

WHITHER DITHER: Experience with High-Order Dithering Algorithms in the Studio. By: James A. Moorer Julia C. Wen. Sonic Solutions San Rafael, CA USA

LCR CIRCUITS Institute of Lifelong Learning, University of Delhi

THE USE OF ARTIFICIAL NEURAL NETWORKS IN THE ESTIMATION OF THE PERCEPTION OF SOUND BY THE HUMAN AUDITORY SYSTEM

Laboratory Assignment 4. Fourier Sound Synthesis

Active Filter Design Techniques

UNIT-3. Electronic Measurements & Instrumentation

Preview. Sound Section 1. Section 1 Sound Waves. Section 2 Sound Intensity and Resonance. Section 3 Harmonics

Auditory filters at low frequencies: ERB and filter shape

3D Distortion Measurement (DIS)

Fundamentals of Environmental Noise Monitoring CENAC

Auditory modelling for speech processing in the perceptual domain

TEAK Sound and Music

UNIT 2. Q.1) Describe the functioning of standard signal generator. Ans. Electronic Measurements & Instrumentation

Audio Noise Reduction and Masking

Since the advent of the sine wave oscillator

Final Exam Study Guide: Introduction to Computer Music Course Staff April 24, 2015

(1.3.1) (1.3.2) It is the harmonic oscillator equation of motion, whose general solution is: (1.3.3)

Lab 9 AC FILTERS AND RESONANCE

Measurement of weighted harmonic distortion HI-2

Transcription:

The EarSpring Model for the Loudness Response in Unimpaired Human Hearing David McClain, Refined Audiometrics Laboratory, LLC December 2006 Abstract We describe a simple nonlinear differential equation for a harmonic oscillator with power dependent stiffness that adequately describes the main features of unimpaired human loudness perception. Solutions of the model are presented, and a mechanism for relating model vibrational power to Sones loudness measure is described. This model is used for estimating hearing correction targets from impaired levels to what would be heard by unimpaired hearing. Contents 1 Introduction 1 2 Qualitative Description of the Model 2 3 Equation of the Model 4 4 Estimation of Model Parameters 5 5 Frequency Dependence 9 6 A Model for HyperRecruitment 9 1 Introduction EarSpring is the name for our model for the whole-system human loudness response for unimpaired hearing. It is an ad-hoc empirically driven model that captures the essential features of human hearing. It is not specifically a cochlear model. Rather, it models the 1

entire system of hearing, incorporating the combined effects of chochlea, afferent 8th nerve, brain, and efferent 8th nerve components, and their complex interactions. Our aim is to understand how unimpaired human hearing responds to signals at various intensity levels. We utilize this normal hearing loudness response to help derive needed corrections for impaired listeners. EarSpring shows us how loudly sounds ought to appear. And so the model furnishes correction targets. We utilize empirical measurements to establish values for the model parameters based on: The Sones ratio between the hearing threshold level and the reference level of 40 dbspl at 1 khz. The detuning (flattening) of tones by 75 cents as tones range in intensity from 40 dbspl to 90 dbspl near 1 khz. EarSpring is able to successfully demonstrate the main features of loudness perception: Linear response to near-threshold level sounds, although for normal daily environmental sound levels this is of little importance. Nearly cube-root behavior for more common sound levels, with intensities above 40 dbspl near 1 khz. Computer simulations of the EarSpring equation demonstrate the production of odd harmonic distortion products from loud signals, in agreement with Inter-Modulation Distortion (IMD) products measured with loud bass-frequency excitation signals. 2 Qualitative Description of the Model To within experimental accuracy, and to first order, we find that the loudness perception at any one pitch, of unimpaired human hearing, can be adequately described by the measure of average vibration power in a nonlinear damped harmonic oscillator. The nonlinearity is provided by a spring constant whose stiffness grows with increasing vibration power. We propose that its nominal resonant frequency, the pitch sensation it produces, is that frequency at which the oscillator resonates at very low power levels, where its behavior is essentially the same as a simple linear damped harmonic oscillator. As input sound levels grow in intensity, the stiffness of the model s spring constant increases in proportion to the vibration power of the induced oscillations. At high enough input levels this nonlinearity begins to dominate the loudness response. 2

The result is essentially a linear damped harmonic oscillator for threshold level sounds, and which produces a cube-root response over sound levels in the range of common daily sound intensities above 40 dbspl. As sound intensity increases, and the stiffness constant grows, the natural resonant frequency of the oscillator shifts toward higher frequencies, and moves above the excitation frequency. This shifting produces two effects: loudness compression and pitch shifting. As the natural resonant frequency of the oscillator shifts above the excitation frequency, the degree of absorbed excitation diminishes with increasing sound intensity. This gives rise to the compression observed as cube-root, wherein for a doubling of apparent sound loudness we require an increase of 10 db in sound intensity. And if we conceptualize the pitch perception of human hearing as corresponding to a parallel bank of such oscillators, those in adjacent frequency bands gradually shift their own resonant frequencies higher in response to increasing sound intensity. And so the lower bands, whose stimulation corresponds to some lower pitch in our sensation, become increasingly excited at higher sound intensities. As a result one can imagine that the excitation will produce an apparent pitch that is increasingly flatter than what the excitation would produce at low intensity levels. Both of these effects are readily measured in the laboratory. The cube-root behavior is well known and readily verified for common sound levels above 40 dbspl. The pitch flattening has been measured as approximately 75 cents for an overall sound intensity increase from 40 dbspl to 90 dbspl, near 1 khz. Threshold region sounds, those below 40 dbspl, are uncommon in normal daily experience, and can generally only be heard well in isolation booths. While important for demonstrating the linearity of human loudness response at very low sound intensity levels, this response is not very important for the description of common daily hearing experiences. We further postulate that our model is the same for all frequency bands when expressed in terms of Phon excitation. The ear canal modifies the sensitivity of our hearing as sounds range in frequency due to resonant enhancement for sounds near 4 khz. Higher order modifications occur as a result of pinnea and head shadowing directional effects. Transmission of sound into the cochlea and its coupling to the basilar membrane also varies in efficiency with frequency. Our assumption is that there is a basic mechanism at work, whose description at different frequencies produces similar results for similar excitation levels reaching the EarSpring. Hence, in absolute terms, and to first order, one must first apply an Equal-Loudness correction to absolute intensities expressed in dbspl, obtain equivalent excitation levels as expressed in Phon, before applying the results of our model. 3

Bear in mind that this is a whole-system model, not a cochlear model. It is not known specifically where such a spring constant could be located, how the oscillator s power is sensed, nor what might produce the rising stiffness with increasing excitation levels. It is conceivable that a good portion of this behavior does reside in the cochlea, where power perception might be aided by stimulation of inner hair cells in response to vibration of the basilar membrane, and where some of the increasing stiffness may arise from afferent nerve stimulation of outer hair cells to diminish the flexibility of the basilar membrane. But we are not so much interested in cochlea specific mechanisms, but rather in the overall loudness response. Cochlear behavior is only one component of the entire system composed of cochlea - efferent nerves - brain - afferent nerves. Hearing is the result of individual behaviors from, and interactions between, the subcomponents of this system. We find that our model is the simplest that adequately describes this whole system. 3 Equation of the Model The EarSpring model is a second order nonlinear differential equation relating the amplitude of vibration of a mechanical system to a driving force. We view the hearing mechanism as a parallel bank of EarSpring systems, each tuned to a different resonant frequency. The equation is very nearly the same as that for a conventional damped harmonic oscillator, except that it incorporates a stiffness term which increases in magnitude with the average power of vibration: where: [ d 2 dt 2 + 2β d dt + k ( 1 + γ y 2 ) ] y(t) = F (t) t = time y(t) = amplitude of vibration F (t) = driving force y 2 = mean power of vibration β = damping constant k = spring constant γ = coefficient of power dependence of spring constant For any fixed level of excitation, the average power of vibration y 2 in the steady state will be a constant. Hence, we see that this describes a harmonic oscillator with poles in its transfer function at iβ ± ω 0, where ω 0 = k γ β 2, and k γ = k ( 1 + γ y 2 ) : 4

H(ω) = 1 ω 2 + 2iβω + k (1 + γ y 2 ) While increasing damping β moves the poles off the real frequency axis and decreases the natural resonant frequency, an increase in the power of vibration moves the poles, in a direction parallel to the frequency axis, toward frequencies further from the origin. Spring stiffness increases, which raises the resonant frequency. 4 Estimation of Model Parameters In the following discussion we will view our system as being excited by a sinusoidal force at the undamped natural frequency k. We can make a frequency independent equation by dividing the EarSpring equation through by k. We will be measuring the system response at this same frequency, and so the reduced transfer function, at that frequency, becomes: H(ω 0 = k) = 1 2i ˆβ + γ y 2 where, now, ˆβ is the frequency normalized damping constant ( ˆβ = β/ k). Since increasing vibrational amplitude shifts the resonant frequency of the EarSpring system, excitation applied with increasing intensity at the undamped resonant frequency will have a gradually lessened effect as the system resonance moves further from the excitation frequency. We can estimate the γ parameter on the basis of tone flattening as excitation intensity rises from 40 to 90 dbspl 1. We might view tone flattening as the consequence of adjacent channel EarSprings shifting their resonant frequencies higher, thereby causing increasing excitation of the next lower channel, and hence the excitation appears ever flatter. 1 For frequencies near 1 khz. Units of dbspl are inappropriate for use at other frequencies. We need the excitation presented to the oscillator. And so units of Phon, derived from dbspl intensities adjusted by a frequency dependent Equal-Loudness correction, are appropriate at all frequencies. At 1 khz these systems of measurement are defined such that 40 Phon = 40 dbspl, a difference of 10 dbspl equals a difference of 10 Phon, and where 0 dbspl is the nominal threshold for hearing which corresponds to an RMS pressure amplitude of 0.0002 dyne/cm 2. At any other frequency the threshold in dbspl will differ, but will always be denoted as 0 Phon, and the scaling of Phon differences to dbspl differences will also be different. 5

In frequency normalized terms, the expression for the resonant frequency of our EarSpring transfer function is: ˆω = (1 + γ y 2 ) ˆβ 2 Taking the ratio of this equation for the two different excitation levels and then solving for the γ power term gives us: Γ 40 1 2 γ a 40 2 = ( F 2 90 1 ) ( (1 S90 F90) 2 ˆβ 2) where, a 40, ˆω 40 = vibration amplitude and resonant frequency at 40 dbspl driving force a 90, ˆω 90 = vibration amplitude and resonant frequency at 90 dbspl driving force S 90 a 90 2 / a 40 2 F 90 ˆω 90 /ˆω 40 and where we have assumed a sinusoidal driving force. The leading 1 2 factor arises from the fact that the mean power of a sinusoid of amplitude a is proportional to 1 2 a 2. The pitch flattening ratio F 90 has been measured in the lab as approximately 75 cents 1.044. We don t yet know the individual values for a 40 and a 90, but we can estimate their magnitude ratio by using the approximate cube-root behavior of loudness sensation with rising sound intensity. Hence a difference of 50 db in sound intensities should produce a ratio S 90 17 db. From these expressions, we can utilize the ratio of the transfer function at two input levels to find the amplitude of vibration for any other input signal level in terms of its ratio to a 40 as: ( S P a P 2 ( ) P 2 4 ˆβ ) 2 + Γ 2 a 40 2 = 40 100P thr (4 ˆβ 2 + (Γ 40 S P ) 2) where P is the excitation pressure level, and P thr is the threshold pressure level corresponding to 0 dbspl. This is a cubic equation in the amplitude magnitude-squared ratio, S P. 6

Benade 2 reports that the ratio S 0 for threshold level stimuli, where P = P thr, ought to be 1/ (22) 2 0.002. So, from the two equations above, we can solve for ˆβ. With this initial estimate for ˆβ we can solve for an improved ratio S 90, and then iterate this process by recomputing Γ 40 and then ˆβ until we converge to the value for ˆβ. Once the value of ˆβ has been found, we can map the amplitude magnitude ratio S P over all input dbspl levels and obtain the solution shown in Figure 1. Figure 1: EarSpring Amplitude versus Driving Force. Linear behavior is apparent near threshold level intensities. At normal sound levels above 40 dbspl we see cube-root compression. As we see, the behavior of the system near threshold sound levels is nearly linear. At sound intensities above 40 dbspl, where most normal daily sounds are found, the behavior is nearly cube-root. The average slope of the graphed solution, over the range from 40 to 100 dbspl, is 1 2 α 0.335.3 2 Benade, Arthur H., Fundamentals of Musical Acoustics - Second, Revised Edition, Dover, 1990, pg. 252. 3 See the next page for the meaning of α. 7

Figure 2 shows the slope of the curve in Figure 1. Near threshold levels its slope is very close to unity, indicating linear behavior. At sound intensity levels much above 40 dbspl, the slope has settled to a nearly constant value of 1/3. Figure 2: EarSpring Slope versus Driving Force. These slope values are 1/2 of the characteristic exponent α discussed in the text. We make the identification of the EarSpring vibration power ratio, S P, with the Sones level for each source intensity level. At threshold sound levels Benade 4 gives the expression for Sones as: ( P Sones = 22P thr ) 2 for P P thr, near threshold levels Exponent α arises in the expression for Sones at intensity levels above 40 dbspl: ( ) P α Sones = for P/P thr > 40 dbspl 100P thr Benade gives the value for α 0.6, but this value is an average over many individuals whose measured values ranged anywhere from 0.5 to 0.7. From our EarSpring model we derive an average value for α 0.669 over the intensity range from 40 to 100 dbspl. 4 ibid 8

5 Frequency Dependence The identification of Sones with the EarSpring vibrational power ratio S P versus the sound intensity P in dbspl can only be correct at 1 khz. The Sones scale is calibrated as 1 Sone for a 40 dbspl sinewave at 1 khz. At other frequencies, the outer ear modifies the base relationship in a manner first shown by Fletcher and Munson. Unimpaired hearing is more sensitive near 4 khz than anywhere else, and this arises from the resonant enhancement provided by the ear canal. Hence to be more correct, the driving force in the EarSpring equation should be identified with units of Phons. These units, defined as identical to dbspl at 1 khz, vary in both their zero point and scaling relative to dbspl at other frequencies. But iso-phon contours in the dbspl-frequency plane correspond to the sensation of equal loudness, and so we make the assumption that these units also describe the degree of coupling of sound to our EarSpring oscillators. When EarSpring is viewed in terms of excitation levels presented in this manner the solution obtained above becomes independent of frequency. In order to utilize the EarSpring solution for hearing correction targets, we must first estimate the sound intensity level in each frequency band in dbspl. We then apply the frequency dependent Equal-Loudness correction to these sound intensities to derive Phon levels above threshold presented to the EarSpring. After this conversion the EarSpring model can furnish sensible target apparent loudness levels, in Sones, for use in hearing corrections. 6 A Model for HyperRecruitment It should be clear that if the γ term in the model stiffness were too small, the suppression of excessive vibration amplitude at high sound field intensities becomes impaired. Figure 3 shows the consequence of a γ term that is too small by a factor of 20. The essential feature of HyperRecruitment at high sound-field intensities is the perception of sound being too loud. That is shown by the height of the red curve above the nominal unimpaired hearing curve shown in green. Such an insufficiency in γ might be caused by the deterioration of outer hair cells in the cochlea, or a failure of the brain to control vibrational amplitudes of the basilar membrane through excitation of the outer hair cells. Diminished hearing, through the destruction of inner hair cells alone, is not able to account for HyperRecruitment because at high enough sound-field intensities the sensation 9

of sound produced by even a limited number of inner hair cells should be sufficient to correct excessive basilar membrane vibration amplitudes. Figure 3: EarSpring model for HyperRecruitment induced by an insufficiency in the γ term of the model stiffness. Unimpaired hearing response in green, HyperRecruitment in red. We suggest that a γ insufficiency is only part of the mechanism for HyperRecruitment. Another paper discusses the more general cases and models for Recruitment, HyperRecruitment, and Decruitment, and their correction to EarSpring target behaviors. 10