INTRODUCTION. General Structure
|
|
- Isaac Blake
- 5 years ago
- Views:
Transcription
1 Transposed carrier and envelope reconstruction Haptic feature substitution Pitch and Envelope extraction EMD decomposition (mus. features) Spatial vibrotactile display Synth acoustic signal Auditory EMD Analysis V S A pitch extraction on/offset detection timbre analysis Beat analysis loudness estimation polyphonic analysis A V S acoustic signal flow vibrational signal flow symbolic information flow FIGURE : General model structure. INTRODUCTION Traditionally, specialized haptic displays have only played a minor role in music performance, although current haptic technology shows great potential for both normal-hearing and hearingimpaired listeners. In this paper, we report on the potential of haptic devices for telematic music performances with a strong focus on a sensory-substitution device for hearing impaired musicians that can be used within telepresence scenarios. To our experience, haptic signals have been proven useful for (i) the reproduction of floor vibrations of music performances [, 2], (ii) to use haptics as a side channel for additional communication within a music performance e.g., using the Vibrobyte, a hand-held wireless haptic display [3], and (iii) to use haptics to substitute for sound with hearing-impaired users. Sensory Substitution has been described in numerous publications [4, 5, 6, 7, 8], and can be defined as the act of presenting information from one modality using another, often to make this information directly accessible to sensory-impaired users. It is well known that vibrations can be extremely important for hearing impaired musicians/listeners. The world-renowned percussionist, Evelyn Glennie, for example, profoundly deaf since the age of 2, always performs barefoot to better sense the floor vibrations that result from the sounding musical instruments. Glennie is known for her preference to feel the music through vibration rather then relying on her residual hearing as it is processed through hearing aids [9]. While a few investigations have dealt with augmenting music through haptic interfaces [,, 2, 3] these papers did not focus on impaired listeners. SENSORY SUBSTITUTION SYSTEM General Structure One of our initial aims was to understand to what extent the haptic and auditory sensory systems are comparable from a perceptual viewpoint. Western traditional music theory is based on several fundamental parameters including melody, rhythm, harmony, texture and form. While
2 FIGURE 2: Haptic Sensory Substitution Prototype. there is little doubt that the temporal resolution of touch is fine enough (and much better than vision) to resolve rhythmic patterns and form, the frequency resolution is not high enough to directly extract pitch information from haptically-displayed audio signals. Numerous studies have investigated tactile frequency selectivity, mostly in the context of work safety or fundamental psychophysical research e.g., see [4, 5, 6]. Interestingly, to the best of our knowledge, none of these studies asked the question that is fundamental for haptic music perception: Can haptic frequency intervals have the same perceptual quality as acoustic frequency intervals, and thus do we perceive them as musical intervals? To our surprise, this was to some extent the case in the low-frequency range [7]. To investigate this normal hearing listeners wore ear muffs and in-ear headphones with pink noise to avoid stimulating the auditory system via bone conduction. Our developed prototype, see Fig. [8], acknowledges the substantial differences between the senses of touch and audition. Unlike he case of the auditory organ, we do not have mechanisms to segregate individual frequency components for touch. While the somatosensory system has different touch receptors, initial test showed that one cannot discriminate between the lower and higher frequencies the Meisner Corpuscles and the Pacinian Corpuscles are tuned to. Further, the most sensitive region is lower in frequency for touch compared to the auditory sense. Consequently, our sensory-substitution prototype transposes the digitized acoustic music signal into the low-frequency range after removing the overtone spectrum, which is distracting for haptic sensation. Further, the information can be spatially distributed using an 8-channel vibrotactile board (see Fig. 2). There are at least two ways of coding polyphonic pitch for a haptic display. Both procedures take advantage of the excellent spatial resolution of the haptic sensory system as well as its inability to segregate an incoming signal into different frequency bands. The first method is to capture monophonic musical instruments individually. This can be done using closely positioned microphones and then processing and presenting each instrument with a different actuator on the 8-channel device. In the second approach, a complex music signal is analyzed using a polyphonic pitch model and then each pitch channel is processed and sent to a different actuator. In both cases, the user can use different fingers for each actuator to perceive different pitches simultaneously. POLYPHONIC PITCH PERCEPTION MODEL Methods The pitch model builds on a functional model of the auditory periphery and previous pitch perception models [9, 2, 2, 22]. In the first step, to simulate the behavior of the basilar membrane, the signal is sent through a Gammatone filterbank with 28 bands to segregate the sound
3 into different auditory bands. Then, the signal frequency f n in each band n is estimated using auto-correlation, measuring the delay τ n between the main and the largest side peak: f n = τ n. () A novel aspect of the model is that not only the frequency is measured in each frequency band, but also the pitch strength by calculating the amplitude ratio a between the largest side peak and the main peak. Further, the deviation b between the estimated frequency f n and the center of the frequency band f c,n is calculated. Next, all results are grouped into four categories:. a >.9, b.3 octaves ( + symbols) 2. a >.9, b >.3 octaves ( symbols) 3. a.9, b.3 octaves ( symbols) 4. a.9, b >.3 octaves ( symbols) FIGURE 3: Pitch estimation for a 44-Hz sinusoidal signal (left graph) and a 44-Hz tone complex (right graph). Results The graphs on the left in Fig. 3 show the results of the pitch model for a 44-Hz sinusoid. The top graph shows the broadband autocorrelation function, the center graph the Fourier Transformation of the signal, and the bottom graph depicts the excitiation of the auditory bands (solid black curve). For the curve, the energy in each of the 28 bands was measured and plotted at the center frequency of the band. All values for the Group (a >.9, b.3, + symbols) are located at 44 Hz, the frequency of the sinusoid, as indicated by the grey curve. The height of the values represent the energy of the band in which the frequency was measured. The values for Group 2 (a >.9, b >.3 octaves, symbols), also point to 44 Hz. They were measured in the adjacent side bands. All other values (Groups 3 and 4) were measured from the noise spectrum at low energies and do not represent the frequency of the sinusoid. The right graphs of Fig. 3 depict the same context but this time for a tone complex with eight higher harmonics at integer multiples of the fundamental frequency: f = n f. The amplitude of the tone complex rolls off with /f. Again, all values for Groups and 2 ( + and symbols) point to the fundamental frequency of 44 Hz, even those that belong to the higher harmonics.
4 FIGURE 4: Same as Fig. 3, but for 22-Hz signals Figure 4 shows the results for a /f tone complex at a lower fundamental frequency of 22 Hz (left graphs). Again, the results in all harmonics point to the fundamental, with the exception of two values in the octave region (44 Hz). It is not clear why in this case the octave is recognized but not for the 44-Hz tone complex. This will be further investigated. For higher harmonics, more than one overtone falls into the same auditory band. The overtones interfere with each other, and based on this interference the autocorrelation method identifies the common fundamental f. For the same reason, the algorithm is able to detect a missing fundamental. The right graphs show the results for the same tone complex, but this time the fundamental of 22 Hz was removed. Still, most values point to 22 Hz. Clearly those values belong to Group 2 (a >.9, b >.3 octaves, symbols) since there is no energy around 22 Hz and the values were computed at higher frequency bands FIGURE 5: Same as Figs. 2 and 4, but for polyphonic tone clusters. Left: Sinusoids with frequencies of 22 Hz, 262 Hz and 33 Hz, right: tone complexes with frequencies of 22 Hz, 262 Hz, 33 Hz and 88 Hz. Finally, chord complexes were analyzed using the model as depicted in Fig. 5. The left graph shows a triad of sinusoids with frequencies of 22 Hz, 262 Hz and 33 Hz. The model correctly identifies all tones. The right graphs shows a cluster of / f tone complexes with the following fundamental frequencies: 22 Hz, 262 Hz, 33 Hz and 88 Hz. The model identifies all fundamental frequencies correctly, but also a number of octaves, for example at 56 Hz which is the octave of the 262-Hz tone.
5 ACKNOWLEDGMENT This material is based upon work supported by the National Science Foundation under Grant No REFERENCES [] C. Abercrombie and J. Braasch, A method for multimodal auralization of audio-tactile stimuli from acoustic and structural measurements, in Proc. of the Convention of the Audio Eng. Soc. New York, NY, volume 27 (29), paper Number [2] C. L. Abercrombie, Influence of vibration and stage construction on the perception of musical performance, Master s thesis, Rensselaer Polytechnic Institute (29). [3] K. McDonald, D. Kouttron, C. Bahn, J. Braasch, and P. Oliveros, The vibrobyte: A haptic interface for co-located performance, in Procceding of the 9th International conference on New Interfaces for Musical Expression (NIME) (Pittsburgh, PA) (29), paper ID 6. [4] K. Kaczmarek and J. Webster, Electrotactile and vibrotactile displays for sensory substitution systems, IEEE Transactions on Biomedical Engineering 38, 6 (99). [5] M. Massimino and T. Sheridan, Sensory substitution for force feedback in teleoperation, Presence 2, (994). [6] P. Bach-y-Rita and S. W. Kercel, Sensory substitution and the human machine interface, Trends in Cognitive Sciences 7, (23). [7] J. Zelek, S. Bromley, D. Asmar, and D. Thompson, A haptic glove as a tactile-vision sensory substitution for wayfinding, Journal of Visual Impairment & Blindness (JVIB) 97, (23). [8] P. Bach-y-Rita, Tactile sensory substitution studies, Annals of the New York Academy of Sciences 3, 83 9 (24). [9] H. Finch, Rain on my parade, please: Evelyn Glennie tells why she relishes getting a soaking but is angry about the future, The Times, July 9 (24). [] M. S. O Modhrain, Playing by feel: incorporating haptic feedback into computer-based musical instruments, Ph.D. thesis, Stanford, CA, USA (2), adviser-chafe, Chris. [] A. Chang, S. O Modhrain, R. Jacob, E. Gunther, and H. Ishii, ComTouch: design of a vibrotactile communication device, in Proceedings of the 4th conference on Designing interactive systems: processes, practices, methods, and techniques, (ACM) (22). [2] E. Gunther, G. Davenport, and S. O Modhrain, Cutaneous grooves: composing for the sense of touch, in NIME 2: Proceedings of the 22 conference on New interfaces for musical expression, 6 (National University of Singapore, Singapore, Singapore) (22). [3] S. O Modhrain and G. Essl, Pebblebox and crumblebag: tactile interfaces for granular synthesis, in NIME 4: Proceedings of the 24 conference on New interfaces for musical expression, (National University of Singapore, Singapore, Singapore) (24). [4] M. E. Altinsoy, Audiotaktile interaktion in virtuellen umgebungen, Ph.D. thesis, Ruhr- Universität Bochum (26).
6 [5] M. J. Zigmond, F. E. Bloom, S. Landis, J. L. Roberts, and L. R. Squire, Fundamental Neuroscience (Peninsula Publishing) (999). [6] E. Kandel, J. Schwartz, and T. Jessell, Principles of neural science (McGraw-Hill New York) (2). [7] D. Egloff, A vibrotactile music system based on sensory substitution, Master s thesis, Rensselaer Polytechnic Institute (2). [8] D. Egloff, J. Braasch, P. Robinson, D. Van Nort, and T. Krueger, A vibrotactile music system based on sensory substitution, The Journal of the Acoustical Society of America 29, 2582 (2). [9] A. de Cheveigné and H. Kawahara, Yin, a fundamental frequency estimator for speech and music, J. Acoust. Soc. Am., (22). [2] A. Cheveigné, Pitch perception models, in The Psychophysics of Pitch, edited by C. J. Plack and A. J. Oxenham, (Springer, Berlin, Heidelberg, New York) (25). [2] J. Licklider, A duplex theory of pitch perception, Cellular and Molecular Life Sciences 7, (95). [22] R. Meddis and L. O Mard, A unitary model of pitch perception, The Journal of the Acoustical Society of America 2, 8 (997).
Perception of pitch. Importance of pitch: 2. mother hemp horse. scold. Definitions. Why is pitch important? AUDL4007: 11 Feb A. Faulkner.
Perception of pitch AUDL4007: 11 Feb 2010. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence Erlbaum, 2005 Chapter 7 1 Definitions
More informationPerception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb A. Faulkner.
Perception of pitch BSc Audiology/MSc SHS Psychoacoustics wk 5: 12 Feb 2009. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence
More informationThe Effect of Frequency Shifting on Audio-Tactile Conversion for Enriching Musical Experience
The Effect of Frequency Shifting on Audio-Tactile Conversion for Enriching Musical Experience Ryuta Okazaki 1,2, Hidenori Kuribayashi 3, Hiroyuki Kajimioto 1,4 1 The University of Electro-Communications,
More informationPerception of pitch. Definitions. Why is pitch important? BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb A. Faulkner.
Perception of pitch BSc Audiology/MSc SHS Psychoacoustics wk 4: 7 Feb 2008. A. Faulkner. See Moore, BCJ Introduction to the Psychology of Hearing, Chapter 5. Or Plack CJ The Sense of Hearing Lawrence Erlbaum,
More information19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007
19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007 MODELING SPECTRAL AND TEMPORAL MASKING IN THE HUMAN AUDITORY SYSTEM PACS: 43.66.Ba, 43.66.Dc Dau, Torsten; Jepsen, Morten L.; Ewert,
More informationHearing and Deafness 2. Ear as a frequency analyzer. Chris Darwin
Hearing and Deafness 2. Ear as a analyzer Chris Darwin Frequency: -Hz Sine Wave. Spectrum Amplitude against -..5 Time (s) Waveform Amplitude against time amp Hz Frequency: 5-Hz Sine Wave. Spectrum Amplitude
More informationSound is the human ear s perceived effect of pressure changes in the ambient air. Sound can be modeled as a function of time.
2. Physical sound 2.1 What is sound? Sound is the human ear s perceived effect of pressure changes in the ambient air. Sound can be modeled as a function of time. Figure 2.1: A 0.56-second audio clip of
More informationFrom Encoding Sound to Encoding Touch
From Encoding Sound to Encoding Touch Toktam Mahmoodi King s College London, UK http://www.ctr.kcl.ac.uk/toktam/index.htm ETSI STQ Workshop, May 2017 Immersing a person into the real environment with Very
More informationThe psychoacoustics of reverberation
The psychoacoustics of reverberation Steven van de Par Steven.van.de.Par@uni-oldenburg.de July 19, 2016 Thanks to Julian Grosse and Andreas Häußler 2016 AES International Conference on Sound Field Control
More informationMusical Acoustics, C. Bertulani. Musical Acoustics. Lecture 14 Timbre / Tone quality II
1 Musical Acoustics Lecture 14 Timbre / Tone quality II Odd vs Even Harmonics and Symmetry Sines are Anti-symmetric about mid-point If you mirror around the middle you get the same shape but upside down
More informationAuditory-Tactile Interaction Using Digital Signal Processing In Musical Instruments
IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 2, Issue 6 (Jul. Aug. 2013), PP 08-13 e-issn: 2319 4200, p-issn No. : 2319 4197 Auditory-Tactile Interaction Using Digital Signal Processing
More informationPreeti Rao 2 nd CompMusicWorkshop, Istanbul 2012
Preeti Rao 2 nd CompMusicWorkshop, Istanbul 2012 o Music signal characteristics o Perceptual attributes and acoustic properties o Signal representations for pitch detection o STFT o Sinusoidal model o
More informationDistortion products and the perceived pitch of harmonic complex tones
Distortion products and the perceived pitch of harmonic complex tones D. Pressnitzer and R.D. Patterson Centre for the Neural Basis of Hearing, Dept. of Physiology, Downing street, Cambridge CB2 3EG, U.K.
More informationCOM325 Computer Speech and Hearing
COM325 Computer Speech and Hearing Part III : Theories and Models of Pitch Perception Dr. Guy Brown Room 145 Regent Court Department of Computer Science University of Sheffield Email: g.brown@dcs.shef.ac.uk
More informationGlasgow eprints Service
Hoggan, E.E and Brewster, S.A. (2006) Crossmodal icons for information display. In, Conference on Human Factors in Computing Systems, 22-27 April 2006, pages pp. 857-862, Montréal, Québec, Canada. http://eprints.gla.ac.uk/3269/
More informationMusical Acoustics, C. Bertulani. Musical Acoustics. Lecture 13 Timbre / Tone quality I
1 Musical Acoustics Lecture 13 Timbre / Tone quality I Waves: review 2 distance x (m) At a given time t: y = A sin(2πx/λ) A -A time t (s) At a given position x: y = A sin(2πt/t) Perfect Tuning Fork: Pure
More informationINDIANA UNIVERSITY, DEPT. OF PHYSICS P105, Basic Physics of Sound, Spring 2010
Name: ID#: INDIANA UNIVERSITY, DEPT. OF PHYSICS P105, Basic Physics of Sound, Spring 2010 Midterm Exam #2 Thursday, 25 March 2010, 7:30 9:30 p.m. Closed book. You are allowed a calculator. There is a Formula
More informationAN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES
Proceedings of the COST G-6 Conference on Digital Audio Effects (DAFX-), Verona, Italy, December 7-9,2 AN AUDITORILY MOTIVATED ANALYSIS METHOD FOR ROOM IMPULSE RESPONSES Tapio Lokki Telecommunications
More informationSignals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend
Signals & Systems for Speech & Hearing Week 6 Bandpass filters & filterbanks Practical spectral analysis Most analogue signals of interest are not easily mathematically specified so applying a Fourier
More informationHCS 7367 Speech Perception
HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Power spectrum model of masking Assumptions: Only frequencies within the passband of the auditory filter contribute to masking. Detection is based
More informationWhat is Sound? Part II
What is Sound? Part II Timbre & Noise 1 Prayouandi (2010) - OneOhtrix Point Never PSYCHOACOUSTICS ACOUSTICS LOUDNESS AMPLITUDE PITCH FREQUENCY QUALITY TIMBRE 2 Timbre / Quality everything that is not frequency
More informationHaptic presentation of 3D objects in virtual reality for the visually disabled
Haptic presentation of 3D objects in virtual reality for the visually disabled M Moranski, A Materka Institute of Electronics, Technical University of Lodz, Wolczanska 211/215, Lodz, POLAND marcin.moranski@p.lodz.pl,
More informationThe Human Auditory System
medial geniculate nucleus primary auditory cortex inferior colliculus cochlea superior olivary complex The Human Auditory System Prominent Features of Binaural Hearing Localization Formation of positions
More informationPrinciples of Musical Acoustics
William M. Hartmann Principles of Musical Acoustics ^Spr inger Contents 1 Sound, Music, and Science 1 1.1 The Source 2 1.2 Transmission 3 1.3 Receiver 3 2 Vibrations 1 9 2.1 Mass and Spring 9 2.1.1 Definitions
More informationSpectral and temporal processing in the human auditory system
Spectral and temporal processing in the human auditory system To r s t e n Da u 1, Mo rt e n L. Jepsen 1, a n d St e p h a n D. Ew e r t 2 1Centre for Applied Hearing Research, Ørsted DTU, Technical University
More informationPsycho-acoustics (Sound characteristics, Masking, and Loudness)
Psycho-acoustics (Sound characteristics, Masking, and Loudness) Tai-Shih Chi ( 冀泰石 ) Department of Communication Engineering National Chiao Tung University Mar. 20, 2008 Pure tones Mathematics of the pure
More informationA CLOSER LOOK AT THE REPRESENTATION OF INTERAURAL DIFFERENCES IN A BINAURAL MODEL
9th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, -7 SEPTEMBER 7 A CLOSER LOOK AT THE REPRESENTATION OF INTERAURAL DIFFERENCES IN A BINAURAL MODEL PACS: PACS:. Pn Nicolas Le Goff ; Armin Kohlrausch ; Jeroen
More informationVIRTUAL ACOUSTICS: OPPORTUNITIES AND LIMITS OF SPATIAL SOUND REPRODUCTION
ARCHIVES OF ACOUSTICS 33, 4, 413 422 (2008) VIRTUAL ACOUSTICS: OPPORTUNITIES AND LIMITS OF SPATIAL SOUND REPRODUCTION Michael VORLÄNDER RWTH Aachen University Institute of Technical Acoustics 52056 Aachen,
More informationMUSC 316 Sound & Digital Audio Basics Worksheet
MUSC 316 Sound & Digital Audio Basics Worksheet updated September 2, 2011 Name: An Aggie does not lie, cheat, or steal, or tolerate those who do. By submitting responses for this test you verify, on your
More informationINTRODUCTION. FIGURE 1: Vibration chair with an electro-dynamic exciter.
INTRODUCTION Perceptible whole-body vibrations, which show a strong correlation with the sound, have been measured in real concert venues [1, 2, 3]. Taking into account the rising threshold of vibration
More informationMath and Music: Understanding Pitch
Math and Music: Understanding Pitch Gareth E. Roberts Department of Mathematics and Computer Science College of the Holy Cross Worcester, MA Topics in Mathematics: Math and Music MATH 110 Spring 2018 March
More informationEnhancing 3D Audio Using Blind Bandwidth Extension
Enhancing 3D Audio Using Blind Bandwidth Extension (PREPRINT) Tim Habigt, Marko Ðurković, Martin Rothbucher, and Klaus Diepold Institute for Data Processing, Technische Universität München, 829 München,
More informationINFLUENCE OF FREQUENCY DISTRIBUTION ON INTENSITY FLUCTUATIONS OF NOISE
INFLUENCE OF FREQUENCY DISTRIBUTION ON INTENSITY FLUCTUATIONS OF NOISE Pierre HANNA SCRIME - LaBRI Université de Bordeaux 1 F-33405 Talence Cedex, France hanna@labriu-bordeauxfr Myriam DESAINTE-CATHERINE
More informationAuditory modelling for speech processing in the perceptual domain
ANZIAM J. 45 (E) ppc964 C980, 2004 C964 Auditory modelling for speech processing in the perceptual domain L. Lin E. Ambikairajah W. H. Holmes (Received 8 August 2003; revised 28 January 2004) Abstract
More information"From Dots To Shapes": an auditory haptic game platform for teaching geometry to blind pupils. Patrick Roth, Lori Petrucci, Thierry Pun
"From Dots To Shapes": an auditory haptic game platform for teaching geometry to blind pupils Patrick Roth, Lori Petrucci, Thierry Pun Computer Science Department CUI, University of Geneva CH - 1211 Geneva
More informationTactile Vision Substitution with Tablet and Electro-Tactile Display
Tactile Vision Substitution with Tablet and Electro-Tactile Display Haruya Uematsu 1, Masaki Suzuki 2, Yonezo Kanno 2, Hiroyuki Kajimoto 1 1 The University of Electro-Communications, 1-5-1 Chofugaoka,
More informationGlasgow eprints Service
Brown, L.M. and Brewster, S.A. and Purchase, H.C. (2005) A first investigation into the effectiveness of Tactons. In, First Joint Eurohaptics Conference and Symposium on Haptic Interfaces for Virtual Environment
More informationSpeech/Music Change Point Detection using Sonogram and AANN
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 6, Number 1 (2016), pp. 45-49 International Research Publications House http://www. irphouse.com Speech/Music Change
More informationBEAT DETECTION BY DYNAMIC PROGRAMMING. Racquel Ivy Awuor
BEAT DETECTION BY DYNAMIC PROGRAMMING Racquel Ivy Awuor University of Rochester Department of Electrical and Computer Engineering Rochester, NY 14627 rawuor@ur.rochester.edu ABSTRACT A beat is a salient
More informationExploring Surround Haptics Displays
Exploring Surround Haptics Displays Ali Israr Disney Research 4615 Forbes Ave. Suite 420, Pittsburgh, PA 15213 USA israr@disneyresearch.com Ivan Poupyrev Disney Research 4615 Forbes Ave. Suite 420, Pittsburgh,
More informationDesigning Audio and Tactile Crossmodal Icons for Mobile Devices
Designing Audio and Tactile Crossmodal Icons for Mobile Devices Eve Hoggan and Stephen Brewster Glasgow Interactive Systems Group, Department of Computing Science University of Glasgow, Glasgow, G12 8QQ,
More informationAUDL GS08/GAV1 Signals, systems, acoustics and the ear. Loudness & Temporal resolution
AUDL GS08/GAV1 Signals, systems, acoustics and the ear Loudness & Temporal resolution Absolute thresholds & Loudness Name some ways these concepts are crucial to audiologists Sivian & White (1933) JASA
More informationVirtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback
Virtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback Taku Hachisu The University of Electro- Communications 1-5-1 Chofugaoka, Chofu, Tokyo 182-8585, Japan +81 42 443 5363
More informationA Tactile Display using Ultrasound Linear Phased Array
A Tactile Display using Ultrasound Linear Phased Array Takayuki Iwamoto and Hiroyuki Shinoda Graduate School of Information Science and Technology The University of Tokyo 7-3-, Bunkyo-ku, Hongo, Tokyo,
More informationAcoustics, signals & systems for audiology. Week 4. Signals through Systems
Acoustics, signals & systems for audiology Week 4 Signals through Systems Crucial ideas Any signal can be constructed as a sum of sine waves In a linear time-invariant (LTI) system, the response to a sinusoid
More informationSpectral analysis of different harmonies Implemented by Equal temperament, Just, and Overtone ratio based tuning system
Spectral analysis of different harmonies Implemented by Equal temperament, Just, and Overtone ratio based tuning system Physics 406, Prof. Steven M Errede Dongryul Lee 1. Introduction Human started enjoying
More informationTesting of Objective Audio Quality Assessment Models on Archive Recordings Artifacts
POSTER 25, PRAGUE MAY 4 Testing of Objective Audio Quality Assessment Models on Archive Recordings Artifacts Bc. Martin Zalabák Department of Radioelectronics, Czech Technical University in Prague, Technická
More informationReflections on a WYFIWIF Tool for Eliciting User Feedback
Reflections on a WYFIWIF Tool for Eliciting User Feedback Oliver Schneider Dept. of Computer Science University of British Columbia Vancouver, Canada oschneid@cs.ubc.ca Karon MacLean Dept. of Computer
More informationANALYSIS AND EVALUATION OF IRREGULARITY IN PITCH VIBRATO FOR STRING-INSTRUMENT TONES
Abstract ANALYSIS AND EVALUATION OF IRREGULARITY IN PITCH VIBRATO FOR STRING-INSTRUMENT TONES William L. Martens Faculty of Architecture, Design and Planning University of Sydney, Sydney NSW 2006, Australia
More informationEXPERIMENTAL AND NUMERICAL ANALYSIS OF THE MUSICAL BEHAVIOR OF TRIANGLE INSTRUMENTS
11th World Congress on Computational Mechanics (WCCM XI) 5th European Conference on Computational Mechanics (ECCM V) 6th European Conference on Computational Fluid Dynamics (ECFD VI) E. Oñate, J. Oliver
More informationHaptic Cueing of a Visual Change-Detection Task: Implications for Multimodal Interfaces
In Usability Evaluation and Interface Design: Cognitive Engineering, Intelligent Agents and Virtual Reality (Vol. 1 of the Proceedings of the 9th International Conference on Human-Computer Interaction),
More informationIII. Publication III. c 2005 Toni Hirvonen.
III Publication III Hirvonen, T., Segregation of Two Simultaneously Arriving Narrowband Noise Signals as a Function of Spatial and Frequency Separation, in Proceedings of th International Conference on
More informationAn introduction to physics of Sound
An introduction to physics of Sound Outlines Acoustics and psycho-acoustics Sound? Wave and waves types Cycle Basic parameters of sound wave period Amplitude Wavelength Frequency Outlines Phase Types of
More informationSpatialization and Timbre for Effective Auditory Graphing
18 Proceedings o1't11e 8th WSEAS Int. Conf. on Acoustics & Music: Theory & Applications, Vancouver, Canada. June 19-21, 2007 Spatialization and Timbre for Effective Auditory Graphing HONG JUN SONG and
More informationLecture 7: Human haptics
ME 327: Design and Control of Haptic Systems Winter 2018 Lecture 7: Human haptics Allison M. Okamura Stanford University types of haptic sensing kinesthesia/ proprioception/ force cutaneous/ tactile Related
More informationA Pilot Study: Introduction of Time-domain Segment to Intensity-based Perception Model of High-frequency Vibration
A Pilot Study: Introduction of Time-domain Segment to Intensity-based Perception Model of High-frequency Vibration Nan Cao, Hikaru Nagano, Masashi Konyo, Shogo Okamoto 2 and Satoshi Tadokoro Graduate School
More informationSONIC: Transcription of Polyphonic Piano Music with Neural Networks
SONIC: Transcription of Polyphonic Piano Music with Neural Networks Matija Marolt Faculty of Computer and Information Science, University of Ljubljana matija.marolt@fri.uni-lj.si, http://lgm.fri.uni-lj.si/~matic
More informationCollaboration in Multimodal Virtual Environments
Collaboration in Multimodal Virtual Environments Eva-Lotta Sallnäs NADA, Royal Institute of Technology evalotta@nada.kth.se http://www.nada.kth.se/~evalotta/ Research question How is collaboration in a
More informationConvention Paper 7024 Presented at the 122th Convention 2007 May 5 8 Vienna, Austria
Audio Engineering Society Convention Paper 7024 Presented at the 122th Convention 2007 May 5 8 Vienna, Austria This convention paper has been reproduced from the author's advance manuscript, without editing,
More informationDiscrimination of Virtual Haptic Textures Rendered with Different Update Rates
Discrimination of Virtual Haptic Textures Rendered with Different Update Rates Seungmoon Choi and Hong Z. Tan Haptic Interface Research Laboratory Purdue University 465 Northwestern Avenue West Lafayette,
More informationDigitally controlled Active Noise Reduction with integrated Speech Communication
Digitally controlled Active Noise Reduction with integrated Speech Communication Herman J.M. Steeneken and Jan Verhave TNO Human Factors, Soesterberg, The Netherlands herman@steeneken.com ABSTRACT Active
More informationEffect of filter spacing and correct tonotopic representation on melody recognition: Implications for cochlear implants
Effect of filter spacing and correct tonotopic representation on melody recognition: Implications for cochlear implants Kalyan S. Kasturi and Philipos C. Loizou Dept. of Electrical Engineering The University
More informationFigure 2. Haptic human perception and display. 2.2 Pseudo-Haptic Feedback 2. RELATED WORKS 2.1 Haptic Simulation of Tapping an Object
Virtual Chromatic Percussions Simulated by Pseudo-Haptic and Vibrotactile Feedback Taku Hachisu 1 Gabriel Cirio 2 Maud Marchal 2 Anatole Lécuyer 2 Hiroyuki Kajimoto 1,3 1 The University of Electro- Communications
More informationSPEECH TO SINGING SYNTHESIS SYSTEM. Mingqing Yun, Yoon mo Yang, Yufei Zhang. Department of Electrical and Computer Engineering University of Rochester
SPEECH TO SINGING SYNTHESIS SYSTEM Mingqing Yun, Yoon mo Yang, Yufei Zhang Department of Electrical and Computer Engineering University of Rochester ABSTRACT This paper describes a speech-to-singing synthesis
More informationTranscription of Piano Music
Transcription of Piano Music Rudolf BRISUDA Slovak University of Technology in Bratislava Faculty of Informatics and Information Technologies Ilkovičova 2, 842 16 Bratislava, Slovakia xbrisuda@is.stuba.sk
More informationComplex Sounds. Reading: Yost Ch. 4
Complex Sounds Reading: Yost Ch. 4 Natural Sounds Most sounds in our everyday lives are not simple sinusoidal sounds, but are complex sounds, consisting of a sum of many sinusoids. The amplitude and frequency
More informationForce versus Frequency Figure 1.
An important trend in the audio industry is a new class of devices that produce tactile sound. The term tactile sound appears to be a contradiction of terms, in that our concept of sound relates to information
More informationMonaural and Binaural Speech Separation
Monaural and Binaural Speech Separation DeLiang Wang Perception & Neurodynamics Lab The Ohio State University Outline of presentation Introduction CASA approach to sound separation Ideal binary mask as
More informationAudio Engineering Society Convention Paper Presented at the 110th Convention 2001 May Amsterdam, The Netherlands
Audio Engineering Society Convention Paper Presented at the th Convention May 5 Amsterdam, The Netherlands This convention paper has been reproduced from the author's advance manuscript, without editing,
More informationGraphical User Interfaces for Blind Users: An Overview of Haptic Devices
Graphical User Interfaces for Blind Users: An Overview of Haptic Devices Hasti Seifi, CPSC554m: Assignment 1 Abstract Graphical user interfaces greatly enhanced usability of computer systems over older
More informationSOUND SOURCE RECOGNITION AND MODELING
SOUND SOURCE RECOGNITION AND MODELING CASA seminar, summer 2000 Antti Eronen antti.eronen@tut.fi Contents: Basics of human sound source recognition Timbre Voice recognition Recognition of environmental
More informationPhase and Feedback in the Nonlinear Brain. Malcolm Slaney (IBM and Stanford) Hiroko Shiraiwa-Terasawa (Stanford) Regaip Sen (Stanford)
Phase and Feedback in the Nonlinear Brain Malcolm Slaney (IBM and Stanford) Hiroko Shiraiwa-Terasawa (Stanford) Regaip Sen (Stanford) Auditory processing pre-cosyne workshop March 23, 2004 Simplistic Models
More informationComputational Perception. Sound localization 2
Computational Perception 15-485/785 January 22, 2008 Sound localization 2 Last lecture sound propagation: reflection, diffraction, shadowing sound intensity (db) defining computational problems sound lateralization
More informationCombining Subjective and Objective Assessment of Loudspeaker Distortion Marian Liebig Wolfgang Klippel
Combining Subjective and Objective Assessment of Loudspeaker Distortion Marian Liebig (m.liebig@klippel.de) Wolfgang Klippel (wklippel@klippel.de) Abstract To reproduce an artist s performance, the loudspeakers
More informationROBUST MULTIPITCH ESTIMATION FOR THE ANALYSIS AND MANIPULATION OF POLYPHONIC MUSICAL SIGNALS
ROBUST MULTIPITCH ESTIMATION FOR THE ANALYSIS AND MANIPULATION OF POLYPHONIC MUSICAL SIGNALS Anssi Klapuri 1, Tuomas Virtanen 1, Jan-Markus Holm 2 1 Tampere University of Technology, Signal Processing
More informationIDENTIFYING AND COMMUNICATING 2D SHAPES USING AUDITORY FEEDBACK. Javier Sanchez
IDENTIFYING AND COMMUNICATING 2D SHAPES USING AUDITORY FEEDBACK Javier Sanchez Center for Computer Research in Music and Acoustics (CCRMA) Stanford University The Knoll, 660 Lomita Dr. Stanford, CA 94305,
More informationComparison of Haptic and Non-Speech Audio Feedback
Comparison of Haptic and Non-Speech Audio Feedback Cagatay Goncu 1 and Kim Marriott 1 Monash University, Mebourne, Australia, cagatay.goncu@monash.edu, kim.marriott@monash.edu Abstract. We report a usability
More informationRhythmic Similarity -- a quick paper review. Presented by: Shi Yong March 15, 2007 Music Technology, McGill University
Rhythmic Similarity -- a quick paper review Presented by: Shi Yong March 15, 2007 Music Technology, McGill University Contents Introduction Three examples J. Foote 2001, 2002 J. Paulus 2002 S. Dixon 2004
More informationTHE CITADEL THE MILITARY COLLEGE OF SOUTH CAROLINA. Department of Electrical and Computer Engineering. ELEC 423 Digital Signal Processing
THE CITADEL THE MILITARY COLLEGE OF SOUTH CAROLINA Department of Electrical and Computer Engineering ELEC 423 Digital Signal Processing Project 2 Due date: November 12 th, 2013 I) Introduction In ELEC
More informationProceedings of Meetings on Acoustics
Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Architectural Acoustics Session 2aAAa: Adapting, Enhancing, and Fictionalizing
More informationHuman Auditory Periphery (HAP)
Human Auditory Periphery (HAP) Ray Meddis Department of Human Sciences, University of Essex Colchester, CO4 3SQ, UK. rmeddis@essex.ac.uk A demonstrator for a human auditory modelling approach. 23/11/2003
More informationA Parametric Model for Spectral Sound Synthesis of Musical Sounds
A Parametric Model for Spectral Sound Synthesis of Musical Sounds Cornelia Kreutzer University of Limerick ECE Department Limerick, Ireland cornelia.kreutzer@ul.ie Jacqueline Walker University of Limerick
More informationSalient features make a search easy
Chapter General discussion This thesis examined various aspects of haptic search. It consisted of three parts. In the first part, the saliency of movability and compliance were investigated. In the second
More informationVibrotactile Apparent Movement by DC Motors and Voice-coil Tactors
Vibrotactile Apparent Movement by DC Motors and Voice-coil Tactors Masataka Niwa 1,2, Yasuyuki Yanagida 1, Haruo Noma 1, Kenichi Hosaka 1, and Yuichiro Kume 3,1 1 ATR Media Information Science Laboratories
More informationLCC for Guitar - Introduction
LCC for Guitar - Introduction In order for guitarists to understand the significance of the Lydian Chromatic Concept of Tonal Organization and the concept of Tonal Gravity, one must first look at the nature
More informationHARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS
HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS Sean Enderby and Zlatko Baracskai Department of Digital Media Technology Birmingham City University Birmingham, UK ABSTRACT In this paper several
More informationDesign and Evaluation of Tactile Number Reading Methods on Smartphones
Design and Evaluation of Tactile Number Reading Methods on Smartphones Fan Zhang fanzhang@zjicm.edu.cn Shaowei Chu chu@zjicm.edu.cn Naye Ji jinaye@zjicm.edu.cn Ruifang Pan ruifangp@zjicm.edu.cn Abstract
More informationPhysics 101. Lecture 21 Doppler Effect Loudness Human Hearing Interference of Sound Waves Reflection & Refraction of Sound
Physics 101 Lecture 21 Doppler Effect Loudness Human Hearing Interference of Sound Waves Reflection & Refraction of Sound Quiz: Monday Oct. 18; Chaps. 16,17,18(as covered in class),19 CR/NC Deadline Oct.
More informationMULTIPLE F0 ESTIMATION IN THE TRANSFORM DOMAIN
10th International Society for Music Information Retrieval Conference (ISMIR 2009 MULTIPLE F0 ESTIMATION IN THE TRANSFORM DOMAIN Christopher A. Santoro +* Corey I. Cheng *# + LSB Audio Tampa, FL 33610
More informationAn Investigation on Vibrotactile Emotional Patterns for the Blindfolded People
An Investigation on Vibrotactile Emotional Patterns for the Blindfolded People Hsin-Fu Huang, National Yunlin University of Science and Technology, Taiwan Hao-Cheng Chiang, National Yunlin University of
More informationAnalysis/Synthesis of Stringed Instrument Using Formant Structure
192 IJCSNS International Journal of Computer Science and Network Security, VOL.7 No.9, September 2007 Analysis/Synthesis of Stringed Instrument Using Formant Structure Kunihiro Yasuda and Hiromitsu Hama
More informationTouch & Haptics. Touch & High Information Transfer Rate. Modern Haptics. Human. Haptics
Touch & Haptics Touch & High Information Transfer Rate Blind and deaf people have been using touch to substitute vision or hearing for a very long time, and successfully. OPTACON Hong Z Tan Purdue University
More informationDesign of New Micro Actuator for Tactile Display
Proceedings of the 17th World Congress The International Federation of Automatic Control Design of New Micro Actuator for Tactile Display Tae-Heon Yang*, Sang Youn Kim**, and Dong-Soo Kwon*** * Department
More informationTHE MATLAB IMPLEMENTATION OF BINAURAL PROCESSING MODEL SIMULATING LATERAL POSITION OF TONES WITH INTERAURAL TIME DIFFERENCES
THE MATLAB IMPLEMENTATION OF BINAURAL PROCESSING MODEL SIMULATING LATERAL POSITION OF TONES WITH INTERAURAL TIME DIFFERENCES J. Bouše, V. Vencovský Department of Radioelectronics, Faculty of Electrical
More informationPerception-based control of vibrato parameters in string instrument synthesis
Perception-based control of vibrato parameters in string instrument synthesis Hanna Järveläinen DEI University of Padova, Italy Helsinki University of Technology, Laboratory of Acoustics and Audio Signal
More informationPsychology in Your Life
Sarah Grison Todd Heatherton Michael Gazzaniga Psychology in Your Life FIRST EDITION Chapter 5 Sensation and Perception 2014 W. W. Norton & Company, Inc. Section 5.1 How Do Sensation and Perception Affect
More informationIN a natural environment, speech often occurs simultaneously. Monaural Speech Segregation Based on Pitch Tracking and Amplitude Modulation
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 15, NO. 5, SEPTEMBER 2004 1135 Monaural Speech Segregation Based on Pitch Tracking and Amplitude Modulation Guoning Hu and DeLiang Wang, Fellow, IEEE Abstract
More informationFeel the beat: using cross-modal rhythm to integrate perception of objects, others, and self
Feel the beat: using cross-modal rhythm to integrate perception of objects, others, and self Paul Fitzpatrick and Artur M. Arsenio CSAIL, MIT Modal and amodal features Modal and amodal features (following
More informationSpectro-Temporal Methods in Primary Auditory Cortex David Klein Didier Depireux Jonathan Simon Shihab Shamma
Spectro-Temporal Methods in Primary Auditory Cortex David Klein Didier Depireux Jonathan Simon Shihab Shamma & Department of Electrical Engineering Supported in part by a MURI grant from the Office of
More informationHapticArmrest: Remote Tactile Feedback on Touch Surfaces Using Combined Actuators
HapticArmrest: Remote Tactile Feedback on Touch Surfaces Using Combined Actuators Hendrik Richter, Sebastian Löhmann, Alexander Wiethoff University of Munich, Germany {hendrik.richter, sebastian.loehmann,
More information