University of Washington Department of Electrical Engineering Computer Speech Processing EE516 Winter 2005
|
|
- Melvin Cobb
- 5 years ago
- Views:
Transcription
1 University of Washington Department of Electrical Engineering Computer Speech Processing EE516 Winter 2005 Lecture 5 Slides Jan 26 th, 2005 Outline of Today s Lecture Announcements Filter-bank analysis of speech for spectral representations (linear/log scale) LPC representations of speech (how to compute LPC coefficients), reflections on an acoustic tube. 1
2 books & sources Rabiner&Juang: Fundamentals of speech recognition. Deller et. al. Discrete-time Processing of speech signals Beranek, Acoustics, Flanagan, Speech Analysis Synthesis and Perception Clark & Yallop, An Intro to Phonetics and phonology Ladefoged A Course in Phonetics Lieberman & Blumstein Speech physiology, speech perception, and acoustic phonetics K. Stevens, Acoustic Phonetics Malmberg, Manual of phonetics Rossing, The Science of Sound Linguistics 001, University of Pennsylvania Filter Bank Analysis Goal: Produce a representation of speech that contains the message of what was said, stripping out what was irrelevant. Vocal tract: while carrier signal is such that F0 is from Hz, and its harmonics go up to at least 5kHz, this is only the carrier Speech information is contained in relatively slow moving time-varying vocal tract response function. Recall Dudley s quote from several lectures ago. So, goal: represent VT response (compared to spectral response). 2
3 Vocal Tract Production Vocal tract is a complicated physical system. Recall Homer Dudley s Voder system. Filter Bank Approach rectifier rectifier produces useful compressed representation of speech Each BPF can be seen as convolution. 3
4 Filter Bank Approach where The BPF separates speech into separate spectral components Alternate strategy: Windowed FFT which we will talk about later. First, why rectification? There are various forms of rectification. full-wave rectification, essentially absolute value: half-wave rectification: square rectification: Our analysis will be just the full-wave version, but others are similar. Full-wave analysis note: if s(n)=αsin(ωn) (i.e., a very narrowband signal), then w(n) is a square wave (odd-harmonics only). 4
5 Full-wave Analysis sine wave square wave rectified sine spectrum of sine spectrum of square (rectifier signal) convolved Full-wave analysis Narrow-band original spectrum After band-pass filtering Rectifier Signal After rectification After Low Pass Filter After downsampler, recover full bandwidth 5
6 Filter spacing Filter spacing could be uniform, constant Q, or perceptually inspired. ideal real Alternative: constant-q like Common values are α=2 (octaves) or α=4/3 (1/3 octave) filters. Can also use critical-band center frequencies (as measured on humans) as b i octave spaced, non-overlapping, Hz (Fs=6.67kHz),C=200. This is constant Q=cf/bw 12-band, 1/3-octave, Hz, C=50 7-band critical-band like filter. 6
7 Alternative view of filtering frequency modulated standalone low-frequency filter. = * ω(n) low-pass filter of length L, evenly centered at zero. Alternative view of filtering w(n) is a window that looks like e.g.,: So we re finding scaled (modulated) DTFT of speech signal at window with center at time location n. 7
8 Alternative view of filtering modulated to center frequency ω i slower-frequency amplitude/phase at time n for spectral region ω i Can use fast FFTs to implement filterbank. FFT gives us outputs for linear spaced frequency bins. L, length of w(n) window determines spectral/temporal resolution tradeoff. L long => good spectral resolution (can get good pitch harmonics, bad temporal resolution, can t see overal spectral envelope or desired vocal tract (VT) response) L short => good temporal resolution, bad spectral resolution, goo d overall spectral envelop shape (VT response), but can t see pitch harmonics LIVE Matlab Demo using specgram(s,512,fs,[],512-32) Mel/Bark scale warping Mel-scale (normal frequency) Mel/Bark scale generalization: Approximations: Mel (α = 0.42), Bark (α=0.55) at 16kHz Bark Scale Mel Scale Ultimate goal: better approximate peripheral auditory system and cochlear filter bank. What it thinks is important is probably also what is important for speech recognition
9 Constant-Q or perceptual output from FFT Simple: just bin outputs of FFT accordingly for const-q or critical band frequency. Example mel-scale spectral filters (sum the weighted FFT outputs accordingly). Mel Processing 9
10 Windowed speech This gives us filter-bank processing in a window of speech, so window slides over speech and we get VT estimate for each window (in form of feature vector) Human Vocal-Tract Model Recall acoustic tube model UL ( z) 1+ r H( z) = = U ( z) 2 G n i= 1 G N N /2 z 1 i= 1 N i= 1 yn [ ] = c+ α yn [ i] + xn [ ] i (1 + r ) az i 1 i delayed all-pole model Question: how do we go directly from speech signal to the optimal parameters of all-pole model??? This is a parameter estimation problem 10
11 Basics statistical parameter estimation Training data: x is p-dimenstional, y is a scaller, x is p-d column vector. Goal: find f:x y s.t. f has minimum error. Only part that matters for all x Taking derivatives and setting to zero, we get best solution: Note: f() might be linear/non-linear, we don t know. But we can still find best solution under a linear model, where f(x) = Ax, A is a 1xp vector of regression coefficients, could say Basics statistical parameter estimation Training data: In general, assume f() is parameterized by some parameters A take derivative of E w.r.t. A, set to zero to get: linear assumption: f(x) = A T x So, Y lies in column space of matrix X (linear combinations of columns of X), when Y is being approximated by AX. These are called the normal equations. 11
12 Basics statistical parameter estimation Normal equations: Called normal equations, because X is orthogonal to E the error. We can apply this general setting to all-pole parameter estimation, but in this case, because of the special nature of the problem, there are some computational simplifications that can be made. un [ ] s[ n] glottal excitation Time-varying Vocal tract system function (as an all-pole model) Speech signal, complex waveform. assume that over short enough time-window, this is essentially TI (piece-wise TI) All-pole parameter estimation All-pole models: s(n) is predictable from weight sum of its past samples + the glottal pulse at that time. Approximation: S to Error transfer function is inverse filter of H above. Goal: find the a i coefficients that make e(n) reflect the glottal pulse. This is hard. Instead, lets assume white Gaussian zero-mean unit variance noise source u[n]. Then estimation problem is just error minimization (this is maximum likelihood estimate). 12
13 All-pole parameter estimation There are two methods: 1) least-squares auto-correlation method, where we window the time signal 2) least-squares auto-covariance method, where we window the error function itself. We do Auto-correlation method first Define error cost function: we window the speech signal s[m] with a window function w[m] but first shift the speech signal to the left by n Take derivatives to get vector a_1, a_2,, a_p for nth time window. For now, drop n-subscripts and do it for one window, since each window is processed independently of the others. Other assumptions: ergodicity: window length is long enough to get good estimate but short enough so that time properties can estimate ensemble properties. In other words: All-pole parameter estimation autocorrelation of speech signal (random interpretation). Normal equations (again), but this time for auto-correlation features of the solution: 1) quadratic, so only one solution (unimodal) 2) fast methods to get solution. 13
14 All-pole parameter estimation In auto-correlation method, s(n) is windowed signal s(m)=w(m)s n (m), and we assume stationary ergodic, so that: Symmetric Toeplitz pxp matrix (same elements on diagonal), pos. semidefinite (might be singular) All-pole parameter estimation s(n) is windowed signal s(m)=w(m)s n (m), why valid? w(m)=0, outside of 0 m N-1, same for s(m) e(m) is 0 outside of 0 m N+p-1 To estimate R, we do: Note, s(m)=0, m<0, m>n-1, so the above non-zero only for: 1 I,k p 14
15 All-pole parameter estimation result is a function only of i-k, or R(i-k), so when we truncate the speech signal using the window method, the auto-correlation method for stationary signals is indeed appropriate. Next, outline of Auto-covariance method better for non-stationary signals where R(I,k) R(i-k) windowed error function: Not Toeplitz, but symmetric All-pole parameter estimation auto-correlation method more common because there are fast algorithms to compute the solution (exploits the Toeplitz regular structure of the matrix) Levinson s recursion: Goes from O(p 3 ) required for normal matrix inversion to O(p 2 ) Durbin s Algorithm: reduces by another constant factor of 2 over Levinson s algorithm. 15
16 All-pole parameter estimation So, we now have a set of coefficients a 1,, a p for each window width. Typical: 25ms analysis window, with a 10ms skip 100Hz sampling rate, probably oversampling Coefficients for each window represent peaks of spectrum, attempting to estimate vocal-tract transfer function within that window Gaussian noise assumption used to minimize error not always correct (e.g., during voiced speech) Ideally, want error function to equal glottal pulse (but this is impossible to get normally). Methods other than LP analysis are now being used: Minimum variance distortionless response spectrum (MVDR) method is better in some cases (particularly high-pitched speakers, children/females) MVDR Basic idea: filter-bank approach, filter speech s(n) with BP p th order filter h l (n) (Murthi&Rao,2000) Filter designed to minimize output power subject to constraint that distortion at given center frequency ω l is zero (meaning output power at that point is unity) Solution is not too different from auto-correlation method: is square (p+1)x(p+1) Toeplitz autocorrelation matrix from before. 16
speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals
16 3. SPEECH ANALYSIS 3.1 INTRODUCTION TO SPEECH ANALYSIS Many speech processing [22] applications exploits speech production and perception to accomplish speech analysis. By speech analysis we extract
More informationEE482: Digital Signal Processing Applications
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 12 Speech Signal Processing 14/03/25 http://www.ee.unlv.edu/~b1morris/ee482/
More informationAdaptive Filters Linear Prediction
Adaptive Filters Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Institute of Electrical and Information Engineering Digital Signal Processing and System Theory Slide 1 Contents
More informationAdvanced audio analysis. Martin Gasser
Advanced audio analysis Martin Gasser Motivation Which methods are common in MIR research? How can we parameterize audio signals? Interesting dimensions of audio: Spectral/ time/melody structure, high
More informationSpeech Synthesis using Mel-Cepstral Coefficient Feature
Speech Synthesis using Mel-Cepstral Coefficient Feature By Lu Wang Senior Thesis in Electrical Engineering University of Illinois at Urbana-Champaign Advisor: Professor Mark Hasegawa-Johnson May 2018 Abstract
More informationE : Lecture 8 Source-Filter Processing. E : Lecture 8 Source-Filter Processing / 21
E85.267: Lecture 8 Source-Filter Processing E85.267: Lecture 8 Source-Filter Processing 21-4-1 1 / 21 Source-filter analysis/synthesis n f Spectral envelope Spectral envelope Analysis Source signal n 1
More informationCommunications Theory and Engineering
Communications Theory and Engineering Master's Degree in Electronic Engineering Sapienza University of Rome A.A. 2018-2019 Speech and telephone speech Based on a voice production model Parametric representation
More informationPerformance Analysis of MFCC and LPCC Techniques in Automatic Speech Recognition
www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume - 3 Issue - 8 August, 2014 Page No. 7727-7732 Performance Analysis of MFCC and LPCC Techniques in Automatic
More informationMel Spectrum Analysis of Speech Recognition using Single Microphone
International Journal of Engineering Research in Electronics and Communication Mel Spectrum Analysis of Speech Recognition using Single Microphone [1] Lakshmi S.A, [2] Cholavendan M [1] PG Scholar, Sree
More informationChapter 4 SPEECH ENHANCEMENT
44 Chapter 4 SPEECH ENHANCEMENT 4.1 INTRODUCTION: Enhancement is defined as improvement in the value or Quality of something. Speech enhancement is defined as the improvement in intelligibility and/or
More informationLinguistic Phonetics. Spectral Analysis
24.963 Linguistic Phonetics Spectral Analysis 4 4 Frequency (Hz) 1 Reading for next week: Liljencrants & Lindblom 1972. Assignment: Lip-rounding assignment, due 1/15. 2 Spectral analysis techniques There
More informationSpeech Signal Analysis
Speech Signal Analysis Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 2&3 14,18 January 216 ASR Lectures 2&3 Speech Signal Analysis 1 Overview Speech Signal Analysis for
More informationSpeech Synthesis; Pitch Detection and Vocoders
Speech Synthesis; Pitch Detection and Vocoders Tai-Shih Chi ( 冀泰石 ) Department of Communication Engineering National Chiao Tung University May. 29, 2008 Speech Synthesis Basic components of the text-to-speech
More informationLecture 6: Speech modeling and synthesis
EE E682: Speech & Audio Processing & Recognition Lecture 6: Speech modeling and synthesis 1 2 3 4 5 Modeling speech signals Spectral and cepstral models Linear Predictive models (LPC) Other signal models
More informationLecture 5: Speech modeling. The speech signal
EE E68: Speech & Audio Processing & Recognition Lecture 5: Speech modeling 1 3 4 5 Modeling speech signals Spectral and cepstral models Linear Predictive models (LPC) Other signal models Speech synthesis
More informationOverview of Code Excited Linear Predictive Coder
Overview of Code Excited Linear Predictive Coder Minal Mulye 1, Sonal Jagtap 2 1 PG Student, 2 Assistant Professor, Department of E&TC, Smt. Kashibai Navale College of Engg, Pune, India Abstract Advances
More informationSpeech Compression Using Voice Excited Linear Predictive Coding
Speech Compression Using Voice Excited Linear Predictive Coding Ms.Tosha Sen, Ms.Kruti Jay Pancholi PG Student, Asst. Professor, L J I E T, Ahmedabad Abstract : The aim of the thesis is design good quality
More informationSignal Processing for Speech Applications - Part 2-1. Signal Processing For Speech Applications - Part 2
Signal Processing for Speech Applications - Part 2-1 Signal Processing For Speech Applications - Part 2 May 14, 2013 Signal Processing for Speech Applications - Part 2-2 References Huang et al., Chapter
More information(i) Understanding the basic concepts of signal modeling, correlation, maximum likelihood estimation, least squares and iterative numerical methods
Tools and Applications Chapter Intended Learning Outcomes: (i) Understanding the basic concepts of signal modeling, correlation, maximum likelihood estimation, least squares and iterative numerical methods
More informationEE 225D LECTURE ON MEDIUM AND HIGH RATE CODING. University of California Berkeley
University of California Berkeley College of Engineering Department of Electrical Engineering and Computer Sciences Professors : N.Morgan / B.Gold EE225D Spring,1999 Medium & High Rate Coding Lecture 26
More informationCepstrum alanysis of speech signals
Cepstrum alanysis of speech signals ELEC-E5520 Speech and language processing methods Spring 2016 Mikko Kurimo 1 /48 Contents Literature and other material Idea and history of cepstrum Cepstrum and LP
More informationAdaptive Filters Application of Linear Prediction
Adaptive Filters Application of Linear Prediction Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Technology Digital Signal Processing
More informationEE482: Digital Signal Processing Applications
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 14 Quiz 04 Review 14/04/07 http://www.ee.unlv.edu/~b1morris/ee482/
More informationEVALUATION OF MFCC ESTIMATION TECHNIQUES FOR MUSIC SIMILARITY
EVALUATION OF MFCC ESTIMATION TECHNIQUES FOR MUSIC SIMILARITY Jesper Højvang Jensen 1, Mads Græsbøll Christensen 1, Manohar N. Murthi, and Søren Holdt Jensen 1 1 Department of Communication Technology,
More informationEvaluation of MFCC Estimation Techniques for Music Similarity Jensen, Jesper Højvang; Christensen, Mads Græsbøll; Murthi, Manohar; Jensen, Søren Holdt
Aalborg Universitet Evaluation of MFCC Estimation Techniques for Music Similarity Jensen, Jesper Højvang; Christensen, Mads Græsbøll; Murthi, Manohar; Jensen, Søren Holdt Published in: Proceedings of the
More informationReading: Johnson Ch , Ch.5.5 (today); Liljencrants & Lindblom; Stevens (Tues) reminder: no class on Thursday.
L105/205 Phonetics Scarborough Handout 7 10/18/05 Reading: Johnson Ch.2.3.3-2.3.6, Ch.5.5 (today); Liljencrants & Lindblom; Stevens (Tues) reminder: no class on Thursday Spectral Analysis 1. There are
More informationSignal Analysis. Peak Detection. Envelope Follower (Amplitude detection) Music 270a: Signal Analysis
Signal Analysis Music 27a: Signal Analysis Tamara Smyth, trsmyth@ucsd.edu Department of Music, University of California, San Diego (UCSD November 23, 215 Some tools we may want to use to automate analysis
More informationVocoder (LPC) Analysis by Variation of Input Parameters and Signals
ISCA Journal of Engineering Sciences ISCA J. Engineering Sci. Vocoder (LPC) Analysis by Variation of Input Parameters and Signals Abstract Gupta Rajani, Mehta Alok K. and Tiwari Vebhav Truba College of
More informationAnalysis/synthesis coding
TSBK06 speech coding p.1/32 Analysis/synthesis coding Many speech coders are based on a principle called analysis/synthesis coding. Instead of coding a waveform, as is normally done in general audio coders
More informationSpeech Coding using Linear Prediction
Speech Coding using Linear Prediction Jesper Kjær Nielsen Aalborg University and Bang & Olufsen jkn@es.aau.dk September 10, 2015 1 Background Speech is generated when air is pushed from the lungs through
More informationDigital Speech Processing and Coding
ENEE408G Spring 2006 Lecture-2 Digital Speech Processing and Coding Spring 06 Instructor: Shihab Shamma Electrical & Computer Engineering University of Maryland, College Park http://www.ece.umd.edu/class/enee408g/
More informationL19: Prosodic modification of speech
L19: Prosodic modification of speech Time-domain pitch synchronous overlap add (TD-PSOLA) Linear-prediction PSOLA Frequency-domain PSOLA Sinusoidal models Harmonic + noise models STRAIGHT This lecture
More informationAPPLICATIONS OF DSP OBJECTIVES
APPLICATIONS OF DSP OBJECTIVES This lecture will discuss the following: Introduce analog and digital waveform coding Introduce Pulse Coded Modulation Consider speech-coding principles Introduce the channel
More informationInternational Journal of Modern Trends in Engineering and Research e-issn No.: , Date: 2-4 July, 2015
International Journal of Modern Trends in Engineering and Research www.ijmter.com e-issn No.:2349-9745, Date: 2-4 July, 2015 Analysis of Speech Signal Using Graphic User Interface Solly Joy 1, Savitha
More informationLab 8. Signal Analysis Using Matlab Simulink
E E 2 7 5 Lab June 30, 2006 Lab 8. Signal Analysis Using Matlab Simulink Introduction The Matlab Simulink software allows you to model digital signals, examine power spectra of digital signals, represent
More informationT Automatic Speech Recognition: From Theory to Practice
Automatic Speech Recognition: From Theory to Practice http://www.cis.hut.fi/opinnot// September 27, 2004 Prof. Bryan Pellom Department of Computer Science Center for Spoken Language Research University
More informationPattern Recognition. Part 6: Bandwidth Extension. Gerhard Schmidt
Pattern Recognition Part 6: Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Institute of Electrical and Information Engineering Digital Signal Processing and System Theory
More informationChapter 7. Frequency-Domain Representations 语音信号的频域表征
Chapter 7 Frequency-Domain Representations 语音信号的频域表征 1 General Discrete-Time Model of Speech Production Voiced Speech: A V P(z)G(z)V(z)R(z) Unvoiced Speech: A N N(z)V(z)R(z) 2 DTFT and DFT of Speech The
More informationSpeech Enhancement using Wiener filtering
Speech Enhancement using Wiener filtering S. Chirtmay and M. Tahernezhadi Department of Electrical Engineering Northern Illinois University DeKalb, IL 60115 ABSTRACT The problem of reducing the disturbing
More informationCS 188: Artificial Intelligence Spring Speech in an Hour
CS 188: Artificial Intelligence Spring 2006 Lecture 19: Speech Recognition 3/23/2006 Dan Klein UC Berkeley Many slides from Dan Jurafsky Speech in an Hour Speech input is an acoustic wave form s p ee ch
More informationLinear Predictive Coding *
OpenStax-CNX module: m45345 1 Linear Predictive Coding * Kiefer Forseth This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 1 LPC Implementation Linear
More informationStructure of Speech. Physical acoustics Time-domain representation Frequency domain representation Sound shaping
Structure of Speech Physical acoustics Time-domain representation Frequency domain representation Sound shaping Speech acoustics Source-Filter Theory Speech Source characteristics Speech Filter characteristics
More informationBiomedical Signals. Signals and Images in Medicine Dr Nabeel Anwar
Biomedical Signals Signals and Images in Medicine Dr Nabeel Anwar Noise Removal: Time Domain Techniques 1. Synchronized Averaging (covered in lecture 1) 2. Moving Average Filters (today s topic) 3. Derivative
More informationQuantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation
Quantification of glottal and voiced speech harmonicsto-noise ratios using cepstral-based estimation Peter J. Murphy and Olatunji O. Akande, Department of Electronic and Computer Engineering University
More informationAUDL GS08/GAV1 Auditory Perception. Envelope and temporal fine structure (TFS)
AUDL GS08/GAV1 Auditory Perception Envelope and temporal fine structure (TFS) Envelope and TFS arise from a method of decomposing waveforms The classic decomposition of waveforms Spectral analysis... Decomposes
More informationDigital Signal Processing
COMP ENG 4TL4: Digital Signal Processing Notes for Lecture #27 Tuesday, November 11, 23 6. SPECTRAL ANALYSIS AND ESTIMATION 6.1 Introduction to Spectral Analysis and Estimation The discrete-time Fourier
More informationSignals, Sound, and Sensation
Signals, Sound, and Sensation William M. Hartmann Department of Physics and Astronomy Michigan State University East Lansing, Michigan Л1Р Contents Preface xv Chapter 1: Pure Tones 1 Mathematics of the
More informationInternational Journal of Engineering and Techniques - Volume 1 Issue 6, Nov Dec 2015
RESEARCH ARTICLE OPEN ACCESS A Comparative Study on Feature Extraction Technique for Isolated Word Speech Recognition Easwari.N 1, Ponmuthuramalingam.P 2 1,2 (PG & Research Department of Computer Science,
More informationSuggested Solutions to Examination SSY130 Applied Signal Processing
Suggested Solutions to Examination SSY13 Applied Signal Processing 1:-18:, April 8, 1 Instructions Responsible teacher: Tomas McKelvey, ph 81. Teacher will visit the site of examination at 1:5 and 1:.
More informationCOMMUNICATION SYSTEMS
COMMUNICATION SYSTEMS 4TH EDITION Simon Hayhin McMaster University JOHN WILEY & SONS, INC. Ш.! [ BACKGROUND AND PREVIEW 1. The Communication Process 1 2. Primary Communication Resources 3 3. Sources of
More informationEC 6501 DIGITAL COMMUNICATION UNIT - II PART A
EC 6501 DIGITAL COMMUNICATION 1.What is the need of prediction filtering? UNIT - II PART A [N/D-16] Prediction filtering is used mostly in audio signal processing and speech processing for representing
More informationComparison of CELP speech coder with a wavelet method
University of Kentucky UKnowledge University of Kentucky Master's Theses Graduate School 2006 Comparison of CELP speech coder with a wavelet method Sriram Nagaswamy University of Kentucky, sriramn@gmail.com
More informationA LPC-PEV Based VAD for Word Boundary Detection
14 A LPC-PEV Based VAD for Word Boundary Detection Syed Abbas Ali (A), NajmiGhaniHaider (B) and Mahmood Khan Pathan (C) (A) Faculty of Computer &Information Systems Engineering, N.E.D University of Engg.
More informationSPEECH AND SPECTRAL ANALYSIS
SPEECH AND SPECTRAL ANALYSIS 1 Sound waves: production in general: acoustic interference vibration (carried by some propagation medium) variations in air pressure speech: actions of the articulatory organs
More informationMATLAB SIMULATOR FOR ADAPTIVE FILTERS
MATLAB SIMULATOR FOR ADAPTIVE FILTERS Submitted by: Raja Abid Asghar - BS Electrical Engineering (Blekinge Tekniska Högskola, Sweden) Abu Zar - BS Electrical Engineering (Blekinge Tekniska Högskola, Sweden)
More informationDifferent Approaches of Spectral Subtraction Method for Speech Enhancement
ISSN 2249 5460 Available online at www.internationalejournals.com International ejournals International Journal of Mathematical Sciences, Technology and Humanities 95 (2013 1056 1062 Different Approaches
More informationThe source-filter model of speech production"
24.915/24.963! Linguistic Phonetics! The source-filter model of speech production" Glottal airflow Output from lips 400 200 0.1 0.2 0.3 Time (in secs) 30 20 10 0 0 1000 2000 3000 Frequency (Hz) Source
More informationPerceptive Speech Filters for Speech Signal Noise Reduction
International Journal of Computer Applications (975 8887) Volume 55 - No. *, October 22 Perceptive Speech Filters for Speech Signal Noise Reduction E.S. Kasthuri and A.P. James School of Computer Science
More informationSignal segmentation and waveform characterization. Biosignal processing, S Autumn 2012
Signal segmentation and waveform characterization Biosignal processing, 5173S Autumn 01 Short-time analysis of signals Signal statistics may vary in time: nonstationary how to compute signal characterizations?
More informationSpeech Coding in the Frequency Domain
Speech Coding in the Frequency Domain Speech Processing Advanced Topics Tom Bäckström Aalto University October 215 Introduction The speech production model can be used to efficiently encode speech signals.
More informationSpeech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter
Speech Enhancement in Presence of Noise using Spectral Subtraction and Wiener Filter 1 Gupteswar Sahu, 2 D. Arun Kumar, 3 M. Bala Krishna and 4 Jami Venkata Suman Assistant Professor, Department of ECE,
More informationProject 0: Part 2 A second hands-on lab on Speech Processing Frequency-domain processing
Project : Part 2 A second hands-on lab on Speech Processing Frequency-domain processing February 24, 217 During this lab, you will have a first contact on frequency domain analysis of speech signals. You
More informationSOUND SOURCE RECOGNITION AND MODELING
SOUND SOURCE RECOGNITION AND MODELING CASA seminar, summer 2000 Antti Eronen antti.eronen@tut.fi Contents: Basics of human sound source recognition Timbre Voice recognition Recognition of environmental
More informationSynthesis Algorithms and Validation
Chapter 5 Synthesis Algorithms and Validation An essential step in the study of pathological voices is re-synthesis; clear and immediate evidence of the success and accuracy of modeling efforts is provided
More informationSPEech Feature Toolbox (SPEFT) Design and Emotional Speech Feature Extraction
SPEech Feature Toolbox (SPEFT) Design and Emotional Speech Feature Extraction by Xi Li A thesis submitted to the Faculty of Graduate School, Marquette University, in Partial Fulfillment of the Requirements
More informationSpeech Enhancement Based On Noise Reduction
Speech Enhancement Based On Noise Reduction Kundan Kumar Singh Electrical Engineering Department University Of Rochester ksingh11@z.rochester.edu ABSTRACT This paper addresses the problem of signal distortion
More informationFilter Banks I. Prof. Dr. Gerald Schuller. Fraunhofer IDMT & Ilmenau University of Technology Ilmenau, Germany. Fraunhofer IDMT
Filter Banks I Prof. Dr. Gerald Schuller Fraunhofer IDMT & Ilmenau University of Technology Ilmenau, Germany 1 Structure of perceptual Audio Coders Encoder Decoder 2 Filter Banks essential element of most
More informationApplications of Music Processing
Lecture Music Processing Applications of Music Processing Christian Dittmar International Audio Laboratories Erlangen christian.dittmar@audiolabs-erlangen.de Singing Voice Detection Important pre-requisite
More informationThe Channel Vocoder (analyzer):
Vocoders 1 The Channel Vocoder (analyzer): The channel vocoder employs a bank of bandpass filters, Each having a bandwidth between 100 Hz and 300 Hz. Typically, 16-20 linear phase FIR filter are used.
More informationIntroducing COVAREP: A collaborative voice analysis repository for speech technologies
Introducing COVAREP: A collaborative voice analysis repository for speech technologies John Kane Wednesday November 27th, 2013 SIGMEDIA-group TCD COVAREP - Open-source speech processing repository 1 Introduction
More informationSPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN. Yu Wang and Mike Brookes
SPEECH ENHANCEMENT USING A ROBUST KALMAN FILTER POST-PROCESSOR IN THE MODULATION DOMAIN Yu Wang and Mike Brookes Department of Electrical and Electronic Engineering, Exhibition Road, Imperial College London,
More informationDigital Signal Processing
Digital Signal Processing Fourth Edition John G. Proakis Department of Electrical and Computer Engineering Northeastern University Boston, Massachusetts Dimitris G. Manolakis MIT Lincoln Laboratory Lexington,
More informationAuditory Based Feature Vectors for Speech Recognition Systems
Auditory Based Feature Vectors for Speech Recognition Systems Dr. Waleed H. Abdulla Electrical & Computer Engineering Department The University of Auckland, New Zealand [w.abdulla@auckland.ac.nz] 1 Outlines
More informationChapter 2: Signal Representation
Chapter 2: Signal Representation Aveek Dutta Assistant Professor Department of Electrical and Computer Engineering University at Albany Spring 2018 Images and equations adopted from: Digital Communications
More informationINTRODUCTION TO ACOUSTIC PHONETICS 2 Hilary Term, week 6 22 February 2006
1. Resonators and Filters INTRODUCTION TO ACOUSTIC PHONETICS 2 Hilary Term, week 6 22 February 2006 Different vibrating objects are tuned to specific frequencies; these frequencies at which a particular
More informationIsolated Digit Recognition Using MFCC AND DTW
MarutiLimkar a, RamaRao b & VidyaSagvekar c a Terna collegeof Engineering, Department of Electronics Engineering, Mumbai University, India b Vidyalankar Institute of Technology, Department ofelectronics
More informationSpeech synthesizer. W. Tidelund S. Andersson R. Andersson. March 11, 2015
Speech synthesizer W. Tidelund S. Andersson R. Andersson March 11, 2015 1 1 Introduction A real time speech synthesizer is created by modifying a recorded signal on a DSP by using a prediction filter.
More informationHCS 7367 Speech Perception
HCS 7367 Speech Perception Dr. Peter Assmann Fall 212 Power spectrum model of masking Assumptions: Only frequencies within the passband of the auditory filter contribute to masking. Detection is based
More informationRobust Algorithms For Speech Reconstruction On Mobile Devices
Robust Algorithms For Speech Reconstruction On Mobile Devices XU SHAO A Thesis presented for the degree of Doctor of Philosophy Speech Group School of Computing Sciences University of East Anglia England
More informationComplex Sounds. Reading: Yost Ch. 4
Complex Sounds Reading: Yost Ch. 4 Natural Sounds Most sounds in our everyday lives are not simple sinusoidal sounds, but are complex sounds, consisting of a sum of many sinusoids. The amplitude and frequency
More informationEE228 Applications of Course Concepts. DePiero
EE228 Applications of Course Concepts DePiero Purpose Describe applications of concepts in EE228. Applications may help students recall and synthesize concepts. Also discuss: Some advanced concepts Highlight
More informationSpeech Processing. Undergraduate course code: LASC10061 Postgraduate course code: LASC11065
Speech Processing Undergraduate course code: LASC10061 Postgraduate course code: LASC11065 All course materials and handouts are the same for both versions. Differences: credits (20 for UG, 10 for PG);
More informationSimulation of Conjugate Structure Algebraic Code Excited Linear Prediction Speech Coder
COMPUSOFT, An international journal of advanced computer technology, 3 (3), March-204 (Volume-III, Issue-III) ISSN:2320-0790 Simulation of Conjugate Structure Algebraic Code Excited Linear Prediction Speech
More informationFFT analysis in practice
FFT analysis in practice Perception & Multimedia Computing Lecture 13 Rebecca Fiebrink Lecturer, Department of Computing Goldsmiths, University of London 1 Last Week Review of complex numbers: rectangular
More informationSpeech Recognition using FIR Wiener Filter
Speech Recognition using FIR Wiener Filter Deepak 1, Vikas Mittal 2 1 Department of Electronics & Communication Engineering, Maharishi Markandeshwar University, Mullana (Ambala), INDIA 2 Department of
More informationSinging Voice Detection. Applications of Music Processing. Singing Voice Detection. Singing Voice Detection. Singing Voice Detection
Detection Lecture usic Processing Applications of usic Processing Christian Dittmar International Audio Laboratories Erlangen christian.dittmar@audiolabs-erlangen.de Important pre-requisite for: usic segmentation
More informationSAMPLING THEORY. Representing continuous signals with discrete numbers
SAMPLING THEORY Representing continuous signals with discrete numbers Roger B. Dannenberg Professor of Computer Science, Art, and Music Carnegie Mellon University ICM Week 3 Copyright 2002-2013 by Roger
More informationSpectral analysis of seismic signals using Burg algorithm V. Ravi Teja 1, U. Rakesh 2, S. Koteswara Rao 3, V. Lakshmi Bharathi 4
Volume 114 No. 1 217, 163-171 ISSN: 1311-88 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu Spectral analysis of seismic signals using Burg algorithm V. avi Teja
More informationAcoustics, signals & systems for audiology. Week 4. Signals through Systems
Acoustics, signals & systems for audiology Week 4 Signals through Systems Crucial ideas Any signal can be constructed as a sum of sine waves In a linear time-invariant (LTI) system, the response to a sinusoid
More informationMFCC AND GMM BASED TAMIL LANGUAGE SPEAKER IDENTIFICATION SYSTEM
www.advancejournals.org Open Access Scientific Publisher MFCC AND GMM BASED TAMIL LANGUAGE SPEAKER IDENTIFICATION SYSTEM ABSTRACT- P. Santhiya 1, T. Jayasankar 1 1 AUT (BIT campus), Tiruchirappalli, India
More informationVoice Excited Lpc for Speech Compression by V/Uv Classification
IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 6, Issue 3, Ver. II (May. -Jun. 2016), PP 65-69 e-issn: 2319 4200, p-issn No. : 2319 4197 www.iosrjournals.org Voice Excited Lpc for Speech
More informationLecture Schedule: Week Date Lecture Title
http://elec3004.org Sampling & More 2014 School of Information Technology and Electrical Engineering at The University of Queensland Lecture Schedule: Week Date Lecture Title 1 2-Mar Introduction 3-Mar
More informationESE531 Spring University of Pennsylvania Department of Electrical and System Engineering Digital Signal Processing
University of Pennsylvania Department of Electrical and System Engineering Digital Signal Processing ESE531, Spring 2017 Final Project: Audio Equalization Wednesday, Apr. 5 Due: Tuesday, April 25th, 11:59pm
More informationIMPROVED CHANNEL ESTIMATION FOR OFDM BASED WLAN SYSTEMS. G.V.Rangaraj M.R.Raghavendra K.Giridhar
IMPROVED CHANNEL ESTIMATION FOR OFDM BASED WLAN SYSTEMS GVRangaraj MRRaghavendra KGiridhar Telecommunication and Networking TeNeT) Group Department of Electrical Engineering Indian Institute of Technology
More informationAntennas and Propagation. Chapter 6b: Path Models Rayleigh, Rician Fading, MIMO
Antennas and Propagation b: Path Models Rayleigh, Rician Fading, MIMO Introduction From last lecture How do we model H p? Discrete path model (physical, plane waves) Random matrix models (forget H p and
More informationtechniques are means of reducing the bandwidth needed to represent the human voice. In mobile
8 2. LITERATURE SURVEY The available radio spectrum for the wireless radio communication is very limited hence to accommodate maximum number of users the speech is compressed. The speech compression techniques
More informationAn Improved Voice Activity Detection Based on Deep Belief Networks
e-issn 2455 1392 Volume 2 Issue 4, April 2016 pp. 676-683 Scientific Journal Impact Factor : 3.468 http://www.ijcter.com An Improved Voice Activity Detection Based on Deep Belief Networks Shabeeba T. K.
More informationSignals & Systems for Speech & Hearing. Week 6. Practical spectral analysis. Bandpass filters & filterbanks. Try this out on an old friend
Signals & Systems for Speech & Hearing Week 6 Bandpass filters & filterbanks Practical spectral analysis Most analogue signals of interest are not easily mathematically specified so applying a Fourier
More informationPeriodic Component Analysis: An Eigenvalue Method for Representing Periodic Structure in Speech
Periodic Component Analysis: An Eigenvalue Method for Representing Periodic Structure in Speech Lawrence K. Saul and Jont B. Allen lsaul,jba @research.att.com AT&T Labs, 180 Park Ave, Florham Park, NJ
More informationAN AUTOREGRESSIVE BASED LFM REVERBERATION SUPPRESSION FOR RADAR AND SONAR APPLICATIONS
AN AUTOREGRESSIVE BASED LFM REVERBERATION SUPPRESSION FOR RADAR AND SONAR APPLICATIONS MrPMohan Krishna 1, AJhansi Lakshmi 2, GAnusha 3, BYamuna 4, ASudha Rani 5 1 Asst Professor, 2,3,4,5 Student, Dept
More information