Quality-enhanced Voice Morphing using Maximum Likelihood Transformations

Similar documents
Direct F 0 Control of an Electrolarynx based on Statistical Excitation Feature Prediction and its Evaluation through Simulation

Statistical Singing Voice Conversion based on Direct Waveform Modification with Global Variance

Statistical Singing Voice Conversion with Direct Waveform Modification based on the Spectrum Differential

Adaptive Harmonic IIR Notch Filter with Varying Notch Bandwidth and Convergence Factor

A Digital Signal Processor Implementation of Silent/Electrolaryngeal Speech Enhancement based on Real-Time Statistical Voice Conversion

The NU-NAIST voice conversion system for the Voice Conversion Challenge 2016

DSI3 Sensor to Master Current Threshold Adaptation for Pattern Recognition

SECURITY AND BER PERFORMANCE TRADE-OFF IN WIRELESS COMMUNICATION SYSTEMS APPLICATIONS

EQUALIZED ALGORITHM FOR A TRUCK CABIN ACTIVE NOISE CONTROL SYSTEM

ELEC2202 Communications Engineering Laboratory Frequency Modulation (FM)

APPLICATION OF THE FAN-CHIRP TRANSFORM TO HYBRID SINUSOIDAL+NOISE MODELING OF POLYPHONIC AUDIO

POWER QUALITY ASSESSMENT USING TWO STAGE NONLINEAR ESTIMATION NUMERICAL ALGORITHM

Kalman Filtering for NLOS Mitigation and Target Tracking in Indoor Wireless Environment

Alternative Encoding Techniques for Digital Loudspeaker Arrays

Additive Synthesis, Amplitude Modulation and Frequency Modulation

Notes on Orthogonal Frequency Division Multiplexing (OFDM)

Non-Linear Weighting Function for Non-stationary Signal Denoising

Relation between C/N Ratio and S/N Ratio

New Adaptive Linear Combination Structure for Tracking/Estimating Phasor and Frequency of Power System

UNIT - II CONTROLLED RECTIFIERS (Line Commutated AC to DC converters) Line Commutated Converter

PREDICTING SOUND LEVELS BEHIND BUILDINGS - HOW MANY REFLECTIONS SHOULD I USE? Apex Acoustics Ltd, Gateshead, UK

Department of Mechanical and Aerospace Engineering, Case Western Reserve University, Cleveland, OH, 2

Efficient Non-linear Changed Mel-filter Bank VAD Algorithm

A soft decision decoding of product BCH and Reed-Müller codes for error control and peak-factor reduction in OFDM

TESTING OF ADCS BY FREQUENCY-DOMAIN ANALYSIS IN MULTI-TONE MODE

Parameter Identification of Transfer Functions Using MATLAB

Speech Enhancement using Temporal Masking and Fractional Bark Gammatone Filters

A New Localization and Tracking Algorithm for Wireless Sensor Networks Based on Internet of Things

Wavelet-based Voice Morphing

Performance Analysis of Atmospheric Field Conjugation Adaptive Arrays

Windowing High-Resolution ADC Data Part 2

OTC Statistics of High- and Low-Frequency Motions of a Moored Tanker. sensitive to lateral loading such as the SAL5 and

Cross-correlation tracking for Maximum Length Sequence based acoustic localisation

A NEW APPROACH TO UNGROUNDED FAULT LOCATION IN A THREE-PHASE UNDERGROUND DISTRIBUTION SYSTEM USING COMBINED NEURAL NETWORKS & WAVELET ANALYSIS

Overlapping Signal Separation in DPX Spectrum Based on EM Algorithm. Chuandang Liu 1, a, Luxi Lu 1, b

FORWARD MASKING THRESHOLD ESTIMATION USING NEURAL NETWORKS AND ITS APPLICATION TO PARALLEL SPEECH ENHANCEMENT

A comparison of LSF and ISP representations for wideband LPC parameter coding using the switched split vector quantiser

Performance Analysis of an AMC System with an Iterative V-BLAST Decoding Algorithm

LOW COST PRODUCTION PHASE NOISE MEASUREMENTS ON MICROWAVE AND MILLIMETRE WAVE FREQUENCY SOURCES

Track-Before-Detect for an Active Towed Array Sonar

Selective Harmonic Elimination for Multilevel Inverters with Unbalanced DC Inputs

Power Improvement in 64-Bit Full Adder Using Embedded Technologies Er. Arun Gandhi 1, Dr. Rahul Malhotra 2, Er. Kulbhushan Singla 3

Iterative Receiver Signal Processing for Joint Mitigation of Transmitter and Receiver Phase Noise in OFDM-Based Cognitive Radio Link

NINTH INTERNATIONAL CONGRESS ON SOUND AND VIBRATION, ICSV9 PASSIVE CONTROL OF LAUNCH NOISE IN ROCKET PAYLOAD BAYS

PARAMETER OPTIMIZATION OF THE ADAPTIVE MVDR QR-BASED BEAMFORMER FOR JAMMING AND MULTIPATH SUPRESSION IN GPS/GLONASS RECEIVERS

An orthogonal multi-beam based MIMO scheme. for multi-user wireless systems

A Preprocessing Method to Increase High Frequency Response of A Parametric Loudspeaker

We are IntechOpen, the world s leading publisher of Open Access books Built by scientists, for scientists. International authors and editors

RAKE Receiver. Tommi Heikkilä S Postgraduate Course in Radio Communications, Autumn II.

Yield Enhancement Techniques for 3D Memories by Redundancy Sharing among All Layers

LETTER Adaptive Multi-Stage Parallel Interference Cancellation Receiver for Multi-Rate DS-CDMA System

Precise Indoor Localization System For a Mobile Robot Using Auto Calibration Algorithm

Modeling Beam forming in Circular Antenna Array with Directional Emitters

Power Optimal Signaling for Fading Multi-access Channel in Presence of Coding Gap

SIG: Signal-Processing

Transmit Beamforming and Iterative Water-Filling Based on SLNR for OFDMA Systems

Detection of Faults in Power System Using Wavelet Transform and Independent Component Analysis

ES 442 Homework #8 Solutions (Spring 2018 Due April 16, 2018 ) Print out homework and do work on the printed pages.. Problem 1 ASCII Code (20 points)

Using Adaptive Modulation in a LEO Satellite Communication System

Enhanced Waveform Interpolative Coding at 4 kbps

Mitigation of GPS L 2 signal in the H I observation based on NLMS algorithm Zhong Danmei 1, a, Wang zhan 1, a, Cheng zhu 1, a, Huang Da 1, a

WIPL-D Pro: What is New in v12.0?

ROBUST UNDERWATER LOCALISATION OF ULTRA LOW FREQUENCY SOURCES IN OPERATIONAL CONTEXT

Comparison Between PLAXIS Output and Neural Network in the Guard Walls

Part 9: Basic AC Theory

A Robust Noise Spectral Estimation Algorithm for Speech Enhancement in Voice Devices

EXPERIMENTAL VERIFICATION OF SINUSOIDAL APPROXIMATION IN ANALYSIS OF THREE-PHASE TWELVE-PULSE OUTPUT VOLTAGE TYPE RECTIFIERS

Ignition and monitoring technique for plasma processing of multicell superconducting radio frequency cavities

Allocation of Multiple Services in Multi-Access Wireless Systems

Fundamental study for measuring microflow with Michelson interferometer enhanced by external random signal

Energy-Efficient Cellular Communications Powered by Smart Grid Technology

Design and Implementation of Block Based Transpose Form FIR Filter

EFFECTS OF MASKING ANGLE AND MULTIPATH ON GALILEO PERFORMANCES IN DIFFERENT ENVIRONMENTS

Phase Noise Modelling and Mitigation Techniques in OFDM Communications Systems

Overlapped frequency-time division multiplexing

Secondary-side-only Simultaneous Power and Efficiency Control in Dynamic Wireless Power Transfer System

Intermediate-Node Initiated Reservation (IIR): A New Signaling Scheme for Wavelength-Routed Networks with Sparse Conversion

A Frequency Domain Approach to Design Constrained Amplitude Spreading Sequences for DS-CDMA Systems for Frequency Selective Fading Channels

Optimal Modulation Index of the Mach-Zehnder Modulator in a Coherent Optical OFDM System Employing Digital Predistortion

A Novel NLOS Mitigation Approach for Wireless Positioning System

Design of an Arrayed Waveguide Grating with flat spectral response

Keywords Frequency-domain equalization, antenna diversity, multicode DS-CDMA, frequency-selective fading

AN OPTIMAL DESIGN PROCESS FOR AN ADEQUATE PRODUCT?

Transmit Power and Bit Allocations for OFDM Systems in a Fading Channel

COMBINED FREQUENCY AND SPATIAL DOMAINS POWER DISTRIBUTION FOR MIMO-OFDM TRANSMISSION

Ruohua Zhou, Josh D Reiss ABSTRACT KEYWORDS INTRODUCTION

Radio Resource Management in a Coordinated Cellular Distributed Antenna System By Using Particle Swarm Optimization

Smarter Balanced Assessment Consortium Claims, Targets, and Standard Alignment for Math

Evaluation of Steady-State and Dynamic Performance of a Synchronized Phasor Measurement Unit

ANALYSIS AND OPTIMIZATION OF SYNTHETIC APERTURE ULTRASOUND IMAGING USING THE EFFECTIVE APERTURE APPROACH. Milen Nikolov, Vera Behar

Implementation of Adaptive Viterbi Decoder

Improved Codebook-based Speech Enhancement based on MBE Model

ELECTROMAGNETIC COVERAGE CALCULATION IN GIS

Keywords: Equivalent Instantaneous Inductance, Finite Element, Inrush Current.

Speech Synthesis using Mel-Cepstral Coefficient Feature

NONLINEAR WAVELET PACKET DENOISING OF IMPULSIVE VIBRATION SIGNALS NIKOLAOS G. NIKOLAOU, IOANNIS A. ANTONIADIS

Introduction Traditionally, studying outage or cellular systes has been based on the signal-to-intererence ratio (SIR) dropping below a required thres

651 Analysis of LSF frame selection in voice conversion

Distributed Resource Allocation Assisted by Intercell Interference Mitigation in Downlink Multicell MC DS-CDMA Systems

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

Transcription:

1 Quality-enhanced Voice Morphing using Maxiu Likelihood Transforations Hui Ye, Student Meber, IEEE, and Steve Young, Meber, IEEE Abstract Voice orphing is a technique for odifying a source speaker s speech to sound as if it was spoken by soe designated target speaker. The core process in a voice orphing syste is the transforation of the spectral envelope of the source speaker to atch that of the target speaker and linear transforations estiated fro tie-aligned parallel training data are coonly used to achieve this. However, the naive application of envelope transforation cobined with the necessary pitch and duration odifications will result in noticeable artifacts. This paper studies the linear transforation approach to voice orphing and investigates these two specific issues. Firstly, a general axiu likelihood fraework is proposed for transfor estiation which avoids the need for parallel training data inherent in conventional least ean square approaches. Secondly, the ain causes of artifacts are identified as being due to glottal coupling, unnatural phase dispersion and the high spectral variance of unvoiced sounds, and copensation techniques are developed to itigate these. The resulting voice orphing syste is evaluated using both subjective and objective easures. These tests show that the proposed approaches are capable of effectively transforing speaker identity whilst aintaining high quality. Furtherore, they do not require carefully prepared parallel training data. Index Ters Voice orphing, voice conversion, linear transforation, phase dispersion I. INTRODUCTION Voice orphing which is also referred to as voice transforation and voice conversion is a technique for odifying a source speaker s speech to sound as if it was spoken by soe designated target speaker. There are any applications of voice orphing including custoising voices for TTS systes, transforing voice-overs in adverts and fils to sound like that of a well-known celebrity, and enhancing the speech of ipaired speakers such as laryngectoees. Two key requireents of any of these applications are that firstly they should not rely on large aounts of parallel training data where both speakers recite identical texts, and secondly, the high audio quality of the source should be preserved in the transfored speech. The core process in a voice orphing syste is the transforation of the spectral envelope of the source speaker to atch that of the target speaker and various approaches have been proposed for doing this such as codebook apping [1], [2], forant apping [3] and linear transforations [4], [5], [6]. Codebook apping, however, typically leads to discontinuities in the transfored speech. Although soe discontinuities can be resolved by soe for of interpolation technique [2], the conversion approach can still suffer fro a lack of robustness as well as degraded quality. On the other hand, forant apping is prone to forant tracking errors. Hence, transforation-based approaches are now the ost popular. In particular, the continuous probabilistic transforation approach introduced by Stylianou et al. [4] provides the baseline for odern systes. In this approach, a Gaussian ixture odel (GMM) is used to classify each incoing speech frae, and a set of linear transforations weighted by the continuous GMM probabilities are applied to give a soothly varying target output. The linear transforations are typically estiated fro tie-aligned parallel training data using least ean squares. More recently, Kain has proposed a variant of this ethod in which the GMM classification is based on a joint density odel[5]. However, like the original Stylianou approach, it still relies on parallel training data. Although the requireent for parallel training data is often acceptable, there are applications which require voice transforation for nonparallel training data. Exaples can be found in the entertainent and edia industries where recordings of unknown speakers need to be transfored to sound like well-known personalities. Further uses are envisaged in applications where the provision of parallel data is ipossible such as when the source and target speaker speak different languages. This paper begins by expressing the continuous probabilistic transfor of Stylianou as a siple interpolated linear transfor. Expressed in a copact for, this representation then leads straightforwardly to the realisation of the conventional training and conversion algoriths. In analogy to the transfor-based adaptation ethods used in recognition [7], [8], the estiation of the interpolated transfor is then extended to a axiu likelihood forulation which does not require that the source and training data be parallel. Although interpolated linear transfors are effective in transforing speaker identity, the direct transforation of successive source speech fraes to yield the required target speech will result in a nuber artifacts. The reasons for this are as follows. Firstly, the reduced diensionality of the spectral vector used to represent the spectral envelope and the averaging effect of the linear transforation result in forant broadening and a loss of spectral detail. Secondly, unnatural phase dispersion in the target speech can lead to audible artifacts and this effect is aggravated when pitch and duration are odified. Thirdly, unvoiced sounds have very high variance and are typically not transfored. However, in that case, residual voicing fro the source is carried over to the target speech resulting in a disconcerting background whispering effect. To achieve high quality of voice conversion, all these issues have to be taken into account and in this paper, we identify and present solutions for each of the. These include a spectral

2 refineent approach to copensate the spectral distortion, a phase prediction ethod for natural phase coupling and an unvoiced sounds transforation schee. Each of these techniques is assessed individually and the overall perforance of the coplete solution evaluated using listening tests. Overall it is found that the enhanceents significantly iprove speaker identification scores and perceived audio quality. The reainder of the paper is organised as follows. First, the transfor-based voice orphing fraework is outlined in Section II, followed by a description of the interpolated linear transfor and its estiation under different training conditions. In Section III, the various probles discussed above and their corresponding solutions are presented. The perforance of the enhanced syste with these new techniques integrated is evaluated in Section IV and finally, overall conclusions are presented in Section V. II. TRANSFORM BASED VOICE MORPHING SYSTEM A. Overall Fraework Transfor-based voice orphing technology converts the speaker identity by odifying the paraeters of an acoustic representation of the speech signal. It norally includes two parts, the training procedure and the transforation procedure. The training procedure operates on exaples of speech fro the source and the target speakers. The input speech exaples are first analyzed to extract the spectral paraeters that represent the speaker identity. Usually these paraeters encode the short-ter acoustic features, such as the spectru shape and the forant structure. After the feature extraction, a conversion function is trained to capture the relationship between the source paraeters and the corresponding target paraeters. In the transforation procedure, the new spectral paraeters are obtained by applying the trained conversion functions to the source paraeters. Finally, the orphed speech is synthesized fro the converted paraeters. Although it is outside the scope of this paper, apping the prosody of the source speaker to be like the target speaker is an equally iportant and challenging proble. In all of the work reported in this paper, the source pitch is siply shifted and scaled to atch the ean and variance of the target speaker. This is just about adequate for siilar speakers such as those used in the evaluations reported later in the paper but it is clearly not a general solution. There are three inter-dependent issues that ust be decided before building a voice orphing syste. Firstly, a atheatical odel ust be chosen which allows the speech signal to be anipulated and regenerated with iniu distortion. Previous research [9], [4], [5] suggests that the sinusoidal odel is a good candidate since, in principle at least, this odel can support odifications to both the prosody and the spectral characteristics of the source signal without inducing significant artifacts[10]. However, in practice, conversion quality is always coproised by phase incoherency in the regenerated signal, and to iniise this proble, a pitch synchronous sinusoidal odel is used in our syste [5], [11]. Secondly, the acoustic features which enable huans to identify speakers ust be extracted and coded. These features should be independent of the essage and the environent so that whatever and wherever the source speaker speaks, his/her voice characteristics can be successfully transfored to sound like the target speaker. Clearly the changes applied to these features ust be capable of straightforward realization by the speech odel. Thirdly, the type of conversion function and the ethod of training and applying the conversion function ust be decided. More details on these two latter issues are presented below. B. Spectral Paraeters As indicated above, the overall shape of the spectral envelope provides an effective representation of the vocal tract characteristics of the speaker and the forant structure of voiced sounds. Generally, there are several ways to estiate the spectral envelope, such as using LPC [12], cepstral coefficients [13] and line spectral frequencies (LSF) [15]. In Stylianou s syste [4], a set of discrete MFCC coefficients is used to represent the spectral envelope. They concluded that this ethod provides a better envelope fit at the specified frequency points than LPC-based ethods. Whilst Kain in [5] used line spectral frequencies (LSF) converted fro the LPC filter paraeters for the reason that LSFs have better linear interpolation attributes. Both ethods have been studied in our previous research in [6] and [11]. LSF is the final choice for our syste as it requires less coefficients to efficiently capture the forant structure. For cases with liited training data, this is rather crucial. Furtherore the robust interpolation properties of LSF are advantageous when using linear transforations for the conversion function. The ain steps in estiating the LSF envelope for each speech frae are as follows, 1) Use the aplitudes of the haronics a k (k = 1,,K) deterined by the pitch synchronous sinusoidal odel to represent the agnitude spectru. K is deterined by the fundaental frequency F 0, its value can typically range fro 50 to 200. 2) Resaple the agnitude spectru non-uniforly according to the bark scale frequency warping using cubic spline interpolation [14]. 3) Copute the LPC coefficients by applying the Levinson- Durbin algorith to the autocorrelation sequence of the warped power spectru. 4) Convert the LPC coefficients to LSF. In order to aintain adequate encoding of the forant structure, LSF spectral vectors with an order of p = 15 were used throughout our voice conversion experients. C. Linear Transfors We now turn to the key proble of finding an appropriate conversion function to transfor the spectral paraeters. Assue that the training data contains two sets of spectral vectors X and Y which respectively encode the speech of the source speaker and the target speaker, X = [x 1,x 2,,x τ ]; Y = [y 1,y 2,,y T ]; (1)

3 where each vector x i (or y j ) is of diension p. A straightforward ethod to convert the source vectors is to use a linear transfor. In the general case, the linear transforation of a p-diensional vector x is represented by a p (p + 1) diensional atrix W applied to the extended vector x = [x, 1]. Since there are a wide variety speech sounds, a single global transfor is not sufficient to capture the variability in huan speech. Therefore, a coonly used technique is to classify the speech sounds into classes using a statistical classifier such as a Gaussian Mixture Model (GMM) and then apply a class-specific transfor. Thus, in this case, the source data set X would be first grouped into N classes using a GMM, and then a class-specific transfor W n would be estiated for each speech class C n for n = 1,,N. However, in practice, the selection of a single transfor fro a finite set of N transforations can lead to discontinuities in the output signal. In addition, the selected transfor ay not be appropriate for source vectors that fall in the overlap area between classes. Hence, in order to generate ore robust transforations, a soft classification is preferred in which all N transforations contribute to the conversion of the source vector. The contribution degree of each transforation atrix depends on the degree to which that source vector belongs to the corresponding speech class. Thus the conversion function applied to each source vector has the following general interpolation for, N F(x) = ( λ n (x)w n ) x (2) n=1 where λ n is the interpolation weight of transforation atrix W n, and its value is given by the probability of vector x falling in speech class C n, i.e. λ n (x) = P(C n x) = α nn(x; µ n, Σ n ) N i=1 α in(x; µ i, Σ i ) where {α n }, {µ n } and {Σ n } are the weights, eans and covariances of the GMM odel respectively, and N() denotes the noral distribution. It should be noted that if λ n (x) is set as { ( 1 for n = argax P(Cn x) ) λ n (x) = 0 otherwise then a hard classification is applied to the conversion function in equation (2). The conversion function F is entirely defined by the p (p+ 1) diensional atrices W n, for n = 1,,N. Two different estiation ethods can be used to train these transforation atrices. 1) Least Square Error Estiation: When parallel training data is available, the transforation atrices can be estiated directly using the least square error (LSE) criterion. In this case, the source and target vectors are tie aligned such that each source training vector x i corresponds to a target training vector y i. For ease of anipulation, the general for of the interpolated transforation in (2) can be rewritten copactly as, (3) (4) where and F(x) = [W 1..W2....WN ] λ 1 (x) x λ 2 (x) x. λ N (x) x = WΛ(x) (5).. W = [W 1.W2...W ] N Λ(x) = λ 1 (x) x λ 2 (x) x. λ N (x) x p ( N (p+1) ( ) N (p+1) 1 ) (6) Gathering all the training vectors into single atrices X and Y as above gives the following set of siultaneous equations for estiating W, Y = WΛ(X) (8) The standard least-squares solution to equation (8) is then W = YΛ(X) ( Λ(X)Λ(X) ) 1 In practice, we use the pseudo inverse in equation (9), since for any cases where the nuber of ixtures is large and the aount of training data is liited, Λ(X)Λ(X) will becoe non-positive definite due to nuerical errors. This LSE training approach is essentially equivalent to Stylianou s approach in [4] but with a ore interpretable and flexible forulation. The accurate alignent of source and target vectors in the training set is crucial for a robust estiation of the transforation atrices. Norally a Dynaic Tie Warping (DTW) algorith is used to obtain the required tie alignent where the local cost function is the spectral distance between source and target vectors. However, the alignent obtained using this ethod will soeties be distorted when the source and target speakers are very different, this is especially a proble in cross gender transforation. Where the orthography of the training data is available, a ore robust approach is to use a speech recogniser in forced alignent ode to find corresponding phone or sub-phone boundaries. A DTW algorith can then be eployed to align the corresponding segents between the source and target utterances. In the work described here, the HTK recogniser is used [18] with a set of speaker independent onophone HMMs. The recogniser is used to force align both the source (7) (9)

4 and the corresponding target utterance, after which the utterances can be labelled into tie-arked segents where each segent corresponds to one HMM state. 2) Maxiu Likelihood Estiation: As noted in the introduction, the provision of parallel training data is not always feasible and hence it would be useful if the required transforation atrices could be estiated fro non-parallel data. The for of equation (5) suggests that, analogous to the use of transfors for adaptation in speech recognition [7], [8], axiu likelihood (ML) should provide a fraework for doing this. First consider the siple case of one global linear transfor W and assue that there is a statistical odel M that has been trained to well-represent the target speaker s speech. Then the optial linear transfor Ŵ applied to the source vectors X = {x t } would be the one that results in the converted vectors having axiu log likelihood with respect to the target speech odel, i.e. Ŵ = argax W T logp(w x t M) (10) = argax W L(W X M) (11) where, in our case, the statistical odel M is a Hidden Markov Model (HMM). There is no closed-for solution for Ŵ, but an efficient iterative solution is possible using Expectation-Maxiisation (EM). Consider the source data set X transfored at each iteration step k by W (k) to give a converted data set X (k) = { x (k) t }, where x (k) t = W (k) x t, (note that k > 0 and x (0) t = x t ), the log likelihood can then be decoposed as, L( X (k) M) = = where T logp(w (k) x t M) P(q (t) x (k 1) t, M) logp(w (k) x t M) (12) = P(q (t) x (k 1) t, M) log P( x(k) t, q (t) M) P(q (t) x (k) t, M) = Q( X (k 1), X (k) ) K( X (k 1), X (k) ) (13) Q( X (k 1), X (k) ) = K( X (k 1), X (k) ) = P(q (t) x (k 1) t, M) logp( x (k) t, q (t) M) (14) P(q (t) x (k 1) t, M) logp(q (t) x (k) t, M). (15) Here q (t) indicates Gaussian coponent of the target HMM M at tie t, and the su is taken over all coponents which can be aligned with x t. Hence P(q (t) x (k 1) t, M) = 1 which justifies the expansion in equation (12). Noting that the likelihood in equation (13) only depends on the second paraeter of Q and K, it follows that L( X (k 1) M)=Q( X (k 1), X (k 1) ) K( X (k 1), X (k 1) ) (16) and by Jensen s Inequality, K( X (k 1), X (k) ) K( X (k 1), X (k 1) ). (17) Hence if the auxiliary function Q( X (k 1), X (k) ) is axiised such that Q( X (k 1), X (k) ) Q( X (k 1), X (k 1) ), then it follows fro equations (13), (16) and (17) that L( X (k) M) L( X (k 1) M). Thus, repeated axiisation of equation (14) to find W (k), each tie updating the Gaussian coponent occupation probabilities to use the previous transfor, leads eventually to Ŵ. In practice, it is found that convergence occurs quickly and only a few iterations are required. Indeed, often just one iteration is sufficient for siilar speakers. The required axiisation at each step k proceeds by rewriting the auxiliary function in (14) (with the constant ters suppressed) as, Q( X (k 1), X (k) ) = 1 β (t) (18) 2 [ ] (W x t µ ) Σ 1 (W x t µ ) where W = W (k) is the transfor at step k, and µ and Σ are the ean vector and covariance atrix of Gaussian coponent in M, and β (t) is, β (t) = β (k 1) (t) = P(q (t) x (k 1) t, M). (19) Note that the initial value of β (0) (t) = P(q (t) x t, M). Differentiating Q in equation (18) with respect to W and equating to zero gives, β (t)σ 1 µ x t = β (t)σ 1 W x t x t. (20) The left-hand side of equation (20) is independent of W so call this Z. Introducing variables, V (t) = β (t)σ 1 (21) D (t) = x t x t (22) equation (20) can then be rewritten as Z = T V (t) WD (t). (23) Assuing that M has diagonal covariance atrices, a closed for solution can be derived by defining a new atrix G (i) with eleents[7], g (i) jq = T v (t) ii d(t) jq j, q = 1,,(d + 1) (24)

5 W is then calculated row by row using w i = G (i) 1 z i (25) where w i and z i are the i-th row of W and Z, respectively. To estiate ultiple transfors using this schee, a source GMM is used to assign the source vectors to classes via equation (4) as in the LSE estiation schee. A transfor atrix is then estiated separately for each class using the above ML schee applied to just the data for that class. Though it is theoretically possible to estiate ultiple transfors using soft classification, in practice, atrices D and G will becoe too large to invert. Hence the sipler hard classification approach is used here. As with the least ean squares ethod using parallel data, perforance is greatly iproved if sub-phone segent boundaries can be accurately deterined in the source data using the target HMM M and forced alignent recognition ode. This enables the set of Gaussians evaluated for each source frae to be liited to just those associated with the HMM state corresponding to the associated sub-phone. This does, of course, require that the orthography of the source utterances be known. Siilarly, knowing the orthography of the target training data akes training the target HMM sipler and ore effective. More details on ipleentation issues are given in the following subsection. D. Evaluation 1) Data: The VOICES database fro OGI is used for evaluation[5]. This corpus contains recorded speech fro 12 different speakers reading 50 phonetically rich sentences. Each sentence is spoken 3 ties by each speaker. The speech data was recorded at 22K Hz sapling rate using a 16 bit encoding in a professional sound-booth with high quality headphones. The recording procedure involved a iicking approach which resulted in a high degree of natural tie-alignent between different speakers. Pitch period inforation for each utterance is also provided and this was used for our pitch synchronous speech representation. In our experients, four different voice conversion tasks were investigated: ale-toale, ale-to-feale, feale-to-ale and feale-to-feale conversion. For each speaker-pair, the first 120 utterances are used as training data, and the reaining 30 utterances for the test set. 2) Objective Measure: Objective easures seek to evaluate the differences between two speech signals. Since any perceived sound differences can be interpreted in ters of differences of spectral features [16], spectral distortion is considered to be a reasonable etric both atheatically and subjectively. In speech processing, a log spectral easure is often used to deterine the distance between two spectra [17]. Siilarly in this paper, the log spectral distortion between two spectral envelopes was used to provide an objective easure of the conversion perforance d(s 1, S 2 ) = 1 K K (10log 10 a 1 k 10log 10a 2 k )2 (26) k=1 log spectral distortion ratio % 65.5 65 64.5 64 63.5 63 62.5 62 61.5 61 60.5 (a) within gender 1 2 4 6 8 10 12 nuber of tranfors LSE ML log spectral distortion ratio % 58 57.5 57 56.5 56 55.5 55 54.5 54 53.5 53 (b) cross gender 1 2 4 6 8 10 12 nuber of tranfors Fig. 1. Spectral distortion ratio for LSE and ML transfors, (a) withingender voice conversion, (b) cross-gender voice conversion. where {a k } are the aplitudes resapled fro the noralised spectral envelope S at K uniforly spread frequencies, and K is set to 100 throughout our experients. A distortion ratio is then used to copare the converted-to-target distortion with the source-to-target distortion, which is defined as, L D = d(s tgt(t), S conv (t)) L d(s 100% (27) tgt(t), S src (t)) where S tgt (t), S src (t) and S conv (t) are the target spectral envelope, source spectral envelope and the converted spectral envelope at tie t respectively. The suation in each case is coputed over tie-aligned data and L is the total nuber of test vectors after tie alignent. It should be noted that since the spectral distortion also depends on the degree to which the tie-alignent process can align siilar vectors, it is typically quite large, even when applied to the sae speaker. For exaple, the average log spectral distortion d between two utterances with identical content and spoken by the sae speaker can vary fro 5 to 10 db, whilst the distortion between two different speakers would norally be in the range fro 13 db to 20 db. So in practice, a distortion ratio of D = 50% would represent acceptable conversion perforance. Note also that a 100% distortion ratio corresponds to the distortion between the source and target spectru. 3) LSE and ML Coparison: The training of LSE transfors is straightforward. First, a GMM odel is trained on the source vectors and the interpolation weights are coputed according to equation (3). Second, a forced alignent of all utterances is coputed and sub-phone boundaries are arked. Third, DTW-based tie alignent is applied constrained by these sub-phone boundaries to produce a set of aligned sourcetarget vector pairs. In the case of the OGI Voice corpus, around 30,000 vector pairs are obtained for each speaker pair. Once the training data has been extracted, the transforation atrices can be coputed using equation (9). The ML training schee is a little ore coplex. First, the orthography of the target speaker s training data is known LSE ML

6 and used to train a onophone HMM set with 4 Gaussian ixture coponents per state. Since the data is sparse, a tiedixture technique is eployed such that the HMM states share Gaussian coponents but with different weights for different states. The sae source GMM as for LSE is used to classify the source vectors so that ultiple ML transfors can be estiated. As suggested above, the source utterances were force-aligned to ap every source training vector to a specific HMM sub-phone state, which therefore required that the orthography of the source training data is also known. The Gaussian coponent occupation probabilities were then coputed as per equation (19) and then the required transforation atrices estiated using equation (25). The nuber of iterations required depends on the source and target data. One iteration is typically sufficient for withingender conversion. However, for cross-gender conversion, two or ore iterations are necessary. Fig.1 shows the spectral distortion ratio using LSE and ML transfors. For both ethods, the distortion decreases as the nuber of transfors is increased until data sparsity results in over-training. For these experients with approxiately 30,000 training vectors per speaker, the results suggest that around 10 transfors is optial. This corresponds to 10 (15 16) = 2400 paraeters. The difference between LSE and ML transfors in the within gender voice conversion is very sall as shown by Fig.1(a), however the difference is larger for the cross gender conversion case as shown in Fig. 1(b). However, defining the signal-to-noise ratio between the LSE and ML transfored utterances as SNR = 10 log 10 N n=1 s lse(n) 2 N n=1 [s lse(n) s l (n)] 2 (28) Table I shows that the signal to noise ratio is actually very high even in the cross-gender case and should be iperceptable to huan listeners. To test this further, a foral listening test was conducted whereby listeners were presented with pairs of utterances generated by the LSE and the ML ethod respectively, and asked to select the one with the highest perceived quality. Note that in this experient only the quality of the converted speech is of concern, not the transforation accuracy of speaker identity. The latter aspect is evaluated in section IV. Table II indicates that the listeners show alost equal preference for the ML and LSE converted utterances and a two tailed t-test indicates that the difference is indeed insignificant (p=0.499 in support of the null hypothesis). TABLE I The SNR ratio in db between LSE and ML transfored utterances. within gender cross gender SNR 30.4 24.1 Note that although the distortion ratio of the cross gender conversion sees uch lower than that of the within gender conversion as shown in Fig. 1, the average log spectral distortion value is actually higher (8.83 db for cross gender and 8.37 db for within gender). This is siply because the TABLE II The result of a preference test to copare LSE and ML transfored log spectral distortion ratio % 66 64 62 60 58 56 54 utterances. ML LSE preference 48.3% 51.7% 20 40 60 80 100 120 140 nuber of Gaussian coponents in the target HMM cross gender within gender Fig. 2. Spectral distortion ratio over different nubers of Gaussian coponents in the target HMM. (10 ML transfors) source to target distortion of cross gender conversion is uch larger than that of within gender conversion. Although the differences are sall and are subjectively iperceptable, distortion is nevertheless consistently lower in all cases for LSE derived transfors copared to ML derived transfors. This ay be because the use of tiealigned parallel data in the LSE case allows the evolution of the spectral vectors to be captured whereas in the ML case, the spectral evolution is only approxiately odelled by the HMM state transitions. This suggests that iproving the odelling accuracy of the target HMM should iprove the ML transfors. Fig. 2 shows that increasing the nuber of Gaussian coponents in the target HMM can reduce the spectral distortion ratio, however this is liited by data sparsity. Fig. 3 shows that when increasing the nuber of training vectors, the spectral distortion ratio decreases for both the ML and LSE cases. Thus, not surprisingly, both ethods can benefit fro ore training data but the ML ethod can benefit fro having ore target training data even when the source data is liited. This latter point can be iportant for applications where there is a very large aount of data available for the target but only liited data for the source [21]. TABLE III The spectral distortion ratios of LSE, parallel ML and non-parallel ML transfors. LSE parallel ML non-parallel ML within gender 65.1% 67.4% 68.0% cross gender 57.1% 61.8% 61.1% The above evaluation was conducted using entirely parallel training data in order to be able to copare the LSE and

7 72 71 70 (a) within gender LSE ML 72 70 68 (b) cross gender LSE ML log spectru 2 4 6 8 LSE conversion function src lse tgt log spectral distortion ratio % 69 68 67 log spectral distortion ratio % 66 64 62 10 0 5500 11000 2 4 frequency (Hz) ML conversion function src l tgt 66 65 60 58 log spectru 6 8 64 1 2 4 8 16 all nuber of training vectors X1000 56 1 2 4 8 16 all nuber of training vectors X1000 Fig. 3. Spectral distortion ratio for single LSE and ML transfors over different nubers of training vectors, (a) within gender voice conversion, (b) cross gender voice conversion. 10 0 5500 11000 frequency (Hz) Fig. 4. Exaples of spectral envelope conversion using ML and LSE estiated linear transfors. (a) Spectral envelope conversion using LSE estiated transfors. (b) Spectral envelope conversion using ML estiated transfors. (dotted line: the source spectral envelope; lighter solid line: the target spectru; dark solid line: the converted spectral envelope.) ML approaches. However, the use of parallel data for the ML approach ay flatter the results copared to what would have been obtained with truly non-parallel training data. To test this a further experient was conducted in which the 120 training utterances for each speaker were divided into two equal sets. For the LSE estiation, the first 60 utterances of both source and target speaker were used for training. For ML estiation, however, the first 60 utterances of the target speaker were used to train the tied-ixture HMM. Then both sets of source utterances were used to generate two different ML transfors: the parallel ML transfors, and the nonparallel ML transfors. Since the training data was only half the size of the previous experients, only 4 transfors were estiated in each case. As shown in Table III, the parallel ML and the non-parallel ML transfors gave very siilar perforance, although both are worse than the LSE transfors. The latter is alost certainly because the target HMM was badly undertrained with only 60 utterances. Finally, an exaple of spectral envelope conversion using LSE and ML transfors is shown in Fig. 4. Both ethods have converted the source spectral envelope to atch the target, however any spectral details have been lost and this is a ajor cause of the spectral distortion. Moreover, listeners report that overall the converted speech is not high quality with any artifacts including a uffled effect. In the following section, these artifacts are analysed and solutions presented. III. SYSTEM ENHANCEMENT The converted speech produced by the baseline syste described above will often contain artifacts. This section discusses these artifacts in ore detail and describes the solutions developed to itigate the. A. Phase Prediction As is well known, the spectral agnitude and phase of huan speech are highly correlated. In the baseline syste, when only spectral agnitudes are odified and the original phase is preserved, a harsh quality is introduced into the converted speech. However, to siultaneously odel the agnitude and phase and then convert the both via a single unified transfor is extreely difficult. Since phase dispersion actually deterines wavefor shape, if we can predict the wavefor shape based on the spectral envelope then we can also predict the phases. Inspired by this idea, the following phase prediction approach has been developed. A GMM odel is first trained to cluster the target spectral envelopes coded via LSF coefficients into M classes (C 1,, C M ) such as in the ML estiation. Then for each target envelope y t we have a set of posterior probabilities P(C y t ). The vector P(y t ) coposed fro these probabilities can then be regarded as another for of representation of the spectral shape, P(y t ) = [P(C 1 y t ),, P(C M y t )] (29) Each eleent P(C i y t ) of this vector can be regarded as the weight of a codebook entry S i and the set of M codebook entries T = [S 1,,S M ] (30) can be chosen to iniise the coding error over the training data. That is, T can be chosen to iniize the following least square error criterion, E = N (s(t) T P(y t )) (s(t) T P(y t )) (31) where s(t) is the t th speech frae in the target training data noralized to a certain pitch value, say 100Hz. The standard solution to equation (31) is then ( N T = s(t)p(y t ) )( N P(y t )P(y t ) ) 1 (32)

8 Angle 25 20 15 10 5 0 5 10 15 target predict iniu codebook 20 0 1000 2000 3000 4000 5000 6000 7000 8000 Frequency (Hz) Fig. 5. Exaple of the unwrapped phase spectra generated by iniu phase, phase codebook and phase prediction. Light solid line: the target phase spectru. Dark solid line:the phase spectru generated by phase prediction. Dashed line: the phase spectru generated by iniu phase. Dotted line: the phase spectru generated by phase codebook. Having estiated T fro the training data, the wavefor shape of any converted spectral envelope can be predicted as s(t) = T P( x t ); (33) The required phases can then be obtained fro the predicted wavefor s(t) using the analysis routine and pitch-scale odification algorith of sinusoidal odelling. This phase prediction ethod has been copared with two other popular phase coding ethods: the iniu phase and the phase codebook approach[19]. The experients were conducted as follows. First the original signal was analyzed using the pitch synchronous sinusoidal odel, then the original phase spectra were replaced by the synthetic phase spectra generated respectively by the iniu phase ethod, the phase codebook ethod and the phase prediction ethod. In our experients, the nuber of speech classes M for each speaker is 64, depending on the nuber of training vectors that can be obtained. Additionally, to reduce the odelling error, the pitch synchronous sinusoidal odel used in our experients has autoatically adjusted the end points of each pitch period to be positioned at the zero-crossing points, and each speech frae was noralized by energy before odelling the phases. Fig. 5 shows an exaple of the unwrapped phase spectra generated by the iniu phase ethod, phase codebook and the phase prediction ethod. Clearly, the phase prediction spectra ore closely fits the target phase spectra. Table IV shows the signal to noise ratio (SNR) using the above three different phase coding ethods. The phase prediction approach outperfors the other two approaches and furtherore the iproveent in audio quality is noticeable in listening tests. TABLE IV The SNR ratio in db of three different phase coding ethods. iniu phase codebook phase phase prediction 5.7 12.3 14.4 B. Spectral Refineent As noted earlier in Fig.4, although the forant structure of the source speech has been transfored to atch the target, the spectral detail has been lost as a result of reducing the diensionality of the envelope representation during the transfor. Another clearly visible effect is the broadening of the spectral peaks caused, at least in part, by the averaging effect of the estiation ethod. All these degradations lead to uffled effects in the converted speech. To solve this proble, a straightforward idea is to reintroduce the lost spectral details to the converted envelopes. A spectral residual prediction approach has been developed to do this based on the residual codebook ethod proposed in [5], where the codebook is trained using a GMM odel. The log agnitude spectru of the spectral residual r t is calculated via r t = 20log 10 H(t) sin 20log 10 H(t) env (34) where H(t) sin is the aplitude contour of the sinusoidal coponents of speech frae t and H(t) env is the spectral envelope represented by the LSF coefficients. In our experients, r t is a 100 diensional vector resapled fro the residual contour. Each spectral residual r t is associated with an LSF vector y t, and is therefore associated with a set of posterior probabilities as in equation (29). Siilar to the phase prediction approach, a residual codebook R = [R 1, R 2,, R M ] is trained. The prediction error on the training data is defined as follows, E = T (r t RP(y t )) (r t RP(y t )) (35) and the solution to R is ( T R = r t P(y t ) )( T P(y t )P(y t ) ) 1 (36) After the residual codebook R is obtained, the spectral residual needed to copensate each converted spectral envelope can be predicted straightforwardly based on the posterior probabilities. TABLE V Effect of residual prediction as easured by log spectral distortion ratios coputed over the real spectru. within gender cross gender before RP 74.4% 73.0% after RP 54.3% 53.8% Table V shows the log spectral distortion ratio before and after residual prediction (RP). Here the log spectral distortion was coputed over the real spectru instead of the spectral envelope. As can be seen, the use of residual prediction results in a 20% absolute decrease in the spectral distortion ratio for both cross and within gender conversions. As entioned earlier, transfor-based voice conversion systes have a tendency to broaden the forants in the converted speech. To itigate this effect and suppress noise in the spectral valleys, a further spectral refineent is to apply

9 a perceptual filter to the regenerated spectral envelope of all voiced sounds. The perceptual filter is defined as, H(ω) = A(z/β) A(z/γ), 0 < γ < β 1 (37) where A(z) is the LPC filter and the choice of paraeters in our syste is β = 1.0 and γ = 0.94. This filter is popular in speech coding [20] and its ore general use in voice conversion is discussed in [6]. C. Transforing Unvoiced Sounds Unvoiced sounds contain very little vocal tract inforation and their inclusion in the envelope transforation process results in noticeable degradation. Hence, in coon with other transfor-based systes, unvoiced sounds in the baseline syste are siply copied fro the source. Many unvoiced sounds do, however, have soe vocal tract colouring and siply copying the source to the target affects the converted speech characteristics, especially in cross gender conversion. A typical effect is the perception of another speaker whispering behind the target speaker. Since ost unvoiced sounds have no obvious vocal tract structure and cannot be regarded as short ter stationary signals, their spectral envelopes show large variations. Therefore it is not effective to convert the using the sae solution as for voiced sounds. However it can be shown epirically that randoly deleting, replicating and concatenating segents of the sae unvoiced sound does not induce significant artifacts. This observation suggests a possible solution based on unit selection and concatenation to transfor unvoiced sounds. In this approach, the target training data is first labelled using the forced alignent technique entioned in the ML estiation schee, so that each speech frae is given an HMM state label together with a voiced/unvoiced decision. All these labels and the target speech fraes are then gathered together into a database. When a segent of unvoiced speech fro the source speaker needs to be transfored, each frae in the segent is first labelled with its corresponding HMM state using the sae forced alignent technique. According to the labels, target unvoiced fraes are then chosen fro the database using a criterion that encourages the selection of fraes which were adjacent in the original target data. This is done by successively selecting the longest atching HMM state sequence. For exaple, if the sequence of source labels is 1 1 1 3 3 2 1, and the longest atching sequence in the target database is 1 1 1 3 then the speech fraes corresponding to this subsequence are extracted. The procedure then repeats looking for a atch for 3 2 1 and so on until the whole of the source segent is atched. The extracted target fraes are then concatenated and their aplitudes are odified to atch the original source fraes. IV. EVALUATION OF ENHANCED SYSTEM In order to test the overall subjective quality of the voice orphing syste, listening tests were conducted to assess both the perceptual accuracy of the transforation, i.e. does the transfored source sound like the target speaker, and the audio quality. For the forer, an ABX-style preference test was perfored whereby a panel of 23 listeners were asked to judge whether an utterance X sounded closer to utterance A or B in ters of speaker identity, where X was the converted speech and A and B were either the source speech or the target speech. The source and target were chosen randoly fro both ale and feale speakers. There were 32 transfored utterances in total, equally split between within-gender and cross-gender transforations. Table VI gives the percentage of the converted utterances that were labelled as closer to the target for each case, where the baseline syste refers to the syste that only transfors the spectral envelopes and enhanced syste refers to the syste that integrates all of the refineents described in section III. The results clearly show that the enhanced syste outperfors the baseline syste in ters of transforing the speaker identity. This is probably ostly due to the inclusion of the spectral residual which contains speaker specific inforation. It is also interesting but perhaps not surprising to note that alost all the errors occurred in the within-gender transforations. TABLE VI Results fro the ABX test. baseline syste enhanced syste ABX 86.4% 91.8% To assess speech quality between the baseline syste and the enhanced syste, a second preference test was conducted whereby listeners were presented with pairs of utterances generated by the baseline syste and the new syste respectively, and then listeners were asked to judge which one has the better speech quality. Table VII indicates that ost listeners prefer the converted speech generated by the enhanced syste. Moreover, as the p-value of this t-test is 0.023, uch lower than the significance level 0.05, the difference between the enhanced syste and the baseline syste in Table VII is statistically significant. This is consistent with the previous objective evaluations. Although the relative contribution of each individual refineent is very difficult to easure, inforal tests suggest that the spectral refineent described in Section III B above contributes the ost to quality enhanceent. TABLE VII Results fro the preference test. baseline syste enhanced syste preference 38.9% 61.1% V. CONCLUSION This paper has presented a study of voice orphing based on interpolated linear transforations. The study has focussed on two ain issues. Firstly, a Maxiu Likelihood ethod of estiating the required transforation functions has been developed which does not depend on the availability of parallel training data. Coparative tests have shown that this ethod

10 is equal in perforance to least ean square estiators using parallel data however it is uch ore flexible. Secondly, the ain causes of artifacts in the converted speech have been identified as excessive spectral soothing, unnatural phase prediction and conversion of unvoiced speech. Solutions to these probles have been proposed and shown to be effective using a variety of objective and subjective easures. Overall, the results show that transfor-based voice conversion can produce the required identity change whilst aintaining acceptable quality. In particular, the flexibility of the ML training technique cobined with the described quality enhanceents offer the proise of iediate application in telephone-based applications such as custoising voice output, novelty voice-essaging, etc. Nevertheless, there is still considerable scope for further work. The ost serious weakness in the current syste is the prosodic odelling. Shifting and scaling the pitch to atch the ean and variance of the target speaker is only adequate when the speakers are siilar. When the speakers are very different (e.g. when converting a British English speaker to an Aerican English speaker), the resulting perception of identity is abiguous. Also, although the enhanceents described in this paper give a substantial iproveent in overall audio quality, there is still residual distortion aking it unsuitable for applications where studio quality is required in the converted speech. [13] O. Cappe, J. Laroche and E. Moulines, Regularized estiation of cepstru envelope fro discrete frequency points, In Proceedings of IEEE ASSP Workshop on Applications of Signal Processing to Audio and Acoustics, New York, 1995. [14] M. Unser, A. Aldroubi and M. Eden, B-Spline Signal Processing, IEEE Trans. on Signal Proc., vol. 41, no. 2, pp. 821-848, February 1993. [15] F. Itakura, Line Spectru Representation of Linear Predictive Coefficients., J Acoust Soc A, vol. 57, no. 4, pp. 535, 1975 [16] Rabiner, L.R. and B. H. Juang, Fundaental of Speech Recognition, Prentice hall, 1993 [17] A. Gray and J.D. Markel, Distance easures for speech processing, IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. ASSP-24, no.5, pp.380-391, October 1976. [18] S. Young, D. Kershaw, J. Odell, D. Ollason, V. Valtchev, P. Woodland. The HTK Book V3, Cabridge University, 2000 [19] R.J. McAulay and T.F. Quatieri, Phase Modelling and Its Application to Sinusoidal Transfor Coding, Proc. IEEE ICASSP, pp. 1713-1715, 1986. [20] J.H. Chen and A. Gersho, Real-tie vector APC speech coding at 48000 bps with adaptive postfiltering, in Proc. of the IEEE Intl. Conf. on Acoustics, Speech and Signal Processing, 1987. [21] H. Ye and S. Young, Voice Conversion for Unknown Speakers, in Proc ICSLP, Jeju, Korea, 2004. ACKNOWLEDGMENT This work was supported by a grant fro Anthropics Technology Ltd. The authors thank the volunteers of the perceptual tests for their assistance. REFERENCES [1] M. Abe, S. Nakaura, K. Shikano and H. Kuwabara, Voice conversion through vector quantization, Proc. IEEE ICASSP, 1988. [2] L. Arslan, D. Talkin, Speaker Transforation Algorith using Segental Codebooks (STASC), Speech Counication, 1999. [3] C.-H. Ho, D. Rentzos, S. Vaseghi, Forant Model estiation and transforation for Voice Morphing, Proc. ICSLP, 2002. [4] Y. Stylianou, O. Cappe and E. Moulines, Continuous probabilistic transfor for voice conversion, IEEE Trans. on Speech and Audio Processing, vol. 6, no. 2, pp. 131-142, 1998. [5] A. Kain, High resolution voice transforation, PhD dissertation, OGI, 2001. [6] H. Ye and S. Young, Perceptually Weighted Linear Transforation for Voice Conversion, Eurospeech 2003. [7] C.J. Leggetter and P.C. Woodland, Maxiu likelihood linear regression for speaker adaptation of continuous density hidden Markov odel. Coputer Speech and Language, vol. 9, pp. 171-185, 1995. [8] M.J.F. Gales, Maxiu Likelihood Linear Transforations for HMMbased Speech Recognition. Coputer Speech and Language, vol. 12, 1998. [9] E. B. George and M. J. T. Sith, Speech Analysis/synthesis and Modification Using an Analysis-by-synthesis/overlap-add Sinusoidal Model, IEEE Trans. on Speech and Audio Proc., vol. 5, no. 5, pp. 389-406, Septeber 1997. [10] T. F. Quatieri and R. J. McAulay, Shape Invariant Tie-scale and Pitch Modification of Speech, IEEE Trans. on Signal Proc., vol. 40, pp. 497-510, March 1992. [11] H. Ye and S. Young, High Quality Voice Morphing, In Proceedings of ICASSP 2004. [12] J. Wouter and W. Macon, Control of Spectral Dynaics in Concatenative Speech Synthesis, IEEE Trans. on Speech and Audio Proc., vol. 9, no. 1, pp. 30-38, January 2001.