Frequency-domain. Time-domain. time-aliasing. Time. Frequency. Frequency. Time. Time. Frequency

Size: px
Start display at page:

Download "Frequency-domain. Time-domain. time-aliasing. Time. Frequency. Frequency. Time. Time. Frequency"

Transcription

1 IEEE TRASACTIOS O SPEECH AD AUDIO PROCESSIG, VOL. XX, O. Y, MOTH Synthesis of sinusoids via non-overlapping inverse Fourier transform Jean Laroche Abstract Additive synthesis is a powerful tool for the analysis/modication/synthesis of complex audio or speech signals. However, the cost of wavetable sinusoidal synthesis can become prohibitive for large numbers of sinusoids (more than a few hundred). In that case, techniques based on the inverse Fourier transform oer an attractive alternative, being 200% to 300% more ecient than wavetable synthesis depending on the number of sinusoids. This paper presents an improved technique based on the concatenation of shortterm signals obtained by inverse Fourier transforms. By contrast with the standard overlap-add technique, the new algorithm requires synthesizing sinusoids in the frequency domain whose time-domain amplitudes vary linearly within the synthesis frame. The technique is shown to achieve higher quality than the standard overlap-add technique, at the cost of a small increase in computation. Keywords I. Introduction MAY musical and audio applications require the calculation of large numbers of sinusoids to synthesize a signal. Examples include sinusoidal analysis/synthesis [1], [2], [3], [4], sinusoidal coding [5], [6]. In most cases, the sinusoids have time-varying parameters to account for the non-stationary nature of the signal to be modeled, and their frequencies, amplitudes and phases are controlled over time. A traditional way of synthesizing sinusoids consists of using a "wavetable" synthesis technique [7], [8] in which one cycle of the sinusoid (or one half-cycle) is pre-calculated and stored in a table. At synthesis time, the signal is read from the table at a rate which depends on the desired frequency, and linear interpolation is used optionally to improve the accuracy of the synthesized signal. The technique is easy to implement, lends itself very well to VLSI applications, but is moderately ecient in terms of computation. Another sinusoidal synthesis technique consists of using a highly-resonant second-order lter excited by a very short impulse-like signal [9], [10], [11]. While this technique has the additional advantage of enabling source-lter synthesis (in which an audio signal is fed into the resonant lter), it is usually much more costly than wavetable synthesis and is therefore limited to the synthesis of small numbers of sinusoids. When large numbers of sinusoids must be synthesized (for example a few thousand), the computational cost of wavetable synthesis becomes prohibitive too, and alternative techniques must be used. One such technique uses an inverse Fourier transform to generate a -point frame containing a sum of several sinusoids [12], [13], [14], [15], [16]. The technique consists of synthesizing each sinusoid's trace in the frequency do- Jean Laroche is with the Joint Creative/Emu Technology Center, Scotts Valley, CA, USA. jeanl@emu.com main, then taking an inverse Fast Fourier Transform (FFT) of the result, and normalizing and overlapping the resulting frame with the preceding frame in a process known as overlap-add synthesis. This is advantageous because a large number of samples of one sinusoid can be relatively accurately represented by a few points in the frequencydomain (typically four to nine points). As a result, when large numbers of sinusoids are needed, the overhead of the inverse-fft/overlap-add stages is easily oset by the low cost of synthesizing a few points per sinusoid in the frequency domain. The technique allows the synthesis of a few thousand sinusoids with time-varying parameters on current general-purpose processors while wavetable synthesis would only allow the synthesis of a few hundred. The cost saving increases with the size of the Fourier transform, as the overhead of the inverse-fft/overlap-add stages is distributed among a larger number of output samples. The signals synthesized by overlap-add inverse-fft techniques are usually not perfectly sinusoidal: The fact that only a few points are synthesized in the frequency domain generates distortions which can be minimized (but not eliminated) by an appropriate choice of the window template. In addition, for sinusoids with time-varying frequencies, in order to avoid amplitude modulation resulting from incoherence during the overlap-add stage, one needs to synthesize sweeping sinusoids, i.e., sinusoids with a frequency that varies linearly within an FFT frame, or better yet, sinusoids with two sweeping rates within one FFT frame [15]. This results in a more complex, more expensive algorithm because the frequency-domain trace of the sinusoid is less simple. In this paper, we present an alternative technique based on non-overlapping inverse-fft synthesis, in which each synthesized frame is simply concatenated with the preceding one, instead of overlap-added. Getting rid of the overlapadd stage eliminates the modulation problems mentioned above. The drawback of this technique is that the linearlyvarying amplitude of the sinusoid must be synthesized in the frequency domain, which makes the algorithm slightly more complex. The rst section briey describes the standard overlap-add technique. In the second section, the non-overlapping technique is introduced and various quality/eciency tradeos are discussed. II. Standard overlap-add inverse-fft synthesis of sinusoids In this section, we briey present the standard overlapadd inverse-fft synthesis technique [12], [13], [14], [15], [16] and discuss the modulation problems arising during the overlap-add stage.

2 2 IEEE TRASACTIOS O SPEECH AD AUDIO PROCESSIG, VOL. XX, O. Y, MOTH 1999 A. Technique The standard overlap-add inverse-fft synthesis requires the choice of what is called a "spectral motif" H(), the representation of a sinusoid in the frequency-domain, given a time-domain window h(n). Specically, we have: H() = 1X i=?1 h(i)e?ji The window h(n) is chosen for its spectral properties: Its main lobe must be as narrow as possible and its side lobes as low as possible so it can be represented with as few points as possible. Typical choices include the Blackman-Harris family, the Kaiser window and so on [17]. For real-time synthesis purposes, the spectral motif H() is precalculated, sampled and stored in a table. Because this spectral motif will later be positioned arbitrarily in the frequency domain, depending on the frequency of the sinusoid to be synthesized, it should be sampled nely enough to make the corresponding frequency inaccuracy inaudible. The size of the spectral motif table does not need to be very large though, since only the positive-frequency half of the main lobe and optionally the rst side lobe need to be stored. In order to synthesize a sinusoid with an amplitude A, a phase at the center of the FFT frame, and a constant frequency!, the spectral motif is copied at the right frequency, with the appropriate amplitude and phase. Denoting Y ( k ) the frequency-domain -point FFT frame with k = 2k=, we have Y ( k ) = A 2 ej H( k?!) for j k?!j < K 2 (1) where K controls the number of FFT channels that are dened for each sinusoid. K ranges typically from 3 to 4. This process is repeated for each sinusoid and the resulting frequency-domain frame Y ( k ) is then inverse- Fourier transformed, yielding the time-domain signal y u (n) at frame u. This short-term signal is then normalized by the inverse of the analysis window 1=h(n) and multiplied by a triangular window g(n). This ensures that the amplitude of each sinusoid will have a triangular shape within the FFT frame, and therefore that the amplitude of each sinusoid after the overlap-add will vary linearly from one frame to the next. This stage is performed in one step, the ratio g(n)=h(n) being stored in a table. Care must be taken for g(n)=h(n) not to take large values at the beginning and the end of the FFT frame (i.e., g(n) must decrease faster than h(n)). Finally, successive normalized frames are overlapadded with a 50% overlap, yielding the output synthesized signal. B. Problems in the overlap-add stage In the simplest method described above, a sinusoid with a time-varying frequency is synthesized by overlap-adding short-term sinusoids with constant frequencies within each FFT frame. This means that in the region where frames u and u + 1 overlap, a sinusoid with frequency! u is added to a sinusoid with a frequency! u+1 close but not exactly equal to! u. As a result, regardless of the dierence of phases at the centers of frames u and u + 1, some degree of amplitude modulation will occur, due to the beating between the two close-frequency sinusoids. The solution proposed in [13] to minimize this problem consists of making the phase of the sinusoid in frame u match that of the sinusoid in frame u + 1 halfway in the overlap region. This only minimizes the artifact, but does not require any additional computation. A more accurate solution consists of synthesizing a sinusoid with a frequency that varies linearly within an FFT frame, with two dierent slopes in the rst and second half-frames [15]. The slope of the second half of frame u is constrained to match that of the rst half of frame u + 1, and as a result, the instantaneous frequencies of the sinusoids in the overlap region coincide perfectly. With the right choice of phases at the center of the FFT frames, the two sinusoids overlap-add perfectly (i.e., with no phase-mismatch amplitude modulation). Unfortunately, synthesizing sinusoids with two dierent sweep rates within an FFT frame adds signicantly to the complexity of the algorithm. Several spectral motifs must be stored for a series of sweep rates, the synthesis of each sinusoid's frequency-domain pattern requires more calculations, and each frame must be synthesized by two half-size inverse Fourier transforms, undermining the cost saving of the overall algorithm. The algorithm presented below eliminates the overlap-add stage by simply allowing successive frames to be concatenated. Overlap-add modulation artifacts are therefore eliminated. III. on-overlapped inverse-fft synthesis of sinusoids A. Underlying idea If the successive frames obtained by inverse Fourier transforms are concatenated instead of overlap-added, one must make sure that amplitudes and phases are "continuous" from frame to frame. Because amplitudes generally vary from frame to frame, ensuring amplitude continuity requires synthesizing sinusoids with time-varying amplitudes within the FFT frame, instead of constant amplitudes as in the overlap-add synthesis technique. In practice, the easiest way to achieve this is to synthesize sinusoids with a linear amplitude variation within the FFT frame, which can be done relatively easily in the frequency domain. ote that it is still possible to use sinusoids with constant frequencies within the FFT frame, because frequency jumps from frame to frame are not too objectionable if the frame rate is high enough. Another important point is that because only a few frequency points are synthesized for each sinusoid, the time-domain short-term signal exhibits distortion, especially near the beginning and the end of the frame. As a result, a few boundary samples must be discarded before the concatenation. The following sections presents the details of the algorithm.

3 LAROCHE: SYTHESIS OF SIUSOIDS VIA O-OVERLAPPIG IVERSE FOURIER TRASFORM 3 B. Synthesizing linear-amplitude sinusoids The spectral motif of a sinusoid with a linearly varying amplitude (more precisely, an ane amplitude) includes two terms, one corresponding to a constant amplitude, and one corresponding to a linear amplitude. It is easier to derive the exact shape of the spectral motif for a null-frequency sinusoid: Given an analysis window h(n), the Fourier transform of a weighted ane function x(n) = h(n)(a + n) is easily obtained from standard Fourier transform results. If u(n) $ U() ($ indicates that U() is the Fourier transform of u(n)), then nu(n) $ j du() d [18]. As a result, the Fourier transform of x(n) is X() = AH() + j dh() d where H() is the Fourier transform of h(n). We see from Eq. (2) that the spectral motif is simply a linear combination of two parts, and therefore we now need to store two spectral motif tables, one corresponding to H() and another one corresponding to its frequency derivative: H 0 () = dh() d It is useful to notice that since h(n) usually is a real symmetrical window, H() is real symmetrical, and H 0 () is real odd-symmetrical, and therefore only the positivefrequency halves ( 0) must be stored. In order to synthesize a sinusoid of frequency!, with a phase, a constant amplitude A and a linear amplitude at the center of the frequency-domain frame, one needs to generate a sinusoidal spectral pattern from the stored spectral motif tables: Y ( k ) = e j A 2 H( k?!) + 2 H0 ( k?!) for (2) j k?!j K 2 (3) in which is the size of the Fourier transform and k = k2=. K controls how many frequency bins are synthesized in the frequency domain around the sinusoidal peak. More will be said later on how K should be selected. The 1=2 terms in Eq. (3) come from the fact that a sinusoid with a strictly positive frequency is represented by two halfamplitude spectral motifs at!. In practice, the signals to be synthesized are real, and the inverse Fast Fourier Transform algorithm only uses the positive-frequency half spectrum, so only one of the two spectral motifs must be synthesized. If! < K 2, then k will become negative in Eq. (3), and the corresponding part of the motif should be reected back into the positive frequencies with a complex conjugation, to account for the fact that the (not synthesized) negative-frequency spectral motif spills into the positive-frequency axis. More specically, denoting G( k ;!) = e j A 2 H( k?!) + 2 H0 ( k?!) Eq. (3) must be replaced by Y ( k ) = 8>< >: 2< fg(0;!)g for k = 0 G( k ;!) + G( k ;?!) for 0 < k?! + 2K G( k ;!) for?! + 2K < k! + 2K where < fzg is the real part of complex z. As in the overlap-add synthesis technique, H() and H 0 () are precalculated, sampled and stored in a table. Because the motif is sampled, it cannot be placed at arbitrary frequencies!: if H() and H 0 () are sampled for = 2=M, then the motif can be placed at discrete frequencies ^! = 2n=M. If M is large enough, the small frequency error can be conned within a range where it is inaudible. The quantized value ^! of the frequency! must be used, however, during the calculation of the frame-toframe phase increment (see below). The above procedure is repeated for every sinusoid to be synthesized, the spectral patterns being added together where they overlap. C. Obtaining the short-term frame Once the spectral patterns corresponding to all the sinusoids have been synthesized, the inverse Fourier transform of the frequency-domain frame Y ( k ) is calculated, usually by means of a Fast Fourier Transform (FFT) algorithm. The result is a short-term signal y u (n), where the subscript u indicates the frame index. In y u (n), the sinusoids have an ane amplitude multiplied by the analysis window h(n) corresponding to the spectral motif H(). To restore the desired ane amplitude, y u (n) is normalized by the inverse of h(n), which requires h(n) to never take null values: y0 u (n) = y u(n) h(n) The distortions introduced by the limited number of points in the spectral patterns are amplied by the normalization Eq. (5), especially where h(n) takes on small values (typically h(n) is a tapered window whose extremities can become quite small). As a result, the samples at the beginning and at the end of the short-term signal exhibit a higher level of distortion than samples located around the middle of the frame. For that reason, D samples are discarded from the beginning and the end of each short-term frame before the frames are concatenated. More will be said below on how D should be chosen, but a typical value is one tenth of the frame size. D. Ensuring amplitude and phase coherence Each short-term frame is concatenated with the preceding one, once D boundary samples have been discarded from both ends of the frame. As a result, the centers of two consecutive frames are?2d samples apart, and this must be taken into account when calculating the phase (4) (5)

4 4 IEEE TRASACTIOS O SPEECH AD AUDIO PROCESSIG, VOL. XX, O. Y, MOTH 1999 and amplitude terms A and for each sinusoid. Specically, the phase u+1 at the center of frame u + 1 must be equal to u+1 = u + ^! u? 2D 2 + ^! u+1? 2D 2 where ^! u is the quantized value of the sinusoid's frequency in frame u. Eq. (6) accounts for the fact that the frequency in the second half-frame u is ^! u while that in the rst halfframe u + 1 is ^! u+1. Similarly, the amplitude a e u of the sinusoid at the end of frame u must match the amplitude a b u+1 of the same sinusoid at the beginning of frame u + 1 (the superscripts e and b respectively denote the end and the beginning of the frame). A u and u in Eq. (3) can be derived from a b u and a e u by ( A u = 1 2 (ae u + a b u) u = 1?2D (ae u? ab u) and If the above continuity equations hold, the signal resulting from the concatenation of the short-term frames will be free of frame-synchronous artifacts. In the following section, we investigate how the various parameters of the algorithm should be set for optimal results. IV. Choosing the right parameters The quality of the signal synthesized by the algorithm described above greatly depends on the choice of the analysis window h(n) and of its length, on the frequency at which the spectral motifs H(!) and H 0 () are sampled, on the number 2K + 1 of frequency points synthesized per sinusoid, and on the number D of samples discarded at frame boundaries prior to the concatenation stage. A. Choosing the frame length The choice of the frame length is mainly a tradeo between computational cost and the rate at which the sinusoidal frequencies are allowed to vary. Since the frequencies of the sinusoids are constant during each frame, and since? 2D samples are synthesized per frame, the sinusoidal frequencies cannot change faster than every? 2D samples. If the frequencies vary slowly in time, large values of can be used without generating audible artifacts. By contrast, for rapidly varying sinusoids, using large frame lengths would yield noticeable frequency "jumps". Unfortunately, there is no precise psycho-acoustical data available on the perceptual detectability of frequency "discontinuities" in sinusoids with piece-wise constant frequencies. Informal tests have shown that for speech signals, the frequencies should be updated at least every 10 to 20 ms, which imposes an upper bound on how large can be. Large values of are desirable, because the computational cost related to the inverse Fourier transform is nearly independent of : the cost of one length- inverse Fourier transform is roughly proportional to log 2, and each inverse FFT yields? 2D samples. The cost of the inverse Fourier transform per output sample is therefore roughly proportional to log?2d 2 which increases very slowly (6) (7) with for standard values (from = 1024 to 4096 the cost increases by about 20% for D = =10). While the cost of the FFT is nearly independent of, the cost per output sample of synthesizing the spectral patterns is roughly inversely proportional to since? 2D samples are synthesized at each frame. As a result, for eciency, should be chosen as large as the rate of frequency variations can allow. B. Choosing the spectral motif, K and D The choice of the spectral motif, or equivalently of the analysis window h(n) is crucial in that it can strongly in- uence the quality of the synthesized signal. Its inuence is tightly tied to the choice of the number 2K + 1 of synthesized frequency bins and the number D of samples discarded at frame boundaries. To better understand the roles played by these parameters, it is useful to examine the kind of distortion introduced by the fact that only 2K +1 points are synthesized per sinusoid in the frequency domain. The easiest way to analyze the distortion consists of considering that a continuous-frequency, band-limited spectral motif ~H() is used: ( H() for jj K ~H() 2 = 0 for jj > K 2 And because H() ~ is band-limited, it corresponds to a non time-limited analysis window h(n). ~ Synthesizing the spectral pattern of a sinusoid consists of shifting H() ~ to the proper center frequency ^! and then sampling it in the frequency-domain at frequencies k = 2k (as is done implicitly in the FFT). Shifting H() ~ corresponds to modu- lating h(n) ~ by a innite-duration complex exponential of frequency ^!. Sampling it at frequencies k = 2k causes the modulated h(n) ~ to be replicated or imaged every samples, and because h(n) ~ is not time-limited, this timedomain imaging causes what is called time-aliasing: portions of the images leek into the "useful" part of the window as shown in Fig. 1. Because of this time-aliasing, the output of the inverse Fourier transform is not h(n) ~ modulated by a complex exponential but includes a sum of (low-amplitude) undesirable components modulated at the same frequency ^!. Unfortunately, how these undesirable components add up depends on ^! and therefore the distortion is a function of the sinusoidal frequency. This means that it is not possible to simply cancel it during the normalization stage Eq. (5) by normalizing by an appropriate window. An important point is that these additive aliased components can be severely amplied during the normalization stage, wherever h(n) becomes small. This is the reason why D samples are discarded from the edge of each frame (where h(n) usually becomes small). The additive time-aliased signals cause the signal in the "useful" frame?=2 < n =2 to be amplitude-modulated (since all the time-aliased signal have the same frequency but dierent phases). Even though relatively large levels of amplitude modulation can pass unnoticed when the modulation is "continuous" (does not exhibit jumps), the situation here is more stringent

5 LAROCHE: SYTHESIS OF SIUSOIDS VIA O-OVERLAPPIG IVERSE FOURIER TRASFORM 5 Frequency-domain Time-domain following section presents an example of such a procedure. π π π 2πK/M π π π Frequency Frequency Frequency /2 -/2 /2 /2 time-aliasing Fig. 1. Time-aliasing resulting from "band-limiting" the spectral motif. Stage 1 (top) shows the continuous-frequency spectral motif (left) and its time-limited corresponding signal (right). Band limiting the spectral motif in stage 2 yields a time-domain signal which extends outside of?=2 =2. Sampling the spectral motif in the frequency-domain in stage 3 corresponds to imaging the time-domain signal, yielding time-aliasing. because the successive frames are concatenated. As a result, sudden discontinuities in the amplitude can occur at the boundary between two frames with dierent patterns of modulation. These discontinuities must be kept below a very small level to remain unnoticeable. The choice of h(n), K and D inuences the amount of distortion in the following manner: h(n) should be chosen such that h(n) ~ (which results from band-limiting H()) has as little an amplitude outside of?=2 < n =2 (so time-aliasing is minimized) and as large an amplitude inside of?=2 < n =2 (so the normalizing stage amplies the aliased components as little as possible). Of course, these are conicting requirements! K must be as small as possible (to reduce the cost of synthesizing each sinusoidal spectral pattern), but the smaller K, the larger the time-aliasing and therefore the larger the distortion. D should also be chosen as small as possible (discarding samples means wasting computations) while keeping the amplication as small as possible in the normalizing stage Eq. (5). ote that the above discussion can be applied to H 0 (). Since its band-limited version H ~ 0 () corresponds in the time-domain to n h(n), ~ it is clear that time-aliasing is even more severe because the multiplicative term n gets larger outside of?=2 < n =2. Because of the large number of parameters, there doesn't seem to be an easy way to obtain an optimal choice of h(n), K and D. In practice, it is convenient to pick a value of K, for example K = 3. Then one can choose one of various standard candidate windows h(n). Finally, D can be picked to keep the distortion below a desired level. The Time Time Time C. Examples of parameter settings In this section, we compare three standard windows from the point of view of the output distortion. The windows are the Hanning window, the Kaiser window with a parameters of 9 and a Chebyshev equiripple window with a -80dB stopband level. To measure the amount of distortion, we measure the level of the components that time-alias back between?=2 < n =2. Specically, given a window h(n), H() is calculated by use of a highly oversampled Fourier transform: If h(n) is a -point window, we use a Fourier transform of size M = O to calculate H() (O can range from 8 to 32). Then H() ~ is obtained by zeroing H() for jj > K 2. Then h(n) ~ is calculating as the inverse Fourier transform of H(). ~ ow contrary to h(n), ~h(n) is not null outside of [?=2 =2], which is the reason why time-aliasing occurs. The amount of time-aliasing in [?=2 =2] is calculated by summing the absolute values of images falling into [?=2 =2] and normalizing by ~h(n): E(n) = k=o=2?1 X k=?o=2+1;k6=0 j ~ h(n)j j ~ h(n? k)j The absolute value in Eq. (8) yields a conservative upper bound for the aliasing distortion. Fig. 2 displays E(n) in db for the three windows, = 1024 and K = 3. Based on Fig. 2, one can set the number of discarded samples D so the level of aliasing is guaranteed to never exceed a given level. The Hanning window is shown to yield the worst results: to keep E(n) under a given level, more samples must be discarded for the Hanning window than for any of the two other windows. This can be explained by the fact that the Hanning window has fairly high side-lobes in the frequency domain, and discarding them by picking a small value for K yields a signicant amount of time-aliasing. In practice, it is sucient to ensure that E(n) is below -40dB to avoid any audible artifact. For the Kaiser window and the Chebyshev window, this requires discarding D = 90 samples from both ends of the frame. To achieve -50dB or more, the Kaiser window would require discarding 170 samples while the Chebyshev window would require discarding 110 samples. ote that for a larger FFT size, these results can simply be scaled proportionally. V. Discussion and conclusion To conclude this paper, we discuss the relative merits of both the standard overlap-add inverse Fourier transform synthesis technique and the new, non-overlapping technique with regard to computation cost, signal distortion, and ease of implementation. A. Computational cost As was mentioned above for the non-overlapping technique, if the Kaiser window is used, one can use D =10 (8)

6 6 IEEE TRASACTIOS O SPEECH AD AUDIO PROCESSIG, VOL. XX, O. Y, MOTH non-overlapping technique shows a very minimal amount. ote that as mentioned above, the overlapping technique 0 10 Hanning E in db Kaiser Chebyshev Sample index Fig. 2. E(n) in db for a Hanning window (solid line), a Kaiser window (dash-dotted line) and a Chebyshev window (dashed line) as a function of the sample index. Only the rst half of the frame is represented. and obtain high-quality results. This means that at each frame,? 2D = 0:8 samples are synthesized. For the overlapping technique =2 samples are synthesized at each frame (a 50% overlap must be used). As a result, for a given output length, about 62% fewer frames are calculated in the non-overlapping technique than in the overlapping technique. Counteracting this saving is the fact that each frame requires more calculations in the non-overlapping technique because the linear amplitude must be synthesized in the frequency domain, requiring the additional term H 0 () in Eq. (3). This amounts roughly to twice as many operations per sinusoid than in the overlapping technique where only the rst half of the equation is used. If the cost of the synthesis is dominated by the cost related to the sinusoidal pattern synthesis (Eq. (3)) (as opposed to the cost of the inverse Fourier transform), then each frame in the non-overlapping technique requires about twice as many computations than in the overlapping technique. In that case, the new technique ends up being about 25% more costly than the overlap-add technique. B. Signal distortion Because the short-term signals are concatenated rather than overlap-added, the new technique is free of the undesirable amplitude modulation that results from overlapadding sinusoids with dierent frequencies. This is a signicant advantage over the overlapping technique. Fig. 3 compares the outputs of the two algorithms, when synthesizing a constant-amplitude complex sinusoid with a sweeping frequency. The normalized frequency started at 2 0:015 radian/sample and increased at a rate of 2 5:10?6 radian/sample 2. The FFT size was = 1024, a Kaiser window with parameter = 9 was used in both techniques, and the parameters were K = 3 and D = 100. The gure shows the magnitude of the output complex sinusoid as a function of time. The overlap-add technique exhibits a very large amount of amplitude modulation while the Amplitude Time in sample Fig. 3. Comparison of the overlapping (top) and non-overlapping (bottom) synthesis techniques for the synthesis of one complex sinusoid with a sweeping frequency. The magnitude of the complex output is shown as a function of the sample index. can be improved to reduce such modulation problems [15], but the computational cost roughly doubles and the algorithm becomes much more complex. An important point is that contrary to the overlapping technique, the nonoverlapping technique requires the sinusoid amplitudes to start from 0 when the sinusoid rst appears, and end at 0 when the sinusoid disappears, to avoid any signal discontinuity. The overlapping technique relies on the overlapadd stage to achieve that, and consequently sinusoids can disappear from or appear in a frame without any adverse consequences. C. Ease of implementation The non-overlapping technique, if anything, is easier to implement than the overlapping technique, because the overlap-add stage is discarded. The additional term in Eq. (3) which codes the linear amplitude in the frequency domain is very similar to the rst one, and does not pose any specic implementation problems, other than requiring one additional table for H 0 (). By contrast, the technique described in [15] to solve amplitude modulation problems in the overlapping technique is signicantly more complex, requiring several tables to be stored for various frequencysweep factors, and an additional linear interpolation stage. D. Conclusion The technique presented in this paper diers from the standard technique in that the short-term signals are concatenated rather than overlap-added, and that sinusoids with linear amplitudes rather than constant amplitudes are synthesized in each frame. The technique oers a signicant quality improvement over the standard technique as it eliminates amplitude modulation problems arising from overlap-adding sinusoids with close frequencies. The new technique is slightly more costly (at most 25%) than the

7 LAROCHE: SYTHESIS OF SIUSOIDS VIA O-OVERLAPPIG IVERSE FOURIER TRASFORM 7 standard technique, but in high-quality synthesis applications the quality improvement can justify this small additional cost. References [1] L.B. Almeida and F.M. Silva, \Variable-frequency synthesis: an improved harmonic coding scheme," in Proc. IEEE Int. Conf. Acoust., Speech, Signal Processing, 1984, pp { [2] R. J. McAulay and T. F. Quatieri, \Speech analysis/synthesis based on a sinusoidal representation," IEEE Trans. Acoust., Speech, Signal Processing, vol. ASSP-34, no. 4, pp. 744{754, Aug [3] E. B. George and M. J. T. Smith, \Analysis-bysynthesis/Overlap-add sinusoidal modeling applied to the analysis and synthesis of musical tones," J. Audio Eng. Soc., vol. 40, no. 6, pp. 497{516, [4] X. Serra and J. Smith, \Spectral modeling synthesis: A sound analysis/synthesis system based on a deterministic plus stochastic decomposition," Computer Music J., vol. 14, no. 4, pp. 12{24, Winter [5] R. J. McAulay and T. F. Quatieri, \Low-rate speech coding based on the sinusoidal model," in Advances in Speech Signal Processing, S. Furui and M. Sondhi, Eds., chapter 6, pp. 165{ 208. Marcel Dekker, [6] J.S. Marques, L.B. Almeida, and J.M. Tribolet, \Harmonic coding at 4.8 kb/s," in Proc. IEEE Int. Conf. Acoust., Speech, Signal Processing, 1990, vol. 1, pp. 17{20. [7] M. Kahrs and K. Brandenburg, Applications of Digital Signal Processing to Audio and Acoustics, Kluwer, orwell, MA, [8] J. Snell, \Design of a digital oscillator which will generate up to 256 low distortion sine waves in real time," Computer Music J., vol. 1, no. 2, pp. 4{25, [9] Y. Potard, P. F. Baisnee, and J. B. Barriere, \Experimenting with models of resonance produced by a new technique for the analysis of impulsive sounds," in Proc. Intern. computer music conf., [10] J. Laroche and J.L. Meillier, \Multichannel excitation/lter modeling of percussive sounds with application to the piano," IEEE Trans. Speech and Audio Processing, vol. 2, no. 2, pp. 329{344, Apr [11] J. Laroche, \Using resonant lters for the synthesis of timevarying sinusoids," Proc. 105th A.E.S Convention, San Francisco, 1998, preprint 4782 (F-6). [12] X. Rodet and P. Depalle, \A new additive synthesis method using inverse Fourier transform and spectral envelope.," in Proc. Intern. computer music conf., San Jose, CA, [13] X. Rodet and P. Depalle, \Spectral envelopes and inverse FFT synthesis," Proc. 93rd A.E.S Convention, San Francisco, 1992, preprint 3393 (H-3). [14] M. Goodwin and X. Rodet, \Ecient Fourier synthesis of nonstationary sinusoids," in Proc. Intern. computer music conf., San Francisco, CA, [15] M. Goodwin and A. Kogon, \Overlap-add synthesis of nonstationary sinusoids," in Proc. Intern. computer music conf., Ban, Canada, [16] M. Goodwin, Adaptive Signal Models: Theory, algorithms and audio applications, Kluwer International Seires in Engineering and Computer Science, orwell, MA, [17] J. Harris, \On the use of windows for harmonic analysis with the discrete Fourier transform," Proc. IEEE, vol. 66, no. 1, pp. 51{83, [18] A. V. Oppenheim and R. W. Schafer, Discrete-Time Signal Processing, Prentice Hall, Englewood Clis, ew Jersey, 1989.

Timbral Distortion in Inverse FFT Synthesis

Timbral Distortion in Inverse FFT Synthesis Timbral Distortion in Inverse FFT Synthesis Mark Zadel Introduction Inverse FFT synthesis (FFT ) is a computationally efficient technique for performing additive synthesis []. Instead of summing partials

More information

FREQUENCY-DOMAIN TECHNIQUES FOR HIGH-QUALITY VOICE MODIFICATION. Jean Laroche

FREQUENCY-DOMAIN TECHNIQUES FOR HIGH-QUALITY VOICE MODIFICATION. Jean Laroche Proc. of the 6 th Int. Conference on Digital Audio Effects (DAFx-3), London, UK, September 8-11, 23 FREQUENCY-DOMAIN TECHNIQUES FOR HIGH-QUALITY VOICE MODIFICATION Jean Laroche Creative Advanced Technology

More information

SINOLA: A New Analysis/Synthesis Method using Spectrum Peak Shape Distortion, Phase and Reassigned Spectrum

SINOLA: A New Analysis/Synthesis Method using Spectrum Peak Shape Distortion, Phase and Reassigned Spectrum SINOLA: A New Analysis/Synthesis Method using Spectrum Peak Shape Distortion, Phase Reassigned Spectrum Geoffroy Peeters, Xavier Rodet Ircam - Centre Georges-Pompidou Analysis/Synthesis Team, 1, pl. Igor

More information

Audio Engineering Society Convention Paper Presented at the 110th Convention 2001 May Amsterdam, The Netherlands

Audio Engineering Society Convention Paper Presented at the 110th Convention 2001 May Amsterdam, The Netherlands Audio Engineering Society Convention Paper Presented at the th Convention May 5 Amsterdam, The Netherlands This convention paper has been reproduced from the author's advance manuscript, without editing,

More information

Non-stationary Analysis/Synthesis using Spectrum Peak Shape Distortion, Phase and Reassignment

Non-stationary Analysis/Synthesis using Spectrum Peak Shape Distortion, Phase and Reassignment Non-stationary Analysis/Synthesis using Spectrum Peak Shape Distortion, Phase Reassignment Geoffroy Peeters, Xavier Rodet Ircam - Centre Georges-Pompidou, Analysis/Synthesis Team, 1, pl. Igor Stravinsky,

More information

A Parametric Model for Spectral Sound Synthesis of Musical Sounds

A Parametric Model for Spectral Sound Synthesis of Musical Sounds A Parametric Model for Spectral Sound Synthesis of Musical Sounds Cornelia Kreutzer University of Limerick ECE Department Limerick, Ireland cornelia.kreutzer@ul.ie Jacqueline Walker University of Limerick

More information

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters FIR Filter Design Chapter Intended Learning Outcomes: (i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters (ii) Ability to design linear-phase FIR filters according

More information

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters

(i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters FIR Filter Design Chapter Intended Learning Outcomes: (i) Understanding of the characteristics of linear-phase finite impulse response (FIR) filters (ii) Ability to design linear-phase FIR filters according

More information

TIME DOMAIN ATTACK AND RELEASE MODELING Applied to Spectral Domain Sound Synthesis

TIME DOMAIN ATTACK AND RELEASE MODELING Applied to Spectral Domain Sound Synthesis TIME DOMAIN ATTACK AND RELEASE MODELING Applied to Spectral Domain Sound Synthesis Cornelia Kreutzer, Jacqueline Walker Department of Electronic and Computer Engineering, University of Limerick, Limerick,

More information

Amplitude Modulated Sinusoidal Models for Audio Modeling and Coding

Amplitude Modulated Sinusoidal Models for Audio Modeling and Coding Amplitude Modulated Sinusoidal Models for Audio Modeling and Coding Mads Græsbøll Christensen, Søren Vang Andersen, and Søren Holdt Jensen Department of Communication Technology, Aalborg University, Denmark

More information

Sound Synthesis Methods

Sound Synthesis Methods Sound Synthesis Methods Matti Vihola, mvihola@cs.tut.fi 23rd August 2001 1 Objectives The objective of sound synthesis is to create sounds that are Musically interesting Preferably realistic (sounds like

More information

ADDITIVE SYNTHESIS BASED ON THE CONTINUOUS WAVELET TRANSFORM: A SINUSOIDAL PLUS TRANSIENT MODEL

ADDITIVE SYNTHESIS BASED ON THE CONTINUOUS WAVELET TRANSFORM: A SINUSOIDAL PLUS TRANSIENT MODEL ADDITIVE SYNTHESIS BASED ON THE CONTINUOUS WAVELET TRANSFORM: A SINUSOIDAL PLUS TRANSIENT MODEL José R. Beltrán and Fernando Beltrán Department of Electronic Engineering and Communications University of

More information

FFT analysis in practice

FFT analysis in practice FFT analysis in practice Perception & Multimedia Computing Lecture 13 Rebecca Fiebrink Lecturer, Department of Computing Goldsmiths, University of London 1 Last Week Review of complex numbers: rectangular

More information

Interpolation Error in Waveform Table Lookup

Interpolation Error in Waveform Table Lookup Carnegie Mellon University Research Showcase @ CMU Computer Science Department School of Computer Science 1998 Interpolation Error in Waveform Table Lookup Roger B. Dannenberg Carnegie Mellon University

More information

ME scope Application Note 01 The FFT, Leakage, and Windowing

ME scope Application Note 01 The FFT, Leakage, and Windowing INTRODUCTION ME scope Application Note 01 The FFT, Leakage, and Windowing NOTE: The steps in this Application Note can be duplicated using any Package that includes the VES-3600 Advanced Signal Processing

More information

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS

WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS NORDIC ACOUSTICAL MEETING 12-14 JUNE 1996 HELSINKI WARPED FILTER DESIGN FOR THE BODY MODELING AND SOUND SYNTHESIS OF STRING INSTRUMENTS Helsinki University of Technology Laboratory of Acoustics and Audio

More information

MUS421/EE367B Applications Lecture 9C: Time Scale Modification (TSM) and Frequency Scaling/Shifting

MUS421/EE367B Applications Lecture 9C: Time Scale Modification (TSM) and Frequency Scaling/Shifting MUS421/EE367B Applications Lecture 9C: Time Scale Modification (TSM) and Frequency Scaling/Shifting Julius O. Smith III (jos@ccrma.stanford.edu) Center for Computer Research in Music and Acoustics (CCRMA)

More information

HIGH ACCURACY FRAME-BY-FRAME NON-STATIONARY SINUSOIDAL MODELLING

HIGH ACCURACY FRAME-BY-FRAME NON-STATIONARY SINUSOIDAL MODELLING HIGH ACCURACY FRAME-BY-FRAME NON-STATIONARY SINUSOIDAL MODELLING Jeremy J. Wells, Damian T. Murphy Audio Lab, Intelligent Systems Group, Department of Electronics University of York, YO10 5DD, UK {jjw100

More information

Hungarian Speech Synthesis Using a Phase Exact HNM Approach

Hungarian Speech Synthesis Using a Phase Exact HNM Approach Hungarian Speech Synthesis Using a Phase Exact HNM Approach Kornél Kovács 1, András Kocsor 2, and László Tóth 3 Research Group on Artificial Intelligence of the Hungarian Academy of Sciences and University

More information

An Introduction to the FDM-TDM Digital Transmultiplexer: Appendix C *

An Introduction to the FDM-TDM Digital Transmultiplexer: Appendix C * OpenStax-CNX module: m32675 1 An Introduction to the FDM-TDM Digital Transmultiplexer: Appendix C * John Treichler This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution

More information

Pitch Detection Algorithms

Pitch Detection Algorithms OpenStax-CNX module: m11714 1 Pitch Detection Algorithms Gareth Middleton This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 1.0 Abstract Two algorithms to

More information

Final Exam Practice Questions for Music 421, with Solutions

Final Exam Practice Questions for Music 421, with Solutions Final Exam Practice Questions for Music 4, with Solutions Elementary Fourier Relationships. For the window w = [/,,/ ], what is (a) the dc magnitude of the window transform? + (b) the magnitude at half

More information

HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS

HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS HARMONIC INSTABILITY OF DIGITAL SOFT CLIPPING ALGORITHMS Sean Enderby and Zlatko Baracskai Department of Digital Media Technology Birmingham City University Birmingham, UK ABSTRACT In this paper several

More information

REAL-TIME BROADBAND NOISE REDUCTION

REAL-TIME BROADBAND NOISE REDUCTION REAL-TIME BROADBAND NOISE REDUCTION Robert Hoeldrich and Markus Lorber Institute of Electronic Music Graz Jakoministrasse 3-5, A-8010 Graz, Austria email: robert.hoeldrich@mhsg.ac.at Abstract A real-time

More information

FOURIER analysis is a well-known method for nonparametric

FOURIER analysis is a well-known method for nonparametric 386 IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, VOL. 54, NO. 1, FEBRUARY 2005 Resonator-Based Nonparametric Identification of Linear Systems László Sujbert, Member, IEEE, Gábor Péceli, Fellow,

More information

Speech Enhancement Using Spectral Flatness Measure Based Spectral Subtraction

Speech Enhancement Using Spectral Flatness Measure Based Spectral Subtraction IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 7, Issue, Ver. I (Mar. - Apr. 7), PP 4-46 e-issn: 9 4, p-issn No. : 9 497 www.iosrjournals.org Speech Enhancement Using Spectral Flatness Measure

More information

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991

RASTA-PLP SPEECH ANALYSIS. Aruna Bayya. Phil Kohn y TR December 1991 RASTA-PLP SPEECH ANALYSIS Hynek Hermansky Nelson Morgan y Aruna Bayya Phil Kohn y TR-91-069 December 1991 Abstract Most speech parameter estimation techniques are easily inuenced by the frequency response

More information

IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES. Q. Meng, D. Sen, S. Wang and L. Hayes

IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES. Q. Meng, D. Sen, S. Wang and L. Hayes IMPULSE RESPONSE MEASUREMENT WITH SINE SWEEPS AND AMPLITUDE MODULATION SCHEMES Q. Meng, D. Sen, S. Wang and L. Hayes School of Electrical Engineering and Telecommunications The University of New South

More information

Lecture 7 Frequency Modulation

Lecture 7 Frequency Modulation Lecture 7 Frequency Modulation Fundamentals of Digital Signal Processing Spring, 2012 Wei-Ta Chu 2012/3/15 1 Time-Frequency Spectrum We have seen that a wide range of interesting waveforms can be synthesized

More information

FIR/Convolution. Visulalizing the convolution sum. Convolution

FIR/Convolution. Visulalizing the convolution sum. Convolution FIR/Convolution CMPT 368: Lecture Delay Effects Tamara Smyth, tamaras@cs.sfu.ca School of Computing Science, Simon Fraser University April 2, 27 Since the feedforward coefficient s of the FIR filter are

More information

NOISE ESTIMATION IN A SINGLE CHANNEL

NOISE ESTIMATION IN A SINGLE CHANNEL SPEECH ENHANCEMENT FOR CROSS-TALK INTERFERENCE by Levent M. Arslan and John H.L. Hansen Robust Speech Processing Laboratory Department of Electrical Engineering Box 99 Duke University Durham, North Carolina

More information

Frequency slope estimation and its application for non-stationary sinusoidal parameter estimation

Frequency slope estimation and its application for non-stationary sinusoidal parameter estimation Frequency slope estimation and its application for non-stationary sinusoidal parameter estimation Preprint final article appeared in: Computer Music Journal, 32:2, pp. 68-79, 2008 copyright Massachusetts

More information

Band-Limited Simulation of Analog Synthesizer Modules by Additive Synthesis

Band-Limited Simulation of Analog Synthesizer Modules by Additive Synthesis Band-Limited Simulation of Analog Synthesizer Modules by Additive Synthesis Amar Chaudhary Center for New Music and Audio Technologies University of California, Berkeley amar@cnmat.berkeley.edu March 12,

More information

FIR/Convolution. Visulalizing the convolution sum. Frequency-Domain (Fast) Convolution

FIR/Convolution. Visulalizing the convolution sum. Frequency-Domain (Fast) Convolution FIR/Convolution CMPT 468: Delay Effects Tamara Smyth, tamaras@cs.sfu.ca School of Computing Science, Simon Fraser University November 8, 23 Since the feedforward coefficient s of the FIR filter are the

More information

Chapter 5 Window Functions. periodic with a period of N (number of samples). This is observed in table (3.1).

Chapter 5 Window Functions. periodic with a period of N (number of samples). This is observed in table (3.1). Chapter 5 Window Functions 5.1 Introduction As discussed in section (3.7.5), the DTFS assumes that the input waveform is periodic with a period of N (number of samples). This is observed in table (3.1).

More information

Speech Coding in the Frequency Domain

Speech Coding in the Frequency Domain Speech Coding in the Frequency Domain Speech Processing Advanced Topics Tom Bäckström Aalto University October 215 Introduction The speech production model can be used to efficiently encode speech signals.

More information

Fourier Theory & Practice, Part I: Theory (HP Product Note )

Fourier Theory & Practice, Part I: Theory (HP Product Note ) Fourier Theory & Practice, Part I: Theory (HP Product Note 54600-4) By: Robert Witte Hewlett-Packard Co. Introduction: This product note provides a brief review of Fourier theory, especially the unique

More information

Identification of Nonstationary Audio Signals Using the FFT, with Application to Analysis-based Synthesis of Sound

Identification of Nonstationary Audio Signals Using the FFT, with Application to Analysis-based Synthesis of Sound Identification of Nonstationary Audio Signals Using the FFT, with Application to Analysis-based Synthesis of Sound Paul Masri, Prof. Andrew Bateman Digital Music Research Group, University of Bristol 1.4

More information

Performance Analysis of FIR Digital Filter Design Technique and Implementation

Performance Analysis of FIR Digital Filter Design Technique and Implementation Performance Analysis of FIR Digital Filter Design Technique and Implementation. ohd. Sayeeduddin Habeeb and Zeeshan Ahmad Department of Electrical Engineering, King Khalid University, Abha, Kingdom of

More information

A Digital Signal Processor for Musicians and Audiophiles Published on Monday, 09 February :54

A Digital Signal Processor for Musicians and Audiophiles Published on Monday, 09 February :54 A Digital Signal Processor for Musicians and Audiophiles Published on Monday, 09 February 2009 09:54 The main focus of hearing aid research and development has been on the use of hearing aids to improve

More information

Single Channel Speaker Segregation using Sinusoidal Residual Modeling

Single Channel Speaker Segregation using Sinusoidal Residual Modeling NCC 2009, January 16-18, IIT Guwahati 294 Single Channel Speaker Segregation using Sinusoidal Residual Modeling Rajesh M Hegde and A. Srinivas Dept. of Electrical Engineering Indian Institute of Technology

More information

CMPT 468: Delay Effects

CMPT 468: Delay Effects CMPT 468: Delay Effects Tamara Smyth, tamaras@cs.sfu.ca School of Computing Science, Simon Fraser University November 8, 2013 1 FIR/Convolution Since the feedforward coefficient s of the FIR filter are

More information

Abstract Dual-tone Multi-frequency (DTMF) Signals are used in touch-tone telephones as well as many other areas. Since analog devices are rapidly chan

Abstract Dual-tone Multi-frequency (DTMF) Signals are used in touch-tone telephones as well as many other areas. Since analog devices are rapidly chan Literature Survey on Dual-Tone Multiple Frequency (DTMF) Detector Implementation Guner Arslan EE382C Embedded Software Systems Prof. Brian Evans March 1998 Abstract Dual-tone Multi-frequency (DTMF) Signals

More information

The Fundamentals of FFT-Based Signal Analysis and Measurement Michael Cerna and Audrey F. Harvey

The Fundamentals of FFT-Based Signal Analysis and Measurement Michael Cerna and Audrey F. Harvey Application ote 041 The Fundamentals of FFT-Based Signal Analysis and Measurement Michael Cerna and Audrey F. Harvey Introduction The Fast Fourier Transform (FFT) and the power spectrum are powerful tools

More information

THE BEATING EQUALIZER AND ITS APPLICATION TO THE SYNTHESIS AND MODIFICATION OF PIANO TONES

THE BEATING EQUALIZER AND ITS APPLICATION TO THE SYNTHESIS AND MODIFICATION OF PIANO TONES J. Rauhala, The beating equalizer and its application to the synthesis and modification of piano tones, in Proceedings of the 1th International Conference on Digital Audio Effects, Bordeaux, France, 27,

More information

FIR window method: A comparative Analysis

FIR window method: A comparative Analysis IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 1, Issue 4, Ver. III (Jul - Aug.215), PP 15-2 www.iosrjournals.org FIR window method: A

More information

When and How to Use FFT

When and How to Use FFT B Appendix B: FFT When and How to Use FFT The DDA s Spectral Analysis capability with FFT (Fast Fourier Transform) reveals signal characteristics not visible in the time domain. FFT converts a time domain

More information

Spectrum. Additive Synthesis. Additive Synthesis Caveat. Music 270a: Modulation

Spectrum. Additive Synthesis. Additive Synthesis Caveat. Music 270a: Modulation Spectrum Music 7a: Modulation Tamara Smyth, trsmyth@ucsd.edu Department of Music, University of California, San Diego (UCSD) October 3, 7 When sinusoids of different frequencies are added together, the

More information

Signal segmentation and waveform characterization. Biosignal processing, S Autumn 2012

Signal segmentation and waveform characterization. Biosignal processing, S Autumn 2012 Signal segmentation and waveform characterization Biosignal processing, 5173S Autumn 01 Short-time analysis of signals Signal statistics may vary in time: nonstationary how to compute signal characterizations?

More information

Quantized Coefficient F.I.R. Filter for the Design of Filter Bank

Quantized Coefficient F.I.R. Filter for the Design of Filter Bank Quantized Coefficient F.I.R. Filter for the Design of Filter Bank Rajeev Singh Dohare 1, Prof. Shilpa Datar 2 1 PG Student, Department of Electronics and communication Engineering, S.A.T.I. Vidisha, INDIA

More information

2.1 BASIC CONCEPTS Basic Operations on Signals Time Shifting. Figure 2.2 Time shifting of a signal. Time Reversal.

2.1 BASIC CONCEPTS Basic Operations on Signals Time Shifting. Figure 2.2 Time shifting of a signal. Time Reversal. 1 2.1 BASIC CONCEPTS 2.1.1 Basic Operations on Signals Time Shifting. Figure 2.2 Time shifting of a signal. Time Reversal. 2 Time Scaling. Figure 2.4 Time scaling of a signal. 2.1.2 Classification of Signals

More information

Module 5. DC to AC Converters. Version 2 EE IIT, Kharagpur 1

Module 5. DC to AC Converters. Version 2 EE IIT, Kharagpur 1 Module 5 DC to AC Converters Version 2 EE IIT, Kharagpur 1 Lesson 37 Sine PWM and its Realization Version 2 EE IIT, Kharagpur 2 After completion of this lesson, the reader shall be able to: 1. Explain

More information

A Faster Method for Accurate Spectral Testing without Requiring Coherent Sampling

A Faster Method for Accurate Spectral Testing without Requiring Coherent Sampling A Faster Method for Accurate Spectral Testing without Requiring Coherent Sampling Minshun Wu 1,2, Degang Chen 2 1 Xi an Jiaotong University, Xi an, P. R. China 2 Iowa State University, Ames, IA, USA Abstract

More information

Signal Processing for Digitizers

Signal Processing for Digitizers Signal Processing for Digitizers Modular digitizers allow accurate, high resolution data acquisition that can be quickly transferred to a host computer. Signal processing functions, applied in the digitizer

More information

Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau

Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau (Also see: Lecture ADSP, Slides 06) In discrete, digital signal we use the normalized frequency, T = / f s =: it is without a

More information

Real-time fundamental frequency estimation by least-square fitting. IEEE Transactions on Speech and Audio Processing, 1997, v. 5 n. 2, p.

Real-time fundamental frequency estimation by least-square fitting. IEEE Transactions on Speech and Audio Processing, 1997, v. 5 n. 2, p. Title Real-time fundamental frequency estimation by least-square fitting Author(s) Choi, AKO Citation IEEE Transactions on Speech and Audio Processing, 1997, v. 5 n. 2, p. 201-205 Issued Date 1997 URL

More information

Chapter 4 SPEECH ENHANCEMENT

Chapter 4 SPEECH ENHANCEMENT 44 Chapter 4 SPEECH ENHANCEMENT 4.1 INTRODUCTION: Enhancement is defined as improvement in the value or Quality of something. Speech enhancement is defined as the improvement in intelligibility and/or

More information

MITOCW MITRES_6-007S11lec18_300k.mp4

MITOCW MITRES_6-007S11lec18_300k.mp4 MITOCW MITRES_6-007S11lec18_300k.mp4 [MUSIC PLAYING] PROFESSOR: Last time, we began the discussion of discreet-time processing of continuous-time signals. And, as a reminder, let me review the basic notion.

More information

I-Hao Hsiao, Chun-Tang Chao*, and Chi-Jo Wang (2016). A HHT-Based Music Synthesizer. Intelligent Technologies and Engineering Systems, Lecture Notes

I-Hao Hsiao, Chun-Tang Chao*, and Chi-Jo Wang (2016). A HHT-Based Music Synthesizer. Intelligent Technologies and Engineering Systems, Lecture Notes I-Hao Hsiao, Chun-Tang Chao*, and Chi-Jo Wang (2016). A HHT-Based Music Synthesizer. Intelligent Technologies and Engineering Systems, Lecture Notes in Electrical Engineering (LNEE), Vol.345, pp.523-528.

More information

Topic 2. Signal Processing Review. (Some slides are adapted from Bryan Pardo s course slides on Machine Perception of Music)

Topic 2. Signal Processing Review. (Some slides are adapted from Bryan Pardo s course slides on Machine Perception of Music) Topic 2 Signal Processing Review (Some slides are adapted from Bryan Pardo s course slides on Machine Perception of Music) Recording Sound Mechanical Vibration Pressure Waves Motion->Voltage Transducer

More information

Biomedical Signals. Signals and Images in Medicine Dr Nabeel Anwar

Biomedical Signals. Signals and Images in Medicine Dr Nabeel Anwar Biomedical Signals Signals and Images in Medicine Dr Nabeel Anwar Noise Removal: Time Domain Techniques 1. Synchronized Averaging (covered in lecture 1) 2. Moving Average Filters (today s topic) 3. Derivative

More information

Time-Frequency Distributions for Automatic Speech Recognition

Time-Frequency Distributions for Automatic Speech Recognition 196 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 9, NO. 3, MARCH 2001 Time-Frequency Distributions for Automatic Speech Recognition Alexandros Potamianos, Member, IEEE, and Petros Maragos, Fellow,

More information

PARSHL: An Analysis/Synthesis Program for Non-Harmonic Sounds Based on a Sinusoidal Representation

PARSHL: An Analysis/Synthesis Program for Non-Harmonic Sounds Based on a Sinusoidal Representation PARSHL: An Analysis/Synthesis Program for Non-Harmonic Sounds Based on a Sinusoidal Representation Julius O. Smith III (jos@ccrma.stanford.edu) Xavier Serra (xjs@ccrma.stanford.edu) Center for Computer

More information

Music 270a: Modulation

Music 270a: Modulation Music 7a: Modulation Tamara Smyth, trsmyth@ucsd.edu Department of Music, University of California, San Diego (UCSD) October 3, 7 Spectrum When sinusoids of different frequencies are added together, the

More information

for Single-Tone Frequency Tracking H. C. So Department of Computer Engineering & Information Technology, City University of Hong Kong,

for Single-Tone Frequency Tracking H. C. So Department of Computer Engineering & Information Technology, City University of Hong Kong, A Comparative Study of Three Recursive Least Squares Algorithms for Single-Tone Frequency Tracking H. C. So Department of Computer Engineering & Information Technology, City University of Hong Kong, Tat

More information

Measurement of RMS values of non-coherently sampled signals. Martin Novotny 1, Milos Sedlacek 2

Measurement of RMS values of non-coherently sampled signals. Martin Novotny 1, Milos Sedlacek 2 Measurement of values of non-coherently sampled signals Martin ovotny, Milos Sedlacek, Czech Technical University in Prague, Faculty of Electrical Engineering, Dept. of Measurement Technicka, CZ-667 Prague,

More information

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking The 7th International Conference on Signal Processing Applications & Technology, Boston MA, pp. 476-480, 7-10 October 1996. Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic

More information

Window Functions And Time-Domain Plotting In HFSS And SIwave

Window Functions And Time-Domain Plotting In HFSS And SIwave Window Functions And Time-Domain Plotting In HFSS And SIwave Greg Pitner Introduction HFSS and SIwave allow for time-domain plotting of S-parameters. Often, this feature is used to calculate a step response

More information

PART I: The questions in Part I refer to the aliasing portion of the procedure as outlined in the lab manual.

PART I: The questions in Part I refer to the aliasing portion of the procedure as outlined in the lab manual. Lab. #1 Signal Processing & Spectral Analysis Name: Date: Section / Group: NOTE: To help you correctly answer many of the following questions, it may be useful to actually run the cases outlined in the

More information

Laboratory Assignment 4. Fourier Sound Synthesis

Laboratory Assignment 4. Fourier Sound Synthesis Laboratory Assignment 4 Fourier Sound Synthesis PURPOSE This lab investigates how to use a computer to evaluate the Fourier series for periodic signals and to synthesize audio signals from Fourier series

More information

Topic 6. The Digital Fourier Transform. (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith)

Topic 6. The Digital Fourier Transform. (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith) Topic 6 The Digital Fourier Transform (Based, in part, on The Scientist and Engineer's Guide to Digital Signal Processing by Steven Smith) 10 20 30 40 50 60 70 80 90 100 0-1 -0.8-0.6-0.4-0.2 0 0.2 0.4

More information

Auditory modelling for speech processing in the perceptual domain

Auditory modelling for speech processing in the perceptual domain ANZIAM J. 45 (E) ppc964 C980, 2004 C964 Auditory modelling for speech processing in the perceptual domain L. Lin E. Ambikairajah W. H. Holmes (Received 8 August 2003; revised 28 January 2004) Abstract

More information

8.3 Basic Parameters for Audio

8.3 Basic Parameters for Audio 8.3 Basic Parameters for Audio Analysis Physical audio signal: simple one-dimensional amplitude = loudness frequency = pitch Psycho-acoustic features: complex A real-life tone arises from a complex superposition

More information

Two-channel Separation of Speech Using Direction-of-arrival Estimation And Sinusoids Plus Transients Modeling

Two-channel Separation of Speech Using Direction-of-arrival Estimation And Sinusoids Plus Transients Modeling Two-channel Separation of Speech Using Direction-of-arrival Estimation And Sinusoids Plus Transients Modeling Mikko Parviainen 1 and Tuomas Virtanen 2 Institute of Signal Processing Tampere University

More information

METHODS FOR SEPARATION OF AMPLITUDE AND FREQUENCY MODULATION IN FOURIER TRANSFORMED SIGNALS

METHODS FOR SEPARATION OF AMPLITUDE AND FREQUENCY MODULATION IN FOURIER TRANSFORMED SIGNALS METHODS FOR SEPARATION OF AMPLITUDE AND FREQUENCY MODULATION IN FOURIER TRANSFORMED SIGNALS Jeremy J. Wells Audio Lab, Department of Electronics, University of York, YO10 5DD York, UK jjw100@ohm.york.ac.uk

More information

SGN Audio and Speech Processing

SGN Audio and Speech Processing Introduction 1 Course goals Introduction 2 SGN 14006 Audio and Speech Processing Lectures, Fall 2014 Anssi Klapuri Tampere University of Technology! Learn basics of audio signal processing Basic operations

More information

Linear Frequency Modulation (FM) Chirp Signal. Chirp Signal cont. CMPT 468: Lecture 7 Frequency Modulation (FM) Synthesis

Linear Frequency Modulation (FM) Chirp Signal. Chirp Signal cont. CMPT 468: Lecture 7 Frequency Modulation (FM) Synthesis Linear Frequency Modulation (FM) CMPT 468: Lecture 7 Frequency Modulation (FM) Synthesis Tamara Smyth, tamaras@cs.sfu.ca School of Computing Science, Simon Fraser University January 26, 29 Till now we

More information

6 Sampling. Sampling. The principles of sampling, especially the benefits of coherent sampling

6 Sampling. Sampling. The principles of sampling, especially the benefits of coherent sampling Note: Printed Manuals 6 are not in Color Objectives This chapter explains the following: The principles of sampling, especially the benefits of coherent sampling How to apply sampling principles in a test

More information

Chapter Three. The Discrete Fourier Transform

Chapter Three. The Discrete Fourier Transform Chapter Three. The Discrete Fourier Transform The discrete Fourier transform (DFT) is one of the two most common, and powerful, procedures encountered in the field of digital signal processing. (Digital

More information

ADAPTIVE NOISE LEVEL ESTIMATION

ADAPTIVE NOISE LEVEL ESTIMATION Proc. of the 9 th Int. Conference on Digital Audio Effects (DAFx-6), Montreal, Canada, September 18-2, 26 ADAPTIVE NOISE LEVEL ESTIMATION Chunghsin Yeh Analysis/Synthesis team IRCAM/CNRS-STMS, Paris, France

More information

Butterworth Window for Power Spectral Density Estimation

Butterworth Window for Power Spectral Density Estimation Butterworth Window for Power Spectral Density Estimation Tae Hyun Yoon and Eon Kyeong Joo The power spectral density of a signal can be estimated most accurately by using a window with a narrow bandwidth

More information

Monophony/Polyphony Classification System using Fourier of Fourier Transform

Monophony/Polyphony Classification System using Fourier of Fourier Transform International Journal of Electronics Engineering, 2 (2), 2010, pp. 299 303 Monophony/Polyphony Classification System using Fourier of Fourier Transform Kalyani Akant 1, Rajesh Pande 2, and S.S. Limaye

More information

United Codec. 1. Motivation/Background. 2. Overview. Mofei Zhu, Hugo Guo, Deepak Music 422 Winter 09 Stanford University.

United Codec. 1. Motivation/Background. 2. Overview. Mofei Zhu, Hugo Guo, Deepak Music 422 Winter 09 Stanford University. United Codec Mofei Zhu, Hugo Guo, Deepak Music 422 Winter 09 Stanford University March 13, 2009 1. Motivation/Background The goal of this project is to build a perceptual audio coder for reducing the data

More information

VIBRATO DETECTING ALGORITHM IN REAL TIME. Minhao Zhang, Xinzhao Liu. University of Rochester Department of Electrical and Computer Engineering

VIBRATO DETECTING ALGORITHM IN REAL TIME. Minhao Zhang, Xinzhao Liu. University of Rochester Department of Electrical and Computer Engineering VIBRATO DETECTING ALGORITHM IN REAL TIME Minhao Zhang, Xinzhao Liu University of Rochester Department of Electrical and Computer Engineering ABSTRACT Vibrato is a fundamental expressive attribute in music,

More information

Initial Vectors (random) Filter (Fault-simulation Based Compaction) Yes. done? predict/construct future vectors; append to test set.

Initial Vectors (random) Filter (Fault-simulation Based Compaction) Yes. done? predict/construct future vectors; append to test set. Ecient Spectral Techniques for Sequential ATPG Ashish Giani y, Shuo Sheng y, Michael S. Hsiao y, and Vishwani D. Agrawal z y Department of Electrical and Computer Engineering, Rutgers University, Piscataway,

More information

Design of FIR Filters

Design of FIR Filters Design of FIR Filters Elena Punskaya www-sigproc.eng.cam.ac.uk/~op205 Some material adapted from courses by Prof. Simon Godsill, Dr. Arnaud Doucet, Dr. Malcolm Macleod and Prof. Peter Rayner 1 FIR as a

More information

Finite Word Length Effects on Two Integer Discrete Wavelet Transform Algorithms. Armein Z. R. Langi

Finite Word Length Effects on Two Integer Discrete Wavelet Transform Algorithms. Armein Z. R. Langi International Journal on Electrical Engineering and Informatics - Volume 3, Number 2, 211 Finite Word Length Effects on Two Integer Discrete Wavelet Transform Algorithms Armein Z. R. Langi ITB Research

More information

Carrier Frequency Offset Estimation in WCDMA Systems Using a Modified FFT-Based Algorithm

Carrier Frequency Offset Estimation in WCDMA Systems Using a Modified FFT-Based Algorithm Carrier Frequency Offset Estimation in WCDMA Systems Using a Modified FFT-Based Algorithm Seare H. Rezenom and Anthony D. Broadhurst, Member, IEEE Abstract-- Wideband Code Division Multiple Access (WCDMA)

More information

Since the advent of the sine wave oscillator

Since the advent of the sine wave oscillator Advanced Distortion Analysis Methods Discover modern test equipment that has the memory and post-processing capability to analyze complex signals and ascertain real-world performance. By Dan Foley European

More information

CMPT 468: Frequency Modulation (FM) Synthesis

CMPT 468: Frequency Modulation (FM) Synthesis CMPT 468: Frequency Modulation (FM) Synthesis Tamara Smyth, tamaras@cs.sfu.ca School of Computing Science, Simon Fraser University October 6, 23 Linear Frequency Modulation (FM) Till now we ve seen signals

More information

Formant Synthesis of Haegeum: A Sound Analysis/Synthesis System using Cpestral Envelope

Formant Synthesis of Haegeum: A Sound Analysis/Synthesis System using Cpestral Envelope Formant Synthesis of Haegeum: A Sound Analysis/Synthesis System using Cpestral Envelope Myeongsu Kang School of Computer Engineering and Information Technology Ulsan, South Korea ilmareboy@ulsan.ac.kr

More information

IMPROVING QUALITY OF SPEECH SYNTHESIS IN INDIAN LANGUAGES. P. K. Lehana and P. C. Pandey

IMPROVING QUALITY OF SPEECH SYNTHESIS IN INDIAN LANGUAGES. P. K. Lehana and P. C. Pandey Workshop on Spoken Language Processing - 2003, TIFR, Mumbai, India, January 9-11, 2003 149 IMPROVING QUALITY OF SPEECH SYNTHESIS IN INDIAN LANGUAGES P. K. Lehana and P. C. Pandey Department of Electrical

More information

The Discrete Fourier Transform. Claudia Feregrino-Uribe, Alicia Morales-Reyes Original material: Dr. René Cumplido

The Discrete Fourier Transform. Claudia Feregrino-Uribe, Alicia Morales-Reyes Original material: Dr. René Cumplido The Discrete Fourier Transform Claudia Feregrino-Uribe, Alicia Morales-Reyes Original material: Dr. René Cumplido CCC-INAOE Autumn 2015 The Discrete Fourier Transform Fourier analysis is a family of mathematical

More information

FIR Filter Design by Frequency Sampling or Interpolation *

FIR Filter Design by Frequency Sampling or Interpolation * OpenStax-CX module: m689 FIR Filter Design by Frequency Sampling or Interpolation * C. Sidney Burrus This work is produced by OpenStax-CX and licensed under the Creative Commons Attribution License 2.

More information

Signal Characterization in terms of Sinusoidal and Non-Sinusoidal Components

Signal Characterization in terms of Sinusoidal and Non-Sinusoidal Components Signal Characterization in terms of Sinusoidal and Non-Sinusoidal Components Geoffroy Peeters, avier Rodet To cite this version: Geoffroy Peeters, avier Rodet. Signal Characterization in terms of Sinusoidal

More information

L19: Prosodic modification of speech

L19: Prosodic modification of speech L19: Prosodic modification of speech Time-domain pitch synchronous overlap add (TD-PSOLA) Linear-prediction PSOLA Frequency-domain PSOLA Sinusoidal models Harmonic + noise models STRAIGHT This lecture

More information

SINUSOIDAL MODELING. EE6641 Analysis and Synthesis of Audio Signals. Yi-Wen Liu Nov 3, 2015

SINUSOIDAL MODELING. EE6641 Analysis and Synthesis of Audio Signals. Yi-Wen Liu Nov 3, 2015 1 SINUSOIDAL MODELING EE6641 Analysis and Synthesis of Audio Signals Yi-Wen Liu Nov 3, 2015 2 Last time: Spectral Estimation Resolution Scenario: multiple peaks in the spectrum Choice of window type and

More information

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE

ROBUST PITCH TRACKING USING LINEAR REGRESSION OF THE PHASE - @ Ramon E Prieto et al Robust Pitch Tracking ROUST PITCH TRACKIN USIN LINEAR RERESSION OF THE PHASE Ramon E Prieto, Sora Kim 2 Electrical Engineering Department, Stanford University, rprieto@stanfordedu

More information

NMF, WOLA, And Binary Filtering: Avoiding the Curse of Time-Aliasing James A. Moorer

NMF, WOLA, And Binary Filtering: Avoiding the Curse of Time-Aliasing James A. Moorer NMF, WOLA, And Binary Filtering: Avoiding the Curse of Time-Aliasing James A. Moorer Please be notified that everything in this discussion is completely obvious and follows directly from principles everyone

More information

Application of Fourier Transform in Signal Processing

Application of Fourier Transform in Signal Processing 1 Application of Fourier Transform in Signal Processing Lina Sun,Derong You,Daoyun Qi Information Engineering College, Yantai University of Technology, Shandong, China Abstract: Fourier transform is a

More information