MASTER'S THESIS. Speech Compression and Tone Detection in a Real-Time System. Kristina Berglund. MSc Programmes in Engineering

Size: px
Start display at page:

Download "MASTER'S THESIS. Speech Compression and Tone Detection in a Real-Time System. Kristina Berglund. MSc Programmes in Engineering"

Transcription

1 2004:003 CIV MASTER'S THESIS Speech Compression and Tone Detection in a Real-Time System Kristina Berglund MSc Programmes in Engineering Department of Computer Science and Electrical Engineering Division of Signal Processing 2004:003 CIV - ISSN: ISRN: LTU-EX--04/003--SE

2 Speech Compression and Tone Detection in a Real-Time System Kristina Berglund Master s Thesis Division of Signal Processing

3

4 To Andreas

5

6 Abstract During the night, encrypted spoken newspapers are broadcasted over the terrestrial FM radio network. These papers can be subscribed to by persons with reading disabilities. The subscribers presently have a special receiver that decrypts and stores the newspaper on a cassette tape. A project aiming to design a new receiver, using digital technology, was started during the year This report describes the parts of the project involving speech compression and tone detection. An overview of different compression techniques, with emphasis on waveform coding, is given. A detailed description of Adaptive Differential Pulse Code Modulation (ADPCM), the compression technique chosen for implementation on a Digital Signal Processor (DSP) is also included. ADPCM was first implemented on the ADSP-2181 DSP, with a good result. In the final version of the digital receiver the ADSP-2191 DSP will be used and hence the code was converted to fit this DSP. Due to some problems this implementation could not be completed within the time frame of this thesis. The final part of this thesis consists of finding a method for detecting a tone inserted between articles in the spoken newspaper. The tone detection is composed of two parts, the first part is reducing the amplitude of the speech while maintaining the amplitude of the tone. For this part a digital resonator was chosen and implemented both in Matlab and on the ADSP-2191 DSP. The second part of the tone detection consists of deciding whether the tone is present or not, this part was implemented only in Matlab.

7

8 Preface This Master s thesis is the final work for my Master of Science degree. The work for this thesis was performed at the Division of Signal Processing at Luleå University of Technology. It was performed as part of a project, aiming to design a digital receiver for a radio transmitted encrypted spoken newspaper system. The main purposes of this thesis are to investigate various compression algorithms and to find a detection technique that can be used to detect tones inserted between articles in the spoken newspaper. I would like to thank my examiner James P. LeBlanc for giving me the opportunity of working in this project and for all support and inputs along the way. I would also like to thank my project leader Per Johansson and my fellow colleagues Anders Larsson and Patrik Pääjärvi for their invaluable help and encouraging words. Kristina Berglund Luleå, January 2004

9

10 Contents 1 Introduction Project Overview Basics Sampling Quantization Scalar Quantization Vector Quantization Speech Compression Waveform Coding Pulse Code Modulation Differential Pulse Code Modulation Adaptive Differential Pulse Code Modulation Subband Coding Transform Coding Vocoders Linear Predictive Coding Hybrid Coding Code Excited Linear Prediction Discussion Adaptive Differential Pulse Code Modulation The Encoder The Decoder Implementation and Results Used Tools Tone Detection Finding the Tone Matched Filter

11 Contents Digital Resonator Implementation Results Conclusions 41 Bibliography 44

12 Chapter 1 Introduction This Master s thesis is part of a project involving myself, the M.Sc. students Anders Larsson and Patrik Pääjärvi and our project leader Per Johansson at Luleå University of Technology. The aim of the project is to design a digital receiver for a radio transmitted encrypted newspaper system. 1.1 Project Overview The radio transmitted papers are spoken versions of daily newspapers, which reading disabled persons can subscribe to. The newspapers are distributed during the night between 2 a.m. and 5 a.m., on the radio channel SR P1. Different papers are broadcasted in different regions of the country, today about 90 different daily newspapers have a spoken version broadcasted each night [1]. Since every newspaper is limited to 90 minutes and one paper is sent in mono on each channel, a maximum of four newspapers can be broadcasted in one region during the night. To prevent non-subscribers from listening to the paper, it is encrypted. To receive the newspaper, the subscribers have a receiver set to record one of the papers transmitted. The receiver first decrypts the newspaper and then stores it on a cassette tape. In order to listen to the paper, the subscribers must insert the tape in a regular cassette player. Between the articles in the paper a tone is inserted. When the listeners fast forward the tape this tone is heard, indicating the start of a new article. 1

13 2 Chapter 1. Introduction The project of designing a digital receiver was initiated in 2002 by Taltidningsnämnden, a government authority whose purpose is to improve the availability of spoken newspapers to the reading disabled. The reasons for designing a new receiver are mainly questions of cost and adaptability. The receiver of today uses analog technology and have high costs of maintenance and repair. Often the subscribers to spoken newspapers are elderly, with difficulties in handling a cassette tape. Since the new receiver have a built-in speaker, nothing needs to be moved in order to listen to the paper. An advantage of the digital receiver is the ability to skip between articles by pressing a button. For this reason, the tones inserted between articles must be detected. Another advantage is the fact that additional features, for example improvements of the sound quality, easily can be added to the digital receiver by a change of software. Below, a brief description of the digital receiver is given. As seen in Fig. 1.1, the transmission is received by an FM-receiver and analogto-digital converted, before passed along to the digital signal processor (DSP). In the DSP, the encrypted newspaper is decrypted, searched for tones between the articles, compressed and then stored on a memory chip connected to the DSP. During playback, the process is reversed. The newspaper is read from the memory, decompressed, digital-to-analog converted and sent out through a speaker. In order to decrypt the newspaper, the incoming transmission is sampled at 19 khz. Each sample is represented in the DSP with an accuracy of 16 bits. Since one newspaper is 90 minutes long, about 205 MB worth of memory is required for storing the paper if no compression technique is used. If this receiver is to be commercialized, the costs must be kept small. Since the cost of the memory is higher the larger the memory is, a small memory is desired. To fit the newspaper on a small memory, it must be compressed. The parts of the project described in this thesis are compression/decompression and tone detection.

14 Chapter 1. Introduction 3 Encrypted transmission FM A/D DSP Decryption Tone detection Compression Write to memory Memory Read from memory D/A Decompression Figure 1.1: Simple block diagram of the digital receiver. The data flow during recording and playback are indicated by the black and white arrows respectively.

15 4 Chapter 1. Introduction

16 Chapter 2 Basics In this section, two basic concepts needed for the understanding of digital compression of speech waveforms are given. 2.1 Sampling Only discrete-time signals can be digitally processed, time continuous signals have to be sampled. Sampling is merely taking the value of the signal at discrete points, equally spaced in time, see Fig The discrete-time version of a time continuous signal s c (t), is s d [n] = s c (nt ), (2.1) where n is an integer and T is the sampling period, the time between each sample. The sampling frequency, f s, is how often a signal is sampled, the relationship between the sampling period and the sampling frequency is described by f s = 1 T. (2.2) A bandlimited signal s(t), i.e. a signal with no frequency components higher than some known limit, is uniquely determined by its samples s[n] if the sampling frequency is high enough. The sampling theorem states that the sampling frequency must be at least twice as high as the maximum frequency component in the signal [2]. That is, f s > 2 f max, (2.3) 5

17 6 Chapter 2. Basics Original signal Sampled signal Amplitude Time Figure 2.1: Sampling of a time continuous signal. where f max is the maximum frequency, for s(t) to be uniquely determined and reconstructible. The frequency 2 f max is called the Nyquist rate. If the sampling frequency is below the Nyquist rate, aliasing may occur. Aliasing means that a frequency is misrepresented, the reconstructed signal will have a lower frequency than the original signal, as illustrated in Fig Aliasing can be avoided by bandlimiting the signal before it is sampled. This is done by applying a lowpass filter to the signal prior to sampling. The lowpass filter used must remove the frequency components that are higher than f s / Quantization Every sampled signal is quantized, to be encodable with a finite number of bits. Quantization can be described as rounding off the value of the signal to the nearest value from a discrete set, see Fig The quantization can be performed on raw speech samples as well as on residuals, for example the difference between consecutive samples. If the quantization of one sample depends on previous samples, it is said to have memory. Quantization can be performed on one sample at a time, known as scalar quantization, or on several samples at a time, known as vector quantization.

18 Chapter 2. Basics Amplitude Time (s) Figure 2.2: Original (dotted line) and reconstructed (solid line) signal, sampling frequency below the Nyquist rate. Output value Input value Figure 2.3: Input(dotted line) and output (solid line) from a 3-bit (eight levels) uniform quantizer Scalar Quantization For all quantization methods some distortion is introduced. For scalar quantization this distortion is called quantization noise, which is measured as the difference between the input and the output of the quantizer. How well a quantizer works depends on the step size. The step size is the measure between two adjacent quantization levels. Uniform quantizers have the same step size between all levels, unlike log quantizers (see Section 3.1.1). Log quantizers have a step size that varies between different quantization levels. For uniform as well as log quantizers, the step size must be kept small, to minimize quantization noise. There is a tradeoff

19 8 Chapter 2. Basics between how small the step size is and how many quantization levels are required. A way of keeping the quantization noise small is to allow the step size to vary from one time to another, depending on the amplitude of the input signal. This is called adaptive quantization and it means that the step size is increased in regions with a high variance in amplitude and decreased in regions with a low variance. The step size adaption can be based on future samples, called forward adaptive quantization or on previous samples, called backward adaptive quantization. Adaptive quantization adds delay and additional storage requirements, but enhances the quality of the reconstructed signal. Forward adaptive quantization is shown in Fig Samples, including the current sample, are stored in a buffer and a statistical analysis is performed on them. Based on this analysis the step size is adjusted and the quantization is carried out. Statistical parameters Input signal Buffering Statistical analysis Quantization encoder Interval index Quantization decoder Reconstructed signal Channel Figure 2.4: Forward adaptive quantization (based on a figure from [3]). Since the analysis is performed on samples not available at the receiver, the statistical parameters from the analysis must be sent as side information. There is a tradeoff between how sensitive the quantizer is to local variations in the signal and how often side information must be sent. If the buffer size is small, the adaption to local changes will be effective, but then the side information must be sent very often. If many samples are stored in the buffer, the side information does not have to be sent that often, however the adaption might miss local variations. The more samples the buffer stores, the bigger the delay and the storage requirements are. No side information has to be sent using backward adaptive quantization. The buffering and the statistical analysis is carried out by both the transmitter and the receiver, see Fig Since the analysis is performed on the output of the quantizer and this contains quantization noise, the adaption to the variation of the signal is not as fine as it is when using forward adaptive quantization. Some well-known methods that use scalar quantization are Pulse Code Modulation, Delta Modulation and Adaptive Differential Pulse Code Modulation, they are all described in Section 3.1.

20 Chapter 2. Basics 9 Input signal Quantization encoder Interval index Quantization decoder Reconstructed signal Buffering Buffering Statistical analysis Statistical analysis Channel Figure 2.5: Backward adaptive quantization (based on a figure from [3]) Vector Quantization In scalar quantization the input to the quantizer is just one sample. The input to a vector quantizer is a vector consisting of several consecutive samples. The samples can be both pure speech samples as well as prediction residuals or some coding parameters. The general idea of vector quantization is that a group of samples can be more efficiently encoded together than one by one. Due to the inherent complexity of vector quantization, significant results within this area were not reported until the late 1970 s [4]. In vector quantization, depicted in Fig. 2.6, the quantization consists of finding a codeword that resembles the input vector. Each incoming vector, s i, is compared to codewords in a codebook. The closest word, often measured in mean squared error, is selected for transmission. Channel Codebook Codebook Input vector Encoder Index Index Decoder Reconstructed vector Figure 2.6: Block diagram of vector quantizer (based on a figure from [4]). For a codebook that consists of L different codewords, the M-dimensional vector space is divided into L non-overlapping cells. Each cell has a corresponding centroid, these are the L codewords. This is shown for a two-dimensional case (M=2) in Fig The incoming vector falls into a cell, C n, and the centroid, ŝ n, of the cell is the codeword chosen for transmission. Actually, what is transmitted is the index, u n of the chosen word, not the codeword itself. The decoding process is the

21 10 Chapter 2. Basics encoding reversed. The codebook, the same as the encoder has, is searched to find the codeword matching the index. When finding a match, the decoded word is the centroid corresponding to the index in question. Vector quantization is often used in hybrid coders, described in Section 3.3. n cell centroid, s^ n n 1 Figure 2.7: Cells for two-dimensional vector quantization (based on a figure from [5]).

22 Chapter 3 Speech Compression Speech compression, or speech coding, is needed for many applications. One application is telecommunication, where a low bit rate often is aimed at. Another is storage, where speech needs to be compressed to fit within a specific memory size. The goal of speech coding is to encode the signal using as few bits as possible and still have a sufficient quality of the reconstructed speech. Sufficient quality can mean a variety of different things, from speech with no artifacts and no difficulty in recognizing the speaker, to nearly non-intelligible unnatural speech, all depending on the application. Another important issue is delay, in real-time systems the delay introduced by compression/decompression must be kept small. In this project the aim has been to find a low complexity compression algorithm that reconstructs speech with high quality. Low complexity is the same as needing few instructions to encode and decode the speech. The field of speech coding can be divided into three major parts or classes, waveform coding, vocoding and hybrid coding. Hybrid coding is not really a separate class, more a mixture of vocoding and waveform coding. In the following sections an overview of some basic compression techniques is given, and in Section 3.4 examples of the quality and complexity for algorithms from each class are given and discussed. 11

23 12 Chapter 3. Speech Compression 3.1 Waveform Coding Waveform coders compress the speech signal without any consideration in how the waveform is generated. Waveform coding can be carried out both in the time domain and in the frequency domain. Pulse Code Modulation, Delta Modulation and Adaptive Differential Pulse Code Modulation are all examples of time domain waveform coding, while Subband Coding and Adaptive Transform Coding operate in the frequency domain. The aim is to construct a signal with a waveform that resembles the original signal. This means that waveform coders also work fairly well for non-speech signals. Waveform coders generally produce speech with a high quality, unfortunately with quite a high bit rate as well. For more information on waveform coders than given in this thesis, the reader is referred to [3] and [6] Pulse Code Modulation Pulse Code Modulation (PCM) is the simplest form of scalar quantization. PCM uses uniform or logarithmic quantization. The samples are rounded off to the nearest value from a discrete set. For uniform PCM the set consists of a number of equally spaced discrete values, the step size between each quantization level is constant. The waveform is approximated by quantizing the input speech samples before transmission. PCM produces speech with a high quality, however, it also requires a very high bit rate. PCM can be made more efficient using a non-uniform step size. This type of quantizer is called log quantizer. The difference between a uniform quantizer and a log quantizer is shown in Fig A-law and µ-law are both examples of log. Output value Output value Input value Input value (a) (b) Figure 3.1: (a) Uniform quantizer; (b) Log quantizer.

24 Chapter 3. Speech Compression 13 quantizers. These two techniques are widely used in many speech applications, µ-law is used in telephone networks in North America and in Japan and A-law is used in Europe [7]. The performance of a 12-bit uniform quantizer can be achieved by a 7-bit log quantizer [4] Differential Pulse Code Modulation Since successive speech samples are highly correlated, the difference between adjacent samples generally has a smaller variance than the original signal. This means that the signal can be encoded using fewer bits when encoding the difference than when encoding the original signal and still achieve the same performance. This is the objective of Differential Pulse Code Modulation (DPCM). DPCM quantizes the difference between adjacent samples, instead of the original signal itself. Many DPCM schemes also use a short-time predictor to further decrease the bit rate, an example of this is Delta Modulation. Delta Modulation Fig. 3.2 shows a Delta Modulator. Based on previous samples the current sample s[k] is estimated to s e [k]. The prediction residual, e[k], is the difference between the input sample and an estimate of the input sample. This difference is then quantized and transmitted. The transmitter and the receiver uses the same predictor P. At the receiver the prediction residual is added to s e [k] to reconstruct the speech sample to ŝ[k]. s[k] + Σ e[k] Q e [k] q e [k] q + Σ s[k] ^ [k] s e P + + Σ ^ s[k] [k] s e + P Figure 3.2: Block diagram of a delta modulation encoder (to the left) and decoder (to the right) (based on figures from [8]). The simplest form of the procedure described above is the first-order, one-bit linear delta modulation. The estimate in this method is based on one previous quantized

25 14 Chapter 3. Speech Compression sample and uses only one bit to encode the difference. This means that the following sample can only be smaller or bigger than the previously encoded sample. Hence, adjacent samples always differ with the step size from each other. If adjacent input samples have the same value, they will be encoded differently. This error is known as granular noise. When using a fixed step size like this method does, another error, called slope overload can occur. This can happen when the sample-to-sample change of the signal is greater or smaller than the step size. Both granular noise and slope overload is depicted in Fig This method can be used for highly oversampled speech, where the correlation is strong. Amplitude Slope overload Step size Granular noise original samples coded samples Sample number Figure 3.3: Two types of quantization errors (based on a figure from [5]) Adaptive Differential Pulse Code Modulation Adaptive Differential Pulse Code Modulation (ADPCM) is an extension of Delta Modulation. ADPCM has both an adaptive step size and an adaptive prediction. A detailed description of the ITU-T standard G , 32, 24, 16 kbit/s adaptive differential pulse code modulation (ADPCM) is found in Chapter Subband Coding A subband coder divides the input signal into several frequency subbands, which then are individually encoded as depicted in Fig The encoding process works as follows. Using a bank of bandpass filters, subband coders divide the frequency

26 Chapter 3. Speech Compression 15 band of the signal into subbands. Each subband is then downsampled to its Nyquist rate (see Section 2.1) and encoded, using for example one of the techniques described above. The downsampling can be done since the bandwidth is narrower for the subbands than for the input signal. Before transmission, the signals are multiplexed. Multiplexing means combining multiple signals in order to send them over one single channel. At the receiving end, the signals are demultiplexed, decoded and modulated back to their original spectral bands before they are added together to reproduce the speech. To reduce the number of bits needed to encode the signals, different number of bits are used for different subbands. For speech, most of the energy is within the frequencies between 120 and 2000 Hz. Hence, more bits are often allotted to lower frequency bands than to higher. Downsampling Bandpass filter 1 Encoder 1 Decoder 1 Bandpass filter 1 Input signal Bandpass filter 2... Bandpass filter M Encoder 2... Encoder M M u l t i p l e x e r D e m u l t i p l e x e r Decoder 2... Decoder M Bandpass filter 2... Bandpass filter M Output signal Encoder Channel Decoder Figure 3.4: Block diagram of general subband coder (based on a figure from [6]) Transform Coding In Fig. 3.5, a simple block diagram of a transform coder is shown. A block of speech samples are unitary transformed to the frequency domain by the encoder. The transform coefficients are quantized and encoded before transmission. The receiver decodes and inverse-transforms the coefficients to reproduce the block of speech samples. The objective of transforming the speech is that the transformed signal is less correlated than the original signal. This means that most of the signal energy is contained in a small subset of transform coefficients. Karhunen-Loéve Transform (KLT), Discrete Cosine Transform (DCT) and Discrete Fourier Transform (DFT) are three examples of transforms that are used by transform coders. KLT is the transform that gives the least correlated coefficients [9]. However, DCT is more

27 16 Chapter 3. Speech Compression popular since it is nearly optimal and easy to compute using the Fast Fourier Transform. As for subband coders, bit allocation is used to remove redundancies. Adaptive Transform Coding (ATC) works in a similar way as described above, except it uses adaptive quantization and adaptive bit allocation. This leads to a higher quality of the reconstructed speech. x X T Encoder T 1 ~ X Decoder x~ Receiver Channel Transmitter Figure 3.5: Block diagram of a transform coder (based on a figure from [4]). 3.2 Vocoders Unlike waveform coders, vocoders (parametric coders) do not try to recreate the speech waveform. Vocoders use a mathematical model based on how the human speech is generated, to create synthesized speech instead of a reconstruction of the input. By an analysis of the input speech, parameters corresponding to the vocal cords and the vocal tract are estimated and quantized before transmission. At the receiver, these parameters are used to tune the model in order to construct synthesized speech. Due to the fact that vocoders operate by synthesizing speech they perform poorly on non-speech signals. Vocoders use very few bits to encode the speech, giving a very synthetic speech where the speaker is hard to identify. For more information on vocoding techniques than given below, the reader is referred to [4] and [5] Linear Predictive Coding A well-known example of Linear Predictive Coding (LPC) is the 2.4 kbit/s LPC-10. Due to the very low bit rate, the reconstructed speech sounds very unnatural. The main use for algorithms like this has been secure transmissions, where a big part of the bandwidth is needed for encrypting the speech. An LPC encoder divides the input signal into frames, typically ms long. Each frame is then analyzed to estimate the parameters needed. Linear prediction parameters, the decision if the

28 Chapter 3. Speech Compression 17 input is voiced or unvoiced speech, the pitch period and the gain are the parameters that are estimated and quantized. The quantized parameters are transmitted and a synthesis model, like the one shown in Fig. 3.6, is used to reconstruct the speech. For voiced speech a pulse generator is used for modelling the excitation signal, for unvoiced speech white noise is used. The excitation signal is used to represent the flow of air from the lungs through the glottis and the LPC parameters are used in the filter representing the oral cavity. The signal is multiplied by a gain to achieve the correct amplitude of the synthesized speech. The parameters are updated for each frame, i.e. every ms. Between the updates, the speech is assumed to be stationary. Pitch Pulse generator Noise generator Voiced/ unvoiced Switch LPC parameters Filter Gain Synthesized speech Figure 3.6: LPC synthesis model (from [5]). 3.3 Hybrid Coding Hybrid coders contain features from both waveform coders and vocoders. They try to combine the low bit rate of vocoders with the high speech quality of waveform coders to produce speech with good quality at a medium or low bit rate. Many hybrid coders can be classified as analysis-by-synthesis linear predictive coders, depicted in Fig Using this system an excitation signal is chosen or formed. The excitation signal is then filtered by a long-term linear predictive (LP) filter A L (z) corresponding to the pitch structure of the speech. Then the short-term LP-filter A(z), representing the vocal tract, is applied to the signal. This is the vocoding part of hybrid coders. The waveform part of these coders is the attempt to match the synthesized speech to the original speech. This is done by the perceptual weighting filter, W (z). W (z) is used to shape the error between the input speech s[n] and the synthesized speech ŝ[n], in order to minimize the Mean Squared Error (MSE). For more information on hybrid coding than given below, the reader is referred to [4] and [5].

29 18 Chapter 3. Speech Compression s[n] Select of Form Excitation Gain Σ Σ ^ s[n] Σ A (z) L A(z) MSE W(z) Figure 3.7: Analysis-by-synthesis linear predictive coder (from [4]) Code Excited Linear Prediction Code Excited Linear Prediction, known as CELP, contains a codebook with different excitation signals. The perceptual weighting filter is applied both to the original signal and to the synthesized signal. A difference is calculated and using this, the excitation yielding the minimal error is chosen. As for the general analysisby-synthesis coder described above, both a long-term and short-term LP-filter is applied to the signal for synthesizing the speech. An example of a standard using CELP is the Federal Standard FS 1016 CELP, which is a 4.8 kbit/s hybrid coder previously used in secure communications. Another example is ITU-T standard G.728 Coding of speech at 16 kbit/s using low-delay code excited linear prediction, known as LD-CELP. This standard has a much higher quality of the synthesized speech than FS 1016, but the bit rate is higher as well. 3.4 Discussion As mentioned in Section 1.1, it is desired to fit the newspaper on a small memory. The 90 minutes long paper is sampled at 19 khz using 16 bits to represent each sample. Thus, the uncompressed newspaper requires MB of memory. In order to get the subscribers to accept a new receiver the quality of the speech must be comparable to the quality of the present receiver, therefore ratings from subjective tests is a good thing to compare. Mean Opinion Score, or simply MOS, is a well-known subjective test used to determine the quality of reconstructed or synthesized speech. The scale in MOS tests range from 1 to 5 as follows:

30 Chapter 3. Speech Compression 19 1 Unsatisfactory (Bad) 2 Poor 3 Fair 4 Good 5 Excellent. In the second column of Table 3.1 the result from a MOS test is given for five different compression techniques. Both the waveform coders PCM and ADPCM have high MOS scores as well as the hybrid coder LD-CELP. Many hybrid coders have a high quality of the synthesized speech, but they tend to have a very high complexity and they require a codebook. Due to these reasons a hybrid coder is not a good choice for the compression of the spoken newspaper. Since vocoders produces unnatural speech they are not considered as candidates for use in the digital receiver. The class of speech coders that is left is waveform coders. From Table 3.1, it is clear that if ADPCM is used to compress the spoken newspaper, it would be possible to store the paper on a 64 MB memory chip. For this reason and the fact that ADPCM has a high quality of the reconstructed speech and a low complexity this algorithm was chosen for the project. A detailed description of this algorithm is given in the next section. Table 3.1: Coding algorithms and their performances (based on information found in [4]). Algorithm MOS CR a MIPS b Memory requirement c (in MB) log PCM ADPCM LPC FS 1016 CELP LD-CELP a CR = Compression Ratio is the ratio between the number of bits needed to represent one uncompressed sample and the number of bits needed to represent one compressed sample. b MIPS = Million Instructions Per Second. These figures are only approximate, the complexity is dependent on the implementation. c The amount of memory required to store a 90 minutes long encoded spoken newspaper.

31 20 Chapter 3. Speech Compression

32 Chapter 4 Adaptive Differential Pulse Code Modulation The compression technique chosen for this project is Adaptive Differential Pulse Code Modulation (ADPCM). The reasons for choosing this particular algorithm is that the quality of the reconstructed speech is high and the complexity is low, about 600 instructions are needed to encode or decode one sample of the speech. Compressed with ADPCM, one newspaper requires approximately 51.3 MB of memory, as seen from Table 3.1. Using a 64 MB memory chip, space will be available for storing overhead information. An additional advantage is that ADPCM is free of charge [10]. ADPCM was approved as the international standard, CCITT Recommendation G kbit/s Adaptive Differential Pulse Code Modulation (ADPCM) in October 1984, see [11]. The standard is now a part of ITU-T Recommendation G , 32, 24, 16 kbit/s Adaptive Differential Pulse Code Modulation (ADPCM) [12]. The version used in this project is the 32 kbit/s ADPCM. This bit rate is valid if the input speech is sampled at 8 khz using 16 bits accuracy, but the sampling frequency used in the digital receiver is 19 khz. With a CR of 4 (see Section 3.4) the output bit rate becomes = 76 kbit/s. This will have no effect on the 4 calculations other than that they will be performed at a higher rate. ADPCM makes use of redundancies in speech signals, as correlation between contiguous samples, to minimize the number of bits needed to represent the signal. The difference between the incoming sample and an estimation of that sample is calculated. The difference is encoded, instead of the actual input signal itself. This reduces the variance of the signal and therefore fewer bits are required to encode 21

33 22 Chapter 4. Adaptive Differential Pulse Code Modulation it. A simplified block diagram of the ADPCM encoder/decoder is shown in Fig The estimate of the input signal, s e [k], is subtracted from the actual input signal s[k] and the resulting difference, d[k], is quantized and encoded to I[k] before transmission. The reconstruction of the speech starts by recreating the received word to d q [k]. The difference is then added to the estimate calculated by A(z) and B(z), to produce the reconstructed speech r[k]. G.726 is backward adaptive, Encoder s[k] d[k] Q I[k] Q 1 output. Step size adaption y[k] d [k] q s [k] e B(z) A(z). r[k] Decoder output Figure 4.1: Simplified block diagram of ADPCM encoder/decoder (based on a figure from [11]). meaning that the prediction of the input signal is made from previous samples. Hence, the prediction can be made at the receiver without any side information having to be sent. As shown in Fig. 4.1, the decoder can be regarded as a subset of the encoder. 4.1 The Encoder The encoding process starts by converting the input signal from logarithmic PCM to uniform PCM, as seen in Fig Next, the difference between the input signal and its estimate is calculated and this value is assigned four bits by the adaptive quantizer. This four-bit word, I[k], is the output of the encoder. I[k] is sent both to the decoder and to the inverse quantizer. The output of the inverse quantizer

34 Chapter 4. Adaptive Differential Pulse Code Modulation 23 is the quantized difference signal, which is added to the estimate of the signal to reconstruct the input. This reconstructed signal, as well as the quantized difference signal, is fed to the adaptive predictor for an estimation of the next input sample. Below follows descriptions of what each block in the encoder does. ADPCM output Reconstructed signal calculator s (k) r s(k) Input PCM s (k) Difference l d(k) l(k) format conversion signal computation Adaptive quantizer Inverse adaptive quantizer d (k) q Adaptive predictor s e(k) a (k) 2 Quantizer scale factor adaptation y(k) a (k) 1 Adaptation speed control t (k) r t (k) d Tone and transition detector y (k) l Figure 4.2: Block diagram of ADPCM encoder (from [12]). Input PCM Format Conversion The input to the ADPCM encoder is an A-law or µ-law pulse code modulated signal, with an accuracy of 8 bits. This log quantized signal is converted to uniform PCM before compression. Difference Signal Computation The second step in the encoding process is to calculate the difference signal by subtracting the estimate of the input from the actual input signal: d[k] = s l [k] s e [k]. (4.1)

35 24 Chapter 4. Adaptive Differential Pulse Code Modulation Adaptive Quantizer ADPCM uses backward adaptive quantization, described in Section The quantizer used is a 15-level non-uniform adaptive quantizer. Before quantizing the difference signal, it is converted to a base-2 logarithmic representation and scaled by the scale factor y[k]. The normalized input to the quantizer is then which is quantized to create the encoder output I[k]. log 2 d[k] y[k], (4.2) Inverse Adaptive Quantizer The inverse adaptive quantizer constructs the quantized difference signal, d q [k], by first decoding I[k] and then adding the scale factor y[k]. Finally, the result is transformed from the logarithmic domain. Quantizer Scale Factor Adaption The scale factor y[k], used in the quantizer and in the inverse quantizer, is composed of two parts. One part for fast varying signals and one for slowly varying signals. The fast, unlocked, scale factor is recursively calculated, using the resulting scale factor y[k]: y u [k] = (1 2 5 )y[k] W (I[k]). (4.3) W is a number from a discrete set, depending on which one of the fifteen quantization levels is used for the current sample, see [12]. The second part of the scale factor is the slow, or locked factor y l [k] and it is calculated as follows: y l [k] = (1 2 6 )y l [k 1] y u [k]. (4.4) The resultant scale factor y[k] is a combination of the fast and the slow scale factors: y[k] = a l [k]y u [k 1] + (1 a l [k])y l [k 1] (4.5)

36 Chapter 4. Adaptive Differential Pulse Code Modulation 25 The controlling parameter a l [k] is described below. Adaptation Speed Control a l [k] is a controlling parameter that ranges between 0 and 1. For speech signals this parameter approaches 1, forcing the quantizer towards the fast mode. For slow signals like tones, the parameter tends towards 0, driving the quantizer into the slow mode. For detailed information on how a l [k] is constructed, please see [12]. Adaptive Predictor and Reconstructed Signal Calculator The estimate, s e [k], of the input signal is calculated using two previously reconstructed samples s r and six previous difference signals d q : where s e [k] = 2 a i [k 1]s r [k i] + s ez [k]. (4.6) i=1 s ez [k] = 6 b i [k 1]d q [k i]. (4.7) i=1 The reconstructed signal s r is calculated by adding the quantized difference to the estimate of the signal: The coefficients a i and b i are found in [12]. s r [k i] = s e [k i] + d q [k i]. (4.8) Tone and Transition Detector When the quantizer is in the slow, locked, mode and a stationary signal change to another stationary signal, problems can occur. An example of this type of signal is tones from a frequency shift keying modem. The problems that might occur are prevented by the tone and transition detector. When a transition between different

37 26 Chapter 4. Adaptive Differential Pulse Code Modulation stationary signals is detected the quantizer is forced into the fast, unlocked, mode by setting all the predictor coefficients equal to zero. 4.2 The Decoder As can be seen in Fig. 4.3 the structure of the decoder is for the most part the same as for the encoder. I[k], the four-bit word representing the difference between the input signal and its estimate, is fed to the inverse quantizer. The quantized difference signal is added to the estimate of the input and then converted from uniform PCM to either A-law or µ-law PCM. The blocks in common for both Figure 4.3: Block diagram of an ADPCM decoder (from [12]). the encoder and decoder were described in Section 4.1; the blocks unique for the decoder are described below. Output PCM Format Conversion This is the block that after reconstructing the signal converts it back to A-law or µ-law PCM format from uniform PCM.

38 Chapter 4. Adaptive Differential Pulse Code Modulation 27 Synchronous Coding Adjustment This block have been added to the ADPCM decoder to reduce cumulative distortion that can appear from successive synchronous tandem codings, i.e., ADPCM to PCM to ADPCM to PCM to ADPCM etc. With this feature, any number of synchronous tandem codings is equivalent to one single coding, providing an ideal channel (no transmission errors). Details about the synchronous coding adjustment is found in [12]. 4.3 Implementation and Results ADPCM was first implemented on the ADSP-2181 DSP, described in Section below, using code obtained from Analog Devices. The speech compressed using this program had a high quality, as expected. Since the sampling frequency had to be set exactly to 19 khz in order to decrypt the newspaper, the ADSP-2181 DSP was replaced by the ADSP-2191 DSP. Both DSPs are programmed using assembly language, and they were supposed to be compatible. However, some instructions were defined differently in the new DSP, therefore a conversion of the code had to be made. Unfortunately, due to some problems, the implementation on the ADSP-2191 could not be completed within the time frame of this thesis Used Tools Both the ADSP-2181 DSP and the ADSP-2191 DSP were used during the implementation of ADPCM. First the evaluation kit ADSP-21XX EZ-KIT Lite was used, some of its attributes are as follows: 16-bit fixed-point ADSP-2181 DSP Ability to perform 33 MIPS 1 80 kb of on-chip RAM AD1847 stereo codec Serial port connection 1 MIPS = Million Instructions Per Second

39 28 Chapter 4. Adaptive Differential Pulse Code Modulation Due to a need of setting the sampling frequency in steps of 1 Hz the ADSP-2181 was replaced by the ADSP The evaluation board for this DSP is called ADSP-2191 EZ-KIT Lite and have the following attributes: 16-bit fixed-point ADSP-2191 DSP Ability to perform 160 MIPS 160 kb of on-chip RAM AD kHz AC 97 SoundMAX codec USB version 1.1 connection Ability to set the sampling frequency in steps of 1 Hz The attributes for both evaluation kits are attained from [13].

40 Chapter 5 Tone Detection Between two articles in the spoken newspaper a 50 Hz tone is inserted, indicating the end of one article and the start of the next. Today the subscribers of spoken newspapers use a normal cassette player when listening to their recorded paper. When listening to the newspaper at normal speed the tone is not audible, partly because of its low frequency and partly because the amplitude of the tone is smaller than the amplitude of the speech, see Fig However, when the listeners fast forward the tape, the frequency of the tone is increased and it can be heard as a beep. In the new digital receiver, it is desired to skip between articles just by pressing a button and for this reason the 50 Hz tone must be detected. The idea is to add a bookmark to the location of the tone. This way pressing the forward button on the new receiver will cause a jump to the next article, that is, the place where the next bookmark is located. 5.1 Finding the Tone The tone is added to the newspaper by the person reading it out, this causes the duration of the tone to vary. A couple of tones have been examined and they have had durations between approximately 0.5 and 2.5 seconds. Due to the human factor, the tone is not always located between articles, sometimes it is found in the end of one article hidden under the speech. Thus, the problem can be described as detecting a known signal with unknown duration in non-white, non-stationary noise (i.e. the speech). 29

41 30 Chapter 5. Tone Detection Hz tones Amplitude Time (s) Time (s) Figure 5.1: Speech and 50 Hz tones. Requirements for the tone detection: Due to limited memory on the DSP, the solution to the tone detection problem should not require too much memory. The tone detection should be fast, i.e. not require many instructions. The detection must be robust, meaning as well detecting all tones as a low probability of false alarm. If the detector fails to detect one tone and the next one is detected, skipping forward means missing an article. False alarm occurs when something other than a tone is believed to be the tone searched for. The first step in detecting the tone is to reduce the amplitude of the speech while the amplitude of the tone is maintained or possibly increased. The second step is to decide whether the tone searched for is present in the signal or not.

42 Chapter 5. Tone Detection Matched Filter Since matched filters can be used to search for known signals in noise it might be a good thing to use for the first step in the process of detecting the 50 Hz tone. A matched filter is designed to maximize the output Signal-to-Noise Ratio (SNR) [14]. Thus, it might be possible to use a matched filter to remove the speech. The aim of a matched filter is not to keep the waveform of the signal searched for unchanged, but to maximize the power of the known signal with respect to the power of the noise. The received signal can be described by r(t) = s(t) + n(t), (5.1) where s(t) is the known signal searched for (in this case the 50 Hz tone) and n(t) is the additive colored noise (the speech) corrupting the tone. If S (f) is the Fourier transform of the known signal s(t) and the Power Spectral Density (PSD) [14], of the colored input noise is P n (f), the transfer function of the matched filter is given by: H(f) = K S (f) P n (f) e j2πft 0, (5.2) where K is an arbitrary real nonzero constant and t 0 is the sampling time. A proof of this is given in [14]. Usage of the filter described above requires knowledge about the PSD of the speech. Due to this a simplification is needed. If the assumption that the noise is white is made, P n (f) equals N 0 /2. This reduces (5.2) to H(f) = 2K N 0 S (f)e j2πft 0. (5.3) Hence, if the noise is white, the impulse response of the matched filter becomes h(t) = Cs(t 0 t), (5.4) where C is an arbitrary real positive constant. The proof of this equation is found in [14]. From (5.4) it is clear that the matched filter is a scaled time-reversed version of the known signal itself, delayed by the interval t 0, see Fig For the matched filter to be realizable, it must be causal, i.e., h(t) = 0, if t < 0 (5.5) and for that reason t 0 must be equal to the length of the known signal. The sampling frequency used in the DSP is 19 khz, one period of the 50 Hz tone is therefore represented by 380 samples. For the matched filter to be optimal,

43 32 Chapter 5. Tone Detection the duration of the tone must be known. As stated before, the duration of the tone varies, and hence an assumption of the duration must be made. Thus, the length of the matched filter becomes the same as the number of samples needed to represent one period of the tone, multiplied by the number of periods chosen. However, this procedure requires a large amount of storage, since the same number of samples as the length of the filter must to be stored. A method for reducing the s(t) h(t) T t t 0 t (a) (b) Figure 5.2: (a) s(t) is the known signal; (b) h(t) is the matched filter. length of the filter is to perform a downsampling of the input signal prior to the detection process, and hereby reduce the number of samples needed to represent one period of the tone. Downsampling can not be done unless a lowpass filter first is applied to the input signal, else aliasing (described in Section 2.1) might occur. For example, if it is desired to use a sampling frequency of 400 Hz, the lowpass filter must remove all frequencies above 200 Hz. If the filter is not steep enough, frequencies above 200 Hz will be represented with a lower frequency, maybe as 50 Hz. A steep lowpass filter is of high order, i.e. many samples must be stored and this diminishes the gain of the downsampling. Therefore the matched filter does not seem to fulfill the storage requirements of the tone detection. The assumption that the noise is white is not completely correct, this error together with the large storage requirement leads to the conclusion that another solution than a matched filter might be more suitable Digital Resonator Another possible solution to the tone detection problem is to remove all frequencies in the input signal except frequencies around 50 Hz. This can be done by applying a bandpass filter to the incoming signal. By testing if the power of the output signal is bigger than some threshold value the presence of the tone can be detected. Since most of the energy in human speech is limited to frequencies between 120 and 2000

44 Chapter 5. Tone Detection 33 Hz, the filter used to remove the speech must drop of sharply. Thus, the magnitude response of the filter should look like a very steep, narrow bandpass filter, like the ideal filter shown in Fig Frequency in Hz Figure 5.3: Magnitude response of an ideal bandpass filter with the passband [ ] Hz. A filter known as a digital resonator [15], is a second order digital filter with complex conjugated poles in ae ±jω 0 where a is close to 1. ω 0 is the digital resonance frequency, that is, the frequency of interest. In this case the resonance frequency is π/190, the digital frequency corresponding to 50 Hz. The digital frequency ranges between 0 and π rad/s, with π corresponding to half the sampling frequency. A pole-zero plot of a general digital resonator is shown in Fig Imaginary Part Real Part Figure 5.4: Pole-zero plot of a general digital resonator. The poles and zeros are indicated by crosses and circles respectively.

Overview of Code Excited Linear Predictive Coder

Overview of Code Excited Linear Predictive Coder Overview of Code Excited Linear Predictive Coder Minal Mulye 1, Sonal Jagtap 2 1 PG Student, 2 Assistant Professor, Department of E&TC, Smt. Kashibai Navale College of Engg, Pune, India Abstract Advances

More information

Chapter IV THEORY OF CELP CODING

Chapter IV THEORY OF CELP CODING Chapter IV THEORY OF CELP CODING CHAPTER IV THEORY OF CELP CODING 4.1 Introduction Wavefonn coders fail to produce high quality speech at bit rate lower than 16 kbps. Source coders, such as LPC vocoders,

More information

EC 6501 DIGITAL COMMUNICATION UNIT - II PART A

EC 6501 DIGITAL COMMUNICATION UNIT - II PART A EC 6501 DIGITAL COMMUNICATION 1.What is the need of prediction filtering? UNIT - II PART A [N/D-16] Prediction filtering is used mostly in audio signal processing and speech processing for representing

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 12 Speech Signal Processing 14/03/25 http://www.ee.unlv.edu/~b1morris/ee482/

More information

10 Speech and Audio Signals

10 Speech and Audio Signals 0 Speech and Audio Signals Introduction Speech and audio signals are normally converted into PCM, which can be stored or transmitted as a PCM code, or compressed to reduce the number of bits used to code

More information

APPLICATIONS OF DSP OBJECTIVES

APPLICATIONS OF DSP OBJECTIVES APPLICATIONS OF DSP OBJECTIVES This lecture will discuss the following: Introduce analog and digital waveform coding Introduce Pulse Coded Modulation Consider speech-coding principles Introduce the channel

More information

techniques are means of reducing the bandwidth needed to represent the human voice. In mobile

techniques are means of reducing the bandwidth needed to represent the human voice. In mobile 8 2. LITERATURE SURVEY The available radio spectrum for the wireless radio communication is very limited hence to accommodate maximum number of users the speech is compressed. The speech compression techniques

More information

EE 225D LECTURE ON MEDIUM AND HIGH RATE CODING. University of California Berkeley

EE 225D LECTURE ON MEDIUM AND HIGH RATE CODING. University of California Berkeley University of California Berkeley College of Engineering Department of Electrical Engineering and Computer Sciences Professors : N.Morgan / B.Gold EE225D Spring,1999 Medium & High Rate Coding Lecture 26

More information

Digital Speech Processing and Coding

Digital Speech Processing and Coding ENEE408G Spring 2006 Lecture-2 Digital Speech Processing and Coding Spring 06 Instructor: Shihab Shamma Electrical & Computer Engineering University of Maryland, College Park http://www.ece.umd.edu/class/enee408g/

More information

ECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2

ECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2 ECE 556 BASICS OF DIGITAL SPEECH PROCESSING Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2 Analog Sound to Digital Sound Characteristics of Sound Amplitude Wavelength (w) Frequency ( ) Timbre

More information

EEE 309 Communication Theory

EEE 309 Communication Theory EEE 309 Communication Theory Semester: January 2016 Dr. Md. Farhad Hossain Associate Professor Department of EEE, BUET Email: mfarhadhossain@eee.buet.ac.bd Office: ECE 331, ECE Building Part 05 Pulse Code

More information

Cellular systems & GSM Wireless Systems, a.a. 2014/2015

Cellular systems & GSM Wireless Systems, a.a. 2014/2015 Cellular systems & GSM Wireless Systems, a.a. 2014/2015 Un. of Rome La Sapienza Chiara Petrioli Department of Computer Science University of Rome Sapienza Italy 2 Voice Coding 3 Speech signals Voice coding:

More information

Analysis/synthesis coding

Analysis/synthesis coding TSBK06 speech coding p.1/32 Analysis/synthesis coding Many speech coders are based on a principle called analysis/synthesis coding. Instead of coding a waveform, as is normally done in general audio coders

More information

QUESTION BANK. SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2

QUESTION BANK. SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2 QUESTION BANK DEPARTMENT: ECE SEMESTER: V SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2 BASEBAND FORMATTING TECHNIQUES 1. Why prefilterring done before sampling [AUC NOV/DEC 2010] The signal

More information

The Channel Vocoder (analyzer):

The Channel Vocoder (analyzer): Vocoders 1 The Channel Vocoder (analyzer): The channel vocoder employs a bank of bandpass filters, Each having a bandwidth between 100 Hz and 300 Hz. Typically, 16-20 linear phase FIR filter are used.

More information

Speech Compression Using Voice Excited Linear Predictive Coding

Speech Compression Using Voice Excited Linear Predictive Coding Speech Compression Using Voice Excited Linear Predictive Coding Ms.Tosha Sen, Ms.Kruti Jay Pancholi PG Student, Asst. Professor, L J I E T, Ahmedabad Abstract : The aim of the thesis is design good quality

More information

EEE 309 Communication Theory

EEE 309 Communication Theory EEE 309 Communication Theory Semester: January 2017 Dr. Md. Farhad Hossain Associate Professor Department of EEE, BUET Email: mfarhadhossain@eee.buet.ac.bd Office: ECE 331, ECE Building Types of Modulation

More information

speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals

speech signal S(n). This involves a transformation of S(n) into another signal or a set of signals 16 3. SPEECH ANALYSIS 3.1 INTRODUCTION TO SPEECH ANALYSIS Many speech processing [22] applications exploits speech production and perception to accomplish speech analysis. By speech analysis we extract

More information

Telecommunication Electronics

Telecommunication Electronics Politecnico di Torino ICT School Telecommunication Electronics C5 - Special A/D converters» Logarithmic conversion» Approximation, A and µ laws» Differential converters» Oversampling, noise shaping Logarithmic

More information

Enhanced Waveform Interpolative Coding at 4 kbps

Enhanced Waveform Interpolative Coding at 4 kbps Enhanced Waveform Interpolative Coding at 4 kbps Oded Gottesman, and Allen Gersho Signal Compression Lab. University of California, Santa Barbara E-mail: [oded, gersho]@scl.ece.ucsb.edu Signal Compression

More information

EC 2301 Digital communication Question bank

EC 2301 Digital communication Question bank EC 2301 Digital communication Question bank UNIT I Digital communication system 2 marks 1.Draw block diagram of digital communication system. Information source and input transducer formatter Source encoder

More information

Pulse Code Modulation

Pulse Code Modulation Pulse Code Modulation EE 44 Spring Semester Lecture 9 Analog signal Pulse Amplitude Modulation Pulse Width Modulation Pulse Position Modulation Pulse Code Modulation (3-bit coding) 1 Advantages of Digital

More information

Time division multiplexing The block diagram for TDM is illustrated as shown in the figure

Time division multiplexing The block diagram for TDM is illustrated as shown in the figure CHAPTER 2 Syllabus: 1) Pulse amplitude modulation 2) TDM 3) Wave form coding techniques 4) PCM 5) Quantization noise and SNR 6) Robust quantization Pulse amplitude modulation In pulse amplitude modulation,

More information

Communications Theory and Engineering

Communications Theory and Engineering Communications Theory and Engineering Master's Degree in Electronic Engineering Sapienza University of Rome A.A. 2018-2019 Speech and telephone speech Based on a voice production model Parametric representation

More information

Comparison of CELP speech coder with a wavelet method

Comparison of CELP speech coder with a wavelet method University of Kentucky UKnowledge University of Kentucky Master's Theses Graduate School 2006 Comparison of CELP speech coder with a wavelet method Sriram Nagaswamy University of Kentucky, sriramn@gmail.com

More information

Speech Synthesis; Pitch Detection and Vocoders

Speech Synthesis; Pitch Detection and Vocoders Speech Synthesis; Pitch Detection and Vocoders Tai-Shih Chi ( 冀泰石 ) Department of Communication Engineering National Chiao Tung University May. 29, 2008 Speech Synthesis Basic components of the text-to-speech

More information

Digital Communication (650533) CH 3 Pulse Modulation

Digital Communication (650533) CH 3 Pulse Modulation Philadelphia University/Faculty of Engineering Communication and Electronics Engineering Digital Communication (650533) CH 3 Pulse Modulation Instructor: Eng. Nada Khatib Website: http://www.philadelphia.edu.jo/academics/nkhatib/

More information

Analog and Telecommunication Electronics

Analog and Telecommunication Electronics Politecnico di Torino - ICT School Analog and Telecommunication Electronics D5 - Special A/D converters» Differential converters» Oversampling, noise shaping» Logarithmic conversion» Approximation, A and

More information

Multirate DSP, part 3: ADC oversampling

Multirate DSP, part 3: ADC oversampling Multirate DSP, part 3: ADC oversampling Li Tan - May 04, 2008 Order this book today at www.elsevierdirect.com or by calling 1-800-545-2522 and receive an additional 20% discount. Use promotion code 92562

More information

IMPLEMENTATION OF G.726 ITU-T VOCODER ON A SINGLE CHIP USING VHDL

IMPLEMENTATION OF G.726 ITU-T VOCODER ON A SINGLE CHIP USING VHDL IMPLEMENTATION OF G.726 ITU-T VOCODER ON A SINGLE CHIP USING VHDL G.Murugesan N. Ramadass Dr.J.Raja paul Perinbum School of ECE Anna University Chennai-600 025 Gm1gm@rediffmail.com ramadassn@yahoo.com

More information

Chapter 2: Digitization of Sound

Chapter 2: Digitization of Sound Chapter 2: Digitization of Sound Acoustics pressure waves are converted to electrical signals by use of a microphone. The output signal from the microphone is an analog signal, i.e., a continuous-valued

More information

Fundamentals of Digital Communication

Fundamentals of Digital Communication Fundamentals of Digital Communication Network Infrastructures A.A. 2017/18 Digital communication system Analog Digital Input Signal Analog/ Digital Low Pass Filter Sampler Quantizer Source Encoder Channel

More information

CHAPTER 3 Syllabus (2006 scheme syllabus) Differential pulse code modulation DPCM transmitter

CHAPTER 3 Syllabus (2006 scheme syllabus) Differential pulse code modulation DPCM transmitter CHAPTER 3 Syllabus 1) DPCM 2) DM 3) Base band shaping for data tranmission 4) Discrete PAM signals 5) Power spectra of discrete PAM signal. 6) Applications (2006 scheme syllabus) Differential pulse code

More information

QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold

QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold circuit 2. What is the difference between natural sampling

More information

CODING TECHNIQUES FOR ANALOG SOURCES

CODING TECHNIQUES FOR ANALOG SOURCES CODING TECHNIQUES FOR ANALOG SOURCES Prof.Pratik Tawde Lecturer, Electronics and Telecommunication Department, Vidyalankar Polytechnic, Wadala (India) ABSTRACT Image Compression is a process of removing

More information

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 03 Quantization, PCM and Delta Modulation Hello everyone, today we will

More information

Speech Coding Technique And Analysis Of Speech Codec Using CS-ACELP

Speech Coding Technique And Analysis Of Speech Codec Using CS-ACELP Speech Coding Technique And Analysis Of Speech Codec Using CS-ACELP Monika S.Yadav Vidarbha Institute of Technology Rashtrasant Tukdoji Maharaj Nagpur University, Nagpur, India monika.yadav@rediffmail.com

More information

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued CSCD 433 Network Programming Fall 2016 Lecture 5 Physical Layer Continued 1 Topics Definitions Analog Transmission of Digital Data Digital Transmission of Analog Data Multiplexing 2 Different Types of

More information

Speech Coding in the Frequency Domain

Speech Coding in the Frequency Domain Speech Coding in the Frequency Domain Speech Processing Advanced Topics Tom Bäckström Aalto University October 215 Introduction The speech production model can be used to efficiently encode speech signals.

More information

Chapter-3 Waveform Coding Techniques

Chapter-3 Waveform Coding Techniques Chapter-3 Waveform Coding Techniques PCM [Pulse Code Modulation] PCM is an important method of analog to-digital conversion. In this modulation the analog signal is converted into an electrical waveform

More information

Waveform Encoding - PCM. BY: Dr.AHMED ALKHAYYAT. Chapter Two

Waveform Encoding - PCM. BY: Dr.AHMED ALKHAYYAT. Chapter Two Chapter Two Layout: 1. Introduction. 2. Pulse Code Modulation (PCM). 3. Differential Pulse Code Modulation (DPCM). 4. Delta modulation. 5. Adaptive delta modulation. 6. Sigma Delta Modulation (SDM). 7.

More information

Voice Transmission --Basic Concepts--

Voice Transmission --Basic Concepts-- Voice Transmission --Basic Concepts-- Voice---is analog in character and moves in the form of waves. 3-important wave-characteristics: Amplitude Frequency Phase Telephone Handset (has 2-parts) 2 1. Transmitter

More information

Audio Signal Compression using DCT and LPC Techniques

Audio Signal Compression using DCT and LPC Techniques Audio Signal Compression using DCT and LPC Techniques P. Sandhya Rani#1, D.Nanaji#2, V.Ramesh#3,K.V.S. Kiran#4 #Student, Department of ECE, Lendi Institute Of Engineering And Technology, Vizianagaram,

More information

PULSE CODE MODULATION (PCM)

PULSE CODE MODULATION (PCM) PULSE CODE MODULATION (PCM) 1. PCM quantization Techniques 2. PCM Transmission Bandwidth 3. PCM Coding Techniques 4. PCM Integrated Circuits 5. Advantages of PCM 6. Delta Modulation 7. Adaptive Delta Modulation

More information

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued CSCD 433 Network Programming Fall 2016 Lecture 5 Physical Layer Continued 1 Topics Definitions Analog Transmission of Digital Data Digital Transmission of Analog Data Multiplexing 2 Different Types of

More information

Compression. Encryption. Decryption. Decompression. Presentation of Information to client site

Compression. Encryption. Decryption. Decompression. Presentation of Information to client site DOCUMENT Anup Basu Audio Image Video Data Graphics Objectives Compression Encryption Network Communications Decryption Decompression Client site Presentation of Information to client site Multimedia -

More information

Speech Synthesis using Mel-Cepstral Coefficient Feature

Speech Synthesis using Mel-Cepstral Coefficient Feature Speech Synthesis using Mel-Cepstral Coefficient Feature By Lu Wang Senior Thesis in Electrical Engineering University of Illinois at Urbana-Champaign Advisor: Professor Mark Hasegawa-Johnson May 2018 Abstract

More information

Speech Enhancement using Wiener filtering

Speech Enhancement using Wiener filtering Speech Enhancement using Wiener filtering S. Chirtmay and M. Tahernezhadi Department of Electrical Engineering Northern Illinois University DeKalb, IL 60115 ABSTRACT The problem of reducing the disturbing

More information

Low Bit Rate Speech Coding

Low Bit Rate Speech Coding Low Bit Rate Speech Coding Jaspreet Singh 1, Mayank Kumar 2 1 Asst. Prof.ECE, RIMT Bareilly, 2 Asst. Prof.ECE, RIMT Bareilly ABSTRACT Despite enormous advances in digital communication, the voice is still

More information

Voice Excited Lpc for Speech Compression by V/Uv Classification

Voice Excited Lpc for Speech Compression by V/Uv Classification IOSR Journal of VLSI and Signal Processing (IOSR-JVSP) Volume 6, Issue 3, Ver. II (May. -Jun. 2016), PP 65-69 e-issn: 2319 4200, p-issn No. : 2319 4197 www.iosrjournals.org Voice Excited Lpc for Speech

More information

Audio /Video Signal Processing. Lecture 1, Organisation, A/D conversion, Sampling Gerald Schuller, TU Ilmenau

Audio /Video Signal Processing. Lecture 1, Organisation, A/D conversion, Sampling Gerald Schuller, TU Ilmenau Audio /Video Signal Processing Lecture 1, Organisation, A/D conversion, Sampling Gerald Schuller, TU Ilmenau Gerald Schuller gerald.schuller@tu ilmenau.de Organisation: Lecture each week, 2SWS, Seminar

More information

Problem Sheet 1 Probability, random processes, and noise

Problem Sheet 1 Probability, random processes, and noise Problem Sheet 1 Probability, random processes, and noise 1. If F X (x) is the distribution function of a random variable X and x 1 x 2, show that F X (x 1 ) F X (x 2 ). 2. Use the definition of the cumulative

More information

Department of Electronics and Communication Engineering 1

Department of Electronics and Communication Engineering 1 UNIT I SAMPLING AND QUANTIZATION Pulse Modulation 1. Explain in detail the generation of PWM and PPM signals (16) (M/J 2011) 2. Explain in detail the concept of PWM and PAM (16) (N/D 2012) 3. What is the

More information

Downloaded from 1

Downloaded from  1 VII SEMESTER FINAL EXAMINATION-2004 Attempt ALL questions. Q. [1] How does Digital communication System differ from Analog systems? Draw functional block diagram of DCS and explain the significance of

More information

Simulation of Conjugate Structure Algebraic Code Excited Linear Prediction Speech Coder

Simulation of Conjugate Structure Algebraic Code Excited Linear Prediction Speech Coder COMPUSOFT, An international journal of advanced computer technology, 3 (3), March-204 (Volume-III, Issue-III) ISSN:2320-0790 Simulation of Conjugate Structure Algebraic Code Excited Linear Prediction Speech

More information

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm

Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm International OPEN ACCESS Journal Of Modern Engineering Research (IJMER) Speech Enhancement Based On Spectral Subtraction For Speech Recognition System With Dpcm A.T. Rajamanickam, N.P.Subiramaniyam, A.Balamurugan*,

More information

Communications and Signals Processing

Communications and Signals Processing Communications and Signals Processing Dr. Ahmed Masri Department of Communications An Najah National University 2012/2013 1 Dr. Ahmed Masri Chapter 5 - Outlines 5.4 Completing the Transition from Analog

More information

EXPERIMENT WISE VIVA QUESTIONS

EXPERIMENT WISE VIVA QUESTIONS EXPERIMENT WISE VIVA QUESTIONS Pulse Code Modulation: 1. Draw the block diagram of basic digital communication system. How it is different from analog communication system. 2. What are the advantages of

More information

Page 0 of 23. MELP Vocoder

Page 0 of 23. MELP Vocoder Page 0 of 23 MELP Vocoder Outline Introduction MELP Vocoder Features Algorithm Description Parameters & Comparison Page 1 of 23 Introduction Traditional pitched-excited LPC vocoders use either a periodic

More information

Auditory modelling for speech processing in the perceptual domain

Auditory modelling for speech processing in the perceptual domain ANZIAM J. 45 (E) ppc964 C980, 2004 C964 Auditory modelling for speech processing in the perceptual domain L. Lin E. Ambikairajah W. H. Holmes (Received 8 August 2003; revised 28 January 2004) Abstract

More information

Wideband Speech Coding & Its Application

Wideband Speech Coding & Its Application Wideband Speech Coding & Its Application Apeksha B. landge. M.E. [student] Aditya Engineering College Beed Prof. Amir Lodhi. Guide & HOD, Aditya Engineering College Beed ABSTRACT: Increasing the bandwidth

More information

Vocoder (LPC) Analysis by Variation of Input Parameters and Signals

Vocoder (LPC) Analysis by Variation of Input Parameters and Signals ISCA Journal of Engineering Sciences ISCA J. Engineering Sci. Vocoder (LPC) Analysis by Variation of Input Parameters and Signals Abstract Gupta Rajani, Mehta Alok K. and Tiwari Vebhav Truba College of

More information

Mel Spectrum Analysis of Speech Recognition using Single Microphone

Mel Spectrum Analysis of Speech Recognition using Single Microphone International Journal of Engineering Research in Electronics and Communication Mel Spectrum Analysis of Speech Recognition using Single Microphone [1] Lakshmi S.A, [2] Cholavendan M [1] PG Scholar, Sree

More information

Datenkommunikation SS L03 - TDM Techniques. Time Division Multiplexing (synchronous, statistical) Digital Voice Transmission, PDH, SDH

Datenkommunikation SS L03 - TDM Techniques. Time Division Multiplexing (synchronous, statistical) Digital Voice Transmission, PDH, SDH TM Techniques Time ivision Multiplexing (synchronous, statistical) igital Voice Transmission, PH, SH Agenda Introduction Synchronous (eterministic) TM Asynchronous (Statistical) TM igital Voice Transmission

More information

Voice mail and office automation

Voice mail and office automation Voice mail and office automation by DOUGLAS L. HOGAN SPARTA, Incorporated McLean, Virginia ABSTRACT Contrary to expectations of a few years ago, voice mail or voice messaging technology has rapidly outpaced

More information

Digital Audio. Lecture-6

Digital Audio. Lecture-6 Digital Audio Lecture-6 Topics today Digitization of sound PCM Lossless predictive coding 2 Sound Sound is a pressure wave, taking continuous values Increase / decrease in pressure can be measured in amplitude,

More information

Pulse Code Modulation

Pulse Code Modulation Pulse Code Modulation Modulation is the process of varying one or more parameters of a carrier signal in accordance with the instantaneous values of the message signal. The message signal is the signal

More information

Appendix B. Design Implementation Description For The Digital Frequency Demodulator

Appendix B. Design Implementation Description For The Digital Frequency Demodulator Appendix B Design Implementation Description For The Digital Frequency Demodulator The DFD design implementation is divided into four sections: 1. Analog front end to signal condition and digitize the

More information

UNIT III -- DATA AND PULSE COMMUNICATION PART-A 1. State the sampling theorem for band-limited signals of finite energy. If a finite energy signal g(t) contains no frequency higher than W Hz, it is completely

More information

ON-LINE LABORATORIES FOR SPEECH AND IMAGE PROCESSING AND FOR COMMUNICATION SYSTEMS USING J-DSP

ON-LINE LABORATORIES FOR SPEECH AND IMAGE PROCESSING AND FOR COMMUNICATION SYSTEMS USING J-DSP ON-LINE LABORATORIES FOR SPEECH AND IMAGE PROCESSING AND FOR COMMUNICATION SYSTEMS USING J-DSP A. Spanias, V. Atti, Y. Ko, T. Thrasyvoulou, M.Yasin, M. Zaman, T. Duman, L. Karam, A. Papandreou, K. Tsakalis

More information

Sampling and Reconstruction of Analog Signals

Sampling and Reconstruction of Analog Signals Sampling and Reconstruction of Analog Signals Chapter Intended Learning Outcomes: (i) Ability to convert an analog signal to a discrete-time sequence via sampling (ii) Ability to construct an analog signal

More information

Filter Banks I. Prof. Dr. Gerald Schuller. Fraunhofer IDMT & Ilmenau University of Technology Ilmenau, Germany. Fraunhofer IDMT

Filter Banks I. Prof. Dr. Gerald Schuller. Fraunhofer IDMT & Ilmenau University of Technology Ilmenau, Germany. Fraunhofer IDMT Filter Banks I Prof. Dr. Gerald Schuller Fraunhofer IDMT & Ilmenau University of Technology Ilmenau, Germany 1 Structure of perceptual Audio Coders Encoder Decoder 2 Filter Banks essential element of most

More information

Robust Linear Prediction Analysis for Low Bit-Rate Speech Coding

Robust Linear Prediction Analysis for Low Bit-Rate Speech Coding Robust Linear Prediction Analysis for Low Bit-Rate Speech Coding Nanda Prasetiyo Koestoer B. Eng (Hon) (1998) School of Microelectronic Engineering Faculty of Engineering and Information Technology Griffith

More information

Signals and Systems. Lecture 13 Wednesday 6 th December 2017 DR TANIA STATHAKI

Signals and Systems. Lecture 13 Wednesday 6 th December 2017 DR TANIA STATHAKI Signals and Systems Lecture 13 Wednesday 6 th December 2017 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON Continuous time versus discrete time Continuous time

More information

Signal Processing for Speech Applications - Part 2-1. Signal Processing For Speech Applications - Part 2

Signal Processing for Speech Applications - Part 2-1. Signal Processing For Speech Applications - Part 2 Signal Processing for Speech Applications - Part 2-1 Signal Processing For Speech Applications - Part 2 May 14, 2013 Signal Processing for Speech Applications - Part 2-2 References Huang et al., Chapter

More information

Waveform Coding Algorithms: An Overview

Waveform Coding Algorithms: An Overview August 24, 2012 Waveform Coding Algorithms: An Overview RWTH Aachen University Compression Algorithms Seminar Report Summer Semester 2012 Adel Zaalouk - 300374 Aachen, Germany Contents 1 An Introduction

More information

Michael F. Toner, et. al.. "Distortion Measurement." Copyright 2000 CRC Press LLC. <

Michael F. Toner, et. al.. Distortion Measurement. Copyright 2000 CRC Press LLC. < Michael F. Toner, et. al.. "Distortion Measurement." Copyright CRC Press LLC. . Distortion Measurement Michael F. Toner Nortel Networks Gordon W. Roberts McGill University 53.1

More information

RECOMMENDATION ITU-R BS User requirements for audio coding systems for digital broadcasting

RECOMMENDATION ITU-R BS User requirements for audio coding systems for digital broadcasting Rec. ITU-R BS.1548-1 1 RECOMMENDATION ITU-R BS.1548-1 User requirements for audio coding systems for digital broadcasting (Question ITU-R 19/6) (2001-2002) The ITU Radiocommunication Assembly, considering

More information

Speech Coding using Linear Prediction

Speech Coding using Linear Prediction Speech Coding using Linear Prediction Jesper Kjær Nielsen Aalborg University and Bang & Olufsen jkn@es.aau.dk September 10, 2015 1 Background Speech is generated when air is pushed from the lungs through

More information

Communications I (ELCN 306)

Communications I (ELCN 306) Communications I (ELCN 306) c Samy S. Soliman Electronics and Electrical Communications Engineering Department Cairo University, Egypt Email: samy.soliman@cu.edu.eg Website: http://scholar.cu.edu.eg/samysoliman

More information

17. Delta Modulation

17. Delta Modulation 7. Delta Modulation Introduction So far, we have seen that the pulse-code-modulation (PCM) technique converts analogue signals to digital format for transmission. For speech signals of 3.2kHz bandwidth,

More information

Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau

Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau Multirate Signal Processing Lecture 7, Sampling Gerald Schuller, TU Ilmenau (Also see: Lecture ADSP, Slides 06) In discrete, digital signal we use the normalized frequency, T = / f s =: it is without a

More information

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking

Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic Masking The 7th International Conference on Signal Processing Applications & Technology, Boston MA, pp. 476-480, 7-10 October 1996. Encoding a Hidden Digital Signature onto an Audio Signal Using Psychoacoustic

More information

E : Lecture 8 Source-Filter Processing. E : Lecture 8 Source-Filter Processing / 21

E : Lecture 8 Source-Filter Processing. E : Lecture 8 Source-Filter Processing / 21 E85.267: Lecture 8 Source-Filter Processing E85.267: Lecture 8 Source-Filter Processing 21-4-1 1 / 21 Source-filter analysis/synthesis n f Spectral envelope Spectral envelope Analysis Source signal n 1

More information

Signals and Systems Lecture 9 Communication Systems Frequency-Division Multiplexing and Frequency Modulation (FM)

Signals and Systems Lecture 9 Communication Systems Frequency-Division Multiplexing and Frequency Modulation (FM) Signals and Systems Lecture 9 Communication Systems Frequency-Division Multiplexing and Frequency Modulation (FM) April 11, 2008 Today s Topics 1. Frequency-division multiplexing 2. Frequency modulation

More information

Adaptive Filters Application of Linear Prediction

Adaptive Filters Application of Linear Prediction Adaptive Filters Application of Linear Prediction Gerhard Schmidt Christian-Albrechts-Universität zu Kiel Faculty of Engineering Electrical Engineering and Information Technology Digital Signal Processing

More information

DIGITAL COMMUNICATION

DIGITAL COMMUNICATION DEPARTMENT OF ELECTRICAL &ELECTRONICS ENGINEERING DIGITAL COMMUNICATION Spring 00 Yrd. Doç. Dr. Burak Kelleci OUTLINE Quantization Pulse-Code Modulation THE QUANTIZATION PROCESS A continuous signal has

More information

The Scientist and Engineer's Guide to Digital Signal Processing By Steven W. Smith, Ph.D.

The Scientist and Engineer's Guide to Digital Signal Processing By Steven W. Smith, Ph.D. The Scientist and Engineer's Guide to Digital Signal Processing By Steven W. Smith, Ph.D. Home The Book by Chapters About the Book Steven W. Smith Blog Contact Book Search Download this chapter in PDF

More information

Chapter 2: Signal Representation

Chapter 2: Signal Representation Chapter 2: Signal Representation Aveek Dutta Assistant Professor Department of Electrical and Computer Engineering University at Albany Spring 2018 Images and equations adopted from: Digital Communications

More information

Lesson 8 Speech coding

Lesson 8 Speech coding Lesson 8 coding Encoding Information Transmitter Antenna Interleaving Among Frames De-Interleaving Antenna Transmission Line Decoding Transmission Line Receiver Information Lesson 8 Outline How information

More information

Accurate Delay Measurement of Coded Speech Signals with Subsample Resolution

Accurate Delay Measurement of Coded Speech Signals with Subsample Resolution PAGE 433 Accurate Delay Measurement of Coded Speech Signals with Subsample Resolution Wenliang Lu, D. Sen, and Shuai Wang School of Electrical Engineering & Telecommunications University of New South Wales,

More information

FOURIER analysis is a well-known method for nonparametric

FOURIER analysis is a well-known method for nonparametric 386 IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, VOL. 54, NO. 1, FEBRUARY 2005 Resonator-Based Nonparametric Identification of Linear Systems László Sujbert, Member, IEEE, Gábor Péceli, Fellow,

More information

Sampling and Signal Processing

Sampling and Signal Processing Sampling and Signal Processing Sampling Methods Sampling is most commonly done with two devices, the sample-and-hold (S/H) and the analog-to-digital-converter (ADC) The S/H acquires a continuous-time signal

More information

CHAPTER 4. PULSE MODULATION Part 2

CHAPTER 4. PULSE MODULATION Part 2 CHAPTER 4 PULSE MODULATION Part 2 Pulse Modulation Analog pulse modulation: Sampling, i.e., information is transmitted only at discrete time instants. e.g. PAM, PPM and PDM Digital pulse modulation: Sampling

More information

Laboratory Assignment 2 Signal Sampling, Manipulation, and Playback

Laboratory Assignment 2 Signal Sampling, Manipulation, and Playback Laboratory Assignment 2 Signal Sampling, Manipulation, and Playback PURPOSE This lab will introduce you to the laboratory equipment and the software that allows you to link your computer to the hardware.

More information

UNIT TEST I Digital Communication

UNIT TEST I Digital Communication Time: 1 Hour Class: T.E. I & II Max. Marks: 30 Q.1) (a) A compact disc (CD) records audio signals digitally by using PCM. Assume the audio signal B.W. to be 15 khz. (I) Find Nyquist rate. (II) If the Nyquist

More information

14 fasttest. Multitone Audio Analyzer. Multitone and Synchronous FFT Concepts

14 fasttest. Multitone Audio Analyzer. Multitone and Synchronous FFT Concepts Multitone Audio Analyzer The Multitone Audio Analyzer (FASTTEST.AZ2) is an FFT-based analysis program furnished with System Two for use with both analog and digital audio signals. Multitone and Synchronous

More information

COMPRESSIVE SAMPLING OF SPEECH SIGNALS. Mona Hussein Ramadan. BS, Sebha University, Submitted to the Graduate Faculty of

COMPRESSIVE SAMPLING OF SPEECH SIGNALS. Mona Hussein Ramadan. BS, Sebha University, Submitted to the Graduate Faculty of COMPRESSIVE SAMPLING OF SPEECH SIGNALS by Mona Hussein Ramadan BS, Sebha University, 25 Submitted to the Graduate Faculty of Swanson School of Engineering in partial fulfillment of the requirements for

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 14 Quiz 04 Review 14/04/07 http://www.ee.unlv.edu/~b1morris/ee482/

More information

2: Audio Basics. Audio Basics. Mark Handley

2: Audio Basics. Audio Basics. Mark Handley 2: Audio Basics Mark Handley Audio Basics Analog to Digital Conversion Sampling Quantization Aliasing effects Filtering Companding PCM encoding Digital to Analog Conversion 1 Analog Audio Sound Waves (compression

More information