UNIT I ELEMENTS OF DIGITAL COMMUNICATION SYSTEMS

Size: px
Start display at page:

Download "UNIT I ELEMENTS OF DIGITAL COMMUNICATION SYSTEMS"

Transcription

1 UNIT I ELEMENTS OF DIGITAL COMMUNICATION SYSTEMS The purpose of a Communication System is to transport an information bearing signal from a source to a user destination via a communication channel. MODEL OF A COMMUNICATION SYSTEM Fig. 1.1: Block diagram of Communication System The three basic elements of every communication systems are 1. Transmitter, 2. Receiver and 3. Channel. The Overall purpose of this system is to transfer information from one point (called Source) to another point, the user destination. The message produced by a source, normally, is not electrical. Hence an input transducer is used for converting the message to a time varying electrical quantity called message signal. Similarly, at the destination point, another transducer converts the electrical waveform to the appropriate message. The transmitter is located at one point in space, the receiver is located at some other point separate from the transmitter, and the channel is the medium that provides the electrical connection between them. The purpose of the transmitter is to transform the message signal produced by the source of information into a form suitable for transmission over the channel.

2 The received signal is normally corrupted version of the transmitted signal, which is due to channel imperfections, noise and interference from other sources.the receiver has the task of operating on the received signal so as to reconstruct a recognizable form of the original message signal and to deliver it to the user destination. Communication Systems are divided into 3 categories: 1. Analog Communication Systems are designed to transmit analog information using analog modulation methods. 2. Digital Communication Systems are designed for transmitting digital information using digital modulation schemes, and 3. Hybrid Systems that use digital modulation schemes for transmitting sampled and quantized values of an analog message signal. ELEMENTS OF DIGITAL COMMUNICATION SYSTEMS: The figure 1.2 shows the functional elements of a digital communication system. Source of Information: 1. Analog Information Sources. 2. Digital Information Sources. Analog Information Sources Microphone actuated by a speech, TV Camera scanning a scene, continuous amplitude signals. Digital Information Sources These are teletype or the numerical output of computer which consists of a sequence of discrete symbols or letters. An Analog information is transformed into a discrete information through the process of sampling and quantizing. Digital Communication System Fig. 1.2: Block Diagram of Digital Communication System

3 SOURCE ENCODER / DECODER: The Source encoder ( or Source coder) converts the input i.e. symbol sequence into a binary sequence of 0s and 1s by assigning code words to the symbols in the input sequence. For eg. :-If a source set is having hundred symbols, then the number of bits used to represent each symbol will be 7 because 2 7 =128 unique combinations are available. The important parameters of a source encoder are block size, code word lengths, average data rate and theefficiency of the coder (i.e. Actual output data rate compared to the minimum achievable rate) At the receiver, the source decoder converts the binary output of the channel decoder into a symbol sequence. The decoder for a system using fixed length code words is quite simple, but the decoder for a system using variable length code words will be very complex. Aim of the source coding is to remove the redundancy in the transmitting information, so that bandwidth required for transmission is minimized. Based on the probability of the symbol code word is assigned. Higher the probability, shorter is the codeword. Ex: Huffman coding. CHANNEL ENCODER / DECODER: Error control is accomplished by the channel coding operation that consists of systematically adding extra bits to the output of the source coder. These extra bits do not convey any information but helps the receiver to detect and / or correct some of the errors in the information bearing bits. There are two methods of channel coding: 1. Block Coding: The encoder takes a block of k information bits from the source encoder and adds r error control bits, where r is dependent on k and error control capabilities desired. 2. Convolution Coding: The information bearing message stream is encoded in a continuous fashion by continuously interleaving information bits and error control bits.

4 The Channel decoder recovers the information bearing bits from the coded binary stream. Error detection and possible correction is also performed by the channel decoder. The important parameters of coder / decoder are: Method of coding, efficiency, error control capabilities and complexity of the circuit. MODULATOR: The Modulator converts the input bit stream into an electrical waveform suitable for transmission over the communication channel. Modulator can be effectively used to minimize the effects of channel noise, tomatch the frequency spectrum of transmitted signal with channel characteristics, to provide the capability to multiplex many signals. DEMODULATOR: The extraction of the message from the information bearing waveform produced by the modulation is accomplished by the demodulator. The output of the demodulator is bit stream. The important parameter is the method of demodulation. CHANNEL: The Channel provides the electrical connection between the source and destination. The different channels are: Pair of wires, Coaxial cable, Optical fibre, Radio channel, Satellite channel or combination of any of these. The communication channels have only finite Bandwidth, non-ideal frequency response, the signal often suffers amplitude and phase distortion as it travels over the channel. Also, the signal power decreases due to the attenuation of the channel. The signal iscorrupted by unwanted, unpredictable electrical signals referred to as noise.

5 The important parameters of the channel are Signal to Noise power Ratio(SNR), usable bandwidth, amplitude and phase response and the statistical properties of noise. Certain Issues in Digital Transmission Sometimes, a comparison between two digital transmission systems is needed. There are many parameters that can be used to compare between digital transmission systems, but some of the most important parameters of a digital transmission system are: Transmission Rate (measured in bits per second): This is a measure of the number of bits that can be transmitted over the communication channel per unit time. Bandwidth Requirements (measured in Hz): This is a measure of the spectrum that the communication system requires to transmit the information at the desired transmission rate. Error probability (measured in percentages): This represents the percentage of bits that are in error relative to the overall number of bits that are transmitted by the communication system. Transmission Power (or Bit Energy) (measured in Watts (or Jules/bit)): This represents the amount of power of the transmitted signal that would be required to achieve a particular desired error probability. System Complexity (measured in cost of building or operating the system): This represents that amount of money that a manufacturer will have to spend to build the system and the amount of money that a user will have to pay to use the system. The performance of a digital transmission system is a function of the following factors: 1. Amount of energy in each digital bit (or pulse): Generally, the more energy a digital bit (or pulse) has, the better the performance that the system will have.

6 2. The distance between the transmitter and receiver: Because energy is spread or attenuated as it travels over the channel and more noise is added due to the existence of more noise sources over long channels, generally the longer the path that the digital transmitted signal has to travel, the worse the performance that the system will have. However, you do not always have control over the distance between the transmitted and receiver. 3. Amount of noise that is added to the signal: Certainly, the less the noise that is added to the transmitted signal, the better the performance of the communication system. We usually have limited control over the added noise. 4. Bandwidth of the transmission channel: By using larger bandwidth, we can either transmit at a higher transmission bit rate while keeping the same probability of bit error, or we can transmit at the same transmission bit rate but reduce the probability of bit error. Generally, the larger the bandwidth allocated to a communication system, the better the performance it will have. Advantages of Digital Communication 1. The effect of distortion, noise and interference is less in a digital communication system. This is because the disturbance must be large enough to change the pulse from one state to the other. 2. Regenerative repeaters can be used at fixed distance along the link, to identify and regenerate a pulse before it is degraded to an ambiguous state. 3. Digital circuits are more reliable and cheaper compared to analog circuits. 4. The Hardware implementation is more flexible than analog hardware because of the use of microprocessors, VLSI chips etc. 5. Signal processing functions like encryption, compression can be employed to maintain the secrecy of the information. 6. Error detecting and Error correcting codes improve the system performance by reducing the probability of error. 7. Combining digital signals using TDM is simpler than combining analog signals using FDM. The different types of signals such as data, telephone, TV can be treated as identical signals in transmission and switching in a digital communication system.

7 8. We can avoid signal jamming using spread spectrum technique. Disadvantages of Digital Communication: 1. Large System Bandwidth:- Digital transmission requires a large system bandwidth to communicate the same information in a digital format as compared to analog format. 2. System Synchronization:- Digital detection requires system synchronization whereas the analog signals generally have no such requirement. Channels for Digital Communications The modulation and coding used in a digital communication system depend on the characteristics of the channel. The two main characteristics of the channel are BANDWIDTH and POWER. In addition the other characteristics are whether the channel is linear or nonlinear, and how free the channel is free from the external interference. Five channels are considered in the digital communication, namely: telephone channels, coaxial cables, optical fibers, microwave radio, and satellite channels. Telephone channel: It is designed to provide voice grade communication. Also good for data communication over long distances. The channel has a band-pass characteristic occupying the frequency range 300Hz to 3400hz, a high SNR of about 30db, and approximately linear response. For the transmission of voice signals the channel provides flat amplitude response. But for thetransmission of data and image transmissions, since the phase delay variations are important an equalizer is used to maintain the flat amplitude response and a linear phase response over the required frequency band. Transmission rates upto16.8 kilobits per second have been achieved over the telephone lines. Coaxial Cable: The coaxial cable consists of a single wire conductor centered inside an outer conductor, which is insulated from each other by a dielectric. The main advantages of the coaxial cable are wide bandwidth and low external

8 interference. But closely spaced repeaters are required. With repeaters spaced at 1km intervals the data rates of 274 megabits per second have been achieved. Optical Fibers: An optical fiber consists of a very fine inner core made of silica glass, surrounded by a concentric layer called cladding that is also made of glass. The refractive index of the glass in the core is slightly higher than refractive index of the glass in the cladding. Hence if a ray of light is launched into an optical fiber at the right oblique acceptance angle, it is continually refracted into the core by the cladding. That means the difference between the refractive indices of the core and cladding helps guide the propagation of the ray of light inside the core of the fiber from one end to the other. Compared to coaxial cables, optical fibers are smaller in size and they offer higher transmission bandwidths and longer repeater separations. Microwave radio: A microwave radio, operating on the line-of-sight link, consists basically of a transmitter and a receiver that are equipped with antennas. The antennas are placed on towers at sufficient height to have the transmitter and receiver in line-of-sight of each other. The operating frequencies range from 1 to 30 GHz. Under normal atmospheric conditions, a microwave radio channel is very reliable and provides path for high-speed digital transmission. But during meteorological variations, a severe degradation occurs in the system performance. Satellite Channel: A Satellite channel consists of a satellite in geostationary orbit, an uplink from ground station, and a down link to another ground station. Both link operate at microwave frequencies, with uplink the uplink frequency higher than the down link frequency. In general, Satellite can be viewed as repeater in the sky. It permits communication over long distances at higher bandwidths and relatively low cost. Bandwidth:

9 Bandwidth is simply a measure of frequency range. The range of frequencies contained in a composite signal is its bandwidth. The bandwidth is normally a difference between two numbers. For example, if a composite signal contains frequencies between 1000 and 5000, its bandwidth is or If a range of 2.40 GHz to 2.48 GHz is used by a device, then the bandwidth would be 0.08 GHz (or more commonly stated as 80MHz).It is easy to see that the bandwidth we define here is closely related to the amount of data you can transmit within it - the more room in frequency space, the more data you can fit in at a given moment. The term bandwidth is often used for something we should rather call a data rate, as in my Internet connection has 1 Mbps of bandwidth, meaning it can transmit data at 1 megabit per second. Sampling A message signal may originate from a digital or analog source. If the message signal is analog in nature, then it has to be converted into digital form before it can transmitted by digital means. The process by which the continuous-time signal is converted into a discrete time signal is called Sampling. Sampling operation is performed in accordance with the sampling theorem. Sampling Theorem for Lowpass Signals Part - I If a signal x(t) does not contain any frequency component beyond W Hz, then the signal is completely described by its instantaneous uniform samples with sampling interval (or period ) of Ts< 1/(2W) sec. Part II The signal x(t) can be accurately reconstructed (recovered) from the set of uniform instantaneous samples by passing the samples sequentially through an ideal (brick-wall) lowpass filter with bandwidth B, where W B < fs W and fs = 1/(Ts).

10 As the samples are generated at equal (same) interval (Ts) of time, the process of sampling is called uniform sampling. Uniform sampling, as compared to any nonuniform sampling, is more extensively used in time-invariant systems as the theory of uniform sampling (either instantaneous or otherwise) is well developed and the techniques are easier to implement in practical systems. The concept of instantaneous sampling is more of a mathematical abstraction as no practical sampling device can actually generate truly instantaneous samples (a sampling pulse should have non-zero energy). However, this is not a deterrent in using the theory of instantaneous sampling, as a fairly close approximation of instantaneous sampling is sufficient for most practical systems. To contain our discussion on Nyquist s theorems, we will introduce some mathematical expressions. If x(t) represents a continuous-time signal, the equivalent set of instantaneous uniform samples {x(nts)} may be represented as, {x(nts)} xs(t) = Σ x(t).δ(t- nts) (1) where x(nts) = x(t) t =nts, δ(t) is a unit pulse singularity function and n is an integer Conceptually, one may think that the continuous-time signal x(t) is multiplied by an (ideal) impulse train to obtain {x(nts)} as in equation(1) can be rewritten as, xs(t) = x(t).σ δ(t- nts). (2) Now, let X(f) denote the Fourier Transform F(T) of x(t), i.e. + X ( f )= x (t ).exp( j 2π ft )dt. (3)

11 Now, from the theory of Fourier Transform, we know that the F.T of Σ δ(t- nts), the impulse train in time domain, is an impulse train in frequency domain: F{Σ δ(t- nts)} = (1/Ts).Σ δ(f- n/ts) = fs.σ δ(f- nfs) (4) If Xs(f) denotes the Fourier transform of the energy signal xs(t), we can write using Eq. (1.2.4) and the convolution property: Xs(f) = X(f)* F{Σ δ(t- nts)} = X(f)*[fs.Σ δ(f- nfs)] = fs.x(f)*σ δ(f- nfs) + = fs. X(λ). Σ δ(f nfs λ)dλ (5) = fs. Σ X(λ).δ(f- nfs-λ)dλ = fs.σ X(f- nfs). f)] [By sifting property of δ(t) and considering δ(f) as an even function, i.e. δ(f) = δ(- This equation, when interpreted appropriately, gives an intuitive proof to Nyquist s theorems as stated above and also helps to appreciate their practical implications. Let us note that while writing Eq.(5), we assumed that x(t) is an energy signal so that its Fourier transform exists. Further, the impulse train in time domain may be viewed as a periodic singularity function with almost zero (but finite) energy such that its Fourier Transform [i.e. a train of impulses in frequency domain] exists. With this setting, if we assume that x(t) has no appreciable frequency component greater than W Hz and if fs > 2W, then Eq.(1.2.5) implies that Xs(f), the Fourier Transform of the sampled signal xs(t) consists of infinite number of replicas of X(f),

12 centered at discrete frequencies n.fs, - < n < and scaled by a constant fs= 1/Ts (Fig. 1.3). Fig. 1.3 Spectra of (a) an analog signal x(t) and (b) its sampled version Fig. 1.3 indicates that the bandwidth of this instantaneously sampled wave xs(t) is infinite while the spectrum of x(t) appears in a periodic manner, centered at discrete frequency values n.fs. Now, Part I of the sampling theorem is about the condition fs > 2.W i.e. (fs W) > W and ( fs + W) < W. As seen from Fig. 1.3, when this condition is satisfied, the spectra of xs(t), centered at f = 0 and f = ± fs do not overlap and hence, the spectrum of x(t) is present in xs(t) without any distortion. This implies that xs(t), the appropriately sampled version of x(t), contains all information about x(t) and thus represents x(t). The second part of Nyquist s theorem suggests a method of recovering x(t) from its sampled version xs(t) by using an ideal lowpass filter. As indicated by dotted lines in Fig. 1.3, an ideal lowpass filter (with brick-wall type response) with a bandwidth W B < (fs W), when fed with xs(t), will allow the portion of Xs(f), centered at f = 0 and will reject all its replicas at f = n fs, for n 0. This implies that the shape of the continuous-time signal xs(t), will be retained at the output of the ideal filter.

13 Hartley Shannon Law The theory behind designing and analyzing channel codes is called Shannon s noisy channel coding theorem. It puts an upper limit on the amount of information you can send in a noisy channel using a perfect channel code. This is given by the following equation: where C is the upper bound on the capacity of the channel (bit/s), B is the bandwidth of the channel (Hz) and SNR is the Signal-to-N ise ratio (unitless). Bandwidth-S/N Tradeoff The expression of the channel capacity of the Gaussian channel makes intuitive sense: 1. As the bandwidth of the channel increases, it is possible to make faster changes in the information signal, thereby increasing the information rate. 2. As S/N increases, one can increase the information rate while stillpreventing errors due to noise. 3. For no noise, S/N tends to infinity and an infinite information rate is possibleirrespective of bandwidth. Thus we may trade off bandwidth for SNR. For example, if S/N = 7 and B = 4kHz, then the channel capacity is C = bits/s. If the SNR increases to S/N = 15 and B is decreased to 3kHz, the channel capacity remains the same. However, as B tends to 1, the channel capacity does not become infinite since, with an increase in bandwidth, the noise power also increases. If the noise power spectral density is ɳ/2, then the total noise power is N = ɳB, so the Shannon-Hartley law becomes

14 Pulse Code Modulation Introduction In the simplest model of a telephone speech communication there is a direct, dedicated, physical connection between the two participants in the conversation, and this link is held for the duration of the conversation. The analogue electrical signal produced by the telephone at either end is sent on to connection without modification. In Pulse Amplitude Modulation (PAM), the unmodified electrical signal is not sent on to the connection. Instead, short samples of the signal are taken at regular intervals, and these samples are sent on to the connection. The amplitude of each sample is identical to the signal voltage at the time when the sample was taken. Typically, 8,000 samples are taken per second, so that the interval between samples is 125 s, and the duration of each sample is approximately 4 s. Because each sample is very short (~4 s) there is a lot of time between samples (~121 s). Samples from other conversations are put into this spare time. Usually the samples from 32 separate conversations are put on to a single line. This process is called Time Division Multiplexing (TDM). Each sample is very short, and will be distorted as it travels across a communications network. In order to reconstruct the original analogue signal the only information the receiver needs to have about a sample is its amplitude, but if this is distorted then all information about the sample has been lost. To overcome this problem, the pulse is not transmitted directly, instead its amplitude is measured and converted into an 8 binary number - a sequence of 1s and 0s. At the receiver end, the receiver merely needs to detect if a 1 or a 0 has been received so that it can still recover the amplitude of a PAM pulse even if the 1s and 0s used to describe it have been distorted.

15 The process of converting the amplitude of each pulse into a stream of 1s and 0s is called Pulse Code Modulation (PCM) Note that the process of PAM and PCM (but without the use of TDM) is essentially used to store music and speech on CDs, but with a higher sample rate, more bits per sample and complex error correction mechanisms. Some terms are: Sampling Quantizing Encoding The process of measuring the amplitude of a continuous-time signal at discrete instants. It converts a continuous-time signal to a discretetime signal. Representing the sampled values of the amplitude by a finite set of levels. It converts a continuous-amplitude sample to a discreteamplitude sample. Designating each quantized level by a (binary) code. Sampling and quantizing operations transform an analogue signal to a digital signal. Use of quantizing and encoding distinguishes PCM from analogue pulse modulation methods. The quantizing and encoding operations are usually performed in the same circuit at the transmitter, which is called an Analogue to Digital Converter (ADC). At the receiver end the decoding operation converts the (8 bit) binary representation of the pulse back into an analogue voltage in a Digital to Analogue Converter (DAC) Pulse Code Modulation Pulse Code Modulation (PCM) is an extension of PAM wherein each analogue sample value is quantized into a discrete value for representation as a digital code word. Thus, as shown in Fig. 2.1 a PAM system can be converted into a PCM system by adding a suitable analogue-to-digital (A/D) converter at the source and a digital-toanalogue (D/A) converter at the destination.

16 M o d u la to r A n a lo g u e In p u t S a m p le r A to D C o n v e rte r B in a ry C o d e r P a ra lle l to S e ria l C o n v e rte r D ig ita l P u lse G e n e ra to r P C M O u tp u t D e m o d u la to r P C M In p u t S e ria l to P a ra lle l C o n v e rte r D to A C o n v e rte r L P F A n a lo g u e O u tp u t Fig. 2.1 PCM is a true digital process as compared to PAM. In PCM the speech signal is converted from analogue to digital form. PCM is standardised for telephony by the ITU-T (International Telecommunications Union - Telecoms, a branch of the UN), in a series of recommendations called the G series. For example the ITU-T recommendations for out-of-band signal rejection in PCM voice coders require that 14 db of attenuation is provided at 4 khz. Also, the ITU-T transmission quality specification for telephony terminals require that the frequency response of the handset microphone has a sharp roll-off from 3.4 khz. In quantization the levels are assigned a binary codeword. All sample values falling between two quantization levels are considered to be located at the centre of the quantization interval. In this manner the quantization process introduces a certain amount of error or distortion into the signal samples. This error known as quantization noise, is minimised by establishing a large number of small quantization intervals. Of course, as the number ofquantization intervals increase, so must the number or bits increase to uniquely identify the quantization intervals. For example, if an analogue voltage level is to be converted to a digital system with 8 discrete levels or quantization steps three bits are required. In the ITU-T version there are 256 quantization steps, 128 positive and 128 negative, requiring 8 bits. A positive level is represented by having bit 8 (MSB) at 0, and for a negative level the MSB is 1.

17 Quantization The process of quantizing a signal is the first part of converting an sequence of analog samples to a PCM code. In quantization, an analog sample with an amplitude that may take value in a specific range is converted to a digital sample with an amplitude that takes one of a specific pre defined set of quantization values. This is performed by dividing the range of possible values of the analog samples into L different levels, and assigning the center value of each level to any sample that falls in that quantization interval. The problem with this process is that it approximates the value of an analog sample with the nearest of the quantization values. So, for almost all samples, the quantized samples will differ from the original samples by a small amount. This amount is called the quantization error. To get some idea on the effect of this quantization error, quantizing audio signals results in a hissing noise similar to what you would hear when play a random signal. Assume that a signal with power Psis to be quantized using a quantizer with L = 2 n levels ranging in voltage from mp tomp as shown in the fig. 2.2 A quantization interval Corresponding quantization value m p v L = 2 n L levels n bits 0 0 Ts 2Ts 3Ts 4Ts 5Ts t m p Q uantizer Input Sam ples x Q uantizer O utput Sam ples x q Fig. 2.2 We can define the variable v to be the height of the each of the L levels of the quantizer as shown above. This gives a value of v equal to

18 2 m p v. L Therefore, for a set of quantizers with the same mp, the larger the number of levels of a quantizer, the smaller the size of each quantization interval, and for a set of quantizers with the same number of quantization intervals, the larger mp is the larger the quantization interval length to accommodate all the quantization range. Now if we look at the input output characteristics of the quantizer, it will be similar to the red line in the following figure. Note that as long as the input is within the quantization range of the quantizer, the output of the quantizer represented by the red line follows the input of the quantizer. When the input of the quantizer exceeds the range of mp tomp, the output of the quantizer starts to deviate from the input and the quantization error (difference between an input and the corresponding output sample) increases significantly. Quantizer Output x q x v/2 x q v/2 v/2 v v v v/2 v v/2 v v v v Quantizer Input x v/2 v/2 v/2 m p Fig..2.3 Now let us define the quantization error represented by the difference between the input sample and the corresponding output sample to be q, or q x x. q

19 Plotting this quantization error versus the input signal of a quantizer is seen next. Notice that the plot of the quantization error is obtained by taking the difference between the blow and red lines in the above Fig. 2.3 Quantization Error q v v v v/2 v/2 v v v v v Quantizer Input x m p Fig. 2.4 It is seen from the Fig 2.4 that the quantization error of any sample is restricted between v/2 and v/2 except when the input signal exceeds the range of quantization of mp to mp. Uniform Quantization O u tp u t -m p +m p In p u t We assume that the amplitude of the signal m(t) is confined to the range (-mp, +mp ). This range (2mp) is divided into L levels, each of step size, given by = 2 mp / L A sample amplitude value is approximated by the midpoint of the interval in which it lies. The input/output characteristics of a uniform quantizer is shown in Fig. 2.5 Fig. 2.5

20 Companding -Companding is the process of compressing and then expanding -High amplitude analog signals are compressed prior to txn. and then expanded in the receiver -Higher amplitude analog signals are compressed and Dynamic range is improved -Early PCM systems used analog companding, where as modern systems use digital companding. Fig 2.6 Basic companding process Analog companding

21 2.7 PCM system with analog companding --In the transmitter, the dynamic range of the analog signal is compressed, and then converted o a linear PCM code. --In the receiver, the PCM code is converted to a PAM signal, filtered, and then expanded back to its original dynamic range. -- There are two methods of analog companding currently being used that closely approximate a logarithmic function and are often called log-pcm codes. The two methods are 1) -law and 2) A-law -law companding

22 V out V m a x V ln 1 ln 1 in V m a x Fig. 2.8m-law companding where Vmax = maximum uncompressed analog input amplitude(volts) Vin = amplitude of the input signal at a particular instant of time (volts) = parameter used tio define the amount of compression (unitless) Vout = compressed output amplitude (volts) A-law companding --A-law is superior to -law in terms of small-signal quality --The compression characteristic is given by where y = Vout Vmax y A x, 1 log A 1 log( A x ) 1 log A 1 0 x A 1, x 1 A x = Vin /

23 Digital Companding: --With digital companding, the analog signal is first sampled and converted to a linear PCM code, and then the linear code is digitally compressed. -- In the receiver, the compressed PCM code is expanded and then decoded back to analog. -- The most recent digitally compressed PCM systems use a 12- bit linear PCM code and an 8-bit compressed PCM code. Digital compression error --To calculate the percentage error introduced by digital compression %error=12-bit encoded voltage - 12-bit decoded voltage X bit decoded voltage 1.3 PCM Encoding Process (HDB3) The output from the analogue to digital converter (ADC) has n parallel bits. In the case of telephony n = 8. The most significant bit is the signed bit. If the measured sample is positive then the signed bit is 0. If the measured sample is negative then the signed bit is 1. The remaining 7 bits are used to code the sample value. The ITU- T define a look up table which allocates a particular binary code to each quantified A-law value. The line coding which is used assigns opposite polarities to successive 1 s. This eliminates any DC voltage on the line, and reduces the inter symbol interference if adjacent bits are 1. If there is silence on the PCM channel then the measured samples will be 0 Vrms and the output of the DAC will be A stream of all zeros is not desirable on an active channel because all zeros could also be a fault condition and

24 it is difficult to recover the clock signal from the incoming signal. The coding system HDB3 is used and was developed to eliminate all zeros, and to assign opposite polarities to successive 1 s. This is a bipolar signalling technique (i.e. relies on the transmission of both positive and negative pulses). In AMI positive and negative pulses (of equal amplitude) are used for alternative symbols 1. No pulse is used for symbol 0. In either case the pulse returns to 0 before the end of the bit interval. This eliminates any DC on the line. HDB3 encoding rules follow those for AMI, except that a sequence of four consecutive 0's are encoded using a special "violation" bit. The 4 th 0 bit is given the same polarity as the last 1-bit which was sent using the AMI encoding rule. This prevents long runs of 0's in the data stream which may otherwise prevent a receiver from tracking the centre of each bit. By introducing violations, extra "edges" are introduced, enabling a Digital PLL to reliably reconstruct the clock signal at the receiver. The HDB3 is transparent to the sequence of bits being transmitted (i.e. whatever data is sent, the Digital PLL can reconstruct the data and extract the bits at the receiver). To prevent a DC being introduced by excessive runs of zeros any run of more than four zeros encodes as B00V. The value of B is assigned + or - alternately throughout the bit stream. Example = B BBBBBBB = B 0 B 0 B 0 B = B V 0 0 B = = B V B B 0

25 PCM Timing and Synchronisation The PCM receiver must be able to identify the start and finish of each full sampling sequence and to identify each bit position. The sampling clock needs to be either sent to, or regenerated at, the receiving side to determine when each full sequence of sampling begins and ends. The data clock is also needed to determine exactly when to read each bit of information s d a ta c lo c k 6 4 k b it/s A PCM channel is sampled at 8,000 Hz or fra m e c lo c k s once every 125 s. If there is one channel or 30 TDM channels the sampling period is fixed at 125 s and this period is known as a frame. B1 B2 B3 B4 B5 B6 B7 B8 B1 Therefore the frame clock must have a period ? of 125 s. The rising edge of the frame clock Fig. 2.9 informs the receiver that the next bit will be Bit 1 of a new sample. The falling edge of the data clock informs the receiver that it must read the data bit. When the bit stream is transmitted along a line the pulses become distorted and the rise and fall times become significant. Ideally, a 1 will be high for s. In practice the pulse may only be above the high threshold for a few s so it is very important that the bit is read within a certain time limit of the clock pulse. The simplest way to synchronise a PCM sender to a PCM receiver is to send the clock signals on different circuits to the data This would be done in a self-contained system such as private branch exchange (PBX). Telephony is full duplex so that there is a coder and a decoder at each port, but each would use the same clock. To minimise the number of circuits it is possible to use a line-coding scheme which allows the receiver to extract the clocks from the PCM signal. In this case the receiver will have free running clocks that lock (using a PLL) to the phase and frequency of the transitions in the data stream. The line-coding scheme ensures that there is a transition for every data bit. Differential pulse coding schemes PCM transmits the absolute value of the signal for each frame. Instead we can transmit information about the difference between each sample. The two main differential coding schemes are: Delta Modulation Differential PCM and Adaptive Differential PCM (ADPCM)

26 Delta Modulation Delta modulation converts an analogue signal, normally voice, into a digital signal. 0 The analogue signal is sampled as in the granular noise PCM process. Then the sample is compared with the previous sample. The result of the comparison is quantified 1 0 using a one bit coder. If the sample is greater than the previous sample a 1 is 0 generated. Otherwise a 0 is generated. The advantage of delta modulation over PCM is its simplicity and lower cost. But the noise performance is not as Fig good as PCM. To reconstruct the original from the quantization, if a 1 is received the signal is increased by a step of size q, if a 0 is received the output is reduced by the same size step. Slope overload occurs when the encoded waveform is more than a step size away from the input signal. This condition happens when the rate of change of the input exceeds the maximum change that can be generated by the output. Overload will occur if: dx(t)/dt q /T = q * fs where: x(t) = input signal, q = step size, T = period between samples, fs = sampling frequency Assume that the input signal has maximum amplitude A and maximum frequency F. The most rapidly changing input is provided by x(t) = A * sin (2 * * F * t). For this dx(t)/dt = 2 * * F * A * sin (2 * * F * t). This slope has a maximum value of 2 * * F * A Overload occurs if 2 * * F * A > q * fs To prevent overload we require q * fs> 2 * * F * A Example A = 2 V, F = 3.4 khz, and the signal is sampled 1,000,000 times per second, requires q > 2 * 3.14 * 3,400 * 2 /1,000,000 V > 42.7 mv

27 Granular noise occurs if the slope changes more slowly than the step size. The reconstructed signal oscillates by 1 step size in every sample. It can be reduced by decreasing the step size. This requires that the sample rate be increased. Delta Modulation requires a sampling rate much higher than twice the bandwidth. It requires oversampling in order to obtain an accurate prediction of the next input, since each encoded sample contains a relatively small amount of information. Delta Modulation requires higher sampling rates than PCM. Differential PCM (DPCM) and ADPCM A n a lo g u e In p u t B a n d L im itin g F ilte r D ifferentiator + - Q u a n tis e r E n o d e r A D C E n c o d e d D iffe re n c e S a m p le s A ccum ulator D A C Fig DPCM is also designed to take advantage of the redundancies in a typical speech waveform. In DPCM the differences between samples are quantized with fewer bits that would be used for quantizing an individual amplitude sample. The sampling rate is often the same as for a comparable PCM system, unlike Delta Modulation. Adaptive Differential Pulse Code Modulation ADPCM is standardised by ITU-T recommendations G.721 and G.726. The method uses 32,000 bits/s per voice channel, as compared to standard PCM s 64,000 bits/s. Four bits are used to describe each sample, which represents the difference between two adjacent samples. Sampling is 8,000 times a second. It makes it possible to reduce the bit flow by half while maintaining an acceptable quality. While the use of ADPCM (rather than PCM) is imperceptible to humans, it can significantly reduce the throughput of high speed modems and fax transmissions. The principle of ADPCM is to use our knowledge of the signal in the past time to predict the signal one sample period later, in the future. The predicted signal is then compared with the actual signal. The difference between these is the signal which is sent to line - it is the error in the prediction. However this is not done by making

28 comparisons on the incoming audio signal - the comparisons are done after PCM coding. To implement ADPCM the original (audio) signal is sampled as for PCM to produce a code word. This code word is manipulated to produce the predicted code word for the next sample. The new predicted code word is compared with the code word of the second sample. The result of this comparison is sent to line. Therefore we need to perform PCM before ADPCM. The ADPCM word represents the prediction error of the signal, and has no significance itself. Instead the decoder must be able to predict the voltage of the recovered signal from the previous samples received, and then determine the actual value of the recovered signal from this prediction and the error signal, and then to reconstruct the original waveform. ADPCM is sometimes used by telecom operators to fit two speech channels onto a single 64 kbit/s link. This was very common for transatlantic phone calls via satellite up until a few years ago. Now, nearly all calls use fibre optic channels at 64 kbit/s. 6.9 Delta Modulation Delta modulation, like DPCM is a predictive waveform coding technique and can be considered as a special case of DPCM. It uses the simplest possible quantizer, namely a two level (one bit) quantizer. The price paid for achieving the simplicity of the quantizer is the increased sampling rate (much higher than the Nyquist rate) and the possibility of slope-overload distortion in the waveform reconstruction, as explained in greater detail later on in this section. In DM, the analog signal is highly over-sampled in order to increase the adjacent sample correlation. The implication of this is that there is very little change in two

29 adjacent samples, thereby enabling us to use a simple one bit quantizer, which like in DPCM, acts on the difference (prediction error) signals. In its original form, the DM coder approximates an input time function by a series of linear segments of constant slope. Such a coder is therefore referred to as a Linear (or non-adaptive) Delta Modulator (LDM). Subsequent developments have resulted in delta modulators where the slope of the approximating function is a variable. Such coders are generally classified under Adaptive Delta Modulation (ADM) schemes. We use DM to indicate either of the linear or adaptive variety. Fig. 2.12: Waveforms illustrative of LDM operation Deltamodulation principleofoperation Deltamodulationwasintroducedinthe1940sasasimplifiedformofpulsecodemodulatio n(pcm),whichrequiredadifficult-to-implementanalog-to-digital(a/d)converter. Theoutputofadeltamodulatorisabitstreamofsamples,atarelatively highrate(eg,100kbit/sor

30 moreforaspeechbandwidthof4 khz)thevalueofeachbitbeing determinedaccordingas towhethertheinputmessagesampleamplitudehasincreasedordecreasedrelativetothepr evioussample.itisan exampleofdifferentialpulsecodemodulation(dpcm). Blockdiagram Theoperationofadeltamodulatoristoperiodicallysampletheinputmessage,tomakeac omparisonofthecurrentsamplewiththatprecedingit,andtooutputasinglebit which indicatesthesignofthe differencebetweenthe twosamples.thisinprinciple would requirea sample-and-hold type circuit. DeJager(1952)hitonanideafordispensingwiththeneedforasampleandholdcircuit.Her easonedthatifthesystemwasproducingthedesiredoutputthenthisoutputcouldbesentb acktotheinputandthetwoanalogsignalscomparedinacomparator.theoutput isa delayedversionoftheinput,andsothecomparison isineffectthatofthecurrentbitwiththepreviousbit,asrequiredbythedeltamodulationpri nciple. Figure2.13illustratesthebasicsysteminblockdiagramform,andthiswillbethemodulat oryouwill bemodelling. Thesystemisintheformofafeedbackloop.Thismeansthatitsoperationisnotn ecessaril

31 yobvious,anditsanalysisnontrivial.butyoucanbuildit,andconfirmthatitdoesbehaveinthe manner adelta modulatorshould. Thesystemisacontinuoustimetodiscretetimeconverter.Infact,itisaformofanalogtodi gitalconverter,andis thestarting pointfrom which more sophisticateddeltamodulatorscanbe developed. Thesamplerblockisclocked.Theoutputfromthesamplerisabipolarsignal,intheblockd iagrambeingeither Vvolts.Thisisthedeltamodulatedsignal,thewaveformof whichisshowninfigure 2.Itisfedback,inafeedbackloop,viaanintegrator,toasummer. Theintegratoroutputisasawtooth-likewaveform,alsoillustratedinFigure shownoverlaid uponthemessage,ofwhich itisanapproximation Itis Figure 2.15:integrator output superimposed on the messagewith the delta modulated signal below Thesawtoothwaveformissubtractedfromthemessage,alsoconnectedtothesummer, andthe difference-anerror signal-isthe signalappearingatthe summeroutput. Anamplifierisshowninthefeedbackloop..Thiscontrolstheloopgain.Inpracticeitmaybe aseparateamplifier,partoftheintegrator,orwithinthesummer.itisusedtocontrolthesize ofthe teeth ofthesawtoothwaveform,inconjunctionwiththeintegratortimeconstant. WhenanalysingtheblockdiagramofFigure 2.13itisconvenienttothinkofthesummerhavingunitygainbetweenbothinputsandtheou tput.themessagecomes in at afixedamplitude.thesignalfromtheintegrator,whichisasawtoothapproximationtothe message,isadjustedwiththeamplifiertomatchitasclosely aspossible.

32 stepsizecalculation InthedeltamodulatorofFigure2.13theoutputoftheintegratorisasawtoothlikeapproximationtotheinputmessage.Theteethofthesaw mustbeabletorise(orfall)fastenoughtofollowthemessage.thustheintegratortimecons tantisanimportantparameter. Foragivensampling(clock)ratethestepslope(volt/s)determinesthesize(volts)oftheste p withinthesamplinginterval. Supposetheamplitudeof therectangularwavefromthesampleris±v volt.forachangeofinputsampleto theintegratorfrom(say)negativeto positive,thechangeofintegrator output will be, afteraclockperiodt: wherekisthegainoftheamplifier precedingthe integrator(asinfigure2.13). AnswerTutorialQuestions1and 2beforeattemptingtheexperiment.Youcanlatercheckyouranswerbymeasurement. slopeoverloadandgranularity ThebinarywaveformillustratedinFigure2.15isthesignaltransmitted.Thisisthedeltamo dulatedsignal. Theintegralofthebinarywaveformisthesawtoothapproximationto themessage. IntheexperimententitledDeltademodulation(inthisVolume)youwillseethatthissawto othwave istheprimaryoutputfromthedemodulatoratthereceiver. Lowpassfilteringofthesawtooth(fromthedemodulator)givesabetterapproximationtot hemessage.buttherewillbeaccompanyingnoiseanddistortion,productsoftheapproxi mationprocessatthemodulator. Theunwantedproductsofthemodulationprocess,observedatthereceiver,areoftwo kinds.thesearedue to slopeoverload, and granularity. slopeoverload Thisoccurswhenthesawtoothapproximationcannotkeepupwiththerate-ofchangeoftheinput signalinthe regionsofgreatestslope. Thestepsizeisreasonableforthosesectionsofthesampledwaveformofsmallslope,butt heapproximationispoorelsewhere.thisis slopeoverload,duetotoosmalla step. Slopeoverloadisillustrated infigure2.16.

33 slo p e o v e rlo a d Figure2.16:slopeoverload Toreducethepossibilityofslopeoverloadthestepsizecanbeincreased(forthesamesamp ling rate).this isillustratedin Figure 2.17.Thesawtoothisbetterabletomatchthemessageinthe regionsofsteep slope. Figure2.17:increasedstepsize to reduce slope overload An alternativemethodofslopeoverloadreductionis toincreasethesamplingrate.thisisillustratedinfigure 2.18,wheretheratehasbeenincreasedbyafactorof2.4times, but thestep isthe same size asinfigure2.15.

34 tim e Figure2.18:increasedsampling rate to reduce slope overload 1.4 Granularnoise ReferbacktoFigure 2.16.Thesawtoothfollowsthemessagebeingsampledquitewellintheregionsofsmallslo pe.toreducetheslopeoverloadthestepsizeisincreased,andnow(figure 2.17)thematchovertheregionsofsmallslopehasbeendegraded. Thedegradationshowsup,atthedemodulator,asincreasedquantizingnoise,or granulari ty. 1.5 noiseanddistortionminimization Thereisaconflictbetweentherequirementsforminimizationofslopeoverloadandthegra nularnoise.theonerequiresanincreasedstepsize,theotherareducedstepsize.you shouldrefertoyourtextbook formorediscussion ofwaysandmeansofreachingacompromise.youwillmeetanexampleintheexperimente ntitledadaptivedeltamodulation(inthisvolume).anoptimumstepcanbedeterminedby minimizingthequantizingerroratthesummer output, or thedistortionatthe demodulatoroutput. Adaptive Delata Modulation

35 The Operation Theory of ADM Modulation From previous chapter, we know that the disadvantage of delta modulation is when the input audio signal frequency exceeded the limitation of delta modulator, i.e. Then this situation will produce the occurrence of slope overload and cause signal distortion. However, the adaptive delta modulation (ADM) is the modification of delta modulation to improve the disadvantage of the occurrence of slope overload. Figure 2.20 is the block diagram of ADM modulator. In figure 2.20, we can see that the delta modulator is comprised by comparator, sampler and integrator, then the slope controller and the level detect algorithm comprise a quantization level adjuster, which can control the gain of the integrator in the delta modulator. ADM modulator is the modification of delta modulator, therefore, due to the delta modulator has the problem of slope overload at low and high frequencies. The reason is the magnitude of the Δ(t) of delta modulator is fixed, i.e. the increment of Δ or -Δ is unable to follow the variation of the slope of the input signal. When the variation of the slope of the input signal is large, the magnitude of Δ(t) still can increase by following the variation, then this situation will not occur the problem of slope overload. On the other hand, there is another technique, which is known as continuous variable slope delta (CVSD) modulation. This technique is commonly used in Bluetooth application. CVSD modulator is also the modification of delta modulator, use to improve the occurrence of slope overload. The different between the CVSD and ADM modulators are the quantization level adjuster A. ADM modulator is discrete values and the quantization level adjuster of CVSD modulator is continuous. Simply, the quantization value of ADM modulator is the variation of digital, such as the quantization values of +1, +2, +3, -2, -3 and so on. As for CVSD modulator, the quantization value is the variation of analog, such as the quantization values of +1, +1.1, +1.2, -1.5, -0.3, -0.9 and so on.

36 Fig The Operation Theory of ADM Modulation UNIT - II Digital modulation techniques Modulation is defined as the process by which some characteristics of a carrier is varied in accordance with a modulating wave. In digital communications, the modulating wave consists of binary data or an M-ary encoded version of it and the carrier is sinusoidal wave. Different Shift keying methods that are used in digital modulation techniques are Amplitude shift keying [ASK] Frequency shift keying [FSK] Phase shift keying [PSK]

37 Fig 3.1 Different modulations 1. ASK[Amplitude Shift Keying]: In a binary ASK system symbol 1 and 0 are transmitted as S (t) 2E b Cos2 f t for symbol 1 T b 1 1 S 2 (t) 0 for symbol 0 2. FSK[Frequency Shift Keying]: In a binary FSK system symbol 1 and 0 are transmitted as S (t) 2E b Cos2 f t for symbol T b S 2 (t) 2E b Cos2 f 2 t for symbol 0 T b 3. PSK[Phase Shift Keying]: In a binary PSK system the pair of signals S 1 (t) and S 2 (t) are used to represent binary symbol 1 and 0 respectively. S 1 (t) 2E b Cos2 fc t for Symbol 1 T b S 2 (t) 2E b Cos(2 fc t ) T b 2E b Cos2 fc t for Symbol 0 T b Hierarchy of digital modulation technique

38 Digital Modulation Technique Coherent Non - Coherent Binary M - ary Hybrid Binary M - ary (m) = 2 (m) = 2 * ASK M-ary ASK M-ary APK * ASK M-ary ASK * FSK M-ary FSK M-ary QAM * FSK M-ary FSK * PSK M-ary PSK * DPSK M-ary DPSK (QPSK)

39 Coherent Binary PSK: Binary Data Sequence Non Return to Zero Level Encoder Product Modulator Binary PSK Signal (t) 2 Cos2 f t 1 c T b Fig. 3.2 Block diagram of BPSK transmitter x(t) T b dt 0 x 1 Choose 1 if x 1 >0 Decision Device Choose 0 if x 1 <0 Correlator 1 (t) Threshold λ = 0 Fig 3.3 Coherent binary PSK receiver In a Coherent binary PSK system the pair of signals S 1 (t) and S 2 (t) are used to represent binary symbol 1 and 0 respectively. S 1 (t) 2E b Cos2 fc t for Symbol 1 T b by S 2 (t) 2E b Cos(2 fc t ) T b 2E b Cos2 fc t for Symbol 0 T b E E b0 b1 Where E b = Average energy transmitted per bit Eb 2 In the case of PSK, there is only one basic function of Unit energy which is given

40 (t) 2 Cos2 f t 0 t T 1 T b c Therefore the transmitted signals are given by Symbol S (t) E (t) 0 t T for 1 S 1 2 b 1 (t) E (t) 0 t T for Symbol 0 b 1 A Coherent BPSK is characterized by having a signal space that is one dimensional (N=1) with two message points (M=2) B B b Tb S 11 S 1 (t) 1 (t) dt E b 0 Tb S 21 S 2 (t) 1 (t) dt E b 0 The message point corresponding to S 1 (t) is located at S 11 E b and S 2 (t) is located at S 21 E b. To generate a binary PSK signal we have to represent the input binary sequence in polar form with symbol 1 and 0 represented by constant amplitude levels of Eb & Eb level encoder. respectively. This signal transmission encoding is performed by a NRZ The resulting binary wave [in polar form] and a sinusoidal carrier 1 (t) [whose frequency fc n c ] are applied to a product modulator. The desired BPSK wave Tb is obtained at the modulator output. To detect the original binary sequence of 1 s and 0 s we apply the noisy PSK signal x(t) to a Correlator, which is also supplied with a locally generated coherent reference signal 1 (t) as shown in fig (b). The correlator output x 1 is compared with a threshold of zero volt. If x1 > 0, the receiver decides in favour of symbol 1. If x1 < 0, the receiver decides in favour of symbol 0.

41 Coherent Binary FSK In a binary FSK system symbol 1 and 0 are transmitted as S (t) 2E b Cos2 f t for symbol 1 1 T b S 2 (t) 2E b Cos2 f 2 t for symbol 0 T b 1 The basic functions are given by (t) 2 Cos2 ft And 1 T b 1 (t) 2 Cos2 ft for 0 t T And Zero Otherwise 2 T b 2 b Therefore FSK is characterized by two dimensional signal space with two message points i.e. N=2 and m=2. The two message points are defined by the signal vector Generation and Detection:- (a)

42 (b) Fig. 3.4: (a) FSK transmitter ( b) FSK Receiver A binary FSK Transmitter is as shown in fig. (a). The incoming binary data sequence is applied to on-off level encoder. The output of encoder is E b volts for symbol 1 and 0 volts for symbol 0. When we have symbol 1 the upper channel is switched on with oscillator frequency f 1, for symbol 0, because of inverter the lower channel is switched on with oscillator frequency f 2. These two frequencies are combined using an adder circuit and then transmitted. The transmitted signal is nothing but required BFSK signal. The detector consists of two correlators. The incoming noisy BFSK signal x(t) is common to both correlator. The Coherent reference signal 1 (t) and 2 (t) are supplied to upper and lower correlators respectively. The correlator outputs are then subtracted one from the other and resulting a random vector l (l=x 1 - x 2 ). The output l is compared with threshold of zero volts. If l > 0, the receiver decides in favour of symbol 1. l < 0, the receiver decides in favour of symbol0

43 BINARY ASK SYSTEM:- Binary ON-OFF Product Binary ASK Level Modulator Data Sequence Signal Encoder Fig. 3.5 BASK transmitter (t) 2 Cos2 f t 1 e Tb Tb x(t) Decision X dt Device 0 If x > λ choose symbol 1 1 (t) Threshold λ Fig. 3.6 Coherent binary ASK demodulator In Coherent binary ASK system the basic function is given by (t) 1 2 Cos2 f e t 0 t T b T b The transmitted signals S 1 (t) and S 2 (t) are given by S 1 (t) E b 1 (t) for Symbol 1 S 2 (t) 0 for Symbol 0 If x < λ choose symbol 0 The BASK system has one dimensional signal space with two messages (N=1, M=2)

44 Region E2 Region E1 Message Point 2 Eb 0 E b Message 2 Point 1 Fig. 3.7 Signal Space representation of BASK signal 1 (t) In transmitter the binary data sequence is given to an on-off encoder. Which gives an output E b volts for symbol 1 and 0 volt for symbol 0. The resulting binary wave [in unipolar form] and sinusoidal carrier 1 (t) are applied to a product modulator. The desired BASK wave is obtained at the modulator output. In demodulator, the received noisy BASK signal x(t) is apply to correlator with coherent reference signal 1 (t) as shown in fig. (b). The correlator output x is compared with threshold λ. If x > λ the receiver decides in favour of symbol 1. If x < λ the receiver decides in favour of symbol 0. Incoherent detection: Fig. 3.8 : Envelope detector for OOK BASK

45 Incoherent detection as used in analog communication does not require carrier for reconstruction. The simplest form of incoherent detector is the envelope detector as shown in Fig The output of envelope detector is the baseband signal. Once the baseband signal is recovered, its samples are taken at regular intervals and compared with threshold. If Z(t) is greater than threshold ( ) a decision will be made in favour of symbol 1 If Z(t) the sampled value is less than threshold ( ) a decision will be made in favour of symbol 0. Non- Coherenent FSK Demodulation:- Fig. 3.9 : Incoherent detection of FSK Fig. 3.9 shows the block diagram of incoherent type FSK demodulator. The detector consists of two band pass filters one tuned to each of the two frequencies used to communicate 0 s and 1 s., The output of filter is envelope detected and then baseband detected using an integrate and dump operation. The detector is simply evaluating which of two possible sinusoids is stronger at the receiver. If we take the difference of the outputs of the two envelope detectors the result is bipolar baseband. The resulting envelope detector outputs are sampled at t = kt b and their values are compared with the threshold and a decision will be made infavour of symbol 1 or 0. Differential Phase Shift Keying:- [DPSK]

46 (a) DPSK Transmitter (b) DPSK Receiver Fig DPSK A DPSK system may be viewed as the non coherent version of the PSK. It eliminates the need for coherent reference signal at the receiver by combining two basic operations at the transmitter (1) Differential encoding of the input binary wave and (2) Phase shift keying Hence the name differential phase shift keying [DPSK]. To send symbol 0 we phase advance the current signal waveform by and to send symbol 1 we leave the phase of the current signal waveform unchanged. The differential encoding process at the transmitter input starts with an arbitrary first but, securing as reference and thereafter the differentially encoded sequence{d k } is generated by using the logical equation. d k d k 1 b k d k 1 bk Where b k is the input binary digit at time kt b and d k-1 is the previous value of the

47 differentially encoded digit. Table illustrate the logical operation involved in the generation of DPSK signal. Input Binary Sequence {b K } Differentially Encoded sequence {d K } Transmitted Phase 0 0 Π 0 0 Π Received Sequence (Demodulated Sequence) A DPSK demodulator is as shown in fig(b). The received signal is first passed through a BPF centered at carrier frequency f c to limit noise power. The filter output and its delay version are applied to correlator the resulting output of correlator is proportional to the cosine of the difference between the carrier phase angles in the two correlator inputs. The correlator output is finally compared with threshold of 0 volts. If correlator output is +ve -- A decision is made in favour of symbol 1 If correlator output is -ve --- A decision is made in favour of symbol 0

48 COHERENT QUADRIPHASE SHIFT KEYING Fig. 3.11(a) QPSK Transmitter Fig. 3.11(b) QPSK Receiver In case of QPSK the carrier is given by s (t) 2E Cos[2 f t (2i 1) / 4] 0 t T i 1 to 4 i T c s (t) 2E Cos[(2i 1) / 4]cos( 2 f t) 2E sin[(2i 1) / 4]sin( 2 f t) 0 t T i 1 to 4 i T c T c

49 Fig. 3.11(c) QPSK Waveform In QPSK system the information carried by the transmitted signal is contained in the phase. The transmitted signals are given by

50 Where the carrier frequency f C n C for some fixed integer nc 7 E = the transmitted signal energy per symbol. T = Symbol duration. The basic functions 1 (t) and 2 (t) are given by 1 (t) 2 T b 2 (t) 2 T b cos 2 f t c sin 2 f t c 0 t T 0 t T There are four message points and the associated signal vectors are defined by E cos 2i 1 Si 4 i 1,2,3,4 E sin 2i 1 4 The table shows the elements of signal vectors, namely S i1 & S i2 Table:-

51 Input dibit Phase of Coordinates of message QPSK points signal(radians) Si1 Si2 10 E E E E E E E E Therefore a QPSK signal is characterized by having a two dimensional signal constellation(i.e.n=2)and four message points(i.e. M=4) as illustrated in fig(d).fig 3.11(d) Signal-space diagram of coherent QPSK system. Unit III Base Band Transmission and Optimal Reception of Digital Signal BASEBAND:

52 Pulse Shaping for Optimum Transmissions

53

54 Receiving Filter: Correlative receiver Base Band Reception Techniques

55 For an AWGN channel and for the case when the transmitted signals are equally likely, the optimum receiver consists of two subsystems 1) Receiver consists of a bank of M product-integrator or correlators Φ 1 (t),φ 2 (t).φ M (t) orthonormal function 2) The bank of correlator operate on the received signal x(t) to produce observation vector x Implemented in the form of maximum likelihood detector that operates on observationvector x to produce an estimate of the transmitted symbol m i i = 1 to M, in a way that would minimize the average probability of symbol error. The N elements of the observation vector x are first multiplied by the corresponding N elements of each of the M signal vectors s 1, s 2 s M, and the resulting products are successively summed in accumulator to form the corresponding set of

56 Inner products {(x, s k )} k= 1, 2..M. The inner products are corrected for the fact that the transmitted signal energies may be unequal. Finally, the largest in the resulting set of numbers is selected and a corresponding decision on the transmitted message made. The optimum receiver is commonly referred as a correlation receiver MATCHED FILTER Science each of the orthonormal basic functions are Φ 1 (t),φ 2 (t).φ M (t) is assumed to be zero outside the interval 0<t<T. we can design a linear filter with impulse response h j (t), with the received signal x(t) the fitter output is given by the convolution integral y j (t) = x j Where x j is the j th correlator output produced by the received signal x(t). A filter whose impulse response is time-reversed and delayed version of the input signal is said to be matched to x j (t)correspondingly, the optimum receiver based on this isreferred as the matched filter receiver. For a matched filter operating in real time to be physically realizable, it must be causal. MATCHED FILTER

57 Φ(t) = input signal h(t) = impulse response W(t) =white noise The impulse response of the matched filter is time-reversed and delayed version of the input signal MATCHED FILTER PROPERTIES PROPERTY 1 The spectrum of the output signal of a matched filter with the matched signal as input is, except for a time delay factor, proportional to the energy spectral density of the input signal. PROPERTY 2 The output signal of a Matched Filter is proportional to a shifted version of the autocorrelation function of the input signal to which the filter is matched. PROPERTY 3 The output Signal to Noise Ratio of a Matched filter depends only on the ratio of the signal energy to the power spectral density of the white noise at the filter input. PROPERTY 4 The Matched Filtering operation may be separated into two matching

58 conditions; namely spectral phase matching that produces the desired output peak at time T, and the spectral amplitude matching that gives this peak value its optimum signal to noise density ratio. EYE PATTERN The quality of digital transmission systems are evaluated using the bit error rate. Degradation of quality occurs in each process modulation, transmission, and detection. The eye pattern is experimental method that contains all the information concerning the degradation of quality. Therefore, careful analysis of the eye pattern is important in analyzing the degradation mechanism. Eye patterns can be observed using an oscilloscope. The received wave is applied to the vertical deflection plates of an oscilloscope and the sawtooth wave at a rate equal to transmitted symbol rate is applied to the horizontal deflection plates, resulting display is eye pattern as it resembles human eye. The interior region of eye pattern is called eye opening We get superposition of successive symbol intervals to produce eye pattern as shown below.

59 The width of the eye opening defines the time interval over which the received wave can be sampled without error from ISI The optimum sampling time corresponds to the maximum eye opening The height of the eye opening at a specified sampling time is a measure of the margin over channel noise. The sensitivity of the system to timing error is determined by the rate of closure of the eye as the sampling time is varied. Any non linear transmission distortion would reveal itself in an asymmetric or squinted eye. When the effected of ISI is excessive, traces from the upper portion of the eye pattern cross traces from lower portion with the result that the eye is completely closed. INFORMATION THEORY Information and Entropy

60 Although it is in principle a very old concept, entropy is generally credited to Shannon because it is the fundamental measure in information theory. Entropy is often defined as an expectation: where 0 log(0) = 0. The base of the logarithm is generally 2. When this is the case, the units of entropy are bits. Entropy captures the amount of randomness or uncertainty in a variable. This, in turn,is aeasure of the average length of a message that would have to be sent to describe a sample. Recall our fair coin from 1. It s entropy is: 0.5log log0.5 = 1; that is, thereis one bit of information in the random variable.this means that on average we need to sendone bit per trial to describe a sample. This should fit your intuitions: if I flip a coin 100 times,i ll need 100 numbers to describe those flips, if order matters. By contrast, our two-headedcoin has entropy. Even if I flip this coin 100 times, it doesn t matterbecause the outcome is always heads. I don t need to send any message to describe a sample. This should fit your intuitions: if I flip a coin 100 times, I ll need 100 numbers to describe those flips, if order matters. By contrast, our two-headed coin has entropy 1log1+ 0log0 = 0. Even if I flip this coin 100 times, it doesn t matter because the outcome is always heads. I don t need to send any message to describe a sample. There are other possibilities besides being completely random and completely deter-mined. Imagine a weighted coin, such that heads occurr 75% of the time. The entropy would be: 0.75log log0.25 = After 100 trials, I d only need a message of about 82 bits on average to describe the sample. Shannon showed that there exists a coder thatcan construct messages of length H(X)+1, nearly matching this ideal rate. Just as with probabilities, we can compute joint and conditional entropies. Joint entropy is the randomness contained in two variables, while conditional entropy is a measure of the randomness of one variable given knowledge of another. Joint entropy is defined as:

61 while the conditional entropy is: There are several interesting facts that follow from these definitions. For example, two random variables, X and Y, are considered independent if and only if H Y X = H Y or H X Y = H X +H Y It is also the case that H Y X H Y (knowing more information can never increase our uncertainty). Similarly, H X Y H X +H Y It is alsothe case that H X Y =H Y X +H X =H X Y +H Y These relations hold in thegeneral case of more than two variables. There are several facts about discrete entropy, H(), that do not hold for continuous ordifferential entropy, h(). The most important is that while H X 0 h() can actually be negative. Worse, even a distribution with an entropy of can still have uncertainty. Luckily, for us, even though differential entropy cannot provide us with an absolute measure of randomness, it is still that case that if h X h Y then X has more randomness than Y. Mutual Information Although conditional entropy can tell us when two variables are completely independent, it is not an adequarte measure of dependence. A small value for H Y X may imply that X tells us a great deal about Y or that H(Y) is small to begin with. Thus, we measure dependence using mutual information: I X Y = H Y H Y X Mutual information is a measure of the reduction of randomness of a variable given knowledge of another variable. Using properties of logarithms, we can derive several equivalent definitions: I X Y = H Y H Y X

62 = H X H X Y =H X +H Y H X Y = I Y X In addition to the definitions above, it is useful to realize that mutual information is a particular case of the Kullback-Leibler divergence. The KL divergence is defined as: KL divergence measures the difference between two distributions. It is sometimes called the relative entropy. It is always non-negative and zero only when p=q; however, it is not a distance because it is not symmetic. In terms of KL divergence, mutual information is: In other words, mutual information is a measure of the difference between the joint probability and product of the individual probabilities. These two distributions are equivalent only when X and Y are independent, and diverge as X and Y become more dependent. Shannon-Fano Code Shannon Fano coding, named after Claude Elwood Shannon and Robert Fano, is a technique for constructing a prefix code based on a set of symbols and their probabilities. It is suboptimal in the sense that it does not achieve the lowest possible expected codeword length like Huffman coding; however unlike Huffman coding, it does guarantee that all codeword lengths are within one bit of their theoretical ideal I(x) = log P(x). In Shannon Fano coding, the symbols are arranged in order from most probable to least probable, and then divided into two sets whose total probabilities are as close as possible to

63 being equal. All symbols then have the first digits of their codes assigned; symbols in the first set receive "0" and symbols in the second set receive "1". As long as any sets with more than one member remain, the same process is repeated on those sets, to determine successive digits of their codes. When a set has been reduced to one symbol, of course, this means the symbol's code is complete and will not form the prefix of any other symbol's code. The algorithm works, and it produces fairly efficient variable-length encodings; when the two smaller sets produced by a partitioning are in fact of equal probability, the one bit of information used to distinguish them is used most efficiently. Unfortunately, Shannon Fano does not always produce optimal prefix codes. For this reason, Shannon Fano is almost never used; Huffman coding is almost as computationally simple and produces prefix codes that always achieve the lowest expected code word length. Shannon Fano coding is used in the IMPLODE compression method, which is part of the ZIP file format, where it is desired to apply a simple algorithm with high performance and minimum requirements for programming. Shannon-Fano Algorithm: A Shannon Fano tree is built according to a specification designed to define an effective code table. The actual algorithm is simple: For a given list of symbols, develop a corresponding list of probabilities or frequency counts so that each symbol s relative frequency of occurrence is known. Sort the lists of symbols according to frequency, with the most frequently occurring symbols at the left and the least common at the right. Divide the list into two parts, with the total frequency counts of the left part being as close to the total of the right as possible. The left part of the list is assigned the binary digit 0, and the right part is assigned the digit 1. This means that the codes for the symbols in the first part will all start with 0, and the codes in the second part will all start with 1.

64 Recursively apply the steps 3 and 4 to each of the two halves, subdividing groups and adding bits to the codes until each symbol has become a corresponding code leaf on the tree. Example: The source of information A generates the symbols {A0, A1, A2, A3 and A4} with the corresponding probabilities {0.4, 0.3, 0.15, 0.1 and 0.05}. Encoding the source symbols using binary encoder and Shannon-Fano encoder gives: Source Symbol Pi Binary Code Shannon-Fano A A A A A Lavg H =

65 Shannon-Fano code is a top-down approach. Constructing the code tree, we get

66 Source Coding All source models in information theory may be viewed as random process or random sequence models. Let us consider the example of a discrete memory less source (DMS), which is a simple random sequence model. A DMS is a source whose output is a sequence of letters such that each letter is independently selected from a fixed alphabet consisting of letters; say a1, a2,.ak. The letters in the source output sequence are assumed to be random and statistically independent of each other. A fixed probability assignment for the occurrence of each letter is also assumed. Let us, consider a small example to appreciate the importance of probability assignment of the source letters. Let us consider a source with four letters a1, a2, a3 and a4 with P(a1)=0.5, P(a2)=0.25, P(a3)= 0.13, P(a4)=0.12. Let us decide to go for binary coding of these four source letters. While this can be done in multiple ways, two encoded representations are shown below: Code Representation#1: a1: 00, a2:01, a3:10, a4:11 Code Representation#2: a1: 0, a2:10, a3:001, a4:110 It is easy to see that in method #1 the probability assignment of a source letter has not been considered and all letters have been represented by two bits each. However in the second method only a1 has been encoded in one bit, a2 in two bits and the remaining two in three bits. It is easy to see that the average number of bits to be used per source letter for the two methods are not the same. ( a for method #1=2 bits per letter and a for method #2 < 2 bits per letter). So, if we consider the issue of encoding a long sequence of letters we have to transmit less number of bits following the second method. This is an important aspect of source coding operation in general. At this point, let us note the following:

67 a) We observe that assignment of small number of bits to more probable letters and assignment of larger number of bits to less probable letters (or symbols) may lead to efficient source encoding scheme. b) However, one has to take additional care while transmitting the encoded letters. A careful inspection of the binary representation of the symbols in method #2 reveals that it may lead to confusion (at the decoder end) in deciding the end of binary representation of a letter and beginning of the subsequent letter. So a source-encoding scheme should ensure that 1) The average number of coded bits (or letters in general) required per source letter is as small as possible and 2) The source letters can be fully retrieved from a received encodedsequence. In the following we discuss a popular variable-length source-coding scheme satisfying the above two requirements. Variable length Coding Let us assume that a DMS U has a K- letter alphabet {a1, a 2,.aK} with probabilities P(a 1), P(a2),. P(aK). Each source letter is to be encoded into a codeword made of elements (or letters) drawn from a code alphabet containing D symbols. Often for ease of implementation a binary code alphabet (D = 2) is chosen. As we observed earlier in an example, different codeword may not have same number of code symbols. If nk denotes the number of code symbols corresponding to the source letter ak, the average number of code letters per source letter ( n ) is: K n= P (ak)nk 5.1 k =1

68 Intuitively, if we encode a very long sequence of letters from a DMS, the number of code letters per source letter will be close to n. Now, a code is said to be uniquely decodable if for each source sequence of finite length, the sequence of code letters corresponding to that source sequence is different from the sequence of code letters corresponding to any other possible source sequence. We will briefly discuss about a subclass of uniquely decodable codes, known as prefix condition code. Let the code word in a code be represented as xk=(xk,1,xk,2,...,xk,nk), wherexk,1,xk,2,...,xk,nkdenote the individualcode letters (when D=2, these are 1 or 0). Any sequence made up of an initial part of xk that is xk,1,xk,2,...,xk,i for i nk is called a prefix of x k. any other codeword. A prefix condition code is a code in which no code word is the prefix of Example: consider the following table and find out which code out of the four shown is /are prefix condition code. Also determine n for each code. Source letters:- a1, a2, a3 and a4 P(ak) :- P(a1)=0.5, P(a2)=0.25, P(a3)= 0.125, P(a4)=0.125 Code Representation#1: a1: 00, a2:01, a3:10, a4:11 Code Representation#2: a1: 0, a2:1, a3:00, a4:11 Code Representation#3: a1: 0, a2:10, a3:110, a4:111 Code Representation#4: a1: 0, a2:01, a3:011, a4:0111 A prefix condition code can be decoded easily and uniquely. Start at the beginning of a

69 sequence and decode one word at a time. Finding the end of a code word is not a problem as the present code word is not a prefix to any other codeword. Example: Consider a coded sequence as per Code Representation #3 of theprevious example. See that the corresponding source letter sequence is a1,a4, a2, a1. Now, we state one important theorem known as Kraft Inequality theorem withoutproof. Binary Huffman Coding (an optimum variable-length source coding scheme) In Binary Huffman Coding each source letter is converted into a binary code word. It is a prefix condition code ensuring minimum average length per source letter in bits. Let the source letters a1, a 2,.aK have probabilities P(a1), P(a2),. P(aK) and let us assume that P(a1) P(a2) P(a 3). P(aK). We now consider a simple example to illustrate the steps for Huffman coding. Steps to calculate Huffman Coding Example Let us consider a discrete memoryless source with six letters having P(a1)=0.3,P(a2)=0.2, P(a 3)=0.15, P(a 4)=0.15, P(a5)=0.12 and P(a6)=0.08. Arrange the letters in descending order of their probability (here they are arranged). Consider the last two probabilities. Tie up the last two probabilities. Assign, say, 0 to the last digit of representation for the least probable letter (a6) and 1 to the last digit of representation for the second least probable letter (a5). That is, assign 1 to the upper arm of the tree and 0 to the lower arm.

70 1 P(a5)= P(a6)= (3) Now, add the two probabilities and imagine a new letter, say b1, substituting for a6 and a5. So P(b1) =0.2. Check whether a4 and b1are the least likely letters. If not, reorder the letters as per Step#1 and add the probabilities of two least likely letters. For our example, it leads to: P(a1)=0.3, P(a2)=0.2, P(b1)=0.2, P(a3)=0.15 and P(a4)=0.15 (4) Now go to Step#2 and start with the reduced ensemble consisting of a1, a2, a3, a4 and b1. Our example results in: P(a3)= P(a4)=0.15 Here we imagine another letter b1, with P(b2)=0.3. t Continue till the first digits of the most reduced ensemble of two letters are assigned a 1 and a 0. Again go back to the step (2): P(a1)=0.3, P(b2)=0.3, P(a2)=0.2 and P(b1)=0.2. Now we consider the last two probabilities: P(a2)= P(b1)=0.2

71 So, P(b3)=0.4. Following Step#2 again, we get, P(b3)=0.4, P(a1)=0.3 and P(b2)=0.3. Next two probabilities lead P(a1)=0.3 to: 1 P(b2)= with P(b4) = 0.6. Finally we get only two probabilities P(b4)=0.6 1 P(b3)= Now, read the code tree inward, starting from the root, and construct the codewords. The first digit of a codeword appears first while reading the code tree inward.

72 Hence, the final representation is: a1=11, a2=01, a3=101, a4=100, a5=001, a6=000. A few observations on the preceding example 1. The event with maximum probability has least number of bits. 2. Prefix condition is satisfied. No representation of one letter is prefix for other. Prefix condition says that representation of any letter should not be a part of any other letter. 3. Average length/letter (in bits) after coding is = P (ai )ni = 2.5 bits/letter. i 4. Note that the entropy of the source is: H(X)=2.465 bits/symbol. Average length per source letter after Huffman coding is a little bit more but close to the source entropy. In fact, the following celebrated theorem due to C. E. Shannon sets the limiting value of average length of codewords from a DMS. CONVOLUTIONAL CODES Convolutional codes are commonly described using two parameters: the code rate and the constraint length. The code rate, k/n, is expressed as a ratio of the number of bits into the convolutional encoder (k) to the number of channel symbols output by the convolutional encoder (n) in a given encoder cycle. The constraint length parameter, K, denotes the "length" of the convolutional encoder, i.e. how many k-bit stages are available to feed the combinatorial logic that produces the output symbols. Closely related to K is the parameter m, which indicates how many encoder cycles an input bit is retained and used for encoding after it first appears at the input to the convolutional encoder. The m parameter can be thought of as the memory length of the encoder. Convolutional codes are widely used as channel codes in practical communication systems for error correction. The encoded bits depend on the current k input bits and a few past input bits. The main decoding strategy for convolutional codes is based on the widely used Viterbi algorithm. As a result of the wide acceptance of convolutional codes, there have been several approaches to modify and extend this basic coding scheme. Trellis coded modulation (TCM) and turbo codes are two such examples. In TCM, redundancy is added by combining coding and modulation into a single operation. This is achieved without any reduction in data rate or expansion in bandwidth as required by only error correcting coding schemes.

73 A simple convolutional encoder is shown in Fig The information bits are fed in small groups of k-bits at a time to a shift register. The output encoded bits are obtained by modulo-2 addition (EXCLUSIVE-OR operation) of the input information bits and the contents of the shift registers which are a few previous information bits. If the encoder generates a group of n encoded bits per group of k information bits, the code rate R is commonly defined as R = k/n. In Fig. 7.1, k = 1 and n = 2. The number, K of elements in the shift register which decides for how many codewords one information bit will affect the encoder output, is known as the constraint length of the code. For the present example, K = 3. The shift register of the encoder is initialized to all-zero-state before encoding operation starts. It is easy to verify that encoded sequence is for an input message sequence of

74 Fig. 7.1A convolutional encoder with k=1, n=2 and r=1/2 The operation of a convolutional encoder can be explained in several but equivalent ways such as, by a) state diagram representation, b) tree diagram representation and c) trellis diagram representation. a) State Diagram Representation A convolutional encoder may be defined as a finite state machine. Contents of the rightmost (K-1) shift register stages define the states of the encoder. So, the encoder in Fig. 7.2has four states. The transition of an encoder from one state to another, ascaused by input bits, is depicted in the state diagram. Fig. 7.2 shows the state diagram of the encoder in Fig A new input bit causes a transition from one state to another. The path information between the states, denoted as b/c1c2, represents input information bit b and the corresponding output bits (c1c2). Again, it is not difficult to verify from the state diagram that an input information sequence b = (1011) generates an encoded sequence c = (11, 10, 00, 01). Fig. 7.2 State diagram representation for the encoder in Fig. 7.1

75 b)tree Diagram Representation The tree diagram representation shows all possible information and encoded sequences for the convolutional encoder. Fig. 7.3 shows the tree diagram for the encoder in Fig The encoded bits are labeled on the branches of the tree. Given an input sequence, the encoded sequence can be directly read from the tree. As an example, an input sequence (1011) results in the encoded sequence (11, 10, 00, 01).

76 c) Trellis Diagram Representation The trellis diagram of a convolutional code is obtained from its state diagram. All state transitions at each time step are explicitly shown in the diagram to retain the time dimension, as is present in the corresponding tree diagram. Usually, supporting descriptions on state transitions, corresponding input and output bits etc. are labeled in the trellis diagram. It is interesting to note that the trellis diagram, which describes the operation of the encoder, is very convenient for describing the behavior of thecorresponding decoder, especially when the famous Viterbi Algorithm (VA) is followed. Figure 7.4 shows the trellis diagram for the encoder in Figure 7.1. Fig. 7.3 A tree diagram for the encoder in Fig. 7.1 Fig. 7.4Trellis diagram, used in the decoder corresponding to the encoder in Fig.7.1

77 Hard-Decision and Soft-Decision Decoding Hard-decision and soft-decision decoding are based on the type of quantization used on the received bits. Hard-decision decoding uses 1-bit quantization on the received samples. Soft-decision decoding uses multi-bit quantization (e.g. 3 bits/sample) on the received sample values. Hard-Decision Viterbi Algorithm The Viterbi Algorithm (VA) finds a maximum likelihood (ML) estimate of a transmitted code sequence c from the corresponding received sequence r by maximizing the probability p(r c) that sequence r is received conditioned on the estimated code sequence c. Sequence c must be a valid coded sequence. The Viterbi algorithm utilizes the trellis diagram to compute the path metrics. The channel is assumed to be memory less, i.e. the noise sample affecting a received bit is independent from the noise sample affecting the other bits. The decoding operation starts from state 00, i.e. with the assumption that the initial state of the encoder is 00. With receipt of one noisy codeword, the decoding operation progresses by one step deeper into the trellis diagram. The branches, associated with a state of the trellis tell us about the corresponding codewords that the encoder may generate starting from this state. Hence, upon receipt of a codeword, it is possible to note the branch metric of each branch by determining the Hamming distance of the received codeword from the valid codeword associated with that branch. Path metric of all branches, associated with all the states are calculated similarly Now, at each depth of the trellis, each state also carries some accumulated path metric, which is the addition of metrics of all branches that construct the most likely path to that state. As an example, the trellis diagram of the code shown in Fig. 7.1, has four states and each state has two incoming and two outgoing branches. At any depth of the trellis, each state can be reached through two paths from the previous stage and as per the VA, the path with lower accumulated path metric is chosen. In the process, the accumulated path metric is updated by adding the metric of the incoming branch with the accumulated path metric of the state from where the branch originated. No decision about a received codeword is taken from such operations and the decoding decision is deliberately delayed to reduce the possibility of erroneous decision. The basic operations which are carried out as per the hard-decision Viterbi Algorithm after receiving one codeword are summarized below: All the branch metrics of all the states are determined;

78 . Accumulated metrics of all the paths (two in our example code) leading to a state are calculated taking into consideration the accumulated path metrics of the states from where the most recent branches emerged; Only one of the paths, entering into a state, which has minimum accumulated path metric is chosen as the survivor path for the state (or, equivalently node ); So, at the end of this process, each state has one survivor path. The history of a survivor path is also maintained by the node appropriately ( e.g. by storing the codewords or the information bits which are associated with the branches making the path); (5) Steps a) to d) are repeated and decoding decision is delayed till sufficient number of codewords has been received. Typically, the delay in decision making = Lx k codewords where L is an integer, e.g. 5 or 6. For the code in Fig. 7.1, the decision delay of 5x3 = 15 codewords may be sufficient for most occasions. This means, we decide about the first received codeword after receiving the 16 th codeword. The decision strategy is simple. Upon receiving the 16 th codeword and carrying out steps a) to d), we compare the accumulated path metrics of all the states ( four in our example) and chose the state with minimum overall accumulated path metric as the winning node for the first codeword. Then we trace back the history of the path associated with this winning node to identify the codeword tagged to the first branch of the path and declare this codeword as the most likely transmitted first codeword. The above procedure is repeated for each received codeword hereafter. Thus, the decision for a codeword is delayed but once the decision process starts, we decide once for every received codeword. For most practical applications, including delay-sensitive digital speech coding and transmission, a decision delay of Lx k codewords is acceptable. Soft-Decision Viterbi Algorithm In soft-decision decoding, the demodulator does not assign a 0 or a 1 to each received bit but uses multi-bit quantized values. The soft-decision Viterbi algorithm is very similar to its hard-decision algorithm except that squared Euclidean distance is used in the branch metrics instead of simpler Hamming distance. However, the performance of a soft-decision VA is much more impressive compared to its HDD (Hard Decision Decoding) counterpart [Fig. 7.6 (a) and (b)]. The computational requirement of a Viterbi decoder grows exponentially as a function of the constraint length and hence it is usually limited in practice to constraint lengths of K = 9.

79 Fig. 7.6 (a) Decoded BER vs input BER for the rate half convolutional codes withviterbi Algorithm ; 1) k = 3 (HDD),2) k = 5 (HDD),3) k = 3 (SDD), and 4) k= 5 (SDD). HDD: Hard Decision Decoding; SDD: Soft Decision Decoding.

80 Fig. 7.6 (b) Decoded BER vs Eb/No(in db) for the rate half convolutional codes withviterbi Algorithm ; 1) Uncoded system; 2) with k = 3 (HDD) and 3) k = 3 (SDD). HDD: Hard Decision Decoding; SDD: Soft Decision Decoding. Introduction: Spread Spectrum Modulation Initially developed for military applications during II world war, that was less sensitive to intentional interference or jamming by third parties. Spread spectrum technology has blossomed into one of the fundamental building blocks in current and next-generation wireless systems Problem of radio transmission

81 Narrow band can be wiped out due to interference To disrupt the communication, the adversary needs to do two things, to detect that a transmission is taking place and to transmit a jamming signal which is designed to confuse the receiver. Solution A spread spectrum system is therefore designed to make these tasks as difficult as possible. Firstly, the transmitted signal should be difficult to detect by an adversary/jammer, i.e.,the signal should have a low probability of intercept (LPI). Secondly, the signal should be difficult to disturb with a jamming signal, i.e., thetransmitted signal should possess an anti-jamming (AJ) property Remedy Spread the narrow band signal into a broad band to protect against interference In a digital communication system the primary resources are Bandwidth andpower. The study of digital communication system deals with efficient utilization ofthese two resources, but there are situations where it is necessary to sacrifice their efficient utilization in order to meet certain other design objectives. For example to provide a form of secure communication (i.e. the transmitted signal is not easily detected or recognized by unwanted listeners) the bandwidth of thetransmitted signal is increased in excess of the minimum bandwidth necessary to transmit it. This requirement is catered by a technique known as Spread SpectrumModulation. The primary advantage of a Spread Spectrum communication system is its ability to reject Interference whether it be the unintentional or the intentionalinterference. The definition of Spread Spectrum modulation may be stated in two parts. Spread Spectrum is a mean of transmission in which the data sequence occupies a BW (Bandwidth) in excess of the minimum BW necessary to transmit it. The Spectrum Spreading is accomplished before transmission through the use of a code that is independent of the data sequence. The Same code is used in the receiver to despread the received signal so that the original data sequence may be recovered. s(t) wide band r(t) wide band b(t) + Noise

82 b(t) Narrow Wide Band Band c(t) n(t) c(t) Wide band (noise) Wide band ---- Transmitter Channel Receiver Fig: 8.1 spread spectrum technique. b(t) = Data Sequence to be transmitted (Narrow Band) c(t) = Wide Band code s(t) = c(t) * b(t) (wide Band) Fig: 8.2Spectrum of signal before & after spreading

83 Code division multiple access CDMA is achannel access methodutilized by various radio communication technologies. One of the basic concepts in data communication is the idea of allowing several transmitters to send information simultaneously over a single communication channel. This allows several users to share a bandwidth of frequencies. This concept is called multiplexing. CDMA employs spread-spectrum technology and a special coding scheme (where each transmitter is assigned a code) to allow multiple users to be multiplexed over the same physical channel. By contrast, time division multiple access (TDMA) divides access by time, while frequency-division multiple access (FDMA) divides it by frequency. CDMA is a form of "spread-spectrum" signaling, since the modulated coded signal has a much higher data bandwidth than the data being communicated. An analogy to the problem of multiple access is a room (channel) in which people wish to communicate with each other. To avoid confusion, people could take turns speaking (time division), speak at different pitches (frequency division), or speak in different languages (code division). CDMA is analogous to the last example where people speaking the same language can understand each other, but not other people. Similarly, in radio CDMA, each group of users is given a shared code. Many codes occupy the same channel, but only users associated with a particular code can understand each other. Technical details CDMA is a spread spectrum multiple access technique. In CDMA a locally generated code runs at a much higher rate than the data to be transmitted. Data for transmission is simply logically XOR (exclusive OR) added with the faster code. The figure shows how spread spectrum signal is generated. The data signal with pulse duration of Tb is XOR added with the code signal with pulse duration of Tc. (Note: bandwidth is proportional to 1 / T where T = bit time) Therefore, the bandwidth of the data signal is 1 / Tb and the bandwidth of the spread spectrum signal is 1 / Tc. Since Tc is much smaller than Tb, the bandwidth of the spread spectrum signal is much larger than the bandwidth of the original signal.

84 Fig. 8.3 CDMA uses Direct Sequence spreading, where spreading process isdone by directly combining the baseband information to high chip rate binary code. The Spreading Factor is the ratio of the chips (UMTS = 3.84Mchips/s) to baseband information rate. Spreading factors vary from 4 to 512 in FDD UMTS. Spreading process gain can in expressed in dbs (Spreading factor 128 = 21dB gain). Fig. 8.4

85 Each user in a CDMA system uses a different code to modulate their signal. Choosing the codes used to modulate the signal is very important in the performance of CDMA systems. The best performance will occur when there is good separation between the signal of a desired user and the signals of other users. The separation of the signals is made by correlating the received signal with the locally generated code of the desired user. If the signal matches the desired user's code then the correlation function will be high and the system can extract that signal. If the desired user's code has nothing in common with the signal the correlation should be as close to zero as possible (thus eliminating the signal); this is referred to as cross correlation. If the code is correlated with the signal at any time offset other than zero, the correlation should be as close to zero as possible. This is referred to as auto-correlation and is used to reject multi-path interference. Fig. 8.5

86 PSUEDO-NOISE SEQUENCE: Generation of PN sequence: Clock S0 Shift Shift Shift Register1 Register2 Register3 S3 Output Logic Circuit Fig 8.6: Maximum-length sequence generator for n=3 A feedback shift register is said to be Linear when the feed back logic consists of entirely mod-2-address ( Ex-or gates). In such a case, the zero state is not permitted. The period of a PN sequence produced by a linear feedback shift register with n flip flops cannot exceed 2 n -1. When the period is exactly 2 n -1, the PN sequence is called a maximum length sequence or m-sequence. Example1: Consider the linear feed back shift register as shown in fig 2involve three flip-flops. The input s o is equal to the mod-2 sum of S 1 and S 3. If the initial state of the shift register is 100. Then the succession of states will be as follows. 100,110,011,011,101,010,001, The output sequence (output S 3 ) is therefore Which repeats itself with period = 7 (n=3) Maximal length codes are commonly used PN codes In binary shift register, the maximum length sequence is N = 2 m -1 chips, where m is the number of stages of flip-flops in the shift register.

87 Fig.8.7 At each clock pulse (6) Contents of register shifts one bit right. (7) Contents of required stages are modulo 2 added and fed back to input. Let initial status of shift register be 1000 Fig. 8.8 Initial stages of Shift registers

88 We can see for shift Register of length m=4..at each clock the change in state of flip-flop is shown. Feed back function is modulo two of X3andX4. After 15 clock pulses the sequence repeats. Output sequence is

89 Properties of PN Sequence Randomness of PN sequence is tested by following properties u Balance property v Run length property w Autocorrelation property 1. Balance property In each Period of the sequence, number of binary ones differ from binary zeros by at most one digit. Consider output of shift register Seven zeros and eight ones -meets balance condition. 2. Run length property Among the runs of ones and zeros in each period, it is desirable that about one half the runs of each type are of length 1, one- fourth are of length 2 and one-eighth are of length 3 and soon. Consider output of shift register Number of runs = { No. of agreements No. of disagreements in comparison of one full period } Consider output of shift register for l=1 Yields PN autocorrelation as

90 Range of PN Sequence Lengths Length 0f Shift Register, m PN Sequence Length,

91 A Notion of Spread Spectrum: An important attribute of Spread Spectrum modulation is that it can provide protection against externally generated interfacing signals with finite power. Protection against jamming (interfacing) waveforms is provided by purposely making the information bearing signal occupy a BW far in excess of the minimum BW necessary to transmit it. This has the effect of making the transmitted signal a noise like appearance so as to blend into the background. Therefore Spread Spectrum is a method of camouflaging the information bearing signal. V b(t) m(t).. r(t) z(t) Tb Decisi 0 dt on Device c(t) n(t) c(t) Threshold=0 <----Transmitter-----> --Channel Receiver Let { b K } denotes a binary data sequence. { c K } denotes a PN sequence. b(t) and c(t) denotes their NRZ polar representation respectively. The desired modulation is achieved by applying the data signal b(t) and PN signal c(t) to a product modulator or multiplier. If the message signal b(t) is narrowband and the PN sequence signal c(t) is wide band, the product signal m(t) is also wide band. The PN sequence performs the role of a Spreading Code. For base band transmission, the product signal m(t) represents the transmitted signal. Therefore m(t) = c(t).b(t) The received signal r(t) consists of the transmitted signal m(t) plus an additive interference noise n(t), Hence r(t) = m(t) + n(t) = c(t).b(t) + n(t)

92 a) Data Signal b(t) b)spreading Code c(t) c)product signal or base band transmitted signal m(t) To recover the original message signal b(t), the received signal r(t) is applied to a demodulator that consists of a multiplier followed by an integrator and a decision device. The multiplier is supplied with a locally generated PN sequence that is exact replica of that used in the transmitter. The multiplier output is given by

93 Z(t) = r(t).c(t) 4. [b(t) * c(t) + n(t)] c(t) 5. c 2 (t).b(t) + c(t).n(t) The data signal b(t) is multiplied twice by the PN signal c(t), where as unwanted signal n(t) is multiplied only once. But c 2 (t) = 1, hence the above equation reduces to Z(t) = b(t) + c(t).n(t) Now the data component b(t) is narrowband, where as the spurious component c(t)n(t) is wide band. Hence by applying the multiplier output to a base band (low pass) filter most of the power in the spurious component c(t)n(t) is filtered out. Thus the effect of the interference n(t) is thus significantly reduced at the receiver output. The integration is carried out for the bit interval 0 t T b to provide the sample value V. Finally, a decision is made by the receiver. If V > Threshold Value 0, say binary symbol 1 If V < Threshold Value 0, say binary symbol 0 Frequency Hop Spread Spectrum: In a frequency hop Spread Spectrum technique, the spectrum of data modulated carrier is widened by changing the carrier frequency in a pseudo random manner. The type of spread spectrum in which the carrier hops randomly form one frequency to another is called Frequency Hop (FH) Spread Spectrum. Since frequency hopping does not covers the entire spread spectrum instantaneously. We are led to consider the rate at which the hop occurs. Depending upon this we have two types of frequency hop. 1. Slow frequency hopping:- In which the symbol rate R s of the MFSK signal is an integer multiple of the hop rate R h. That is several symbols are transmitted on each frequency hop. 2. Fast Frequency hopping:- In which the hop rate R h is an integral multiple of the MFSK symbol rate R s. That is the carrier frequency will hoop several times during the transmission of one symbol. A common modulation format for frequency hopping system is that of M- ary frequency shift keying (MFSK).

94 Slow frequency hopping:- Fig a) Shows the block diagram of an FH / MFSK transmitter, which involves frequency modulation followed by mixing. The incoming binary data are applied to an M-ary FSK modulator. The resulting modulated wave and the output from a digital frequency synthesizer are then applied to a mixer that consists of a multiplier followed by a band pass filter. The filter is designed to select the sum frequency component resulting from the multiplication process as the transmitted signal. An k bit segments of a PN sequence drive the frequencysynthesizer, which enables the carrier frequency to hop over 2 n distinct values. Since frequency synthesizers are unable to maintain phase coherence over successive hops, most frequency hops spread spectrum communication system use non coherent M-ary modulation system. Fig 8.12:- Frequency hop spread M-ary Frequency shift keying

UNIT-1. Basic signal processing operations in digital communication

UNIT-1. Basic signal processing operations in digital communication UNIT-1 Lecture-1 Basic signal processing operations in digital communication The three basic elements of every communication systems are Transmitter, Receiver and Channel. The Overall purpose of this system

More information

Chapter-1: Introduction

Chapter-1: Introduction Chapter-1: Introduction The purpose of a Communication System is to transport an information bearing signal from a source to a user destination via a communication channel. MODEL OF A COMMUNICATION SYSTEM

More information

EEE 309 Communication Theory

EEE 309 Communication Theory EEE 309 Communication Theory Semester: January 2016 Dr. Md. Farhad Hossain Associate Professor Department of EEE, BUET Email: mfarhadhossain@eee.buet.ac.bd Office: ECE 331, ECE Building Part 05 Pulse Code

More information

CHAPTER 3 Syllabus (2006 scheme syllabus) Differential pulse code modulation DPCM transmitter

CHAPTER 3 Syllabus (2006 scheme syllabus) Differential pulse code modulation DPCM transmitter CHAPTER 3 Syllabus 1) DPCM 2) DM 3) Base band shaping for data tranmission 4) Discrete PAM signals 5) Power spectra of discrete PAM signal. 6) Applications (2006 scheme syllabus) Differential pulse code

More information

10 Speech and Audio Signals

10 Speech and Audio Signals 0 Speech and Audio Signals Introduction Speech and audio signals are normally converted into PCM, which can be stored or transmitted as a PCM code, or compressed to reduce the number of bits used to code

More information

Fundamentals of Digital Communication

Fundamentals of Digital Communication Fundamentals of Digital Communication Network Infrastructures A.A. 2017/18 Digital communication system Analog Digital Input Signal Analog/ Digital Low Pass Filter Sampler Quantizer Source Encoder Channel

More information

EEE 309 Communication Theory

EEE 309 Communication Theory EEE 309 Communication Theory Semester: January 2017 Dr. Md. Farhad Hossain Associate Professor Department of EEE, BUET Email: mfarhadhossain@eee.buet.ac.bd Office: ECE 331, ECE Building Types of Modulation

More information

Time division multiplexing The block diagram for TDM is illustrated as shown in the figure

Time division multiplexing The block diagram for TDM is illustrated as shown in the figure CHAPTER 2 Syllabus: 1) Pulse amplitude modulation 2) TDM 3) Wave form coding techniques 4) PCM 5) Quantization noise and SNR 6) Robust quantization Pulse amplitude modulation In pulse amplitude modulation,

More information

Digital Communication (650533) CH 3 Pulse Modulation

Digital Communication (650533) CH 3 Pulse Modulation Philadelphia University/Faculty of Engineering Communication and Electronics Engineering Digital Communication (650533) CH 3 Pulse Modulation Instructor: Eng. Nada Khatib Website: http://www.philadelphia.edu.jo/academics/nkhatib/

More information

Pulse Code Modulation

Pulse Code Modulation Pulse Code Modulation EE 44 Spring Semester Lecture 9 Analog signal Pulse Amplitude Modulation Pulse Width Modulation Pulse Position Modulation Pulse Code Modulation (3-bit coding) 1 Advantages of Digital

More information

TE 302 DISCRETE SIGNALS AND SYSTEMS. Chapter 1: INTRODUCTION

TE 302 DISCRETE SIGNALS AND SYSTEMS. Chapter 1: INTRODUCTION TE 302 DISCRETE SIGNALS AND SYSTEMS Study on the behavior and processing of information bearing functions as they are currently used in human communication and the systems involved. Chapter 1: INTRODUCTION

More information

MAHALAKSHMI ENGINEERING COLLEGE TIRUCHIRAPALLI

MAHALAKSHMI ENGINEERING COLLEGE TIRUCHIRAPALLI MAHALAKSHMI ENGINEERING COLLEGE TIRUCHIRAPALLI 621213 QUESTION BANK DEPARTMENT: ECE SEMESTER: V SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT-I DIGITAL COMMUNICATION SYSTEMS PART -A (2 Marks)

More information

Outline / Wireless Networks and Applications Lecture 3: Physical Layer Signals, Modulation, Multiplexing. Cartoon View 1 A Wave of Energy

Outline / Wireless Networks and Applications Lecture 3: Physical Layer Signals, Modulation, Multiplexing. Cartoon View 1 A Wave of Energy Outline 18-452/18-750 Wireless Networks and Applications Lecture 3: Physical Layer Signals, Modulation, Multiplexing Peter Steenkiste Carnegie Mellon University Spring Semester 2017 http://www.cs.cmu.edu/~prs/wirelesss17/

More information

Voice Transmission --Basic Concepts--

Voice Transmission --Basic Concepts-- Voice Transmission --Basic Concepts-- Voice---is analog in character and moves in the form of waves. 3-important wave-characteristics: Amplitude Frequency Phase Telephone Handset (has 2-parts) 2 1. Transmitter

More information

Downloaded from 1

Downloaded from  1 VII SEMESTER FINAL EXAMINATION-2004 Attempt ALL questions. Q. [1] How does Digital communication System differ from Analog systems? Draw functional block diagram of DCS and explain the significance of

More information

Communications and Signals Processing

Communications and Signals Processing Communications and Signals Processing Dr. Ahmed Masri Department of Communications An Najah National University 2012/2013 1 Dr. Ahmed Masri Chapter 5 - Outlines 5.4 Completing the Transition from Analog

More information

EXPERIMENT WISE VIVA QUESTIONS

EXPERIMENT WISE VIVA QUESTIONS EXPERIMENT WISE VIVA QUESTIONS Pulse Code Modulation: 1. Draw the block diagram of basic digital communication system. How it is different from analog communication system. 2. What are the advantages of

More information

CHAPTER 4. PULSE MODULATION Part 2

CHAPTER 4. PULSE MODULATION Part 2 CHAPTER 4 PULSE MODULATION Part 2 Pulse Modulation Analog pulse modulation: Sampling, i.e., information is transmitted only at discrete time instants. e.g. PAM, PPM and PDM Digital pulse modulation: Sampling

More information

PULSE CODE MODULATION (PCM)

PULSE CODE MODULATION (PCM) PULSE CODE MODULATION (PCM) 1. PCM quantization Techniques 2. PCM Transmission Bandwidth 3. PCM Coding Techniques 4. PCM Integrated Circuits 5. Advantages of PCM 6. Delta Modulation 7. Adaptive Delta Modulation

More information

Communications I (ELCN 306)

Communications I (ELCN 306) Communications I (ELCN 306) c Samy S. Soliman Electronics and Electrical Communications Engineering Department Cairo University, Egypt Email: samy.soliman@cu.edu.eg Website: http://scholar.cu.edu.eg/samysoliman

More information

Lecture 3 Concepts for the Data Communications and Computer Interconnection

Lecture 3 Concepts for the Data Communications and Computer Interconnection Lecture 3 Concepts for the Data Communications and Computer Interconnection Aim: overview of existing methods and techniques Terms used: -Data entities conveying meaning (of information) -Signals data

More information

Chapter 2: Fundamentals of Data and Signals

Chapter 2: Fundamentals of Data and Signals Chapter 2: Fundamentals of Data and Signals TRUE/FALSE 1. The terms data and signal mean the same thing. F PTS: 1 REF: 30 2. By convention, the minimum and maximum values of analog data and signals are

More information

QUESTION BANK. SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2

QUESTION BANK. SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2 QUESTION BANK DEPARTMENT: ECE SEMESTER: V SUBJECT CODE / Name: EC2301 DIGITAL COMMUNICATION UNIT 2 BASEBAND FORMATTING TECHNIQUES 1. Why prefilterring done before sampling [AUC NOV/DEC 2010] The signal

More information

UNIT III -- DATA AND PULSE COMMUNICATION PART-A 1. State the sampling theorem for band-limited signals of finite energy. If a finite energy signal g(t) contains no frequency higher than W Hz, it is completely

More information

Introduction to Telecommunications and Computer Engineering Unit 3: Communications Systems & Signals

Introduction to Telecommunications and Computer Engineering Unit 3: Communications Systems & Signals Introduction to Telecommunications and Computer Engineering Unit 3: Communications Systems & Signals Syedur Rahman Lecturer, CSE Department North South University syedur.rahman@wolfson.oxon.org Acknowledgements

More information

Department of Electronics and Communication Engineering 1

Department of Electronics and Communication Engineering 1 UNIT I SAMPLING AND QUANTIZATION Pulse Modulation 1. Explain in detail the generation of PWM and PPM signals (16) (M/J 2011) 2. Explain in detail the concept of PWM and PAM (16) (N/D 2012) 3. What is the

More information

Data Encoding g(p (part 2)

Data Encoding g(p (part 2) Data Encoding g(p (part 2) CSE 3213 Instructor: U.T. Nguyen 10/11/2007 12:44 PM 1 Analog Data, Digital Signals (5.3) 2 1 Analog Data, Digital Signals Digitization Conversion of analog data into digital

More information

CODING TECHNIQUES FOR ANALOG SOURCES

CODING TECHNIQUES FOR ANALOG SOURCES CODING TECHNIQUES FOR ANALOG SOURCES Prof.Pratik Tawde Lecturer, Electronics and Telecommunication Department, Vidyalankar Polytechnic, Wadala (India) ABSTRACT Image Compression is a process of removing

More information

Lecture Fundamentals of Data and signals

Lecture Fundamentals of Data and signals IT-5301-3 Data Communications and Computer Networks Lecture 05-07 Fundamentals of Data and signals Lecture 05 - Roadmap Analog and Digital Data Analog Signals, Digital Signals Periodic and Aperiodic Signals

More information

SEN366 Computer Networks

SEN366 Computer Networks SEN366 Computer Networks Prof. Dr. Hasan Hüseyin BALIK (5 th Week) 5. Signal Encoding Techniques 5.Outline An overview of the basic methods of encoding digital data into a digital signal An overview of

More information

2. By convention, the minimum and maximum values of analog data and signals are presented as voltages.

2. By convention, the minimum and maximum values of analog data and signals are presented as voltages. Chapter 2: Fundamentals of Data and Signals Data Communications and Computer Networks A Business Users Approach 8th Edition White TEST BANK Full clear download (no formatting errors) at: https://testbankreal.com/download/data-communications-computer-networksbusiness-users-approach-8th-edition-white-test-bank/

More information

ECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2

ECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2 ECE 556 BASICS OF DIGITAL SPEECH PROCESSING Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2 Analog Sound to Digital Sound Characteristics of Sound Amplitude Wavelength (w) Frequency ( ) Timbre

More information

QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold

QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold QUESTION BANK EC 1351 DIGITAL COMMUNICATION YEAR / SEM : III / VI UNIT I- PULSE MODULATION PART-A (2 Marks) 1. What is the purpose of sample and hold circuit 2. What is the difference between natural sampling

More information

Announcements : Wireless Networks Lecture 3: Physical Layer. Bird s Eye View. Outline. Page 1

Announcements : Wireless Networks Lecture 3: Physical Layer. Bird s Eye View. Outline. Page 1 Announcements 18-759: Wireless Networks Lecture 3: Physical Layer Please start to form project teams» Updated project handout is available on the web site Also start to form teams for surveys» Send mail

More information

Chapter-3 Waveform Coding Techniques

Chapter-3 Waveform Coding Techniques Chapter-3 Waveform Coding Techniques PCM [Pulse Code Modulation] PCM is an important method of analog to-digital conversion. In this modulation the analog signal is converted into an electrical waveform

More information

ANALOGUE AND DIGITAL COMMUNICATION

ANALOGUE AND DIGITAL COMMUNICATION ANALOGUE AND DIGITAL COMMUNICATION Syed M. Zafi S. Shah Umair M. Qureshi Lecture xxx: Analogue to Digital Conversion Topics Pulse Modulation Systems Advantages & Disadvantages Pulse Code Modulation Pulse

More information

EITF25 Internet Techniques and Applications L2: Physical layer. Stefan Höst

EITF25 Internet Techniques and Applications L2: Physical layer. Stefan Höst EITF25 Internet Techniques and Applications L2: Physical layer Stefan Höst Data vs signal Data: Static representation of information For storage Signal: Dynamic representation of information For transmission

More information

UNIT TEST I Digital Communication

UNIT TEST I Digital Communication Time: 1 Hour Class: T.E. I & II Max. Marks: 30 Q.1) (a) A compact disc (CD) records audio signals digitally by using PCM. Assume the audio signal B.W. to be 15 khz. (I) Find Nyquist rate. (II) If the Nyquist

More information

Comm 502: Communication Theory. Lecture 4. Line Coding M-ary PCM-Delta Modulation

Comm 502: Communication Theory. Lecture 4. Line Coding M-ary PCM-Delta Modulation Comm 502: Communication Theory Lecture 4 Line Coding M-ary PCM-Delta Modulation PCM Decoder PCM Waveform Types (Line Coding) Representation of binary sequence into the electrical signals that enter the

More information

two computers. 2- Providing a channel between them for transmitting and receiving the signals through it.

two computers. 2- Providing a channel between them for transmitting and receiving the signals through it. 1. Introduction: Communication is the process of transmitting the messages that carrying information, where the two computers can be communicated with each other if the two conditions are available: 1-

More information

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued CSCD 433 Network Programming Fall 2016 Lecture 5 Physical Layer Continued 1 Topics Definitions Analog Transmission of Digital Data Digital Transmission of Analog Data Multiplexing 2 Different Types of

More information

Chapter 2: Digitization of Sound

Chapter 2: Digitization of Sound Chapter 2: Digitization of Sound Acoustics pressure waves are converted to electrical signals by use of a microphone. The output signal from the microphone is an analog signal, i.e., a continuous-valued

More information

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued

CSCD 433 Network Programming Fall Lecture 5 Physical Layer Continued CSCD 433 Network Programming Fall 2016 Lecture 5 Physical Layer Continued 1 Topics Definitions Analog Transmission of Digital Data Digital Transmission of Analog Data Multiplexing 2 Different Types of

More information

CHAPTER -15. Communication Systems

CHAPTER -15. Communication Systems CHAPTER -15 Communication Systems COMMUNICATION Communication is the act of transmission and reception of information. COMMUNICATION SYSTEM: A system comprises of transmitter, communication channel and

More information

QUESTION BANK SUBJECT: DIGITAL COMMUNICATION (15EC61)

QUESTION BANK SUBJECT: DIGITAL COMMUNICATION (15EC61) QUESTION BANK SUBJECT: DIGITAL COMMUNICATION (15EC61) Module 1 1. Explain Digital communication system with a neat block diagram. 2. What are the differences between digital and analog communication systems?

More information

Contents. Telecom Service Chae Y. Lee. Data Signal Transmission Transmission Impairments Channel Capacity

Contents. Telecom Service Chae Y. Lee. Data Signal Transmission Transmission Impairments Channel Capacity Data Transmission Contents Data Signal Transmission Transmission Impairments Channel Capacity 2 Data/Signal/Transmission Data: entities that convey meaning or information Signal: electric or electromagnetic

More information

EC 6501 DIGITAL COMMUNICATION UNIT - II PART A

EC 6501 DIGITAL COMMUNICATION UNIT - II PART A EC 6501 DIGITAL COMMUNICATION 1.What is the need of prediction filtering? UNIT - II PART A [N/D-16] Prediction filtering is used mostly in audio signal processing and speech processing for representing

More information

EEE482F: Problem Set 1

EEE482F: Problem Set 1 EEE482F: Problem Set 1 1. A digital source emits 1.0 and 0.0V levels with a probability of 0.2 each, and +3.0 and +4.0V levels with a probability of 0.3 each. Evaluate the average information of the source.

More information

Physical Layer: Outline

Physical Layer: Outline 18-345: Introduction to Telecommunication Networks Lectures 3: Physical Layer Peter Steenkiste Spring 2015 www.cs.cmu.edu/~prs/nets-ece Physical Layer: Outline Digital networking Modulation Characterization

More information

An ISO 9001:2008 Certified Institution EC6501-Digital Communication 1 Unit-1: Sampling & Quantization The purpose of a Communication System is to transport an information bearing signal from a source to

More information

Signal Encoding Techniques

Signal Encoding Techniques 2 Techniques ITS323: to Data Communications CSS331: Fundamentals of Data Communications Sirindhorn International Institute of Technology Thammasat University Prepared by Steven Gordon on 3 August 2015

More information

2. TELECOMMUNICATIONS BASICS

2. TELECOMMUNICATIONS BASICS 2. TELECOMMUNICATIONS BASICS The purpose of any telecommunications system is to transfer information from the sender to the receiver by a means of a communication channel. The information is carried by

More information

Basic Concepts in Data Transmission

Basic Concepts in Data Transmission Basic Concepts in Data Transmission EE450: Introduction to Computer Networks Professor A. Zahid A.Zahid-EE450 1 Data and Signals Data is an entity that convey information Analog Continuous values within

More information

Chapter 4 Digital Transmission 4.1

Chapter 4 Digital Transmission 4.1 Chapter 4 Digital Transmission 4.1 Copyright The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 4-1 DIGITAL-TO-DIGITAL CONVERSION In this section, we see how we can represent

More information

Pulse Code Modulation

Pulse Code Modulation Pulse Code Modulation Modulation is the process of varying one or more parameters of a carrier signal in accordance with the instantaneous values of the message signal. The message signal is the signal

More information

SOME PHYSICAL LAYER ISSUES. Lecture Notes 2A

SOME PHYSICAL LAYER ISSUES. Lecture Notes 2A SOME PHYSICAL LAYER ISSUES Lecture Notes 2A Delays in networks Propagation time or propagation delay, t prop Time required for a signal or waveform to propagate (or move) from one point to another point.

More information

Data Communication. Chapter 3 Data Transmission

Data Communication. Chapter 3 Data Transmission Data Communication Chapter 3 Data Transmission ١ Terminology (1) Transmitter Receiver Medium Guided medium e.g. twisted pair, coaxial cable, optical fiber Unguided medium e.g. air, water, vacuum ٢ Terminology

More information

TELECOMMUNICATION SYSTEMS

TELECOMMUNICATION SYSTEMS TELECOMMUNICATION SYSTEMS By Syed Bakhtawar Shah Abid Lecturer in Computer Science 1 MULTIPLEXING An efficient system maximizes the utilization of all resources. Bandwidth is one of the most precious resources

More information

Signal Characteristics

Signal Characteristics Data Transmission The successful transmission of data depends upon two factors:» The quality of the transmission signal» The characteristics of the transmission medium Some type of transmission medium

More information

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Direct link. Point-to-point.

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Direct link. Point-to-point. Terminology (1) Chapter 3 Data Transmission Transmitter Receiver Medium Guided medium e.g. twisted pair, optical fiber Unguided medium e.g. air, water, vacuum Spring 2012 03-1 Spring 2012 03-2 Terminology

More information

DIGITAL COMMINICATIONS

DIGITAL COMMINICATIONS Code No: R346 R Set No: III B.Tech. I Semester Regular and Supplementary Examinations, December - 23 DIGITAL COMMINICATIONS (Electronics and Communication Engineering) Time: 3 Hours Max Marks: 75 Answer

More information

SUMMER 15 EXAMINATION. 1) The answers should be examined by key words and not as word-to-word as given in the

SUMMER 15 EXAMINATION. 1) The answers should be examined by key words and not as word-to-word as given in the SUMMER 15 EXAMINATION Subject Code: 17535 Model Answer Important Instructions to examiners: 1) The answers should be examined by key words and not as word-to-word as given in the model answer scheme. 2)

More information

Chapter 3 Data Transmission COSC 3213 Summer 2003

Chapter 3 Data Transmission COSC 3213 Summer 2003 Chapter 3 Data Transmission COSC 3213 Summer 2003 Courtesy of Prof. Amir Asif Definitions 1. Recall that the lowest layer in OSI is the physical layer. The physical layer deals with the transfer of raw

More information

Digital Communication - Analog to Digital

Digital Communication - Analog to Digital Unit 26. Digital Communication Digital Communication - Analog to Digital The communication that occurs in our day-to-day life is in the form of signals. These signals, such as sound signals, generally,

More information

Physical Layer. Networks: Physical Layer 1

Physical Layer. Networks: Physical Layer 1 Physical Layer Networks: Physical Layer 1 Physical Layer Part 1 Definitions Nyquist Theorem - noiseless Shannon s Result with noise Analog versus Digital Amplifier versus Repeater Networks: Physical Layer

More information

EC 2301 Digital communication Question bank

EC 2301 Digital communication Question bank EC 2301 Digital communication Question bank UNIT I Digital communication system 2 marks 1.Draw block diagram of digital communication system. Information source and input transducer formatter Source encoder

More information

Data Communications & Computer Networks

Data Communications & Computer Networks Data Communications & Computer Networks Chapter 3 Data Transmission Fall 2008 Agenda Terminology and basic concepts Analog and Digital Data Transmission Transmission impairments Channel capacity Home Exercises

More information

(Refer Slide Time: 3:11)

(Refer Slide Time: 3:11) Digital Communication. Professor Surendra Prasad. Department of Electrical Engineering. Indian Institute of Technology, Delhi. Lecture-2. Digital Representation of Analog Signals: Delta Modulation. Professor:

More information

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay

Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Digital Communication Prof. Bikash Kumar Dey Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 03 Quantization, PCM and Delta Modulation Hello everyone, today we will

More information

Department of Electronics & Telecommunication Engg. LAB MANUAL. B.Tech V Semester [ ] (Branch: ETE)

Department of Electronics & Telecommunication Engg. LAB MANUAL. B.Tech V Semester [ ] (Branch: ETE) Department of Electronics & Telecommunication Engg. LAB MANUAL SUBJECT:-DIGITAL COMMUNICATION SYSTEM [BTEC-501] B.Tech V Semester [2013-14] (Branch: ETE) KCT COLLEGE OF ENGG & TECH., FATEHGARH PUNJAB TECHNICAL

More information

Class 4 ((Communication and Computer Networks))

Class 4 ((Communication and Computer Networks)) Class 4 ((Communication and Computer Networks)) Lesson 5... SIGNAL ENCODING TECHNIQUES Abstract Both analog and digital information can be encoded as either analog or digital signals. The particular encoding

More information

Introduction to Discrete-Time Control Systems

Introduction to Discrete-Time Control Systems Chapter 1 Introduction to Discrete-Time Control Systems 1-1 INTRODUCTION The use of digital or discrete technology to maintain conditions in operating systems as close as possible to desired values despite

More information

ITM 1010 Computer and Communication Technologies

ITM 1010 Computer and Communication Technologies ITM 1010 Computer and Communication Technologies Lecture #14 Part II Introduction to Communication Technologies: Digital Signals: Digital modulation, channel sharing 2003 香港中文大學, 電子工程學系 (Prof. H.K.Tsang)

More information

Islamic University of Gaza. Faculty of Engineering Electrical Engineering Department Spring-2011

Islamic University of Gaza. Faculty of Engineering Electrical Engineering Department Spring-2011 Islamic University of Gaza Faculty of Engineering Electrical Engineering Department Spring-2011 DSP Laboratory (EELE 4110) Lab#4 Sampling and Quantization OBJECTIVES: When you have completed this assignment,

More information

7.1 Introduction 7.2 Why Digitize Analog Sources? 7.3 The Sampling Process 7.4 Pulse-Amplitude Modulation Time-Division i i Modulation 7.

7.1 Introduction 7.2 Why Digitize Analog Sources? 7.3 The Sampling Process 7.4 Pulse-Amplitude Modulation Time-Division i i Modulation 7. Chapter 7 Digital Representation of Analog Signals Wireless Information Transmission System Lab. Institute of Communications Engineering g National Sun Yat-sen University Contents 7.1 Introduction 7.2

More information

Review of Lecture 2. Data and Signals - Theoretical Concepts. Review of Lecture 2. Review of Lecture 2. Review of Lecture 2. Review of Lecture 2

Review of Lecture 2. Data and Signals - Theoretical Concepts. Review of Lecture 2. Review of Lecture 2. Review of Lecture 2. Review of Lecture 2 Data and Signals - Theoretical Concepts! What are the major functions of the network access layer? Reference: Chapter 3 - Stallings Chapter 3 - Forouzan Study Guide 3 1 2! What are the major functions

More information

COMP211 Physical Layer

COMP211 Physical Layer COMP211 Physical Layer Data and Computer Communications 7th edition William Stallings Prentice Hall 2004 Computer Networks 5th edition Andrew S.Tanenbaum, David J.Wetherall Pearson 2011 Material adapted

More information

Point-to-Point Communications

Point-to-Point Communications Point-to-Point Communications Key Aspects of Communication Voice Mail Tones Alphabet Signals Air Paper Media Language English/Hindi English/Hindi Outline of Point-to-Point Communication 1. Signals basic

More information

The quality of the transmission signal The characteristics of the transmission medium. Some type of transmission medium is required for transmission:

The quality of the transmission signal The characteristics of the transmission medium. Some type of transmission medium is required for transmission: Data Transmission The successful transmission of data depends upon two factors: The quality of the transmission signal The characteristics of the transmission medium Some type of transmission medium is

More information

Communication Systems Lecture-12: Delta Modulation and PTM

Communication Systems Lecture-12: Delta Modulation and PTM Communication Systems Lecture-12: Delta Modulation and PTM Department of Electrical and Computer Engineering Lebanese American University chadi.abourjeily@lau.edu.lb October 26, 2017 Delta Modulation (1)

More information

Communications IB Paper 6 Handout 3: Digitisation and Digital Signals

Communications IB Paper 6 Handout 3: Digitisation and Digital Signals Communications IB Paper 6 Handout 3: Digitisation and Digital Signals Jossy Sayir Signal Processing and Communications Lab Department of Engineering University of Cambridge jossy.sayir@eng.cam.ac.uk Lent

More information

In this lecture. System Model Power Penalty Analog transmission Digital transmission

In this lecture. System Model Power Penalty Analog transmission Digital transmission System Model Power Penalty Analog transmission Digital transmission In this lecture Analog Data Transmission vs. Digital Data Transmission Analog to Digital (A/D) Conversion Digital to Analog (D/A) Conversion

More information

Practical Approach of Producing Delta Modulation and Demodulation

Practical Approach of Producing Delta Modulation and Demodulation IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 11, Issue 3, Ver. II (May-Jun.2016), PP 87-94 www.iosrjournals.org Practical Approach of

More information

Chapter-15. Communication systems -1 mark Questions

Chapter-15. Communication systems -1 mark Questions Chapter-15 Communication systems -1 mark Questions 1) What are the three main units of a Communication System? 2) What is meant by Bandwidth of transmission? 3) What is a transducer? Give an example. 4)

More information

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Simplex. Direct link.

Terminology (1) Chapter 3. Terminology (3) Terminology (2) Transmitter Receiver Medium. Data Transmission. Simplex. Direct link. Chapter 3 Data Transmission Terminology (1) Transmitter Receiver Medium Guided medium e.g. twisted pair, optical fiber Unguided medium e.g. air, water, vacuum Corneliu Zaharia 2 Corneliu Zaharia Terminology

More information

Announcement : Wireless Networks Lecture 3: Physical Layer. A Reminder about Prerequisites. Outline. Page 1

Announcement : Wireless Networks Lecture 3: Physical Layer. A Reminder about Prerequisites. Outline. Page 1 Announcement 18-759: Wireless Networks Lecture 3: Physical Layer Peter Steenkiste Departments of Computer Science and Electrical and Computer Engineering Spring Semester 2010 http://www.cs.cmu.edu/~prs/wirelesss10/

More information

DIGITAL COMMUNICATION

DIGITAL COMMUNICATION DIGITAL COMMUNICATION TRAINING LAB Digital communication has emerged to augment or replace the conventional analog systems, which had been used widely a few decades back. Digital communication has demonstrated

More information

CSE 123: Computer Networks Alex C. Snoeren. Project 1 out Today, due 10/26!

CSE 123: Computer Networks Alex C. Snoeren. Project 1 out Today, due 10/26! CSE 123: Computer Networks Alex C. Snoeren Project 1 out Today, due 10/26! Signaling Types of physical media Shannon s Law and Nyquist Limit Encoding schemes Clock recovery Manchester, NRZ, NRZI, etc.

More information

Data Transmission (II)

Data Transmission (II) Agenda Lecture (02) Data Transmission (II) Analog and digital signals Analog and Digital transmission Transmission impairments Channel capacity Shannon formulas Dr. Ahmed ElShafee 1 Dr. Ahmed ElShafee,

More information

3.6 Intersymbol interference. 1 Your site here

3.6 Intersymbol interference. 1 Your site here 3.6 Intersymbol intererence 1 3.6 Intersymbol intererence what is intersymbol intererence and what cause ISI 1. The absolute bandwidth o rectangular multilevel pulses is ininite. The channels bandwidth

More information

Lecture 3: Data Transmission

Lecture 3: Data Transmission Lecture 3: Data Transmission 1 st semester 1439-2017 1 By: Elham Sunbu OUTLINE Data Transmission DATA RATE LIMITS Transmission Impairments Examples DATA TRANSMISSION The successful transmission of data

More information

Nonuniform multi level crossing for signal reconstruction

Nonuniform multi level crossing for signal reconstruction 6 Nonuniform multi level crossing for signal reconstruction 6.1 Introduction In recent years, there has been considerable interest in level crossing algorithms for sampling continuous time signals. Driven

More information

Data Communications and Networks

Data Communications and Networks Data Communications and Networks Abdul-Rahman Mahmood http://alphapeeler.sourceforge.net http://pk.linkedin.com/in/armahmood abdulmahmood-sss twitter.com/alphapeeler alphapeeler.sourceforge.net/pubkeys/pkey.htm

More information

Chapter-2 SAMPLING PROCESS

Chapter-2 SAMPLING PROCESS Chapter-2 SAMPLING PROCESS SAMPLING: A message signal may originate from a digital or analog source. If the message signal is analog in nature, then it has to be converted into digital form before it can

More information

SIGNALS AND SYSTEMS LABORATORY 13: Digital Communication

SIGNALS AND SYSTEMS LABORATORY 13: Digital Communication SIGNALS AND SYSTEMS LABORATORY 13: Digital Communication INTRODUCTION Digital Communication refers to the transmission of binary, or digital, information over analog channels. In this laboratory you will

More information

Theory of Telecommunications Networks

Theory of Telecommunications Networks Theory of Telecommunications Networks Anton Čižmár Ján Papaj Department of electronics and multimedia telecommunications CONTENTS Preface... 5 1 Introduction... 6 1.1 Mathematical models for communication

More information

MODULATION AND MULTIPLE ACCESS TECHNIQUES

MODULATION AND MULTIPLE ACCESS TECHNIQUES 1 MODULATION AND MULTIPLE ACCESS TECHNIQUES Networks and Communication Department Dr. Marwah Ahmed Outlines 2 Introduction Digital Transmission Digital Modulation Digital Transmission of Analog Signal

More information

Jitter in Digital Communication Systems, Part 1

Jitter in Digital Communication Systems, Part 1 Application Note: HFAN-4.0.3 Rev.; 04/08 Jitter in Digital Communication Systems, Part [Some parts of this application note first appeared in Electronic Engineering Times on August 27, 200, Issue 8.] AVAILABLE

More information

Multiplexing Concepts and Introduction to BISDN. Professor Richard Harris

Multiplexing Concepts and Introduction to BISDN. Professor Richard Harris Multiplexing Concepts and Introduction to BISDN Professor Richard Harris Objectives Define what is meant by multiplexing and demultiplexing Identify the main types of multiplexing Space Division Time Division

More information

Wireless Communications

Wireless Communications 2. Physical Layer DIN/CTC/UEM 2018 Periodic Signal Periodic signal: repeats itself in time, that is g(t) = g(t + T ) in which T (given in seconds [s]) is the period of the signal g(t) The number of cycles

More information