Communication Theory II
|
|
- Kelly Robbins
- 5 years ago
- Views:
Transcription
1 Communication Theory II Lecture 14: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 25 th,
2 Previous Lecture: Source Code Generation: Lossless Data Compression Techniques o Limits of Lossless data compression: The entropy of the source establishes the fundamental limit on the removal of redundancy from the data o Prefix code Shannon s 1 st theorem; source coding theorem: L H S Huffman code (a prefix code with optimal shortest expected length) o Lempel-Ziv code: standard algorithm for file compression 2
3 Lecture Outlines osource Code Transmission over a Noisy Channel: Channel Coding Discrete memoryless channel Example: Binary symmetric channel Mutual Information Channel Capacity Channel-Coding Theorem 3
4 Channel coding (encoder/decoder) ochannel coding consists of : Channel encoder: mapping the incoming data sequence into a channel input sequence Channel decoder: inverse mapping the channel output sequence into an output data sequence in such a way that the overall effect of channel noise on the system is minimized 4
5 A simple Diagram for Channel Coding ofor simplicity, we have not include the source encoder/decoder in figure 5
6 A Discrete Memoryless Channel oa statistical model with an input R.V. X and an output R.V. Y that is a noisy version of X oevery unit of time, the channel accepts an input symbol X selected from an alphabet X and, in response, it emits an output symbol Y from an alphabet Y odiscrete: both of the alphabets X and Y have finite sizes omemoryless: the current output symbol depends only on the current input symbol and not any previous or future symbol
7 Example: Binary Symmetric Channel (BSC) Transition Probability Diagram obinary: X and Y of size (cardinality) of 2 osymmetric: the probability of receiving 1 if 0 is sent is the same as the probability of receiving 0 if 1 is sent 7
8 Channel Matrix (Stochastic Matrix) ocompletely specifies a channel: The transmission medium Signal specifications (binary, r-ary, simplex, orthogonal, kind of receiver used: receiver determines the error probabilities) o Defined by a set of transition probabilities: Where is the conditional probability that y k is received when x j is transmitted obsc: Channel Matrix 8
9 Channel Matrix Properties o Each row corresponds to a fixed channel input o Each column corresponds to a fixed channel output o Satisfy the axioms of probability: and The sum of elements along any row is always equal to one 9
10 Average uncertainity about the channel input: H(x) olet P(x i ) be the probability of the event that the channel input X=x i Prior probability: before observing the channel output oh(x): average uncertainty about the channel input (i.e., before observing the channel output op(y j )= i P(y j x i )P(x i )= i P(y j,x i ): marginal probability distribution of the output R.V Y 10
11 Equivocation of x with respect to y : H(x y) olet P(x i y j ) be a conditional probability that x i is transmitted when y j is received There is an uncertainty of log P(x i y j ) about x i when y j is received 1 oh(x y): Equivocation (to use unclear language) of x with respect to y Average uncertainty about the channel input after observing the channel output Average loss of information about a transmitted symbol when a symbol is received Average uncertainty about a transmitted symbol x when a symbol y is received P(x i,y j )=P(x i y j )P(x i ) is the joint probability distribution of the R.Vs X and Y ofor error free channel: average information is H(x) bits Due to noise, we lose an average of H(x y) bits per symbol 11
12 Mutual Information of a channel I(x; y) bits per symbol o Mutual Information of a channel (symmetric) Amount of information that the receiver receives A measure of the uncertainity about the channel input, which is resolved by observing the channel output I(x; y)=h(x) H(x y) bits per symbol and j P x i, y j = P(x i ) A measure of the uncertainity about the channel output, which is resolved by observing the channel input I(x; y)=i(y; x)=h(y) H(y x) bits per symbol For error free channel, no loss of information I(x; y)=i(y; x)=h(x)=h(y) bits per symbol 12
13 ononnegative Some Properties of Mutual Information (why?) H(x y) H(x): We cannot lose information, on the average, by observing the output of a channel more than the average transmitted information H(x) owhen does I(x; y)=0? When we receive no information at the receiver; i.e, lose all input information on the channel: H(x y) = H(x) This is satisfied if the input and output symbols of the channel are statistically independent (why?) If the input is given, this gives no idea about the output received information is zero 13
14 Channel Capacity o Property of a particular physical channel (i.e., function of channel matrix values; function of signal power and channel noise) that measure its intrinsic ability to convey information o Maximum information that can be transmitted by one symbol over the channel o The maximum value of the mutual information of the channel w.r.t P(x i ) r s r c C s o Let a source that produces r s symbols/second the enocoder transmits r c code digit/sec Code rate of the channel encoder R code = r s r c symbol/code digit(bit) o Channel Capacity per second (critical rate)=r c C s bits/second 14
15 Example: Channel Capacity of BSC o Let then: o For a given P e, Ω P e is fixed, and 15
16 Recall the Entropy of Bernoulli Random Variable p 0 1 p 0 H(X) H(X) Note: lim x 0 x log x = 0 Entropy function H(Po) 16
17 Channel Capacity of BSC o o is maximum if \ o For error free communications: C s max=1 bit/symbol We can transmit 1 bit of information per binary digit (symbol) for error free communications One binary bit can convey one of the two equiprobable messages o C s max=1 bit/symbol at P e =0 (noiseless) or P e =1 (good as noiseless. why? Output is known) o C s min=0 bit/symbol at P e =0.5 (why?) The transmitted and received signals are statistically independent (P(x)=P(y) (why?) If 0 is received, 0 and 1 are equally likely to have been transmitted information=0 17
18 Magnitude of Channel Capacity o o I(x; y)=h(x) H(x y) bits per symbol o C S is always less than or equal to the average information per input symbol o Example 1: C S H x bit/binary symbol; H(x y) 0 If we use binary symbols as input: the maximum value of H(x)=1 bit (when P(x o )=P(x 1 )) C S 1 bit/binary symbol o Example 2: If we use r-ary symbols as input: the maximum value of H(x)=1 r-ary unit (in case of equiprobable) C S 1 r-ary unit/binary symbol 18
19 Source Code Transmission over a Noisy Channel o A source with an entropy H(m) bits Can we achieve error free communication? If so, what is the cost? oredundancy can help to compact noise Adding a single parity check code (ensures that the total number of 1 s is always even or odd) can check a single error cost: L new = L+1 which means a decreased efficiency oimmunity against noise can be increased by increasing redundancy Example: adding a number of repeated guards for each binary symbol oshannon shows that it is theoretically possible to achieve error-free communications by adding sufficient redundancy 19
20 Shannon s 2 nd Theorem Channel-Coding Theorem r s r c S H(s) C s o Channel coding theorem: for error free communication (arbitrary small error) Source information rate channel capacity per second r s H(s) bits/sec r c C s bits/sec Otherwise if H(s) > r c C s, then error occurs o Messages from the source with entropy H(S) must be encoded by binary codes with a word length at least H(S)/C s o If condition satisfies, then good codes do exist o For discrete memoryless channels that produces equally likely symbols H(s)= 1 bit/symbol For error free communications R code C s bit/symbol We do not need to lower R code too much (a small reduction of the rate of information transmission is enough) Channel capacity C s Channel Capacity per second=r c C s bits/second r c code symbol/ sec r s symbol/ sec H(s) source entropy bit/symbol Source information rate=r s H(s) bits/sec Code rate R code = r s r c 20
21 Questions 21
Communication Theory II
Communication Theory II Lecture 13: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 22 th, 2015 1 o Source Code Generation Lecture Outlines Source Coding
More informationInformation Theory and Communication Optimal Codes
Information Theory and Communication Optimal Codes Ritwik Banerjee rbanerjee@cs.stonybrook.edu c Ritwik Banerjee Information Theory and Communication 1/1 Roadmap Examples and Types of Codes Kraft Inequality
More information4. Which of the following channel matrices respresent a symmetric channel? [01M02] 5. The capacity of the channel with the channel Matrix
Send SMS s : ONJntuSpeed To 9870807070 To Recieve Jntu Updates Daily On Your Mobile For Free www.strikingsoon.comjntu ONLINE EXMINTIONS [Mid 2 - dc] http://jntuk.strikingsoon.com 1. Two binary random
More informationComputing and Communications 2. Information Theory -Channel Capacity
1896 1920 1987 2006 Computing and Communications 2. Information Theory -Channel Capacity Ying Cui Department of Electronic Engineering Shanghai Jiao Tong University, China 2017, Autumn 1 Outline Communication
More informationA Brief Introduction to Information Theory and Lossless Coding
A Brief Introduction to Information Theory and Lossless Coding 1 INTRODUCTION This document is intended as a guide to students studying 4C8 who have had no prior exposure to information theory. All of
More informationChapter 1 INTRODUCTION TO SOURCE CODING AND CHANNEL CODING. Whether a source is analog or digital, a digital communication
1 Chapter 1 INTRODUCTION TO SOURCE CODING AND CHANNEL CODING 1.1 SOURCE CODING Whether a source is analog or digital, a digital communication system is designed to transmit information in digital form.
More informationComm. 502: Communication Theory. Lecture 6. - Introduction to Source Coding
Comm. 50: Communication Theory Lecture 6 - Introduction to Source Coding Digital Communication Systems Source of Information User of Information Source Encoder Source Decoder Channel Encoder Channel Decoder
More informationIntroduction to Source Coding
Comm. 52: Communication Theory Lecture 7 Introduction to Source Coding - Requirements of source codes - Huffman Code Length Fixed Length Variable Length Source Code Properties Uniquely Decodable allow
More informationCOPYRIGHTED MATERIAL. Introduction. 1.1 Communication Systems
1 Introduction The reliable transmission of information over noisy channels is one of the basic requirements of digital information and communication systems. Here, transmission is understood both as transmission
More informationEntropy, Coding and Data Compression
Entropy, Coding and Data Compression Data vs. Information yes, not, yes, yes, not not In ASCII, each item is 3 8 = 24 bits of data But if the only possible answers are yes and not, there is only one bit
More informationLECTURE VI: LOSSLESS COMPRESSION ALGORITHMS DR. OUIEM BCHIR
1 LECTURE VI: LOSSLESS COMPRESSION ALGORITHMS DR. OUIEM BCHIR 2 STORAGE SPACE Uncompressed graphics, audio, and video data require substantial storage capacity. Storing uncompressed video is not possible
More informationMultimedia Systems Entropy Coding Mahdi Amiri February 2011 Sharif University of Technology
Course Presentation Multimedia Systems Entropy Coding Mahdi Amiri February 2011 Sharif University of Technology Data Compression Motivation Data storage and transmission cost money Use fewest number of
More informationDEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK. Subject Name: Information Coding Techniques UNIT I INFORMATION ENTROPY FUNDAMENTALS
DEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK Subject Name: Year /Sem: II / IV UNIT I INFORMATION ENTROPY FUNDAMENTALS PART A (2 MARKS) 1. What is uncertainty? 2. What is prefix coding? 3. State the
More informationInformation Theory and Huffman Coding
Information Theory and Huffman Coding Consider a typical Digital Communication System: A/D Conversion Sampling and Quantization D/A Conversion Source Encoder Source Decoder bit stream bit stream Channel
More informationCOURSE MATERIAL Subject Name: Communication Theory UNIT V
NH-67, TRICHY MAIN ROAD, PULIYUR, C.F. - 639114, KARUR DT. DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING COURSE MATERIAL Subject Name: Communication Theory Subject Code: 080290020 Class/Sem:
More informationECEn 665: Antennas and Propagation for Wireless Communications 131. s(t) = A c [1 + αm(t)] cos (ω c t) (9.27)
ECEn 665: Antennas and Propagation for Wireless Communications 131 9. Modulation Modulation is a way to vary the amplitude and phase of a sinusoidal carrier waveform in order to transmit information. When
More informationPart A: Question & Answers UNIT I AMPLITUDE MODULATION
PANDIAN SARASWATHI YADAV ENGINEERING COLLEGE DEPARTMENT OF ELECTRONICS & COMMUNICATON ENGG. Branch: ECE EC6402 COMMUNICATION THEORY Semester: IV Part A: Question & Answers UNIT I AMPLITUDE MODULATION 1.
More informationLecture5: Lossless Compression Techniques
Fixed to fixed mapping: we encoded source symbols of fixed length into fixed length code sequences Fixed to variable mapping: we encoded source symbols of fixed length into variable length code sequences
More informationEE303: Communication Systems
EE303: Communication Systems Professor A. Manikas Chair of Communications and Array Processing Imperial College London An Overview of Fundamentals: Channels, Criteria and Limits Prof. A. Manikas (Imperial
More informationCourse Developer: Ranjan Bose, IIT Delhi
Course Title: Coding Theory Course Developer: Ranjan Bose, IIT Delhi Part I Information Theory and Source Coding 1. Source Coding 1.1. Introduction to Information Theory 1.2. Uncertainty and Information
More informationChapter 1 Coding for Reliable Digital Transmission and Storage
Wireless Information Transmission System Lab. Chapter 1 Coding for Reliable Digital Transmission and Storage Institute of Communications Engineering National Sun Yat-sen University 1.1 Introduction A major
More informationLecture 1 Introduction
Lecture 1 Introduction I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw September 22, 2015 1 / 46 I-Hsiang Wang IT Lecture 1 Information Theory Information
More informationModule 8: Video Coding Basics Lecture 40: Need for video coding, Elements of information theory, Lossless coding. The Lecture Contains:
The Lecture Contains: The Need for Video Coding Elements of a Video Coding System Elements of Information Theory Symbol Encoding Run-Length Encoding Entropy Encoding file:///d /...Ganesh%20Rana)/MY%20COURSE_Ganesh%20Rana/Prof.%20Sumana%20Gupta/FINAL%20DVSP/lecture%2040/40_1.htm[12/31/2015
More informationCoding Techniques and the Two-Access Channel
Coding Techniques and the Two-Access Channel A.J. Han VINCK Institute for Experimental Mathematics, University of Duisburg-Essen, Germany email: Vinck@exp-math.uni-essen.de Abstract. We consider some examples
More informationCommunication Theory II
Communication Theory II Lecture 18: Pulse Code Modulation Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt April 19 th, 2015 1 Lecture Outlines opulse Code Modulation (PCM) Sampling and
More informationDigital Television Lecture 5
Digital Television Lecture 5 Forward Error Correction (FEC) Åbo Akademi University Domkyrkotorget 5 Åbo 8.4. Error Correction in Transmissions Need for error correction in transmissions Loss of data during
More informationIMPERIAL COLLEGE of SCIENCE, TECHNOLOGY and MEDICINE, DEPARTMENT of ELECTRICAL and ELECTRONIC ENGINEERING.
IMPERIAL COLLEGE of SCIENCE, TECHNOLOGY and MEDICINE, DEPARTMENT of ELECTRICAL and ELECTRONIC ENGINEERING. COMPACT LECTURE NOTES on COMMUNICATION THEORY. Prof. Athanassios Manikas, version Spring 22 Digital
More informationECE 4400:693 - Information Theory
ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 1: Introduction & Overview Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Information Theory 1 / 26 Outline 1 Course Information 2 Course Overview
More informationChannel Coding/Decoding. Hamming Method
Channel Coding/Decoding Hamming Method INFORMATION TRANSFER ACROSS CHANNELS Sent Received messages symbols messages source encoder Source coding Channel coding Channel Channel Source decoder decoding decoding
More informationRab Nawaz. Prof. Zhang Wenyi
Rab Nawaz PhD Scholar (BL16006002) School of Information Science and Technology University of Science and Technology of China, Hefei Email: rabnawaz@mail.ustc.edu.cn Submitted to Prof. Zhang Wenyi wenyizha@ustc.edu.cn
More informationSolutions to Assignment-2 MOOC-Information Theory
Solutions to Assignment-2 MOOC-Information Theory 1. Which of the following is a prefix-free code? a) 01, 10, 101, 00, 11 b) 0, 11, 01 c) 01, 10, 11, 00 Solution:- The codewords of (a) are not prefix-free
More informationMATHEMATICS IN COMMUNICATIONS: INTRODUCTION TO CODING. A Public Lecture to the Uganda Mathematics Society
Abstract MATHEMATICS IN COMMUNICATIONS: INTRODUCTION TO CODING A Public Lecture to the Uganda Mathematics Society F F Tusubira, PhD, MUIPE, MIEE, REng, CEng Mathematical theory and techniques play a vital
More informationSyllabus. osmania university UNIT - I UNIT - II UNIT - III CHAPTER - 1 : INTRODUCTION TO DIGITAL COMMUNICATION CHAPTER - 3 : INFORMATION THEORY
i Syllabus osmania university UNIT - I CHAPTER - 1 : INTRODUCTION TO Elements of Digital Communication System, Comparison of Digital and Analog Communication Systems. CHAPTER - 2 : DIGITAL TRANSMISSION
More informationComm 502: Communication Theory
Comm 50: Communication Theory Prof. Dean of the faculty of IET The German University in Cairo 1 COMM 50: Communication Theory Instructor: Ahmed El-Mahdy Office : C3.319 Lecture Time: Sat. nd Slot Office
More informationDigital Communication Systems ECS 452
Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th 2. Source Coding 1 Office Hours: BKD, 6th floor of Sirindhralai building Monday 10:00-10:40 Tuesday 12:00-12:40
More informationCOMM901 Source Coding and Compression Winter Semester 2013/2014. Midterm Exam
German University in Cairo - GUC Faculty of Information Engineering & Technology - IET Department of Communication Engineering Dr.-Ing. Heiko Schwarz COMM901 Source Coding and Compression Winter Semester
More informationDEGRADED broadcast channels were first studied by
4296 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 54, NO 9, SEPTEMBER 2008 Optimal Transmission Strategy Explicit Capacity Region for Broadcast Z Channels Bike Xie, Student Member, IEEE, Miguel Griot,
More informationMulticasting over Multiple-Access Networks
ing oding apacity onclusions ing Department of Electrical Engineering and omputer Sciences University of alifornia, Berkeley May 9, 2006 EE 228A Outline ing oding apacity onclusions 1 2 3 4 oding 5 apacity
More informationSpeech Coding in the Frequency Domain
Speech Coding in the Frequency Domain Speech Processing Advanced Topics Tom Bäckström Aalto University October 215 Introduction The speech production model can be used to efficiently encode speech signals.
More informationOutline. Communications Engineering 1
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband channels Signal space representation Optimal
More informationScheduling in omnidirectional relay wireless networks
Scheduling in omnidirectional relay wireless networks by Shuning Wang A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Applied Science
More informationELEC3028 (EL334) Digital Transmission
ELEC3028 (EL334) Digital Transmission Half of the unit: Information Theory MODEM (modulator and demodulator) Professor Sheng Chen: Building 53, Room 4005 E-mail: sqc@ecs.soton.ac.uk Lecture notes from:
More informationECE Advanced Communication Theory, Spring 2007 Midterm Exam Monday, April 23rd, 6:00-9:00pm, ELAB 325
C 745 - Advanced Communication Theory, Spring 2007 Midterm xam Monday, April 23rd, 600-900pm, LAB 325 Overview The exam consists of five problems for 150 points. The points for each part of each problem
More informationCompression. Encryption. Decryption. Decompression. Presentation of Information to client site
DOCUMENT Anup Basu Audio Image Video Data Graphics Objectives Compression Encryption Network Communications Decryption Decompression Client site Presentation of Information to client site Multimedia -
More informationFREDRIK TUFVESSON ELECTRICAL AND INFORMATION TECHNOLOGY
1 Information Transmission Chapter 5, Block codes FREDRIK TUFVESSON ELECTRICAL AND INFORMATION TECHNOLOGY 2 Methods of channel coding For channel coding (error correction) we have two main classes of codes,
More informationKINGS COLLEGE OF ENGINEERING DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING QUESTION BANK
KINGS COLLEGE OF ENGINEERING DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING QUESTION BANK SUB.NAME : COMMUNICATION THEORY SUB.CODE: EC1252 YEAR : II SEMESTER : IV UNIT I AMPLITUDE MODULATION SYSTEMS
More informationA Bit of network information theory
Š#/,% 0/,94%#(.)15% A Bit of network information theory Suhas Diggavi 1 Email: suhas.diggavi@epfl.ch URL: http://licos.epfl.ch Parts of talk are joint work with S. Avestimehr 2, S. Mohajer 1, C. Tian 3,
More information# 12 ECE 253a Digital Image Processing Pamela Cosman 11/4/11. Introductory material for image compression
# 2 ECE 253a Digital Image Processing Pamela Cosman /4/ Introductory material for image compression Motivation: Low-resolution color image: 52 52 pixels/color, 24 bits/pixel 3/4 MB 3 2 pixels, 24 bits/pixel
More informationEE 435/535: Error Correcting Codes Project 1, Fall 2009: Extended Hamming Code. 1 Introduction. 2 Extended Hamming Code: Encoding. 1.
EE 435/535: Error Correcting Codes Project 1, Fall 2009: Extended Hamming Code Project #1 is due on Tuesday, October 6, 2009, in class. You may turn the project report in early. Late projects are accepted
More informationCommunications Overhead as the Cost of Constraints
Communications Overhead as the Cost of Constraints J. Nicholas Laneman and Brian. Dunn Department of Electrical Engineering University of Notre Dame Email: {jnl,bdunn}@nd.edu Abstract This paper speculates
More informationHamming Codes as Error-Reducing Codes
Hamming Codes as Error-Reducing Codes William Rurik Arya Mazumdar Abstract Hamming codes are the first nontrivial family of error-correcting codes that can correct one error in a block of binary symbols.
More informationWireless Network Information Flow
Š#/,% 0/,94%#(.)15% Wireless Network Information Flow Suhas iggavi School of Computer and Communication Sciences, Laboratory for Information and Communication Systems (LICOS), EPFL Email: suhas.diggavi@epfl.ch
More informationDIGITAL COMMINICATIONS
Code No: R346 R Set No: III B.Tech. I Semester Regular and Supplementary Examinations, December - 23 DIGITAL COMMINICATIONS (Electronics and Communication Engineering) Time: 3 Hours Max Marks: 75 Answer
More informationThe ternary alphabet is used by alternate mark inversion modulation; successive ones in data are represented by alternating ±1.
Alphabets EE 387, Notes 2, Handout #3 Definition: An alphabet is a discrete (usually finite) set of symbols. Examples: B = {0,1} is the binary alphabet T = { 1,0,+1} is the ternary alphabet X = {00,01,...,FF}
More informationSummary of Basic Concepts
Transmission Summary of Basic Concepts Sender Channel Receiver Dr. Christian Rohner Encoding Modulation Demodulation Decoding Bits Symbols Noise Terminology Communications Research Group Bandwidth [Hz]
More informationSolutions to Information Theory Exercise Problems 5 8
Solutions to Information Theory Exercise roblems 5 8 Exercise 5 a) n error-correcting 7/4) Hamming code combines four data bits b 3, b 5, b 6, b 7 with three error-correcting bits: b 1 = b 3 b 5 b 7, b
More informationThe idea of similarity is through the Hamming
Hamming distance A good channel code is designed so that, if a few bit errors occur in transmission, the output can still be identified as the correct input. This is possible because although incorrect,
More informationBlock Markov Encoding & Decoding
1 Block Markov Encoding & Decoding Deqiang Chen I. INTRODUCTION Various Markov encoding and decoding techniques are often proposed for specific channels, e.g., the multi-access channel (MAC) with feedback,
More informationQUESTION BANK (VI SEM ECE) (DIGITAL COMMUNICATION)
QUESTION BANK (VI SEM ECE) (DIGITAL COMMUNICATION) UNIT-I: PCM & Delta modulation system Q.1 Explain the difference between cross talk & intersymbol interference. Q.2 What is Quantization error? How does
More informationSHANNON S source channel separation theorem states
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 9, SEPTEMBER 2009 3927 Source Channel Coding for Correlated Sources Over Multiuser Channels Deniz Gündüz, Member, IEEE, Elza Erkip, Senior Member,
More informationECE 8771, Information Theory & Coding for Digital Communications Summer 2010 Syllabus & Outline (Draft 1 - May 12, 2010)
ECE 8771, Information Theory & Coding for Digital Communications Summer 2010 Syllabus & Outline (Draft 1 - May 12, 2010) Instructor: Kevin Buckley, Tolentine 433a, 610-519-5658 (W), 610-519-4436 (F), buckley@ece.vill.edu,
More informationMAS160: Signals, Systems & Information for Media Technology. Problem Set 4. DUE: October 20, 2003
MAS160: Signals, Systems & Information for Media Technology Problem Set 4 DUE: October 20, 2003 Instructors: V. Michael Bove, Jr. and Rosalind Picard T.A. Jim McBride Problem 1: Simple Psychoacoustic Masking
More informationFundamentals of Digital Communications and Data Transmission
Fundamentals of Digital Communications and Data Transmission 29 th October 2008 Abdullah Al-Meshal Overview Introduction Communication systems Digital communication system Importance of Digital transmission
More informationError Control Coding. Aaron Gulliver Dept. of Electrical and Computer Engineering University of Victoria
Error Control Coding Aaron Gulliver Dept. of Electrical and Computer Engineering University of Victoria Topics Introduction The Channel Coding Problem Linear Block Codes Cyclic Codes BCH and Reed-Solomon
More informationphotons photodetector t laser input current output current
6.962 Week 5 Summary: he Channel Presenter: Won S. Yoon March 8, 2 Introduction he channel was originally developed around 2 years ago as a model for an optical communication link. Since then, a rather
More information2.1. General Purpose Run Length Encoding Relative Encoding Tokanization or Pattern Substitution
2.1. General Purpose There are many popular general purpose lossless compression techniques, that can be applied to any type of data. 2.1.1. Run Length Encoding Run Length Encoding is a compression technique
More informationDCSP-1: Introduction. Jianfeng Feng. Department of Computer Science Warwick Univ., UK
DCSP-1: Introduction Jianfeng Feng Department of Computer Science Warwick Univ., UK Jianfeng.feng@warwick.ac.uk http://www.dcs.warwick.ac.uk/~feng/dcsp.html Time Monday (L) 14.00-15.00 CS 1.01 Tuesday
More informationExercises to Chapter 2 solutions
Exercises to Chapter 2 solutions 1 Exercises to Chapter 2 solutions E2.1 The Manchester code was first used in Manchester Mark 1 computer at the University of Manchester in 1949 and is still used in low-speed
More informationCoding for Efficiency
Let s suppose that, over some channel, we want to transmit text containing only 4 symbols, a, b, c, and d. Further, let s suppose they have a probability of occurrence in any block of text we send as follows
More informationTSKS01 Digital Communication Lecture 1
TSKS01 Digital Communication Lecture 1 Introduction, Repetition, Channels as Filters, Complex-baseband representation Emil Björnson Department of Electrical Engineering (ISY) Division of Communication
More informationM.Sc. Thesis. Optimization of the Belief Propagation algorithm for Luby Transform decoding over the Binary Erasure Channel. Marta Alvarez Guede
Circuits and Systems Mekelweg 4, 2628 CD Delft The Netherlands http://ens.ewi.tudelft.nl/ CAS-2011-07 M.Sc. Thesis Optimization of the Belief Propagation algorithm for Luby Transform decoding over the
More informationINSTITUTE OF AERONAUTICAL ENGINEERING (Autonomous) Dundigal, Hyderabad
INSTITUTE OF AERONAUTICAL ENGINEERING (Autonomous) Dundigal, Hyderabad - 500 03 ELECTRONICS AND COMMUNICATION ENGINEERING TUTORIAL QUESTION BANK Name : DIGITAL COMMUNICATIONS Code : A6020 Class : III -
More informationMultiuser Information Theory and Wireless Communications. Professor in Charge: Toby Berger Principal Lecturer: Jun Chen
Multiuser Information Theory and Wireless Communications Professor in Charge: Toby Berger Principal Lecturer: Jun Chen Where and When? 1 Good News No homework. No exam. 2 Credits:1-2 One credit: submit
More informationLecture 13 February 23
EE/Stats 376A: Information theory Winter 2017 Lecture 13 February 23 Lecturer: David Tse Scribe: David L, Tong M, Vivek B 13.1 Outline olar Codes 13.1.1 Reading CT: 8.1, 8.3 8.6, 9.1, 9.2 13.2 Recap -
More informationFundamentals of Wireless Communication
Communication Technology Laboratory Prof. Dr. H. Bölcskei Sternwartstrasse 7 CH-8092 Zürich Fundamentals of Wireless Communication Homework 5 Solutions Problem 1 Simulation of Error Probability When implementing
More informationECE 556 BASICS OF DIGITAL SPEECH PROCESSING. Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2
ECE 556 BASICS OF DIGITAL SPEECH PROCESSING Assıst.Prof.Dr. Selma ÖZAYDIN Spring Term-2017 Lecture 2 Analog Sound to Digital Sound Characteristics of Sound Amplitude Wavelength (w) Frequency ( ) Timbre
More informationIntro to coding and convolutional codes
Intro to coding and convolutional codes Lecture 11 Vladimir Stojanović 6.973 Communication System Design Spring 2006 Massachusetts Institute of Technology 802.11a Convolutional Encoder Rate 1/2 convolutional
More informationCommunication Theory II
Communication Theory II Lecture 17: Conversion of Analog Waveforms into Coded Pulses Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt April 16 th, 2015 1 opulse Modulation Analog Pulse
More informationModulation and Coding Tradeoffs
0 Modulation and Coding Tradeoffs Contents 1 1. Design Goals 2. Error Probability Plane 3. Nyquist Minimum Bandwidth 4. Shannon Hartley Capacity Theorem 5. Bandwidth Efficiency Plane 6. Modulation and
More informationProblem Sheet 1 Probability, random processes, and noise
Problem Sheet 1 Probability, random processes, and noise 1. If F X (x) is the distribution function of a random variable X and x 1 x 2, show that F X (x 1 ) F X (x 2 ). 2. Use the definition of the cumulative
More informationChannel Concepts CS 571 Fall Kenneth L. Calvert
Channel Concepts CS 571 Fall 2006 2006 Kenneth L. Calvert What is a Channel? Channel: a means of transmitting information A means of communication or expression Webster s NCD Aside: What is information...?
More informationRADIO SYSTEMS ETIN15. Channel Coding. Ove Edfors, Department of Electrical and Information Technology
RADIO SYSTEMS ETIN15 Lecture no: 7 Channel Coding Ove Edfors, Department of Electrical and Information Technology Ove.Edfors@eit.lth.se 2016-04-18 Ove Edfors - ETIN15 1 Contents (CHANNEL CODING) Overview
More informationInfo theory and big data
Info theory and big data Typical or not typical, that is the question Han Vinck University Duisburg Essen, Germany September 2016 A.J. Han Vinck, Yerevan, September 2016 Content: big data issues A definition:
More informationIntroduction to Error Control Coding
Introduction to Error Control Coding 1 Content 1. What Error Control Coding Is For 2. How Coding Can Be Achieved 3. Types of Coding 4. Types of Errors & Channels 5. Types of Codes 6. Types of Error Control
More informationDigital Modulation Techniques
Digital Modulation Techniques Digital-to-Analog signals is the next conversion we will discuss in this chapter. These techniques are also called as Digital Modulation techniques. Digital Modulation provides
More informationDCSP-3: Minimal Length Coding. Jianfeng Feng
DCSP-3: Minimal Length Coding Jianfeng Feng Department of Computer Science Warwick Univ., UK Jianfeng.feng@warwick.ac.uk http://www.dcs.warwick.ac.uk/~feng/dcsp.html Automatic Image Caption (better than
More informationETSF15 Physical layer communication. Stefan Höst
ETSF15 Physical layer communication Stefan Höst Physical layer Analog vs digital (Previous lecture) Transmission media Modulation Represent digital data in a continuous world Disturbances, Noise and distortion
More informationHamming Codes and Decoding Methods
Hamming Codes and Decoding Methods Animesh Ramesh 1, Raghunath Tewari 2 1 Fourth year Student of Computer Science Indian institute of Technology Kanpur 2 Faculty of Computer Science Advisor to the UGP
More informationTheory of Telecommunications Networks
Theory of Telecommunications Networks Anton Čižmár Ján Papaj Department of electronics and multimedia telecommunications CONTENTS Preface... 5 1 Introduction... 6 1.1 Mathematical models for communication
More informationLDPC Decoding: VLSI Architectures and Implementations
LDPC Decoding: VLSI Architectures and Implementations Module : LDPC Decoding Ned Varnica varnica@gmail.com Marvell Semiconductor Inc Overview Error Correction Codes (ECC) Intro to Low-density parity-check
More informationENERGY-EFFICIENT ALGORITHMS FOR SENSOR NETWORKS
ENERGY-EFFICIENT ALGORITHMS FOR SENSOR NETWORKS Prepared for: DARPA Prepared by: Krishnan Eswaran, Engineer Cornell University May 12, 2003 ENGRC 350 RESEARCH GROUP 2003 Krishnan Eswaran Energy-Efficient
More informationVARDHAMAN COLLEGE OF ENGINEERING (AUTONOMOUS) Affiliated to JNTUH, Hyderabad ASSIGNMENT QUESTION BANK
VARDHAMAN COLLEGE OF ENGINEERING (AUTONOMOUS) Affiliated to JNTUH, Hyderabad ASSIGNMENT QUESTION BANK Name of the subject: Digital Communications B.Tech/M.Tech/MCA/MBA Subject Code: A1424 Semester: VI
More informationLecture 1: Tue Jan 8, Lecture introduction and motivation
Lecture 1: Tue Jan 8, 2019 Lecture introduction and motivation 1 ECE 6602: Digital Communications GEORGIA INSTITUTE OF TECHNOLOGY, SPRING 2019 PREREQUISITE: ECE 6601. Strong background in probability is
More informationCapacity-Achieving Rateless Polar Codes
Capacity-Achieving Rateless Polar Codes arxiv:1508.03112v1 [cs.it] 13 Aug 2015 Bin Li, David Tse, Kai Chen, and Hui Shen August 14, 2015 Abstract A rateless coding scheme transmits incrementally more and
More informationCoding for the Slepian-Wolf Problem With Turbo Codes
Coding for the Slepian-Wolf Problem With Turbo Codes Jan Bajcsy and Patrick Mitran Department of Electrical and Computer Engineering, McGill University Montréal, Québec, HA A7, Email: {jbajcsy, pmitran}@tsp.ece.mcgill.ca
More informationA Hybrid Technique for Image Compression
Australian Journal of Basic and Applied Sciences, 5(7): 32-44, 2011 ISSN 1991-8178 A Hybrid Technique for Image Compression Hazem (Moh'd Said) Abdel Majid Hatamleh Computer DepartmentUniversity of Al-Balqa
More informationB. Tech. (SEM. VI) EXAMINATION, (2) All question early equal make. (3) In ease of numerical problems assume data wherever not provided.
" 11111111111111111111111111111111111111111111111111111111111111III *U-3091/8400* Printed Pages : 7 TEC - 601! I i B. Tech. (SEM. VI) EXAMINATION, 2007-08 DIGIT AL COMMUNICATION \ V Time: 3 Hours] [Total
More informationSpreading Codes and Characteristics. Error Correction Codes
Spreading Codes and Characteristics and Error Correction Codes Global Navigational Satellite Systems (GNSS-6) Short course, NERTU Prasad Krishnan International Institute of Information Technology, Hyderabad
More informationThe Lempel-Ziv (LZ) lossless compression algorithm was developed by Jacob Ziv (AT&T Bell Labs / Technion Israel) and Abraham Lempel (IBM) in 1978;
Georgia Institute of Technology - Georgia Tech Lorraine ECE 6605 Information Theory Lempel-Ziv Lossless Compresion General comments The Lempel-Ziv (LZ) lossless compression algorithm was developed by Jacob
More informationChannel Coding RADIO SYSTEMS ETIN15. Lecture no: Ove Edfors, Department of Electrical and Information Technology
RADIO SYSTEMS ETIN15 Lecture no: 7 Channel Coding Ove Edfors, Department of Electrical and Information Technology Ove.Edfors@eit.lth.se 2012-04-23 Ove Edfors - ETIN15 1 Contents (CHANNEL CODING) Overview
More information