Lecture5: Lossless Compression Techniques
|
|
- Prosper Golden
- 5 years ago
- Views:
Transcription
1 Fixed to fixed mapping: we encoded source symbols of fixed length into fixed length code sequences Fixed to variable mapping: we encoded source symbols of fixed length into variable length code sequences Source Symbol s k Symbol Probability P k Symbol Code, Word c k Code I Code Word Length l k Symbol Code Word c k Code II Code Word Length l k s 0 / s / s 2 / s 3 /8 2 4 Variable to fixed mapping encode variable length source sequences into a fixed length code words Variable to variable mapping encode variable length source sequences into a variable length code words LECTURES
2 Shannon Fano Code It is a technique for constructing a prefix code based on a set of symbols and their probabilities (estimated or measured) It is suboptimal in the sense that it does not achieve the lowest possible expected code word length The technique was proposed by Claude Shannon and was attributed to Robert Fano Algorithm -Source symbols are listed in order of decreasing probability from left to right 2-Divide the list into two parts, with the total probability (or frequency of occurrence) of the left part being as close to the total of the right as possible 3-The left part of the list is assigned the binary digit 0, and the right part is assigned the digit This means that the codes for the symbols in the first part will all start with 0, and the codes in the second part will all start with 4-Recursively apply the steps 2 and 3 to each of the two halves, subdividing groups and adding bits to the codes until each symbol has become a corresponding code leaf on the tree LECTURES 2
3 Shannon Fano Code Ex: Assume a sequence of alphabet A={ a, b, c, d, e, f } with the following occurrence weights, {9, 8, 6, 5, 4, 2} respectively. Apply Shannon Fano coding and discuss the suboptimality LECTURES 3
4 Shannon Fano Code 7 7 LECTURES 4
5 Shannon Fano Code 5
6 Shannon Fano Code LECTURES 6
7 Shannon Fano Code LECTURES 7
8 Shannon Fano Code 0 0 e 4 f 2 LECTURES 8
9 Shannon Fano Code Symbol Frequency P CODE a b c d e f L K k P k l k L 2x( ) 3x( ) 2.5bits / symbol LECTURES 9
10 Huffman procedure is based on two observations regarding optimum prefix codes: - Symbols that occur more frequently (have a higher probability of occurrence) will have shorter codewords than symbols that occur less frequently 2-The two symbols that occur least frequently will have the same length It is commonly used for lossless compressions Huffman code is a prefix, variable length code that can achieve the shortest average code length for a given input alphabet with pmf In general, Huffman coding is a form of statistical coding as not all characters occur with the same frequency (Probability) The process of finding the optimal code was algorithmized by Huffman A code is an optimal prefix code if it has the same average codeword length as a Huffman code for the given pmf 0
11 Ex: Consider the following short text Eerie eyes seen near lake Build Huffman tree to encode the text -Get the count of each letter - Sort from lowest to highest count
12 - Create a new node as the sum of the lowest two count 2
13 - Create a new node as the sum of the lowest two count - Locate the node at its precedence and create a new one 3
14 - Locate the node at its precedence and create a new one 4
15 5
16 6
17 7
18 8
19 9
20 20
21 2
22 22
23 23
24 The frequency of the root node should equal to the number of characters in the text
25 Least Frequent Same Length More Frequent Shorter Length 25
26 Huffman Algorithm -Source symbols are listed in order of decreasing probability (frequency) 2-The two source symbols of lowest probability are assigned a 0 and (splitting stage ) 3-These two source symbols are combined into a new source symbol with probability equal to the sum of the two original probabilities (The list of source symbols and therefore source statistics is thereby reduced in size by one) 4-The probability of the new symbol is placed in the list in accordance with its value The steps are repeated until we are left with a final list of source statistics of only two for which a 0 and a are assigned 26
27 Ex: Find the Huffman code for the following source given the corresponding probabilities; marginal pmf S o 0. S 0.2 S S S 4 0. S S 0.2 S S o 0. S H( s) (0.4) log ( ) 2x(0.2) log 2( ) 2x(0.) xlog 2( ) L 0.4x 0.2x3 0.2x3 0.x4 0.x4 2.2 H( s) CR K Pk k ( l k L) 2.36 LECTURES 27
28 Alternative solution S o 0. S 0.2 S S S 4 0. S S S S o S 4 0. H( s) (0.4) log ( ) 2x(0.2) log 2( ) 2x(0.) xlog 2( ) L 0.4x2 0.2x2 0.2x2 0.x3 0.x3 2.2 H( s) CR K Pk k ( l L) Hence, to obtain a minimum variance Huffman code, we always put the combined symbol as high in the list as possible LECTURES k
29 Conclusions Huffman encoding process (Huffman tree) is not unique The code with the lowest code length variance is the best because it ensure the least amount of variability in the encoded stream An encoding scheme that minimizes the average length is called optimal coding Huffman code process is optimal if: H( S) L H( S) Looser bound In fact (from literatures), it can be shown that if P max is the largest probability in the probability model, then for P max 0.5, while for P max < 0.5 H( S) L H( S) P max H( S) L H( S) Pmax Tighter bound The average codeword length is equal to the marginal entropy if all marginal probabilities are integer powers of 2 LECTURES 29
30 Huffman vs. Shannon Symbol Frequency P a b c d e f L K k P k l k L 2x( ) 3x0.47 4x( ) bits / symbol The average code length is less than that of Shannon Fano code, Hence Huffman is optimum, but Shannon is suboptimal LECTURES 30
LECTURE VI: LOSSLESS COMPRESSION ALGORITHMS DR. OUIEM BCHIR
1 LECTURE VI: LOSSLESS COMPRESSION ALGORITHMS DR. OUIEM BCHIR 2 STORAGE SPACE Uncompressed graphics, audio, and video data require substantial storage capacity. Storing uncompressed video is not possible
More informationIntroduction to Source Coding
Comm. 52: Communication Theory Lecture 7 Introduction to Source Coding - Requirements of source codes - Huffman Code Length Fixed Length Variable Length Source Code Properties Uniquely Decodable allow
More informationCommunication Theory II
Communication Theory II Lecture 13: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 22 th, 2015 1 o Source Code Generation Lecture Outlines Source Coding
More informationModule 3 Greedy Strategy
Module 3 Greedy Strategy Dr. Natarajan Meghanathan Professor of Computer Science Jackson State University Jackson, MS 39217 E-mail: natarajan.meghanathan@jsums.edu Introduction to Greedy Technique Main
More informationInformation Theory and Communication Optimal Codes
Information Theory and Communication Optimal Codes Ritwik Banerjee rbanerjee@cs.stonybrook.edu c Ritwik Banerjee Information Theory and Communication 1/1 Roadmap Examples and Types of Codes Kraft Inequality
More informationModule 3 Greedy Strategy
Module 3 Greedy Strategy Dr. Natarajan Meghanathan Professor of Computer Science Jackson State University Jackson, MS 39217 E-mail: natarajan.meghanathan@jsums.edu Introduction to Greedy Technique Main
More informationA Brief Introduction to Information Theory and Lossless Coding
A Brief Introduction to Information Theory and Lossless Coding 1 INTRODUCTION This document is intended as a guide to students studying 4C8 who have had no prior exposure to information theory. All of
More informationCoding for Efficiency
Let s suppose that, over some channel, we want to transmit text containing only 4 symbols, a, b, c, and d. Further, let s suppose they have a probability of occurrence in any block of text we send as follows
More informationGreedy Algorithms. Kleinberg and Tardos, Chapter 4
Greedy Algorithms Kleinberg and Tardos, Chapter 4 1 Selecting gas stations Road trip from Fort Collins to Durango on a given route with length L, and fuel stations at positions b i. Fuel capacity = C miles.
More informationComm. 502: Communication Theory. Lecture 6. - Introduction to Source Coding
Comm. 50: Communication Theory Lecture 6 - Introduction to Source Coding Digital Communication Systems Source of Information User of Information Source Encoder Source Decoder Channel Encoder Channel Decoder
More informationHuffman Coding - A Greedy Algorithm. Slides based on Kevin Wayne / Pearson-Addison Wesley
- A Greedy Algorithm Slides based on Kevin Wayne / Pearson-Addison Wesley Greedy Algorithms Greedy Algorithms Build up solutions in small steps Make local decisions Previous decisions are never reconsidered
More informationModule 8: Video Coding Basics Lecture 40: Need for video coding, Elements of information theory, Lossless coding. The Lecture Contains:
The Lecture Contains: The Need for Video Coding Elements of a Video Coding System Elements of Information Theory Symbol Encoding Run-Length Encoding Entropy Encoding file:///d /...Ganesh%20Rana)/MY%20COURSE_Ganesh%20Rana/Prof.%20Sumana%20Gupta/FINAL%20DVSP/lecture%2040/40_1.htm[12/31/2015
More informationInformation Theory and Huffman Coding
Information Theory and Huffman Coding Consider a typical Digital Communication System: A/D Conversion Sampling and Quantization D/A Conversion Source Encoder Source Decoder bit stream bit stream Channel
More informationEntropy, Coding and Data Compression
Entropy, Coding and Data Compression Data vs. Information yes, not, yes, yes, not not In ASCII, each item is 3 8 = 24 bits of data But if the only possible answers are yes and not, there is only one bit
More information1 This work was partially supported by NSF Grant No. CCR , and by the URI International Engineering Program.
Combined Error Correcting and Compressing Codes Extended Summary Thomas Wenisch Peter F. Swaszek Augustus K. Uht 1 University of Rhode Island, Kingston RI Submitted to International Symposium on Information
More informationSolutions to Assignment-2 MOOC-Information Theory
Solutions to Assignment-2 MOOC-Information Theory 1. Which of the following is a prefix-free code? a) 01, 10, 101, 00, 11 b) 0, 11, 01 c) 01, 10, 11, 00 Solution:- The codewords of (a) are not prefix-free
More informationUCSD ECE154C Handout #21 Prof. Young-Han Kim Thursday, April 28, Midterm Solutions (Prepared by TA Shouvik Ganguly)
UCSD ECE54C Handout #2 Prof. Young-Han Kim Thursday, April 28, 26 Midterm Solutions (Prepared by TA Shouvik Ganguly) There are 3 problems, each problem with multiple parts, each part worth points. Your
More information# 12 ECE 253a Digital Image Processing Pamela Cosman 11/4/11. Introductory material for image compression
# 2 ECE 253a Digital Image Processing Pamela Cosman /4/ Introductory material for image compression Motivation: Low-resolution color image: 52 52 pixels/color, 24 bits/pixel 3/4 MB 3 2 pixels, 24 bits/pixel
More informationCOMM901 Source Coding and Compression Winter Semester 2013/2014. Midterm Exam
German University in Cairo - GUC Faculty of Information Engineering & Technology - IET Department of Communication Engineering Dr.-Ing. Heiko Schwarz COMM901 Source Coding and Compression Winter Semester
More informationMAS160: Signals, Systems & Information for Media Technology. Problem Set 4. DUE: October 20, 2003
MAS160: Signals, Systems & Information for Media Technology Problem Set 4 DUE: October 20, 2003 Instructors: V. Michael Bove, Jr. and Rosalind Picard T.A. Jim McBride Problem 1: Simple Psychoacoustic Masking
More informationMultimedia Systems Entropy Coding Mahdi Amiri February 2011 Sharif University of Technology
Course Presentation Multimedia Systems Entropy Coding Mahdi Amiri February 2011 Sharif University of Technology Data Compression Motivation Data storage and transmission cost money Use fewest number of
More informationCHAPTER 5 PAPR REDUCTION USING HUFFMAN AND ADAPTIVE HUFFMAN CODES
119 CHAPTER 5 PAPR REDUCTION USING HUFFMAN AND ADAPTIVE HUFFMAN CODES 5.1 INTRODUCTION In this work the peak powers of the OFDM signal is reduced by applying Adaptive Huffman Codes (AHC). First the encoding
More informationMonday, February 2, Is assigned today. Answers due by noon on Monday, February 9, 2015.
Monday, February 2, 2015 Topics for today Homework #1 Encoding checkers and chess positions Constructing variable-length codes Huffman codes Homework #1 Is assigned today. Answers due by noon on Monday,
More informationLossless Grayscale Image Compression using Blockwise Entropy Shannon (LBES)
Volume No., July Lossless Grayscale Image Compression using Blockwise ntropy Shannon (LBS) S. Anantha Babu Ph.D. (Research Scholar) & Assistant Professor Department of Computer Science and ngineering V
More informationMAS.160 / MAS.510 / MAS.511 Signals, Systems and Information for Media Technology Fall 2007
MIT OpenCourseWare http://ocw.mit.edu MAS.160 / MAS.510 / MAS.511 Signals, Systems and Information for Media Technology Fall 2007 For information about citing these materials or our Terms of Use, visit:
More informationDigital Communication Systems ECS 452
Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th 2. Source Coding 1 Office Hours: BKD, 6th floor of Sirindhralai building Monday 10:00-10:40 Tuesday 12:00-12:40
More informationSOME EXAMPLES FROM INFORMATION THEORY (AFTER C. SHANNON).
SOME EXAMPLES FROM INFORMATION THEORY (AFTER C. SHANNON). 1. Some easy problems. 1.1. Guessing a number. Someone chose a number x between 1 and N. You are allowed to ask questions: Is this number larger
More informationThe Lempel-Ziv (LZ) lossless compression algorithm was developed by Jacob Ziv (AT&T Bell Labs / Technion Israel) and Abraham Lempel (IBM) in 1978;
Georgia Institute of Technology - Georgia Tech Lorraine ECE 6605 Information Theory Lempel-Ziv Lossless Compresion General comments The Lempel-Ziv (LZ) lossless compression algorithm was developed by Jacob
More informationRun-Length Based Huffman Coding
Chapter 5 Run-Length Based Huffman Coding This chapter presents a multistage encoding technique to reduce the test data volume and test power in scan-based test applications. We have proposed a statistical
More informationWednesday, February 1, 2017
Wednesday, February 1, 2017 Topics for today Encoding game positions Constructing variable-length codes Huffman codes Encoding Game positions Some programs that play two-player games (e.g., tic-tac-toe,
More informationGENERIC CODE DESIGN ALGORITHMS FOR REVERSIBLE VARIABLE-LENGTH CODES FROM THE HUFFMAN CODE
GENERIC CODE DESIGN ALGORITHMS FOR REVERSIBLE VARIABLE-LENGTH CODES FROM THE HUFFMAN CODE Wook-Hyun Jeong and Yo-Sung Ho Kwangju Institute of Science and Technology (K-JIST) Oryong-dong, Buk-gu, Kwangju,
More informationHuffman Coding For Digital Photography
Huffman Coding For Digital Photography Raydhitya Yoseph 13509092 Program Studi Teknik Informatika Sekolah Teknik Elektro dan Informatika Institut Teknologi Bandung, Jl. Ganesha 10 Bandung 40132, Indonesia
More informationDigital Communication Systems ECS 452
Digital Communication Systems ECS 452 Asst. Prof. Dr. Prapun Suksompong prapun@siit.tu.ac.th Source Coding 1 Office Hours: BKD 3601-7 Monday 14:00-16:00 Wednesday 14:40-16:00 Noise & Interference Elements
More informationHUFFMAN CODING. Catherine Bénéteau and Patrick J. Van Fleet. SACNAS 2009 Mini Course. University of South Florida and University of St.
Catherine Bénéteau and Patrick J. Van Fleet University of South Florida and University of St. Thomas SACNAS 2009 Mini Course WEDNESDAY, 14 OCTOBER, 2009 (1:40-3:00) LECTURE 2 SACNAS 2009 1 / 10 All lecture
More informationLossless Image Compression Techniques Comparative Study
Lossless Image Compression Techniques Comparative Study Walaa Z. Wahba 1, Ashraf Y. A. Maghari 2 1M.Sc student, Faculty of Information Technology, Islamic university of Gaza, Gaza, Palestine 2Assistant
More informationECE Advanced Communication Theory, Spring 2007 Midterm Exam Monday, April 23rd, 6:00-9:00pm, ELAB 325
C 745 - Advanced Communication Theory, Spring 2007 Midterm xam Monday, April 23rd, 600-900pm, LAB 325 Overview The exam consists of five problems for 150 points. The points for each part of each problem
More informationCSE 100: BST AVERAGE CASE AND HUFFMAN CODES
CSE 100: BST AVERAGE CASE AND HUFFMAN CODES Recap: Average Case Analysis of successful find in a BST N nodes Expected total depth of all BSTs with N nodes Recap: Probability of having i nodes in the left
More informationFAST LEMPEL-ZIV (LZ 78) COMPLEXITY ESTIMATION USING CODEBOOK HASHING
FAST LEMPEL-ZIV (LZ 78) COMPLEXITY ESTIMATION USING CODEBOOK HASHING Harman Jot, Rupinder Kaur M.Tech, Department of Electronics and Communication, Punjabi University, Patiala, Punjab, India I. INTRODUCTION
More informationComparison of Data Compression in Text Using Huffman, Shannon-Fano, Run Length Encoding, and Tunstall Method
Comparison of Data Compression in Text Using Huffman, Shannon-Fano, Run Length Encoding, and Tunstall Method Dea Ayu Rachesti College Student, Faculty of Electrical Engineering, Telkom University, Bandung,
More informationSlides credited from Hsueh-I Lu, Hsu-Chun Hsiao, & Michael Tsai
Slides credited from Hsueh-I Lu, Hsu-Chun Hsiao, & Michael Tsai Mini-HW 6 Released Due on 11/09 (Thu) 17:20 Homework 2 Due on 11/09 (Thur) 17:20 Midterm Time: 11/16 (Thur) 14:20-17:20 Format: close book
More information6.02 Introduction to EECS II Spring Quiz 1
M A S S A C H U S E T T S I N S T I T U T E O F T E C H N O L O G Y DEPARTMENT OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE 6.02 Introduction to EECS II Spring 2011 Quiz 1 Name SOLUTIONS Score Please
More informationKeywords Audio Steganography, Compressive Algorithms, SNR, Capacity, Robustness. (Figure 1: The Steganographic operation) [10]
Volume 4, Issue 5, May 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Audio Steganography
More information6.004 Computation Structures Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 6.004 Computation Structures Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Welcome to 6.004! Course
More informationChapter 1 INTRODUCTION TO SOURCE CODING AND CHANNEL CODING. Whether a source is analog or digital, a digital communication
1 Chapter 1 INTRODUCTION TO SOURCE CODING AND CHANNEL CODING 1.1 SOURCE CODING Whether a source is analog or digital, a digital communication system is designed to transmit information in digital form.
More informationThe idea of similarity is through the Hamming
Hamming distance A good channel code is designed so that, if a few bit errors occur in transmission, the output can still be identified as the correct input. This is possible because although incorrect,
More informationProblem Sheet 1 Probability, random processes, and noise
Problem Sheet 1 Probability, random processes, and noise 1. If F X (x) is the distribution function of a random variable X and x 1 x 2, show that F X (x 1 ) F X (x 2 ). 2. Use the definition of the cumulative
More informationA High-Throughput Memory-Based VLC Decoder with Codeword Boundary Prediction
1514 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 10, NO. 8, DECEMBER 2000 A High-Throughput Memory-Based VLC Decoder with Codeword Boundary Prediction Bai-Jue Shieh, Yew-San Lee,
More informationCommunication Theory II
Communication Theory II Lecture 14: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 25 th, 2015 1 Previous Lecture: Source Code Generation: Lossless
More information6.450: Principles of Digital Communication 1
6.450: Principles of Digital Communication 1 Digital Communication: Enormous and normally rapidly growing industry, roughly comparable in size to the computer industry. Objective: Study those aspects of
More informationPublished by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 1
VHDL design of lossy DWT based image compression technique for video conferencing Anitha Mary. M 1 and Dr.N.M. Nandhitha 2 1 VLSI Design, Sathyabama University Chennai, Tamilnadu 600119, India 2 ECE, Sathyabama
More informationChapter 6: Memory: Information and Secret Codes. CS105: Great Insights in Computer Science
Chapter 6: Memory: Information and Secret Codes CS105: Great Insights in Computer Science Overview When we decide how to represent something in bits, there are some competing interests: easily manipulated/processed
More informationDECISION TREE TUTORIAL
Kardi Teknomo DECISION TREE TUTORIAL Revoledu.com Decision Tree Tutorial by Kardi Teknomo Copyright 2008-2012 by Kardi Teknomo Published by Revoledu.com Online edition is available at Revoledu.com Last
More informationDCSP-3: Minimal Length Coding. Jianfeng Feng
DCSP-3: Minimal Length Coding Jianfeng Feng Department of Computer Science Warwick Univ., UK Jianfeng.feng@warwick.ac.uk http://www.dcs.warwick.ac.uk/~feng/dcsp.html Automatic Image Caption (better than
More information!!"!#$#!%!"""#&#!%!""#&#"%!"# &#!%!# # ##$#!%!"'###&#!%!"(##&#"%!"!#&#!%!""# #!!"!#$#!%)# &#!%*# &#"%(##&#!%!# Base or
/9/ General dea More its and ytes inary umbers & uffman oding!!"#"$%&'()*+,-%$"+).'!/$/$")#'$/',//)/'+,' %/)/'+*'%'/)+-/)+)'%$'%'/"'&%/'%)6' $"-/.'%)6'! /)+-/)%.'&"#$9'-%#)/$"-9'%,#/9'-%9'+&+,9' :,,/)$9';'!!/$/$")#'6//)6'+)'/)+-/)+)'
More information6. FUNDAMENTALS OF CHANNEL CODER
82 6. FUNDAMENTALS OF CHANNEL CODER 6.1 INTRODUCTION The digital information can be transmitted over the channel using different signaling schemes. The type of the signal scheme chosen mainly depends on
More informationWhat You ll Learn Today
CS101 Lecture 18: Image Compression Aaron Stevens 21 October 2010 Some material form Wikimedia Commons Special thanks to John Magee and his dog 1 What You ll Learn Today Review: how big are image files?
More informationEntropy Coding. Outline. Entropy. Definitions. log. A = {a, b, c, d, e}
Outline efinition of ntroy Three ntroy coding techniques: Huffman coding rithmetic coding Lemel-Ziv coding ntroy oding (taken from the Technion) ntroy ntroy of a set of elements e,,e n with robabilities,
More informationImages with (a) coding redundancy; (b) spatial redundancy; (c) irrelevant information
Images with (a) coding redundancy; (b) spatial redundancy; (c) irrelevant information 1992 2008 R. C. Gonzalez & R. E. Woods For the image in Fig. 8.1(a): 1992 2008 R. C. Gonzalez & R. E. Woods Measuring
More informationArithmetic Compression on SPIHT Encoded Images
Arithmetic Compression on SPIHT Encoded Images Todd Owen, Scott Hauck {towen, hauck}@ee.washington.edu Dept of EE, University of Washington Seattle WA, 98195-2500 UWEE Technical Report Number UWEETR-2002-0007
More informationSHANNON S source channel separation theorem states
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 9, SEPTEMBER 2009 3927 Source Channel Coding for Correlated Sources Over Multiuser Channels Deniz Gündüz, Member, IEEE, Elza Erkip, Senior Member,
More informationPooja Rani(M.tech) *, Sonal ** * M.Tech Student, ** Assistant Professor
A Study of Image Compression Techniques Pooja Rani(M.tech) *, Sonal ** * M.Tech Student, ** Assistant Professor Department of Computer Science & Engineering, BPS Mahila Vishvavidyalya, Sonipat kulriapooja@gmail.com,
More informationA SURVEY ON DICOM IMAGE COMPRESSION AND DECOMPRESSION TECHNIQUES
A SURVEY ON DICOM IMAGE COMPRESSION AND DECOMPRESSION TECHNIQUES Shreya A 1, Ajay B.N 2 M.Tech Scholar Department of Computer Science and Engineering 2 Assitant Professor, Department of Computer Science
More informationLectures: Feb 27 + Mar 1 + Mar 3, 2017
CS420+500: Advanced Algorithm Design and Analysis Lectures: Feb 27 + Mar 1 + Mar 3, 2017 Prof. Will Evans Scribe: Adrian She In this lecture we: Summarized how linear programs can be used to model zero-sum
More informationHuffman-Compressed Wavelet Trees for Large Alphabets
Laboratorio de Bases de Datos Facultade de Informática Universidade da Coruña Departamento de Ciencias de la Computación Universidad de Chile Huffman-Compressed Wavelet Trees for Large Alphabets Gonzalo
More informationHuffman Coding with Non-Sorted Frequencies
Huffman Coding with Non-Sorted Frequencies Shmuel T. Klein and Dana Shapira Abstract. A standard way of implementing Huffman s optimal code construction algorithm is by using a sorted sequence of frequencies.
More informationSpeech Coding in the Frequency Domain
Speech Coding in the Frequency Domain Speech Processing Advanced Topics Tom Bäckström Aalto University October 215 Introduction The speech production model can be used to efficiently encode speech signals.
More informationCOURSE MATERIAL Subject Name: Communication Theory UNIT V
NH-67, TRICHY MAIN ROAD, PULIYUR, C.F. - 639114, KARUR DT. DEPARTMENT OF ELECTRONICS AND COMMUNICATION ENGINEERING COURSE MATERIAL Subject Name: Communication Theory Subject Code: 080290020 Class/Sem:
More information4. Which of the following channel matrices respresent a symmetric channel? [01M02] 5. The capacity of the channel with the channel Matrix
Send SMS s : ONJntuSpeed To 9870807070 To Recieve Jntu Updates Daily On Your Mobile For Free www.strikingsoon.comjntu ONLINE EXMINTIONS [Mid 2 - dc] http://jntuk.strikingsoon.com 1. Two binary random
More informationFundamentals of Digital Communication
Fundamentals of Digital Communication Network Infrastructures A.A. 2017/18 Digital communication system Analog Digital Input Signal Analog/ Digital Low Pass Filter Sampler Quantizer Source Encoder Channel
More information[Manisha*, 4.(10): October, 2015] ISSN: (I2OR), Publication Impact Factor: 3.785
IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY A SURVEY OF DATA COMPRESSION TECHNIQUES Ms.Anjikhane Manisha*, Ms.Hannure Asma * M.Tech-CSE Department of Computer Science,Maharashtra.
More information2. REVIEW OF LITERATURE
2. REVIEW OF LITERATURE Digital image processing is the use of the algorithms and procedures for operations such as image enhancement, image compression, image analysis, mapping. Transmission of information
More informationECE 8771, Information Theory & Coding for Digital Communications Summer 2010 Syllabus & Outline (Draft 1 - May 12, 2010)
ECE 8771, Information Theory & Coding for Digital Communications Summer 2010 Syllabus & Outline (Draft 1 - May 12, 2010) Instructor: Kevin Buckley, Tolentine 433a, 610-519-5658 (W), 610-519-4436 (F), buckley@ece.vill.edu,
More information2.1. General Purpose Run Length Encoding Relative Encoding Tokanization or Pattern Substitution
2.1. General Purpose There are many popular general purpose lossless compression techniques, that can be applied to any type of data. 2.1.1. Run Length Encoding Run Length Encoding is a compression technique
More informationDEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK. Subject Name: Information Coding Techniques UNIT I INFORMATION ENTROPY FUNDAMENTALS
DEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK Subject Name: Year /Sem: II / IV UNIT I INFORMATION ENTROPY FUNDAMENTALS PART A (2 MARKS) 1. What is uncertainty? 2. What is prefix coding? 3. State the
More informationSimple, Fast, and Efficient Natural Language Adaptive Compression
Simple, Fast, and Efficient Natural Language Adaptive Compression Nieves R. Brisaboa, Antonio Fariña, Gonzalo Navarro and José R. Paramá Database Lab., Univ. da Coruña, Facultade de Informática, Campus
More informationA Recursive Threshold Visual Cryptography Scheme
A Recursive Threshold Visual Cryptography cheme Abhishek Parakh and ubhash Kak Department of Computer cience Oklahoma tate University tillwater, OK 74078 Abstract: This paper presents a recursive hiding
More informationHamming net based Low Complexity Successive Cancellation Polar Decoder
Hamming net based Low Complexity Successive Cancellation Polar Decoder [1] Makarand Jadhav, [2] Dr. Ashok Sapkal, [3] Prof. Ram Patterkine [1] Ph.D. Student, [2] Professor, Government COE, Pune, [3] Ex-Head
More informationThe Need for Data Compression. Data Compression (for Images) -Compressing Graphical Data. Lossy vs Lossless compression
The Need for Data Compression Data Compression (for Images) -Compressing Graphical Data Graphical images in bitmap format take a lot of memory e.g. 1024 x 768 pixels x 24 bits-per-pixel = 2.4Mbyte =18,874,368
More informationB. Tech. (SEM. VI) EXAMINATION, (2) All question early equal make. (3) In ease of numerical problems assume data wherever not provided.
" 11111111111111111111111111111111111111111111111111111111111111III *U-3091/8400* Printed Pages : 7 TEC - 601! I i B. Tech. (SEM. VI) EXAMINATION, 2007-08 DIGIT AL COMMUNICATION \ V Time: 3 Hours] [Total
More informationEfficient and Compact Representations of Some Non-Canonical Prefix-Free Codes
Efficient and Compact Representations of Some Non-Canonical Prefix-Free Codes Antonio Fariña 1, Travis Gagie 2, Giovanni Manzini 3, Gonzalo Navarro 4, and Alberto Ordóñez 5 1 Database Laboratory, University
More informationComputer Graphics. Si Lu. Fall er_graphics.htm 10/02/2015
Computer Graphics Si Lu Fall 2017 http://www.cs.pdx.edu/~lusi/cs447/cs447_547_comput er_graphics.htm 10/02/2015 1 Announcements Free Textbook: Linear Algebra By Jim Hefferon http://joshua.smcvt.edu/linalg.html/
More informationAn Efficient Approach for Image Compression using Segmented Probabilistic Encoding with Shanon Fano[SPES].
An Efficient Approach for Compression using Segmented Probabilistic Encoding with Shanon Fano[SPES]. Dr. T. Bhaskara Reddy 1, Miss. Hema Suresh Yaragunti 2, Mr. T. Sri Harish Reddy 3, Dr. S. Kiran 4 1
More informationTSTE17 System Design, CDIO. General project hints. Behavioral Model. General project hints, cont. Lecture 5. Required documents Modulation, cont.
TSTE17 System Design, CDIO Lecture 5 1 General project hints 2 Project hints and deadline suggestions Required documents Modulation, cont. Requirement specification Channel coding Design specification
More informationConvolutional Coding Using Booth Algorithm For Application in Wireless Communication
Available online at www.interscience.in Convolutional Coding Using Booth Algorithm For Application in Wireless Communication Sishir Kalita, Parismita Gogoi & Kandarpa Kumar Sarma Department of Electronics
More informationCompression. Encryption. Decryption. Decompression. Presentation of Information to client site
DOCUMENT Anup Basu Audio Image Video Data Graphics Objectives Compression Encryption Network Communications Decryption Decompression Client site Presentation of Information to client site Multimedia -
More informationIndian Institute of Technology, Roorkee, India
Volume-, Issue-, Feb.-7 A COMPARATIVE STUDY OF LOSSLESS COMPRESSION TECHNIQUES J P SATI, M J NIGAM, Indian Institute of Technology, Roorkee, India E-mail: jypsati@gmail.com, mkndnfec@gmail.com Abstract-
More informationCS 540-2: Introduction to Artificial Intelligence Homework Assignment #2. Assigned: Monday, February 6 Due: Saturday, February 18
CS 540-2: Introduction to Artificial Intelligence Homework Assignment #2 Assigned: Monday, February 6 Due: Saturday, February 18 Hand-In Instructions This assignment includes written problems and programming
More informationChapter 12. Cross-Layer Optimization for Multi- Hop Cognitive Radio Networks
Chapter 12 Cross-Layer Optimization for Multi- Hop Cognitive Radio Networks 1 Outline CR network (CRN) properties Mathematical models at multiple layers Case study 2 Traditional Radio vs CR Traditional
More informationProf. Feng Liu. Fall /02/2018
Prof. Feng Liu Fall 2018 http://www.cs.pdx.edu/~fliu/courses/cs447/ 10/02/2018 1 Announcements Free Textbook: Linear Algebra By Jim Hefferon http://joshua.smcvt.edu/linalg.html/ Homework 1 due in class
More informationDEVELOPMENT OF LOSSY COMMPRESSION TECHNIQUE FOR IMAGE
DEVELOPMENT OF LOSSY COMMPRESSION TECHNIQUE FOR IMAGE Asst.Prof.Deepti Mahadeshwar,*Prof. V.M.Misra Department of Instrumentation Engineering, Vidyavardhini s College of Engg. And Tech., Vasai Road, *Prof
More informationMATHEMATICS IN COMMUNICATIONS: INTRODUCTION TO CODING. A Public Lecture to the Uganda Mathematics Society
Abstract MATHEMATICS IN COMMUNICATIONS: INTRODUCTION TO CODING A Public Lecture to the Uganda Mathematics Society F F Tusubira, PhD, MUIPE, MIEE, REng, CEng Mathematical theory and techniques play a vital
More informationA Modified Image Template for FELICS Algorithm for Lossless Image Compression
Research Article International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347-5161 2014 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet A Modified
More informationECEn 665: Antennas and Propagation for Wireless Communications 131. s(t) = A c [1 + αm(t)] cos (ω c t) (9.27)
ECEn 665: Antennas and Propagation for Wireless Communications 131 9. Modulation Modulation is a way to vary the amplitude and phase of a sinusoidal carrier waveform in order to transmit information. When
More information[Srivastava* et al., 5(8): August, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116
IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY COMPRESSING BIOMEDICAL IMAGE BY USING INTEGER WAVELET TRANSFORM AND PREDICTIVE ENCODER Anushree Srivastava*, Narendra Kumar Chaurasia
More informationECE 6640 Digital Communications
ECE 6640 Digital Communications Dr. Bradley J. Bazuin Assistant Professor Department of Electrical and Computer Engineering College of Engineering and Applied Sciences Chapter 8 8. Channel Coding: Part
More informationCSE101: Algorithm Design and Analysis. Ragesh Jaiswal, CSE, UCSD
Longest increasing subsequence Problem Longest increasing subsequence: You are given a sequence of integers A[1], A[2],..., A[n] and you are asked to find a longest increasing subsequence of integers.
More informationReview: Our Approach 2. CSC310 Information Theory
CSC30 Informaton Theory Sam Rowes Lecture 3: Provng the Kraft-McMllan Inequaltes September 8, 6 Revew: Our Approach The study of both compresson and transmsson requres that we abstract data and messages
More informationChapter 4: The Building Blocks: Binary Numbers, Boolean Logic, and Gates
Chapter 4: The Building Blocks: Binary Numbers, Boolean Logic, and Gates Objectives In this chapter, you will learn about The binary numbering system Boolean logic and gates Building computer circuits
More informationSPIHT Algorithm with Huffman Encoding for Image Compression and Quality Improvement over MIMO OFDM Channel
SPIHT Algorithm with Huffman Encoding for Image Compression and Quality Improvement over MIMO OFDM Channel Dnyaneshwar.K 1, CH.Suneetha 2 Abstract In this paper, Compression and improving the Quality of
More informationTopic 23 Red Black Trees
Topic 23 "People in every direction No words exchanged No time to exchange And all the little ants are marching Red and Black antennas waving" -Ants Marching, Dave Matthew's Band "Welcome to L.A.'s Automated
More information