Bounds on Redundancy in Constrained Delay Arithmetic Coding Ofer ShayevitzEado Meron Meir Feder Ram Zamir Tel Aviv University.

Slides:



Advertisements
Similar presentations
Ulams Game and Universal Communications Using Feedback Ofer Shayevitz June 2006.
Advertisements

Capacity Upper Bounds for Deletion Channels Suhas Diggavi Michael Mitzenmacher Henry Pfister.
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
15-583:Algorithms in the Real World
Lecture 4 (week 2) Source Coding and Compression
Lecture 3: Source Coding Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Sampling and Pulse Code Modulation
Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Arithmetic Coding. Gabriele Monfardini - Corso di Basi di Dati Multimediali a.a How we can do better than Huffman? - I As we have seen, the.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Data Compression.
Optimal Partitions of Strings: A new class of Burrows-Wheeler Compression Algorithms Raffaele Giancarlo Marinella Sciortino
Compression & Huffman Codes
SWE 423: Multimedia Systems Chapter 7: Data Compression (3)
Fundamental limits in Information Theory Chapter 10 :
2015/6/15VLC 2006 PART 1 Introduction on Video Coding StandardsVLC 2006 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
SWE 423: Multimedia Systems Chapter 7: Data Compression (2)
A Data Compression Algorithm: Huffman Compression
Fundamentals of Multimedia Chapter 7 Lossless Compression Algorithms Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
EE 3220: Digital Communication Dr Hassan Yousif 1 Dr. Hassan Yousif Ahmed Department of Electrical Engineering College of Engineering at Wadi Aldwasser.
Lossless Data Compression Using run-length and Huffman Compression pages
2015/7/12VLC 2008 PART 1 Introduction on Video Coding StandardsVLC 2008 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
CSI Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability.
Low Complexity Algebraic Multicast Network Codes Sidharth “Sid” Jaggi Philip Chou Kamal Jain.
Source Coding Hafiz Malik Dept. of Electrical & Computer Engineering The University of Michigan-Dearborn
Noise, Information Theory, and Entropy
Noise, Information Theory, and Entropy
Basics of Compression Goals: to understand how image/audio/video signals are compressed to save storage and increase transmission efficiency to understand.
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Huffman Coding Vida Movahedi October Contents A simple example Definitions Huffman Coding Algorithm Image Compression.
Information Theory & Coding…
15-853Page :Algorithms in the Real World Data Compression II Arithmetic Coding – Integer implementation Applications of Probability Coding – Run.
1 Lossless Compression Multimedia Systems (Module 2 Lesson 2) Summary:  Adaptive Coding  Adaptive Huffman Coding Sibling Property Update Algorithm 
Dr.-Ing. Khaled Shawky Hassan
Basics of Data Compression Paolo Ferragina Dipartimento di Informatica Università di Pisa.
Prof. Amr Goneid Department of Computer Science & Engineering
7. 1 Arithmetic vs. Huffman Most of Multimedia methods use Huffman coding. We should check why Huffman has been chosen: Error Resilience. Average Codeword's.
Basics of Data Compression Paolo Ferragina Dipartimento di Informatica Università di Pisa.
Communication Over Unknown Channels: A Personal Perspective of Over a Decade Research* Meir Feder Dept. of Electrical Engineering-Systems Tel-Aviv University.
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
Prof. Amr Goneid, AUC1 Analysis & Design of Algorithms (CSCE 321) Prof. Amr Goneid Department of Computer Science, AUC Part 8. Greedy Algorithms.
Digital Communications I: Modulation and Coding Course Term Catharina Logothetis Lecture 12.
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
DIGITAL COMMUNICATIONS Linear Block Codes
Huffman Code and Data Decomposition Pranav Shah CS157B.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Data Compression Meeting October 25, 2002 Arithmetic Coding.
CS654: Digital Image Analysis Lecture 34: Different Coding Techniques.
Timo O. Korhonen, HUT Communication Laboratory 1 Convolutional encoding u Convolutional codes are applied in applications that require good performance.
Digital Image Processing Lecture 22: Image Compression
1 Source Coding and Compression Dr.-Ing. Khaled Shawky Hassan Room: C3-222, ext: 1204, Lecture 7 (W5)
A Low-Complexity Universal Architecture for Distributed Rate-Constrained Nonparametric Statistical Learning in Sensor Networks Avon Loy Fernandes, Maxim.
Lossless Decomposition and Huffman Codes Sophia Soohoo CS 157B.
Index construction: Compression of documents Paolo Ferragina Dipartimento di Informatica Università di Pisa Reading Managing-Gigabytes: pg 21-36, 52-56,
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Greedy algorithms 2 David Kauchak cs302 Spring 2012.
Search Engines WS 2009 / 2010 Prof. Dr. Hannah Bast Chair of Algorithms and Data Structures Department of Computer Science University of Freiburg Lecture.
UNIT –V INFORMATION THEORY EC6402 : Communication TheoryIV Semester - ECE Prepared by: S.P.SIVAGNANA SUBRAMANIAN, Assistant Professor, Dept. of ECE, Sri.
Information theory Data compression perspective Pasi Fränti
Ch4. Zero-Error Data Compression Yuan Luo. Content  Ch4. Zero-Error Data Compression  4.1 The Entropy Bound  4.2 Prefix Codes  Definition and.
DATA STRUCTURES AND ALGORITHM (CSE 220)
Data Coding Run Length Coding
EE465: Introduction to Digital Image Processing
Digital Image Processing Lecture 20: Image Compression May 16, 2005
Huffman Coding, Arithmetic Coding, and JBIG2
Scalar Quantization – Mathematical Model
Analysis & Design of Algorithms (CSCE 321)
CSE 589 Applied Algorithms Spring 1999
CSE 589 Applied Algorithms Spring 1999
Presentation transcript:

Bounds on Redundancy in Constrained Delay Arithmetic Coding Ofer ShayevitzEado Meron Meir Feder Ram Zamir Tel Aviv University

Arithmetic Coding (AC) Background  Source coding with finite alphabet  Sequential coding at the entropy rate  Encoder -  Sequentially maps growing source strings into nested source intervals in [0,1)  Interval size = Probability of the sequence  Shrinking intervals  Converge to a source point  Outputs bits representing the smallest binary interval containing the source interval  Decoder –  Tracks binary interval  Decodes smallest source interval containing the current binary interval ! ( x 1 ) X = © 0 ; 1 ;:::; K ¡ 1 ª x n 4 = x 1 ;:::; x n

Setting and Notations  was binary encoded – Arbitrary coding scheme  However, only for was decoded  The delay (in symbols) at time n is defined as  Random prefix  The delay is a random variable  The redundancy-delay function for a source is Infimum is taken over all achievable compression ratios obeying the delay constraint D = D ( x n ) = n ¡ k ( x n ) x n x k k = k ( x n ) · n R ( d c ) 4 = i n f C ( C ¡ H ) s. t. D ( x n ) < d c + 1

Previous Work – Delay vs. Redundancy  [Jelinek 1968]  Memoryless source  B 2 V coding  Constant buffer size  Delay in encoded bits  Beautiful analysis of the tradeoff between buffer overflow probability and redundancy  However…  Delay in symbols = Time!  Errs when buffer overflows – not strictly lossless  B 2 V with a finite (symbol ) delay constraint: R ( d c ) = O ( 1 d c )

Delay in AC - An Example  Source symbols each with probability  Assume the source sequence is encoded f 0 ; 1 ; 2 g Binary Interval Encoder’s Interval Decoder’s Interval Encoded symbols : Output Bits: Decoded Symbols: ; 0 ; 2 ;::: 2 symbol delay at time n=3

Ternary Example Cont. Assume now the source sequence is encoded Binary Interval Encoder’s Interval Decoder’s Interval Encoded Symbols : Output Bits: Decoded symbols: 1 1 ; 1 ; 1 ;::: ::: Source point Unbounded delay !!! ! ( x 1 ) = 1 2

Delay in AC  Origin – Discrepancy between source intervals and binary intervals  Infinite number of source sequences with unbounded delay exist  Bounded Delay  Must intervene in the normal AC flow!

Previous Work – The Delay’s Tail Probability  AC for a memoryless source  Bound on the delay’s tail probability [Shayevitz Zamir Feder 2006]  Resulting bound on the expected delay P n D ( x n ) > d o · 4 d ® d l og 1 ® + 4 ® d µ l og l n l n 2 ¶ |{z} K ® 4 = max © p 0 ; p 2 ;:::; p K ¡ 1 ª E ( D ) · ® ( K ( 1 ¡ ® ) + l og 1 ® ) ( 1 ¡ ® ) 2

Our New Result  Goal – Intervene with the AC flow to meet a finite delay constraint  Insertion algorithm – inserts fictitious symbols to nullify the delay  Constructive upper bound on  Bounding the redundancy-delay exponent R ( d c ) E r d 4 = l i m n ! 1 ¡ 1 d c l og R ( d c ) ¸ l og 1 = ® R ( d c ) · k ( ® ) ® d c ( 1 + d c l og 1 = ® ) 2 = O ( ® d c )

The Forbidden Points Concept  Ternary source, assume was encoded  No output bits not decoded yet  Condition for decoding : The binary interval must shrink into  Magnifying … £ 1 3 ; 2 3 ¢ Binary Interval Encoder Source Interval £ 1 3 ; 2 3 ¢ x 1 = 1 x 1 x 1

 not decoded if the Encoder’s interval:  Contains the point  Contains either or  And so on… We call those the forbidden points  Each “step” is the maximal binary step so that the interval ‘s edges are not crossed  To nullify the delay  Avoid the forbidden points! x

Solution - Insertion Algorithm  Append two fictitious symbols to the source’s alphabet  Map each to an interval of length  Shrink all other intervals evenly  The encoder keeps track of decoder’s delay, and inserts a fictitious symbol when the delay constraint is breached  Can we guarantee a delay reduction? "

Solution – Cont.  Yes! – with a suitable fictitious symbol mapping Forbidden free interval Encoder’s Source Interval Binary interval

Analysis  A suitable mapping assures that one of the fictitious symbols is contained in a forbidden free interval  Inserting it will nullify the delay!  Origin of Redundancy  Expected codelength of fictitious symbols  Mismatch of source probabilities (Shrinkage) P ( D ¸ d c ) ¢ l og 1 = " l og 1 1 ¡ 2 "

 Optimizing for we get  And the exponent is bounded by Analysis – Cont. " R ( d c ) · k ( ® ) ® d c ( 1 + d c l og 1 = ® ) 2 E r d 4 = l i m n ! 1 ¡ 1 d c l og R ( d c ) ¸ l og 1 = ®

English Text Compression  Zero order model H ¼ 4 b i t s E ( D ) ¼ 3 : 5 sym b o l s B2VB2VB2VB2V d c > 100

Summary  Arithmetic coding scheme with an insertion algorithm can attain a finite delay  The redundancy in V 2 V coding decays exponentially with the delay constraint  In comparison, for B 2 V coding the redundancy decays like R ( d c ) = O ¡ ® d c ¢ R ( d c ) = O ( 1 d c )

Future Research  Sources with memory  Sharper lower bounds for the redundancy-delay exponent, depending on the entire source’s distribution  Upper bounds ?

Thank You !