ELG5126 Source Coding and Data Compression

Slides:



Advertisements
Similar presentations
T.Sharon-A.Frank 1 Multimedia Compression Basics.
Advertisements

15 Data Compression Foundations of Computer Science ã Cengage Learning.
Data Compression CS 147 Minh Nguyen.
Chapter 7 End-to-End Data
Spring 2003CS 4611 Multimedia Outline Compression RTP Scheduling.
1 Audio Compression Techniques MUMT 611, January 2005 Assignment 2 Paul Kolesnik.
SWE 423: Multimedia Systems
SWE 423: Multimedia Systems Chapter 7: Data Compression (1)
CSCI 3 Chapter 1.8 Data Compression. Chapter 1.8 Data Compression  For the purpose of storing or transferring data, it is often helpful to reduce the.
Spatial and Temporal Data Mining
Computer Science 335 Data Compression.
Losslessy Compression of Multimedia Data Hao Jiang Computer Science Department Sept. 25, 2007.
CSc 461/561 CSc 461/561 Multimedia Systems 0. Introduction.
Fundamentals of Multimedia Chapter 7 Lossless Compression Algorithms Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
T.Sharon-A.Frank 1 Multimedia Image Compression 2 T.Sharon-A.Frank Coding Techniques – Hybrid.
Department of Computer Engineering University of California at Santa Cruz Data Compression (2) Hai Tao.
Image Compression - JPEG. Video Compression MPEG –Audio compression Lossy / perceptually lossless / lossless 3 layers Models based on speech generation.
Software Research Image Compression Mohamed N. Ahmed, Ph.D.
Lossy Compression Based on spatial redundancy Measure of spatial redundancy: 2D covariance Cov X (i,j)=  2 e -  (i*i+j*j) Vertical correlation   
Speech coding. What’s the need for speech coding ? Necessary in order to represent human speech in a digital form Applications: mobile/telephone communication,
Lecture 10 Data Compression.
Compression is the reduction in size of data in order to save space or transmission time. And its used just about everywhere. All the images you get on.
Computer Vision – Compression(2) Hanyang University Jong-Il Park.
 Coding efficiency/Compression ratio:  The loss of information or distortion measure:
1 Section 3. Image Compression Xudong Ni Group Member: Wei Yan,Li Yang,Xudong Ni Computer Science Florida International University.
CS Spring 2011 CS 414 – Multimedia Systems Design Lecture 7 – Basics of Compression (Part 2) Klara Nahrstedt Spring 2011.
Prof. Amr Goneid Department of Computer Science & Engineering
Multimedia Specification Design and Production 2012 / Semester 1 / L3 Lecturer: Dr. Nikos Gazepidis
Chapter 1: Data Storage.
Image Processing and Computer Vision: 91. Image and Video Coding Compressing data to a smaller volume without losing (too much) information.
CIS679: Multimedia Basics r Multimedia data type r Basic compression techniques.
8. 1 MPEG MPEG is Moving Picture Experts Group On 1992 MPEG-1 was the standard, but was replaced only a year after by MPEG-2. Nowadays, MPEG-2 is gradually.
1 Classification of Compression Methods. 2 Data Compression  A means of reducing the size of blocks of data by removing  Unused material: e.g.) silence.
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
Outline Kinds of Coding Need for Compression Basic Types Taxonomy Performance Metrics.
Compression video overview 演講者:林崇元. Outline Introduction Fundamentals of video compression Picture type Signal quality measure Video encoder and decoder.
Computer Vision – Compression(1) Hanyang University Jong-Il Park.
Chapter 17 Image Compression 17.1 Introduction Redundant and irrelevant information  “Your wife, Helen, will meet you at Logan Airport in Boston.
Chapter 1 Data Storage © 2007 Pearson Addison-Wesley. All rights reserved.
Additive White Gaussian Noise
Source Coding Efficient Data Representation A.J. Han Vinck.
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
CS654: Digital Image Analysis
Chapter 1 Background 1. In this lecture, you will find answers to these questions Computers store and transmit information using digital data. What exactly.
JPEG.
Chapter 1 Data Storage © 2007 Pearson Addison-Wesley. All rights reserved.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 7 – Basics of Compression (Part 2) Klara Nahrstedt Spring 2012.
Fundamentals of Multimedia Chapter 6 Basics of Digital Audio Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
1 Part A Multimedia Production Chapter 2 Multimedia Basics Digitization, Coding-decoding and Compression Information and Communication Technology.
An introduction to audio/video compression Prepared by :: Bhatt shivani ( )
Submitted To-: Submitted By-: Mrs.Sushma Rani (HOD) Aashish Kr. Goyal (IT-7th) Deepak Soni (IT-8 th )
Chapter 1: Data Storage.
Computer Science: An Overview Eleventh Edition
IMAGE COMPRESSION.
Data Compression.
Multimedia Outline Compression RTP Scheduling Spring 2000 CS 461.
Data Compression.
Data Compression CS 147 Minh Nguyen.
Context-based Data Compression
Why Compress? To reduce the volume of data to be transmitted (text, fax, images) To reduce the bandwidth required for transmission and to reduce storage.
UNIT IV.
CSE 589 Applied Algorithms Spring 1999
Image Transforms for Robust Coding
Speech coding.
15 Data Compression Foundations of Computer Science ã Cengage Learning.
Govt. Polytechnic Dhangar(Fatehabad)
Chapter 8 – Compression Aims: Outline the objectives of compression.
15 Data Compression Foundations of Computer Science ã Cengage Learning.
Presentation transcript:

ELG5126 Source Coding and Data Compression Eric Dubois

Context: signal transmission binary stream Information Source Encoder Channel Information Receiver Decoder signal binary stream

Context: signal transmission binary stream Information Source Encoder Channel aka ‘data’ Information Receiver Decoder signal binary stream

Context: signal transmission binary stream Information Source Encoder Channel error measure aka ‘data’ Information Receiver Decoder signal binary stream

Examples of information sources Speech Image Video Text file Music Radiograph Binary executable computer program Computer graphics primitives Weather radar map

Examples of channels Airwaves (EM radiation) Cable Telephone line Hard disk CD, DVD Flash memory device Optical path Internet

Examples of information receivers TV screen and viewer Audio system and listener Computer file Image printer and viewer Compute engine

Possible error measures No errors permitted (lossless coding) Numerical measures of error, e.g. mean- squared error (MSE), signal-to-noise ratio (SNR) Numerical measures of perceptual difference Mean opinion scores from human users

Measures of ‘channel rate’ Data rate (bits per second) Transmission time (seconds) File size (bytes) Average number of bits per source symbol

What is compression? There is usually a ‘natural’ representation for the source data at a given level of fidelity and sampling rate. Examples: 8 bits per character in ASCII data 24 bits per RGB color pixel 16 bits audio signal sample This natural representation leads to a certain raw channel rate (which is generally too high). Compression involves reducing the channel rate for a given level of distortion (which may be zero for lossless coding).

Compression ratio Example: HDTV, 1080I Raw channel rate: 1493 Mbit/s (1920*1080*30*24) Compressed channel rate: ~20 Mbit/s Compression ratio: ~75

Sources Categories of sources continuous time or domain: x(t), x(h,v) discrete time or domain: x[n], x[m,n] continuous amplitude or value: xR discrete amplitude or value: x A = {a1, a2, … aM} We will only consider discrete domain sources. We assume that continuous domain signals can be sampled with negligible loss. This is not considered in this course. We will mainly concentrate on one-dimensional signals such as text, speech, audio, etc. Extensions to images are covered in ELG5378. A source signal is a sequence of values drawn from a source alphabet A: x[1], x[2], … , x[n]  A

Source Coder A source coder transforms a source sequence into a coded sequence whose values are drawn from a code alphabet G : u[1], u[2], …, u[i]  G Normally G = {0,1}, and we will limit ourselves to this case. Note that the time indexes for the source sequence x[n] and the coded sequence u[i] do not correspond. The decoder must estimate the source signal on the basis of the received coded sequence û[i]. This may be different from u[i] if there are transmission errors. We will generally assume that there are no transmission errors.

Categories of compression coders Lossless coding: The source sequence has discrete values, and these must be reproduced without error. Examples where this is required is text, data, executables, and some quantized signals such as X-rays. Lossy coding: The source sequence may be either continuous or discrete valued. There exists a distortion criterion. The decoded sequence may be mathematically different from the source sequence, but the distortion should be kept sufficiently small. Examples are speech and images. Often a perceptual distortion criterion is desired. Lossless coding methods are often a component of a lossy coding system.

The compression problem There are two variants of the compression problem For a given source and distortion measure, minimize the channel rate for a given level of distortion D0 (which can be zero). For a given source and distortion measure, minimize the distortion (or maximize the quality) for a given channel rate R0.

Rate versus distortion performance In a coding system, there is typically a tradeoff between rate and distortion R D

Rate versus distortion performance In a coding system, there is typically a tradeoff between rate and distortion R D D0

Rate versus distortion performance In a coding system, there is typically a tradeoff between rate and distortion R R0 D

When can we compress? When there is statistical redundancy. For example, for a sequence of outcomes of a fair 16-sided die, we need 4 bits to represent each outcome. No compression is possible. In English text, some letters occur far more often than others. We can assign shorter codes to the common ones and longer codes to the uncommon ones and achieve compression (e.g., Morse code).

Statistical redundancy There are many types of statistical redundancy. For example, in English text, we are pretty sure that the next letter after a Q will be a U, so we can exploit it. The key to successful compression will be to formulate models that capture the statistical redundancy in the source.

When can we compress? (2) When there is irrelevancy. In many cases, the data is specified more precisely than it needs to be for the intended purpose. The data may be oversampled, or quantized more finely than it needs to be, either everywhere, or in some parts of the signal. This particularly applies to data meant only for consumption and not further processing.

Exploiting irrelevancy To exploit irrelevancy, we need a good model of the requirements of the receiver, e.g., human vision, hearing, etc. We also need a suitable representation of the data, e.g., transform or wavelet representations. Again, the key to success will be the formulation of appropriate models.

The elements of a source coder Change of representation Quantization (not for lossless coding) Binary code assignment All will depend on good models of the source and the receiver.

The Course Outline

Professor Eric Dubois CBY A-512 Tel: 562-5800 X 6400 edubois@uottawa.ca www.eecs.uottawa.ca/~edubois/courses/ELG5126

Textbook Textbook: K. Sayood, Introduction to Data Compression, third edition, Morgan Kaufmann Publishers, 2006.

Prerequisite Basic probability and signal processing as typically obtained in an undergraduate Electrical Engineering program (e.g., at uOttawa, ELG3125 Signal and System Analysis, ELG3126 Random Signals and Systems

Objective The objective of this course is to present the fundamental principles underlying data and waveform compression. The course begins with the study of lossless compression of discrete sources. These techniques are applicable to compression of text, data, programs and any other type of information where no loss is tolerable. They also form an integral part of schemes for lossy compression of waveforms such as audio and video signals, which is the topic of the second part of the course.

Objective The main goal of the course is to provide an understanding of the basic techniques and theories underlying popular compression systems and standards such as ZIP, FAX, MP3, JPEG, MPEG and so on, as well as the principles underlying future systems. Some of the applications will be addressed in student projects.

Course Outline Lossless coding: Discrete sources, binary codes, entropy, Huffman and related codes, Markov models, adaptive coding. Arithmetic coding: Principles, coding and decoding techniques, implementation issues. Dictionary techniques: Principles, static dictionary, adaptive dictionary. Waveform coding: Distortion measures, rate- distortion theory and bounds, models.

Course Outline (2) Quantization: Formulation, performance, uniform and non-uniform quantizers, quantizer optimization, vector quantization. Predictive coding: Prediction theory, differential coding (DPCM), adaptive coding. Transform and subband coding: Change of basis, block transforms and filter banks, bit allocation and quantization. Applications (student projects)

Grading 20% Assignments: Several assignments, to be handed in during class on the due-date specified. There will be a 5% penalty for each day late, and no assignment will be accepted after one week. 30% Project: An individual project on an application of data compression involving some experimental work. A project report and presentation at the end of the course will be required. More details will follow early in the course. 20% Midterm exam: Closed-book exam, 80 minutes in length. 30% Final exam: Closed-book exam, 3 hours in length, covering the whole course.

Enjoy the course