In the beginning was the Word... 情報理論:日本語,英語で隔年開講 今年度は日本語で授業を行うが,スライドは英語のものを 使用 Information Theory: English and Japanese, alternate years the course will.

Slides:



Advertisements
Similar presentations
Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Advertisements

STATISTIC & INFORMATION THEORY (CSNB134) MODULE 12 ERROR DETECTION & CORRECTION.
Information Theory EE322 Al-Sanie.
Enhancing Secrecy With Channel Knowledge
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
1 exercise in the previous class Determine the stationary probabilities. Compute the probability that 010 is produced. A BC 0/0.4 0/0.5 1/0.6 0/0.81/0.5.
Learning Objectives Explain similarities and differences among algorithms, programs, and heuristic solutions List the five essential properties of an algorithm.
Chapter 6 Information Theory
FOUR WEEKS PROJECT TOPIC: ERROR DETECTION AND CORRECTION GROUP MEMBERS: ALEX NYARKO ANTHONY TSETSE BENJAMIN BOATENG SUPERVISOR: ZOLTAN SAFAR DATE: JUNE.
ENGS Lecture 8 ENGS 4 - Lecture 8 Technology of Cyberspace Winter 2004 Thayer School of Engineering Dartmouth College Instructor: George Cybenko,
The 2006 International Seminar of E-commerce Academic and Application Research Tainan, Taiwan, R.O.C. March 1-2,2006 An Application of Coding Theory into.
SWE 423: Multimedia Systems
Fundamental limits in Information Theory Chapter 10 :
JPN 311: Conversation and Composition 許可 (permission)
2015/6/15VLC 2006 PART 1 Introduction on Video Coding StandardsVLC 2006 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
CSCI 3 Chapter 1.8 Data Compression. Chapter 1.8 Data Compression  For the purpose of storing or transferring data, it is often helpful to reduce the.
Spatial and Temporal Data Mining
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Information Theory and Security. Lecture Motivation Up to this point we have seen: –Classical Crypto –Symmetric Crypto –Asymmetric Crypto These systems.
Code and Decoder Design of LDPC Codes for Gbps Systems Jeremy Thorpe Presented to: Microsoft Research
1 Chapter 1 Introduction. 2 Outline 1.1 A Very Abstract Summary 1.2 History 1.3 Model of the Signaling System 1.4 Information Source 1.5 Encoding a Source.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
2015/7/12VLC 2008 PART 1 Introduction on Video Coding StandardsVLC 2008 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
15-853Page :Algorithms in the Real World Error Correcting Codes I – Overview – Hamming Codes – Linear Codes.
Exercise IV-A p.164. What did they say? 何と言ってましたか。 1.I’m busy this month. 2.I’m busy next month, too. 3.I’m going shopping tomorrow. 4.I live in Kyoto.
A Short Course at Tamkang University Taipei, Taiwan, R.O.C. March An Application of Coding Theory into Experimental Design Shigeichi Hirasawa.
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Information Theory and Security
Management Information Systems Lection 06 Archiving information CLARK UNIVERSITY College of Professional and Continuing Education (COPACE)
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Richard W. Hamming Learning to Learn The Art of Doing Science and Engineering Session 10: Coding Theory I Learning to Learn The Art of Doing Science and.
1 INF244 Textbook: Lin and Costello Lectures (Tu+Th ) covering roughly Chapter 1;Chapters 9-19? Weekly exercises: For your convenience Mandatory.
Dr.-Ing. Khaled Shawky Hassan
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 5 Section 1 – Slide 1 of 33 Chapter 5 Section 1 Probability Rules.
Exercise in the previous class (Apr. 26) For s = (|s|=36), compute  2 -values of s for block length with 1, 2, 3.
Information and Coding Theory Linear Block Codes. Basic definitions and some examples. Juris Viksna, 2015.
Exercise in the previous class p: the probability that symbols are delivered correctly C: 1 00 → → → → What is the threshold.
Information Coding in noisy channel error protection:-- improve tolerance of errors error detection: --- indicate occurrence of errors. Source.
exercise in the previous class
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
Image Compression (Chapter 8) CSC 446 Lecturer: Nada ALZaben.
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
Data and Computer Communications by William Stallings Eighth Edition Digital Data Communications Techniques Digital Data Communications Techniques Click.
Summer 2004CS 4953 The Hidden Art of Steganography A Brief Introduction to Information Theory  Information theory is a branch of science that deals with.
Lecture PowerPoint Slides Basic Practice of Statistics 7 th Edition.
Error Control Code. Widely used in many areas, like communications, DVD, data storage… In communications, because of noise, you can never be sure that.
Introduction to Digital and Analog Communication Systems
Information Theory The Work of Claude Shannon ( ) and others.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Essential Statistics Chapter 91 Introducing Probability.
Chapter 10 Introducing Probability BPS - 5th Ed. Chapter 101.
COMP541 Arithmetic Circuits
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
Chapter 10 Algorithmic Thinking. Learning Objectives Explain similarities and differences among algorithms, programs, and heuristic solutions List the.
Prof. Noriyoshi Yamauchi
K p k'k' p'p'  probability amplitude locality,Lorentz inv.gauge inv. spinor   vector A  T  electron quark scattering scattering cross section Feynman.
1 CSCD 433 Network Programming Fall 2013 Lecture 5a Digital Line Coding and other...
Lecture 20 CSE 331 July 30, Longest path problem Given G, does there exist a simple path of length n-1 ?
Error Detecting and Error Correcting Codes
Bootstrapping 2014/4/13 R basic 3 Ryusuke Murakami.
DIGITAL COMMUNICATION. Introduction In a data communication system, the output of the data source is transmitted from one point to another. The rate of.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
1 CSCD 433 Network Programming Fall 2016 Lecture 4 Digital Line Coding and other...
8 Coding Theory Discrete Mathematics: A Concept-based Approach.
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Information Theory Michael J. Watts
Context-based Data Compression
A Brief Introduction to Information Theory
Lecture 2: Basic Information Theory
Presentation transcript:

In the beginning was the Word... 情報理論:日本語,英語で隔年開講 今年度は日本語で授業を行うが,スライドは英語のものを 使用 Information Theory: English and Japanese, alternate years the course will be taught in Japanese in this year video-recorded English classes  Lecture Archives 2011 Slides are in English this slide can be found at test questions are given in both of Japanese and English 1

Information Theory Information Theory (情報理論) is founded by C. E. Shannon in 1948 focuses on mathematical theory of communication gave essential impacts on today’s digital technology wired/wireless communication/broadcasting CD/DVD/HDD data compression cryptography, linguistics, bioinformatics, games,... In this class, we learn basic subjects of information theory. (half undergraduate level + half graduate school level) 2 Claude E. Shannon

class plan This class consists of four chapters (+ this introduction): chapter 0: the summary and the schedule of this course (today) chapter 1: measuring information chapter 2: compact representation of information chapter 3: coding for noisy communication chapter 4: cryptography 3

what’s the problem? To understand our problem, date back to 1940s... Teletype ( 電信 ) was widely used for communication. Morse code: dots ( ∙ ) and dashes ( − ) 4 They already had “digital communication” dot = 1 unit long, dash = 3 units long 1 unit silence between marks 3 units silence between letters, etc.

machinery for information processing No computers yet, but there were “machines”... 5 They could do something complicated. The transmission/recording of messages were... inefficient...messages should be as short as possible unreliable...messages are often disturbed by noises The efficiency and the reliability were two major problems. Teletype model 14-KTR, Enigma machine

the model of communication A communication system can be modeled as; 6 C.E. Shannon, A Mathematical Theory of Communication, The Bell System Technical Journal, 27, pp. 379–423, 623–656, encoder, modulator, codec, etc... channel, storage medium, etc...

what is the “efficiency”? A communication is efficient if the size of B is small. subject to A = D, or A ≈ D with, or without noise (B ≠ C, or B = C) 7 ABCD

problem one: efficiency Example: You need to record the weather of Tokyo everyday. weather = {sunny, cloudy, rainy} You can use “0” and “1”, but you cannot use blank spaces. 8 weather sunny cloudy rainy codeword bit record everyday 200 bits for 100 days Can we shorten the representation?

better code? The code B gives shorter representation than the code A. Can we decode the code B correctly? Yes, as far as the sequence is processed from the beginning. Is there a code which is more compact than this code B? No, and Yes(→ next slide). 9 weather sunny cloudy rainy code A code B code A code B

think average Sometimes, events are not equally likely with the code A, 2.0 bit / event (always) with the code B, 2    0.2 = 1.8 bit / event in average with the code C, 1    0.2 = 1.5 bit / event in average weather sunny cloudy rainy probability code A code B code C

the best code? Can we represent information with bit per event?...No, maybe. It is likely that there is a “limit” which we cannot get over. Shannon investigated the limit mathematically. → For this event set, we need or more bit per event. 11 weather sunny cloudy rainy probability This is the amount of information which must be carried by the code.

class plan in April chapter 0: the summary and the schedule of this course chapter 1: measuring information We establish a mathematical means to measure information in a quantitative manner. chapter 2: compact representation of information We learn several coding techniques which give compact representation of information. chapter 3: coding for noisy communication chapter 4: cryptography 12

what is the “reliability”? A communication is reliable if A = D or A ≈ D. the existence of noise is essential (B ≠ C) How small can we make the size of B? 13 ABCD

problem two: reliability Communication is not always reliable. transmitted information ≠ received information 14 Errors of this kind are unavoidable in real communication. In the usual conversation, we sometimes use phonetic codes. ABCABC ABCADC ABCAlpha, Bravo, Charlie ABC Alpha, Bravo, Charlie あさひの「あ」 いろはの「い」

phonetic code A phonetic code adds redundant information. The redundant part helps correcting possible errors. →use this mechanism over 0-1 data, and we can correct errors! 15 the real information redundant ( 冗長な ) information for correcting possible errors

redundancy Q. Can we add “redundancy” to binary data? A. Yes, use parity bits. A parity bit is... a binary digit which is to make the number of 1’s in data even → (two 1’s → two 1’s) → (three 1’s → four 1’s) One parity bit may tell you that there are odd numbers of errors, but not more than that. 16

to correct error(s) basic idea: use several parity bits to correct errors Example: Add five parity bits to four-bits data (a 0, a 1, a 2, a 3 ). 17 This code corrects one-bit error, but it is too straightforward. a0a0 a2a2 a1a1 a3a3 p0p0 p1p1 q0q0 q1q1 r codeword = (a 0, a 1, a 2, a 3, p 0, p 1, q 0, q 1, r)

class plan in May chapter 0: the summary and the schedule of this course chapter 1: measuring information chapter 2: compact representation of information chapter 3: coding for noisy communication We study practical coding techniques for finding and correcting errors. chapter 4: cryptography We review techniques for protecting information from intensive attacks. 18

schedule April 19 Tue Thu May June × × test: questions given in English/Japanese report (quiz): will be assigned by the end of April (Mon) 0405 × statistics in 2011: A / B / C / did not pass... 13

chapter 1: measuring information 20

motivation “To tell plenty of things, we need more words.”...maybe true, but can you give the proof of this statement? We will need to measure information quantitatively ( 定量的に測る ) 2. observe the relation between the amount of information and its representation. Chapter 1 focuses on the first step above. 21

the uncertainty ( 不確実さ ) Information tells what has happened at the information source. Before you receive information, there is much uncertainty. After you receive information, the uncertainty becomes small. the difference of uncertainty  the amount of information FIRST, we need to measure the uncertainty of information source. 22 much uncertainty Before small uncertainty After this difference indicates the amount of information

the definition of uncertainty The uncertainty is defined according to the statistics ( 統計量 ), BUT, we do not have enough time today.... In the rest of today’s talk, we study two typical information sources. memoryless & stationary information source Markov information source 23

assumption In this class, we assume that... an information source produces one symbol per unit time (discrete-time information source) the set of possible symbols is finite and countable ( 有限可算 ) (digital information source) Note however that, in the real world, there are continuous-time and/or analogue information sources. cf. sampling & quantization 24

Preliminary ( 準備 ) Assume a discrete-time digital information source S: M = {a 1,..., a k }... the set of symbols of S (S is said to be a k-ary information source.) X t...the symbol which S produces at time t The sequence X 1,..., X n is called a message produced by S. Example: S = fair dice 25 if the message is, then

memoryless & stationary information source 26 trial 1 trial 2 trial 3 : ajcgea... gajkfh... wasdas... : the same probability distribution memoryless = 無記憶 stationary = 定常

memoryless & stationary information source 27

Markov information source a simple model of information source with memory The choice of the next symbol depends on at most m previous symbols (m-th order Markov source) 28 Andrey Markov m = 0  memoryless source m = 1  simple Markov source

Example of (simple) Markov source S... memoryless & stationary source with P(0) = q, P(1) = 1 – q 29 S R 1-bit register XtXt if X t–1 = 0, then R = 0: S = 0  X t = 0... P Xt|Xt–1 (0 | 0) = q S = 1  X t = 1... P Xt|Xt–1 (1 | 0) = 1 – q if X t–1 = 1, then R = 1: S = 0  X t = 1... P Xt|Xt–1 (1 | 1) = q S = 1  X t = 0... P Xt|Xt–1 (0 | 1) = 1 – q

01 1 / 1–q 0 / 1–q 0 / q1 / q Markov source as a finite state machine m-th order k-ary Markov source: The next symbols depends on previous m symbols. The model is having one of k m internal states. The state changes when a new symbol is generated.  finite state machine 30 S R 1-bit register XtXt generated symbol probability

31 two important properties irreducible ( 既約 ) Markov source: We can move to any state from any state. A BC this example is NOT irreducible aperiodic ( 非周期的 ) Markov source: We have no periodical behavior (strict discussion needed...). AB this example is NOT aperiodic irreducible + aperiodic = regular

32 example of the regular Markov source AB 0/0.9 1/0.1 0/0.8 1/0.2 converge ( 収束する ) to the same probabilities stationary probabilities time P (state=A) P (state=B) start from the state 0 time P (state=A) P (state=B) start from the state 1

33 computation of the stationary probabilities  t : P(state = A) at time t  t : P(state = B) at time t  t+1 = 0.9  t  t  t+1 = 0.1  t  t  t+1 +  t+1 = 1 AB 0/0.9 1/0.1 0/0.81/0.2 If  t and  t converge to  and , respectively, then we can put  t+1 =  t =  and  t+1 =  t = .  = 0.9    = 0.1    +  = 1  =8/9,  =1/9

34 Markov source as a stationary source After enough time has elapsed... a regular Markov source can be regarded as a stationary source AB 0/0.9 1/0.1 0/0.81/0.2  =8/9,  =1/9 0 will be produced with probability P(0) = 0.9   = will be produced with probability P(1) = 0.1   = 0.111

35 summary of today’s class overview of this course motivation four chapters typical information sources memoryless & stationary source Markov source

36 exercise Determine the stationary probabilities. Compute the probability that 010 is produced. A BC 0/0.4 0/0.5 1/0.6 0/0.81/0.5 1/0.2 This is to check your understanding. This is not a report assignment.