Inf 723 Information & Computing

Slides:



Advertisements
Similar presentations
DCSP-7: Information Jianfeng Feng Department of Computer Science Warwick Univ., UK
Advertisements

Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Information Theory EE322 Al-Sanie.
CSC Dr. Gary Locklair Exam #4 … CSC Dr. Gary Locklair update date on slides 5, 6, 7.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
FIU Chapter 7: Input/Output Jerome Crooks Panyawat Chiamprasert
Michael A. Nielsen University of Queensland Quantum entropy Goals: 1.To define entropy, both classical and quantum. 2.To explain data compression, and.
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
Chapter 6 Information Theory
Interrupts (contd..) Multiple I/O devices may be connected to the processor and the memory via a bus. Some or all of these devices may be capable of generating.
SWE 423: Multimedia Systems
CSCI 3 Chapter 1.8 Data Compression. Chapter 1.8 Data Compression  For the purpose of storing or transferring data, it is often helpful to reduce the.
1 NETWORK CODING Anthony Ephremides University of Maryland - A NEW PARADIGM FOR NETWORKING - February 29, 2008 University of Minnesota.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
CHAPTER 6 FILE PROCESSING. 2 Introduction  The most convenient way to process involving large data sets is to store them into a file for later processing.
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Noise, Information Theory, and Entropy
STATISTIC & INFORMATION THEORY (CSNB134)
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
Chapter 1 Basics of Probability.
CSC 335 Data Communications and Networking Lecture 4c: Communication and Multiplexing Dr. Cheer-Sun Yang.
Foundations of Computer Science Computing …it is all about Data Representation, Storage, Processing, and Communication of Data 10/4/20151CS 112 – Foundations.
(Important to algorithm analysis )
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
ICS 220 – Data Structures and Algorithms Lecture 11 Dr. Ken Cosh.
ALGORITHMS FOR ISNE DR. KENNETH COSH WEEK 13.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
Linawati Electrical Engineering Department Udayana University
Summer 2004CS 4953 The Hidden Art of Steganography A Brief Introduction to Information Theory  Information theory is a branch of science that deals with.
Prof. Amr Goneid, AUC1 Analysis & Design of Algorithms (CSCE 321) Prof. Amr Goneid Department of Computer Science, AUC Part 8. Greedy Algorithms.
“Victor Babes” UNIVERSITY OF MEDICINE AND PHARMACY TIMISOARA DEPARTMENT OF MEDICAL INFORMATICS AND BIOPHYSICS Medical Informatics Division
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
Dept of Bioenvironmental Systems Engineering National Taiwan University Lab for Remote Sensing Hydrology and Spatial Modeling Introduction STATISTICS Introduction.
Coding Theory Efficient and Reliable Transfer of Information
Source Coding Efficient Data Representation A.J. Han Vinck.
Basic Principles (continuation) 1. A Quantitative Measure of Information As we already have realized, when a statistical experiment has n eqiuprobable.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Combinatorics (Important to algorithm analysis ) Problem I: How many N-bit strings contain at least 1 zero? Problem II: How many N-bit strings contain.
Entropy (YAC- Ch. 6)  Introduce the thermodynamic property called Entropy (S)  Entropy is defined using the Clausius inequality  Introduce the Increase.
STATISTIC & INFORMATION THEORY (CSNB134) MODULE 11 COMPRESSION.
1 CSCD 433 Network Programming Fall 2013 Lecture 5a Digital Line Coding and other...
Lecture 12 Huffman Algorithm. In computer science and information theory, a Huffman code is a particular type of optimal prefix code that is commonly.
Huffman Coding (2 nd Method). Huffman coding (2 nd Method)  The Huffman code is a source code. Here word length of the code word approaches the fundamental.
ON HERITAGE AND THE INFORMATION DISCIPLINES By Marcia J. Bates Professor Emerita Department of Information Studies University of California, Los Angeles.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
1 CSCD 433 Network Programming Fall 2016 Lecture 4 Digital Line Coding and other...
Information Theory Information Suppose that we have the source alphabet of q symbols s 1, s 2,.., s q, each with its probability p(s i )=p i. How much.
KHARKOV NATIONAL MEDICAL UNIVERSITY MEDICAL INFORMATICS МЕДИЧНА ІНФОРМАТИКА MEDICAL INFORMATICS.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Shannon Entropy Shannon worked at Bell Labs (part of AT&T)
(Important to algorithm analysis )
Data Compression.
Data Compression CS 147 Minh Nguyen.
Information Theory Michael J. Watts
Context-based Data Compression
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
(Important to algorithm analysis )
Analysis & Design of Algorithms (CSCE 321)
A Brief Introduction to Information Theory
Inf 723 Information & Computing
Human Communication 101.
Dr. Clincy Professor of CS
Data Information Knowledge and Processing
Presenting information as bit patterns
Floating Point Numbers - continuing
Data Compression.
Presentation transcript:

Inf 723 Information & Computing Jagdish S. Gangolly Interdisciplinary PhD Program in Information Science Department of Informatics, College of Computing & Information State University of New York at Albany 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Information Reports Information context (factive, not truth-functional) Information content ( Information carrier Indicating fact 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) An example: The acoustic waves from the speaker carry the information that the announcer said, “Nancy Reagan is irritated.” 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Information context: The acoustic waves from Information carrier: the speaker carry the information Information content: that the announcer said, Indicating fact: “Nancy Reagan is irritated.” 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Principles Facts carry information. The informational content of a fact is a true proposition. The information a fact carries is relative to a constraint. The information a fact carries is not an intrinsic property of it. The informational content of a fact can concern remote things and situations. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Principles Informational content can be specific; the propositions that are informational contents can be about objects that are not part of the indicating fact. Indicating facts contain such information only relative to connecting facts; the information is incremental, given those facts. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Principles Many different facts, involving variations in objects, properties, relations and spatiotemporal locations, can indicate one and the same informational content—relative to the same or different constraints. Information can be stored and transmitted in a variety of forms. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Principles Having information is good; creatures whose behavior is guided or controlled by information (by their information carrying states) are more likely to succeed than those which are not so guided. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) To give form to a message by moulding it into a shape or pattern that can be communicated. Measurement, Quantitative Meaning, Qualitative 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy Measure of information content Communications example Claude Shannon’s work 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy Consider the proposition “it will rain tomorrow”. Let p = Probability{It will rain tomorrow} The information content I of a message with probability of occurrence p is I = log (1/p) = -log p 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy I = log (1/p) = -log p If p = 0, I = infinity. In other words, if a priori you consider an even to be impossible, then the message that the event occurred has information content infinity, since you would be very surprised If p = 1, I = 0. In other words, if a priori you consider an even to be a certainty, then the information content of the message that it occurred is 0 since you are not surprised at all. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy The expected information content of a certain message about the rain is given by: H(p) = p . log (1/p) + (1 - p) . log (1 - p) 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy If p = 0, then H(0) = 0 . log (1/0) + 1 . log (1/1) = 0 If p = 1, then H(1) = 1 . log (1/1) + 0 . log (0) = 0 If p = 1/2, then H(1/2) = 1/2 . log (2) + 1/2 . log (2) = 1 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy The information content is usually measured in bits which (take values 0 or 1), and therefore the logarithms in the formulae are to the base 2 However, one can measure such information content in other measurement units such as 10 (decimal) or e (natural). If it is measured in units of 10, they are called Hartleys, named after the statistician who developed the concept. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding Weather in California: sunny or cloudy Probabilities: P{sunny}=7/8, P{cloudy}=1/8 Weather reports for two days to be sent The weather on the two consequent days are independent, ie., P(w1 & w2) = P(w1) . P(w2) 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding Suppose you use code alphabet for weather {0, 1} where 0 stands for sunny and 1 stands for cloudy. . If the two days weather information is to be sent in a message, the average length of the message code is 2 (00, 01, 10, or 11) bits, or 2 .(8/7) + 2 . (1/7) = 2 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding The entropy of the message is H = (7/8) .log( 1/(7/8)) + (1/8) .log(1/(1/8)) = 0.54 bits The question is, if we can exploit the fact that the weather on two consecutive days is independent, in order to reduce the average code length og the message. This can be accomplished by the following code: 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding Weather Probability Code SS (7/8) .(7/8)=49/64 SC (7/8) .(1/8)=7/64 10 CS (1/8) .(7/8)=7/64 110 CC (1/8) .(1/8)=1/64 11 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding The average length of this code is given by 1/2/2019 Inf 723 Information & Computing (Gangolly)

Entropy - An Application in Coding Therefore, the average length of a daily report is L/2 = 0.5 .(86/64) = 0.68 By considering a larger sequence for the weather report, we have reduced the average length of the report closer to the entropy of 0.54. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy In the transmission of messages over wires, one can achieve higher reliability by repeating each character in the message by repetition and polling. The greater the repetition the greater the reliability However, such repetition adds overheads to the payload message. At the limit, we can achieve perfect reliability by repeating the characters in the message infinite times, but the message except for the first character would never be sent. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Inf 723 Information & Computing (Gangolly) Entropy Shannon showed that it was possible, by appropriate coding schemes, it is possible to achieve reliability without sacrificing efficiency This is accomplished by two strategies assigning shorter codes to highly likely events grouping events 1/2/2019 Inf 723 Information & Computing (Gangolly)

Forms of Information (Bates) Information is the pattern of organization of matter and energy. All information is natural information, in that it exists in the material world of matter and energy. Represented information is natural information that is encoded or embodied. 1/2/2019 Inf 723 Information & Computing (Gangolly)

Forms of Information (Bates) “an organizing mechanism which provides an ability to deal with the environment. It is a symbolic description having modes of interpreting and interacting with the environment” (Goonatilake) 1/2/2019 Inf 723 Information & Computing (Gangolly)

Forms of Information (Bates) Information flow lineages (Goonatilake) Genetic information Neural-cultural (Experienced, enacted, expressed) Exo-somatic (information stored outside the animal as the “externalization of memories”) (embedded, recorded) Residue (trace) 1/2/2019 Inf 723 Information & Computing (Gangolly)

Forms of Information (Bates) Information as a sign (semiotics) Sign, interpretation, reference to object 1/2/2019 Inf 723 Information & Computing (Gangolly)