Information Theory and Security

Slides:



Advertisements
Similar presentations
Copyright © Cengage Learning. All rights reserved.
Advertisements

Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
1 Probability Theory Dr. Deshi Ye
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Probability Simple Events
Basic Terms of Probability Section 3.2. Definitions Experiment: A process by which an observation or outcome is obtained. Sample Space: The set S of all.
Mathematics in Today's World
SI485i : NLP Day 2 Probability Review. Introduction to Probability Experiment (trial) Repeatable procedure with well-defined possible outcomes Outcome.
Describing Probability
MAT 103 Probability In this chapter, we will study the topic of probability which is used in many different areas including insurance, science, marketing,
Independence and the Multiplication Rule
1 Codes, Ciphers, and Cryptography-Ch 2.3 Michael A. Karls Ball State University.
Probability theory and average-case complexity. Review of probability theory.
Chain Rules for Entropy
4.2 Probability Models. We call a phenomenon random if individual outcomes are uncertain but there is nonetheless a regular distribution of outcomes in.
Business Statistics: A Decision-Making Approach, 7e © 2008 Prentice-Hall, Inc. Chap 4-1 Business Statistics: A Decision-Making Approach 7 th Edition Chapter.
Fundamental limits in Information Theory Chapter 10 :
Excursions in Modern Mathematics, 7e: Copyright © 2010 Pearson Education, Inc. 15 Chances, Probabilities, and Odds 15.1Random Experiments and Sample.
Basic Probability. Theoretical versus Empirical Theoretical probabilities are those that can be determined purely on formal or logical grounds, independent.
Ref. Cryptography: theory and practice Douglas R. Stinson
Information Theory and Security pt. 2. Lecture Motivation Previous lecture talked about a way to measure “information”. In this lecture, our objective.
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Information Theory and Security. Lecture Motivation Up to this point we have seen: –Classical Crypto –Symmetric Crypto –Asymmetric Crypto These systems.
June 1, 2004Computer Security: Art and Science © Matt Bishop Slide #32-1 Chapter 32: Entropy and Uncertainty Conditional, joint probability Entropy.
LING 438/538 Computational Linguistics Sandiway Fong Lecture 17: 10/24.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
CryptographyPerfect secrecySlide 1 Today What does it mean for a cipher to be: –Computational secure? Unconditionally secure? Perfect secrecy –Conditional.
Algebra1 Independent and Dependent Events
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Chapter 9 Introducing Probability - A bridge from Descriptive Statistics to Inferential Statistics.
§1 Entropy and mutual information
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
Sets, Combinatorics, Probability, and Number Theory Mathematical Structures for Computer Science Chapter 3 Copyright © 2006 W.H. Freeman & Co.MSCS SlidesProbability.
1 9/8/2015 MATH 224 – Discrete Mathematics Basic finite probability is given by the formula, where |E| is the number of events and |S| is the total number.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
All of Statistics Chapter 5: Convergence of Random Variables Nick Schafer.
EXIT NEXT Click one of the buttons below or press the enter key BACKTOPICSProbability Mayeen Uddin Khandaker Mayeen Uddin Khandaker Ph.D. Student Ph.D.
Simple Mathematical Facts for Lecture 1. Conditional Probabilities Given an event has occurred, the conditional probability that another event occurs.
Copyright © Cengage Learning. All rights reserved. CHAPTER 9 COUNTING AND PROBABILITY.
College Algebra Sixth Edition James Stewart Lothar Redlin Saleem Watson.
BA 201 Lecture 6 Basic Probability Concepts. Topics Basic Probability Concepts Approaches to probability Sample spaces Events and special events Using.
1 Information Theory Nathanael Paul Oct. 09, 2002.
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
PROBABILITY, PROBABILITY RULES, AND CONDITIONAL PROBABILITY
Review Homework pages Example: Counting the number of heads in 10 coin tosses. 2.2/
Sixth lecture Concepts of Probabilities. Random Experiment Can be repeated (theoretically) an infinite number of times Has a well-defined set of possible.
Natural Language Processing Giuseppe Attardi Introduction to Probability IP notice: some slides from: Dan Jurafsky, Jim Martin, Sandiway Fong, Dan Klein.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Probability. What is probability? Probability discusses the likelihood or chance of something happening. For instance, -- the probability of it raining.
+ Chapter 5 Overview 5.1 Introducing Probability 5.2 Combining Events 5.3 Conditional Probability 5.4 Counting Methods 1.
Great Theoretical Ideas in Computer Science for Some.
Bringing Together Paradox, Counting, and Computation To Make Randomness! CS Lecture 21 
Warm Up: Quick Write Which is more likely, flipping exactly 3 heads in 10 coin flips or flipping exactly 4 heads in 5 coin flips ?
Huffman Coding (2 nd Method). Huffman coding (2 nd Method)  The Huffman code is a source code. Here word length of the code word approaches the fundamental.
Probability theory and average-case complexity. Review of probability theory.
Dr. Michael Nasief 1. MathematicsMeasure TheoryProbability TheoryRandom Process theory 2.
Counting and Probability. Imagine tossing two coins and observing whether 0, 1, or 2 heads are obtained. Below are the results after 50 tosses Tossing.
Probability and statistics - overview Introduction Basics of probability theory Events, probability, different types of probability Random variable, probability.
1 What Is Probability?. 2 To discuss probability, let’s begin by defining some terms. An experiment is a process, such as tossing a coin, that gives definite.
1 COMP2121 Discrete Mathematics Principle of Inclusion and Exclusion Probability Hubert Chan (Chapters 7.4, 7.5, 6) [O1 Abstract Concepts] [O3 Basic Analysis.
Essential Ideas for The Nature of Probability
Chapter 4 Probability Concepts
What is Probability? Quantification of uncertainty.
Introduction to Probability
CSCI 5832 Natural Language Processing
PROBABILITY AND STATISTICS
Lecture 2 Basic Concepts on Probability (Section 0.2)
Sets, Combinatorics, Probability, and Number Theory
Presentation transcript:

Information Theory and Security

Lecture Motivation Up to this point we have seen: Classical Crypto Symmetric Crypto Asymmetric Crypto These systems have focused on issues of confidentiality: Ensuring that an adversary cannot infer the original plaintext message, or cannot learn any information about the original plaintext from the ciphertext. But what does “information” mean? In this lecture and the next we will put a more formal framework around the notion of what information is, and use this to provide a definition of security from an information-theoretic point of view.

Lecture Outline Probability Review: Conditional Probability and Bayes Entropy: Desired properties and definition Chain Rule and conditioning Coding and Information Theory Huffman codes General source coding results Secrecy and Information Theory Probabilistic definitions of a cryptosystem Perfect Secrecy

The Basic Idea Suppose we roll a 6-sided dice. Let A be the event that the number of dots is odd. Let B be the event that the number of dots is at least 3. A = {1, 3, 5} B = {3, 4, 5, 6} I tell you: the roll belongs to both A and B then you know there are only two possibilities: {3, 5} In this sense tells you more than just A or just B. That is, there is less uncertainty in than in A or B. Information is closely linked with this idea of uncertainty: Information increases when uncertainty decreases.

Probability Review, pg. 1 A random variable (event) is an experiment whose outcomes are mapped to real numbers. For our discussion we will deal with discrete-valued random variables. Probability: We denote pX(x) = Pr(X = x). For a subset A, Joint Probability: Sometimes we want to consider more than two events at the same time, in which we case we lump them together into a joint random variable, e.g. Z = (X,Y). Independence: We say that two events are independent if

Probability Review, pg. 2 Conditional Probability: We will often ask questions about the probability of events Y given that we have observed X=x. In particular, we define the conditional probability of Y=y given X=x by Independence: We immediately get Bayes’s Theorem: If pX(x)>0 and pY(y)>0 then

Example Example: Suppose we draw a card from a standard deck. Let X be the random variable describing the suit (e.g. clubs, diamonds, hearts, spades). Let Y be the value of the card (e.g. two, three, …, ace). Then Z=(X,Y) gives the 52 possibilities for the card. P( (X,Y) = (x,y) ) = P(X=x, Y=y) = 1/52 P(X=“clubs”) = 13/52 = ¼ P(Y=“3”) = 4/52 = 1/13

Entropy and Uncertainty We are concerned with how much uncertainty a random event has, but how do we define or measure uncertainty? We want our measure to have the following properties: To each set of nonnegative numbers with , we define the uncertainty by . should be a continuous function: A slight change in p should not drastically change for all n>0. Uncertainty increases when there are more outcomes. If 0<q<1, then

Entropy, pg. 2 We define the entropy of a random variable by Example: Consider a fair coin toss. There are two outcomes, with probability ½ each. The entropy is Example: Consider a non-fair coin toss X with probability p of getting heads and 1-p of getting tails. The entropy is The entropy is maximum when p= ½.

Entropy, pg. 3 Entropy may be thought of as the number of yes-no questions needed to accurately determine the outcome of a random event. Example: Flip two coins, and let X be the number of heads. The possibilities are {0,1,2} and the probabilities are {1/4, 1/2, 1/4}. The Entropy is So how can we relate this to questions? First, ask “Is there exactly one head?” You will half the time get the right answer… Next, ask “Are there two heads?” Half the time you needed one question, half you needed two

Entropy, pg. 4 Suppose we have two random variables X and Y, the joint entropy H(X,Y) is given by Conditional Entropy: In security, we ask questions of whether an observation reduces the uncertainty in something else. In particular, we want a notion of conditional entropy. Given that we observe event X, how much uncertainty is left in Y?

Entropy, pg. 5 Chain Rule: The Chain Rule allows us to relate joint entropy to conditional entropy via H(X,Y) = H(Y|X)+H(X). (Remaining details will be provided on the white board) Meaning: Uncertainty in (X,Y) is the uncertainty of X plus whatever uncertainty remains in Y given we observe X.

Entropy, pg. 6 Main Theorem: Entropy is non-negative. where denotes the number of elements in the sample space of X. (Conditioning reduces entropy) with equality if and only if X and Y are independent.

Entropy and Source Coding Theory There is a close relationship between entropy and representing information. Entropy captures the notion of how many “Yes-No” questions are needed to accurately identify a piece of information… that is, how many bits are needed! One of the main focus areas in the field of information theory is on the issue of source-coding: How to efficiently (“Compress”) information into as few bits as possible. We will talk about one such technique, Huffman Coding. Huffman coding is for a simple scenario, where the source is a stationary stochastic process with independence between successive source symbols

Huffman Coding, pg. 1 Suppose we have an alphabet with four letters A, B, C, D with frequencies: We could represent this with A=00, B=01, C=10, D=11. This would mean we use an average of 2 bits per letter. On the other hand, we could use the following representation: A=1, B=01, C=001, D=000. Then the average number of bits per letter becomes (0.5)*1+(0.3)*2+(0.1)*3+(0.1)*3 = 1.7 Hence, this representation, on average, is more efficient. A B C D 0.5 0.3 0.1 0.1

Huffman Coding, pg. 2 Huffman Coding is an algorithm that produces a representation for a source. The Algorithm: List all outputs and their probabilities Assign a 1 and 0 to smallest two, and combine to form an output with probability equal to the sum Sort List according to probabilities and repeat the process The binary strings are then obtained by reading backwards through the procedure A 0.5 1 1.0 1 B 0.3 0.5 1 C 0.1 0.2 D 0.1 Symbol Representations A: 1 B: 01 C: 001 D: 000

Huffman Coding, pg. 3 In the previous example, we used probabilities. We may directly use event counts. Example: Consider 8 symbols, and suppose we have counted how many times they have occurred in an output sample. We may derive the Huffman Tree The corresponding length vector is (2,2,3,3,3,4,5,5) The average codelength is 2.83. If we had used a full-balanced tree representation (i.e. the straight-forward representation) we would have had an average codelength of 3. S1 S2 S3 S4 S5 S6 S7 S8 28 25 20 16 15 8 7 5

Huffman Coding, pg. 4 We would like to quantify the average amount of bits needed in terms of entropy. Theorem: Let L be the average number of bits per output for Huffman encoding of a random variable X, then Here, lx =length of codeword assigned to symbol x. Example: Let’s look back at the 4 symbol example Our average codelength was 1.7 bits.

Next Time We will look at how entropy is related to security Generalized definition of encryption Perfect Secrecy Manipulating entropy relationships The next computer project will also be handed out