P is a subset of NP Let prob be a problem in P

Slides:



Advertisements
Similar presentations
Induction of Decision Trees (IDT)
Advertisements

Introduction to Algorithms NP-Complete
Learning from Observations Chapter 18 Section 1 – 3.
P is a subset of NP Let prob be a problem in P There is a deterministic algorithm alg that solves prob in polynomial time O(n k ), for some constant k.
Machine Learning A Quick look Sources:
Homework. Homework (next class) Read Chapter 2 of the Experience Management book and answer the following questions: Provide an example of something that.
NP-complete and NP-hard problems Transitivity of polynomial-time many-one reductions Concept of Completeness and hardness for a complexity class Definition.
The Theory of NP-Completeness
1 NP-Complete Problems. 2 We discuss some hard problems:  how hard? (computational complexity)  what makes them hard?  any solutions? Definitions 
© The McGraw-Hill Companies, Inc., Chapter 8 The Theory of NP-Completeness.
Learning From Observations
1 CSE 417: Algorithms and Computational Complexity Winter 2001 Lecture 21 Instructor: Paul Beame.
1 Polynomial Time Reductions Polynomial Computable function : For any computes in polynomial time.
LEARNING DECISION TREES
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Learning decision trees
Learning decision trees derived from Hwee Tou Ng, slides for Russell & Norvig, AI a Modern Approachslides Tom Carter, “An introduction to information theory.
Ref: Pfleeger96, Ch.31 NP-Complete Problems Reference: Pfleeger, Charles P., Security in Computing, 2nd Edition, Prentice Hall, 1996.
Induction of Decision Trees (IDT) CSE 335/435 Resources: – –
Themes of Presentations Rule-based systems/expert systems (Catie) Software Engineering (Khansiri) Fuzzy Logic (Mark) Configuration Systems (Sudhan) *
The Theory of NP-Completeness 1. Nondeterministic algorithms A nondeterminstic algorithm consists of phase 1: guessing phase 2: checking If the checking.
Machine Learning CPS4801. Research Day Keynote Speaker o Tuesday 9:30-11:00 STEM Lecture Hall (2 nd floor) o Meet-and-Greet 11:30 STEM 512 Faculty Presentation.
The Theory of NP-Completeness 1. What is NP-completeness? Consider the circuit satisfiability problem Difficult to answer the decision problem in polynomial.
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Inductive learning Simplest form: learn a function from examples
1 The Theory of NP-Completeness 2012/11/6 P: the class of problems which can be solved by a deterministic polynomial algorithm. NP : the class of decision.
Machine Learning A Quick look Sources: Artificial Intelligence – Russell & Norvig Artifical Intelligence - Luger By: Héctor Muñoz-Avila.
Lecture 22 More NPC problems
LEARNING DECISION TREES Yılmaz KILIÇASLAN. Definition - I Decision tree induction is one of the simplest, and yet most successful forms of learning algorithm.
NP Complexity By Mussie Araya. What is NP Complexity? Formal Definition: NP is the set of decision problems solvable in polynomial time by a non- deterministic.
Learning from Observations Chapter 18 Through
CHAPTER 18 SECTION 1 – 3 Learning from Observations.
Week 10Complexity of Algorithms1 Hard Computational Problems Some computational problems are hard Despite a numerous attempts we do not know any efficient.
Decision Trees. What is a decision tree? Input = assignment of values for given attributes –Discrete (often Boolean) or continuous Output = predicated.
NP-Complete Problems. Running Time v.s. Input Size Concern with problems whose complexity may be described by exponential functions. Tractable problems.
1 Chapter 34: NP-Completeness. 2 About this Tutorial What is NP ? How to check if a problem is in NP ? Cook-Levin Theorem Showing one of the most difficult.
Chapter 11 Introduction to Computational Complexity Copyright © 2011 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1.
Chapter 18 Section 1 – 3 Learning from Observations.
Learning From Observations Inductive Learning Decision Trees Ensembles.
1 1. Which of these sequences correspond to Hamilton cycles in the graph? (a) (b) (c) (d) (e)
Decision Tree Learning CMPT 463. Reminders Homework 7 is due on Tuesday, May 10 Projects are due on Tuesday, May 10 o Moodle submission: readme.doc and.
The Theory of NP-Completeness
NP-Completeness A problem is NP-complete if: It is in NP
The NP class. NP-completeness
P & NP.
Chapter 10 NP-Complete Problems.
Learning from Observations
Learning from Observations
Introduce to machine learning
Advanced Algorithms Analysis and Design
Lecture 2-2 NP Class.
Presented By S.Yamuna AP/CSE
NP-Completeness (2) NP-Completeness Graphs 7/23/ :02 PM x x x x
Hard Problems Introduction to NP
NP-Completeness Yin Tat Lee
Intro to Theory of Computation
Analysis and design of algorithm
Chapter 11 Limitations of Algorithm Power
NP-Complete Problems.
CS154, Lecture 13: P vs NP.
Learning from Observations
NP-Completeness Yin Tat Lee
CSE 6408 Advanced Algorithms.
NP-Completeness Reference: Computers and Intractability: A Guide to the Theory of NP-Completeness by Garey and Johnson, W.H. Freeman and Company, 1979.
P, NP and NP-Complete Problems
The Theory of NP-Completeness
CSE 589 Applied Algorithms Spring 1999
P, NP and NP-Complete Problems
Learning from Observations
Machine Learning: Decision Tree Learning
Presentation transcript:

P is a subset of NP Let prob be a problem in P There is a deterministic algorithm alg that solves prob in polynomial time O(nk), for some constant k That same algorithm alg runs in a nondeterminsitc machine (it just do not use the oracle) Thus, alg has the same polynomial complexity, O(nk) Thus, prob is in NP

Why Guess & Check Implies NP If a problem prob satisfies Guess and Check prob …. solution polynomial The nondeterministic version prob …. solution

Induction of Decision Trees CSE 335/435 Resources: http://www.aaai.org/AITopics/html/expert.html (article: Think About It: Artificial Intelligence & Expert Systems) http://www.aaai.org/AITopics/html/trees.html http://www.academicpress.com/expert/leondes_expert_vol1_ch3.pdf

Motivation # 1: Analysis Tool Suppose that a company have a data base of sales data, lots of sales data How can that company’s CEO use this data to figure out an effective sales strategy Safeway, Giant, etc cards: what is that for?

Motivation # 1: Analysis Tool (cont’d) Sales data induction Decision Tree Ex’ple Bar Fri Hun Pat Type Res wait x1 no yes some french x4 full thai x5 x6 x7 x8 x9 x10 x11 “if buyer is male & and age between 24-35 & married then he buys sport magazines”

Motivation # 1: Analysis Tool (cont’d) Decision trees has been frequently used in IDSS Some companies: SGI: provides tools for decision tree visualization Acknosoft (France), Tech:Inno (Germany): combine decision trees with CBR technology Several applications Decision trees are used for Data Mining

Parenthesis: Expert Systems Have been used in (Sweet; How Computers Work 1999): medicine oil and mineral exploration weather forecasting stock market predictions financial credit, fault analysis some complex control systems Two components: Knowledge Base Inference Engine

The Knowledge Base in Expert Systems A knowledge base consists of a collection of IF-THEN rules: if buyer is male & age between 24-50 & married then he buys sport magazines if buyer is male & age between 18-30 then he buys PC games magazines Knowledge bases of fielded expert systems contain hundreds and sometimes even thousands such rules. Frequently rules are contradictory and/or overlap

The Inference Engine in Expert Systems The inference engine reasons on the rules in the knowledge base and the facts of the current problem Typically the inference engine will contain policies to deal with conflicts, such as “select the most specific rule in case of conflict” Some expert systems incorporate probabilistic reasoning, particularly those doing predictions

Expert Systems: Some Examples MYCIN. It encodes expert knowledge to identify kinds of bacterial infections. Contains 500 rules and use some form of uncertain reasoning DENDRAL. Identifies interpret mass spectra on organic chemical compounds MOLGEN. Plans gene-cloning experiments in laboratories. XCON. Used by DEC to configure, or set up, VAX computers. Contained 2500 rules and could handle computer system setups involving 100-200 modules.

Main Drawback of Expert Systems: The Knowledge Acquisition Bottle-Neck The main problem of expert systems is acquiring knowledge from human specialist is a difficult, cumbersome and long activity. Name KB #Rules Const. time (man/years) Maint. time MYCIN KA 500 10 N/A XCON 2500 18 3 KB = Knowledge Base KA = Knowledge Acquisition

Motivation # 2: Avoid Knowledge Acquisition Bottle-Neck GASOIL is an expert system for designing gas/oil separation systems stationed of-shore The design depends on multiple factors including: proportions of gas, oil and water, flow rate, pressure, density, viscosity, temperature and others To build that system by hand would had taken 10 person years It took only 3 person-months by using inductive learning! GASOIL saved BP millions of dollars

Motivation # 2 : Avoid Knowledge Acquisition Bottle-Neck Name KB #Rules Const. time (man/years) Maint. time (man/months) MYCIN KA 500 10 N/A XCON 2500 18 3 GASOIL IDT 2800 1 0.1 BMT KA (IDT) 30000+ 9 (0.3) 2 (0.1) KB = Knowledge Base KA = Knowledge Acquisition IDT = Induced Decision Trees

Example of a Decision Tree Patrons? waitEstimate? no yes 0-10 >60 Full no yes none some Alternate? Reservation? Yes 30-60 no yes Hungry? yes No 10-30 Alternate? Yes no Raining? Fri/Sat? No Yes yes no No no Bar? Yes yes

Definition of A Decision Tree A decision tree is a tree where: The leaves are labeled with classifications (if the classification is “yes” or “no”. The tree is called a boolean tree) The non-leaves nodes are labeled with attributes The arcs out of a node labeled with an attribute A are labeled with the possible values of the attribute A

Some Properties of Decision Trees Decision trees represent rules (or more formally logical sentences): r (patrons(r,full)  waitingTime(r,t)  t > 60  willWait(r)) Decision trees represent functions: F: Patrons × WaitExtimate × Hungry × type × Fri/Sat × Alternate × Raining  {True,False} F(Full, >60, _, _, _, _, _) = No F(Full,10-30,Yes,_,_,Yes,Yes) = Yes …

Some Properties of Decision Trees (II) Lets consider a Boolean function: F: A1 × A2 × … × An {Yes, No} F can obviously be represented in a table: A1 A2 … An y/n Question: What is the maximum number of rows does the table defining F has assuming that each attribute Ai has 2 values?

Some Properties of Decision Trees (III) Answer: 2n because: F: A1 × A2 × … × An {Yes, No} 2 2 … 2 The number of possible combinations is 2 × 2 × … × 2 = 2n This means that representing these tables and processing them requires a lot of effort Question: How many functions F: A1 × A2 × … × An {Yes, No} are there assuming that each attribute Ai has 2 values? 2 # of rows in table = 22 n Answer:

Some Properties of Decision Trees (IV) We observed that given a decision tree, it can be represented as a Boolean function. Question: Given a Boolean function: F: A1 × A2 × … × An {Yes, No} Where each of Ai can take a finite set of attributes. Can F be represented as a decision tree? Answer: Yes!. Make A1 first node, A2 second node, etc. (Brute force)

Some Properties of Decision Trees (V) Representing a table one may have several possible decision trees As an example, the next slide shows a table This table has at least 2 decision trees: The example decision tree of the restaurant of Slide # 12 The brute force Which is the best decision tree, why?

Example Ex’ple Bar Fri Hun Pat Alt Type wait x1 no yes some French x4 full Thai x5 x6 Italian x7 none Burger x8 x9 x10 x11 No

Homework Assignment: We never test the same attribute twice along one branch in a decision tree. Why not? See next slides (Slide # 26 in particular)

Our Current List of NP-Complete Problems Answer is “yes” or “no” (Input)Output Decision Problem “Are there values of variables making the CNF true?” (CNF) values of variables making formula true CNF-sat (Circuit) values of input making the circuit’s output true “Are there values of input making the circuit’s output true?” Circuit-sat

Complete Graphs A complete graph is one where there is an edge between every two nodes G B C A

Clique Clique: A complete subgraph of a graph Problem (Clique): Find the clique with the largest number of nodes in a graph 20 10 G H 16 B 9 F 6 C 13 12 8 24 A 4 6 E D 20

Clique is NP-Complete Proving NP completeness can be difficult CNF-sat Homework Clique Circuit-sat

CNF Can Be Reduced into Clique First we need to formulate Clique as a decision problem: Decision Problem (Clique): Given a graph and a value K, is there a clique of size K (i.e., with K nodes) in the graph? Why we need to formulate the decision problem? Because we want to show a reduction such that the decision problem for CNF-sat is true if and only if the decision problem for the Clique is true