Similarity in CBR (Cont’d) Sources: –Chapter 4 –www.iiia.csic.es/People/enric/AICom.html –www.ai-cbr.org.

Slides:



Advertisements
Similar presentations
Experimental Design Some Terms.
Advertisements

Prepared by Dr. Taha MAhdy
Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
1 Programming for Engineers in Python Autumn Lecture 5: Object Oriented Programming.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Chapter 5 Understanding Entity Relationship Diagrams.
EECS 20 Lecture 3 (January 22, 2001) Tom Henzinger Sets, Tuples, Functions.
Distance Measures Tan et al. From Chapter 2.
Data Flow Analysis Compiler Design Nov. 8, 2005.
Chapter 11: Limitations of Algorithmic Power
Mathematics.
1.Definition of a function 2.Finding function values 3.Using the vertical line test.
2009 Mathematics Standards of Learning Training Institutes Algebra II Virginia Department of Education.
18 Days.  2.1 Definition of a Function 2.1 Definition of a Function  2.2 Graphs of Functions 2.2 Graphs of Functions  2.3 Quadratic Functions 2.3 Quadratic.
Themes of Presentations Rule-based systems/expert systems (Catie) Software Engineering (Khansiri) Fuzzy Logic (Mark) Configuration Systems (Sudhan) *
Math 3121 Abstract Algebra I Lecture 3 Sections 2-4: Binary Operations, Definition of Group.
Constraint Satisfaction Definition. A constraint is a formula of the form: (x = y) (x  y) (x = red) (x  red) Where x and y are variables that can take.
Programming Project (Last updated: August 31 st /2010) Updates: - All details of project given - Deadline: Part I: September 29 TH 2010 (in class) Part.
DATA MINING LECTURE 13 Absorbing Random walks Coverage.
Copyright © 2014, 2010, 2006 Pearson Education, Inc. 1 Chapter 1 Introduction to Functions and Graphs Book cover will go here.
Chapter 1. Functions and Models
Week 7 - Friday.  What did we talk about last time?  Set disproofs  Russell’s paradox  Function basics.
Slide Chapter 1 Test Mean: 84 Median: 89 Breakdown: 22 A’s 12 B’s 4 C’s 4 D’s 3 F’s Copyright © 2009 Pearson Education, Inc.
What multiplication or subtraction problem is being modeled? Take-away Missing addend Repeated addition Repeated subtraction
Source for Information Gain Formula Artificial Intelligence: A Modern Approach by Stuart Russell and Peter Norvig Chapter 18: Learning from Observations.
Similarity in CBR (Cont’d) Sources: –Chapter 4 – –
Copyright © 2009 Pearson Education, Inc. CHAPTER 2: More on Functions 2.1 Increasing, Decreasing, and Piecewise Functions; Applications 2.2 The Algebra.
CSCI 2670 Introduction to Theory of Computing November 23, 2004.
Similarity in CBR Sources: –Chapter 4 – –
Mathematical Background and Linked Lists. 2 Iterative Algorithm for Sum Find the sum of the first n integers stored in an array v : sum (v[], n) temp_sum.
Randomized Composable Core-sets for Submodular Maximization Morteza Zadimoghaddam and Vahab Mirrokni Google Research New York.
1 Relational Algebra and Calculas Chapter 4, Part A.
Design Elements of Graphical Representation, (Factors supporting appearance and functionality of solutions). P0CCUAA.
1 Relational Algebra Chapter 4, Sections 4.1 – 4.2.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
1 CHAPTER 5 : FOURIER SERIES  Introduction  Periodic Functions  Fourier Series of a function  Dirichlet Conditions  Odd and Even Functions  Relationship.
Complete Graphs A complete graph is one where there is an edge between every two nodes A C B G.
Course: Software Engineering Unit 2: Theories and SchemasSlide Number 1 Unit 2: Theories and Schemas This lecture aims to: Recall some concepts in logic:
IDSS: Overview of Themes AI  Introduction  Overview IDT  Attribute-Value Rep.  Decision Trees  Induction CBR  Introduction  Representation  Similarity.
College Algebra Sixth Edition James Stewart Lothar Redlin Saleem Watson.
Chapter 6 - Basic Similarity Topics
Random Variables Example:
Chapter 16 UML Class Diagrams 1CS6359 Fall 2012 John Cole.
Copyright © Cengage Learning. All rights reserved. Line and Angle Relationships 1 1 Chapter.
Chapter 11 Introduction to Computational Complexity Copyright © 2011 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1.
Copyright © Cengage Learning. All rights reserved. CHAPTER 8 RELATIONS.
Evolutionary Computing Chapter 12. / 26 Chapter 12: Multiobjective Evolutionary Algorithms Multiobjective optimisation problems (MOP) -Pareto optimality.
Chapter 6- LINEAR MAPPINGS LECTURE 8 Prof. Dr. Zafer ASLAN.
Database Design, Application Development, and Administration, 6 th Edition Copyright © 2015 by Michael V. Mannino. All rights reserved. Chapter 5 Understanding.
An Inverse Function.
Power Functions (part 2)
Announcements Topics: Work On:
Lecture Notes for Chapter 2 Introduction to Data Mining
Numerical Methods and Analysis
Function Inverse Section 1.5
Virtual University of Pakistan
Introduction To Statistics
Relational Algebra 461 The slides for this text are organized into chapters. This lecture covers relational algebra, from Chapter 4. The relational calculus.
AP Calculus Chapter 1 Section 2
CHAPTER 2: More on Functions
Chapter 2 More on Functions.
Patterns of Thinking and Writing
Introduction To Statistics
Relations and Functions
Concept of a Function.
Inverse Functions and Relations
13.2 Limits and Continuity day 1
1. Defining and Representing Functions
12.12 Lagrange Remainder Mathboat.com.
Visual Algebra for Teachers
Presentation transcript:

Similarity in CBR (Cont’d) Sources: –Chapter 4 – –

Simple-Matching-Coefficient (SMC)  H(X,Y) = n – (A + D) = B + C Another distance-similarity compatible function is f(x) = 1 – x/max (where max is the maximum value for x)  We can define the SMC similarity, sim H : sim H (X,Y) = 1 – ((n – (A+D))/n) = (A+D)/n = 1- ((B+C)/n) Solution (I): Show that f(x) is order inverting: if x f(y) Proportion of the difference

Simple-Matching-Coefficient (SMC) (II) If we use on sim H (X,Y) = 1- ((B+C)/n) = factor(A, B, C, D)  Monotonic:  If A  A’ then:  If B  B’ then:  If C  C’ then:  If D  D’ then: factor(A,B,C,D)  factor(A’,B,C,D) factor(A,B’,C,D)  factor(A,B,C,D) factor(A,B,C’,D)  factor(A,B,C,D) factor(A,B,C,D)  factor(A,B,C,D’)  Symmetric: sim H (X,Y) = sim H (Y,X) Solution(II): Show that sim H (X,Y) is monotonic

Variations of SMC (III) We introduce a weight, , with 0 <  < 1: simH(X,Y) = (A+D)/n = (A+D)/(A+B+C+D) sim  (X,Y) = (  (A+D))/ (  (A+D) + (1 -  )(B+C))  For which  is sim  (X,Y) = sim H (X,Y)?  = 0.5  sim  (X,Y) preserves the monotonic and symmetric conditions Solution(III): Show that sim  (X,Y) is monotonic

Homework (Part IV): Attributes May Have multiple Values X = (X 1, …, X n ) where X i  T i Y = (Y 1, …,Y n ) where Y i  T i Each T i is finite Define a formula for the Hamming distance in this context

Tversky Contrast Model Defines a non monotonic distance Comparison of a situation S with a prototype P (i.e, a case) S and P are sets of features The following sets:  A = S  P  B = P – S  C = S – P A S P C B

Tversky Contrast Model (2) Tversky-distance: Where f: Sets  [0,  ), , , and  are constants f, , , and  are fixed and defined by the user Example:  If f(A) = # elements in A   =  =  = 1  T counts the number of elements in common minus the differences  The Tversky-distance is not symmetric T(P,S) =  f(A) -  f(B) -  f(C)

Local versus Global Similarity Metrics In many situations we have similarity metrics between attributes of the same type (called local similarity metrics). Example: For a complex engine, we may have a similarity for the temperature of the engine In such situations a reasonable approach to define a global similarity sim  (x,y) is to “aggregate” the local similarity metrics sim i (x i,y i ). A widely used practice sim  (x,y) to increate monotonically with each sim i (x i,y i ). What requirements should we give to sim  (x,y) in terms of the use of sim i (x i,y i )?

Local versus Global Similarity Metrics (Formal Definitions) A local similarity metric on an attribute T i is a similarity metric sim i : T i  T i  [0,1] A function  : [0,1] n  [0,1] is an aggregation function if:   (0,0,…,0) = 0   is monotonic non-decreasing on every argument Given a collection of n similarity metrics sim 1, …, sim n, for attributes taken values from T i, a global similarity metric, is a similarity metric sim:V  V  [0,1], V in T 1  …  T n, such that there is an aggregation  function with: sim(X,Y) = sim  (X,Y) =  (sim 1 (X 1,Y 1 ), …,sim n (X n,Y n )) Homework: provide an example of an aggregation function and a non-aggregation function and prove it. Show a global sim. metric

Solution Suppose that cases use an object oriented representation:  Suppose that cases use a taxonomical representation, describe how you would measure similarity and give a concrete example illustrating the process you described to measure similarity  Suppose that cases use a compositional representation, describe how you would measure similarity and give a concrete example illustrating the process you described to measure similarity Suggestion: look at the book!

Frontiers of Knowledge Dealing with numerical and non numerical values  Aggregation of local similarity metrics into a global similarity metric helps  but sometimes we don’t have local similarity metrics

Homework (II) From Chapter 5, what is the difference between completion and adaptation functions? What si their role on adaptation? Provide an example Show that Graph coloring is NP-complete  Assume that Constraint-SAT is NP complete  Definition. A constraint is a formula of the form: –(x = y) –(x  y) Where x and y are variables that can take values from a set (e.g., {yellow, white, black, red, …})  Definition. Constraint-SAT: given a conjunction of constraints, is there an instantiation of the variables that makes the conjunction true?