Uncertainty Management in Rule-based Expert Systems

Slides:



Advertisements
Similar presentations
© Negnevitsky, Pearson Education, Introduction, or what is uncertainty? Introduction, or what is uncertainty? Basic probability theory Basic probability.
Advertisements

Introductory Mathematics & Statistics for Business
FT228/4 Knowledge Based Decision Support Systems
Lahore University of Management Sciences, Lahore, Pakistan Dr. M.M. Awais- Computer Science Department 1 Lecture 12 Dealing With Uncertainty Probabilistic.
Bayesian Network and Influence Diagram A Guide to Construction And Analysis.
 Negnevitsky, Pearson Education, Lecture 3 Uncertainty management in rule- based expert systems n Introduction, or what is uncertainty? n Basic.
Rulebase Expert System and Uncertainty. Rule-based ES Rules as a knowledge representation technique Type of rules :- relation, recommendation, directive,
PROBABILITY. Uncertainty  Let action A t = leave for airport t minutes before flight from Logan Airport  Will A t get me there on time ? Problems :
1 1 Slide STATISTICS FOR BUSINESS AND ECONOMICS Seventh Edition AndersonSweeneyWilliams Slides Prepared by John Loucks © 1999 ITP/South-Western College.
Uncertainty Everyday reasoning and decision making is based on uncertain evidence and inferences. Classical logic only allows conclusions to be strictly.
FT228/4 Knowledge Based Decision Support Systems
Chapter 4: Reasoning Under Uncertainty
B. Ross Cosc 4f79 1 Uncertainty Knowledge can have uncertainty associated with it - Knowledge base: rule premises, rule conclusions - User input: uncertain,
5/17/20151 Probabilistic Reasoning CIS 479/579 Bruce R. Maxim UM-Dearborn.
Certainty and Evidence
AI – CS364 Uncertainty Management Introduction to Uncertainty Management 21 st September 2006 Dr Bogdan L. Vrusias
Introduction to Probabilities Farrokh Alemi, Ph.D. Saturday, February 21, 2004.
AI – CS364 Uncertainty Management 26 th September 2006 Dr Bogdan L. Vrusias
CSNB234 ARTIFICIAL INTELLIGENCE
Slides are based on Negnevitsky, Pearson Education, Lecture 3 Uncertainty management in rule- based expert systems n Introduction, or what is uncertainty?
Topic 6: Introduction to Hypothesis Testing
Computer Intelligence and Soft Computing
1 Bayesian Reasoning Chapter 13 CMSC 471 Adapted from slides by Tim Finin and Marie desJardins.
Lecture 05 Rule-based Uncertain Reasoning
Chapter 4: Basic Probability
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
For Monday after Spring Break Read Homework: –Chapter 13, exercise 6 and 8 May be done in pairs.
Probability, Bayes’ Theorem and the Monty Hall Problem
HAWKES LEARNING SYSTEMS math courseware specialists Copyright © 2010 by Hawkes Learning Systems/Quant Systems, Inc. All rights reserved. Chapter 14 Analysis.
Copyright © Cengage Learning. All rights reserved. 8 Tests of Hypotheses Based on a Single Sample.
Statistical Decision Theory
Chapter 8 Probability Section R Review. 2 Barnett/Ziegler/Byleen Finite Mathematics 12e Review for Chapter 8 Important Terms, Symbols, Concepts  8.1.
Engineering Economic Analysis Canadian Edition
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Testing Theories: The Problem of Sampling Error. The problem of sampling error It is often the case—especially when making point predictions—that what.
1 Reasoning Under Uncertainty Artificial Intelligence Chapter 9.
Statistical Hypotheses & Hypothesis Testing. Statistical Hypotheses There are two types of statistical hypotheses. Null Hypothesis The null hypothesis,
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Soft Computing Lecture 19 Part 2 Hybrid Intelligent Systems.
Correlation Assume you have two measurements, x and y, on a set of objects, and would like to know if x and y are related. If they are directly related,
Uncertainty. Assumptions Inherent in Deductive Logic-based Systems All the assertions we wish to make and use are universally true. Observations of the.
Uncertainty in Expert Systems
Reasoning Under Uncertainty. 2 Objectives Learn the meaning of uncertainty and explore some theories designed to deal with it Find out what types of errors.
Artificial Intelligence CIS 342 The College of Saint Rose David Goldschmidt, Ph.D.
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Reasoning with Uncertainty دكترمحسن كاهاني
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Uncertainty  Most intelligent systems have some degree of uncertainty associated with them.  The types of uncertainty that can occur in KBS may be caused.
International Conference on Fuzzy Systems and Knowledge Discovery, p.p ,July 2011.
Textbook Basics of an Expert System: – “Expert systems: Design and Development,” by: John Durkin, 1994, Chapters 1-4. Uncertainty (Probability, Certainty.
+ Chapter 5 Overview 5.1 Introducing Probability 5.2 Combining Events 5.3 Conditional Probability 5.4 Counting Methods 1.
Sampling Design and Analysis MTH 494 Lecture-21 Ossam Chohan Assistant Professor CIIT Abbottabad.
1 HETEROSCEDASTICITY: WEIGHTED AND LOGARITHMIC REGRESSIONS This sequence presents two methods for dealing with the problem of heteroscedasticity. We will.
UNIVERSITI TENAGA NASIONAL 1 CCSB354 ARTIFICIAL INTELLIGENCE Chapter 8.2 Certainty Factors Chapter 8.2 Certainty Factors Instructor: Alicia Tang Y. C.
Copyright © Cengage Learning. All rights reserved. 2 Probability.
Anifuddin Azis UNCERTAINTY. 2 Introduction The world is not a well-defined place. There is uncertainty in the facts we know: What’s the temperature? Imprecise.
AP Statistics From Randomness to Probability Chapter 14.
Chapter 12 Certainty Theory (Evidential Reasoning) 1.
REASONING UNDER UNCERTAINTY: CERTAINTY THEORY
Bayes’ Theorem Suppose we have estimated prior probabilities for events we are concerned with, and then obtain new information. We would like to a sound.
CHAPTER 5 Handling Uncertainty BIC 3337 EXPERT SYSTEM.
Unit 5: Hypothesis Testing
Basic Probabilistic Reasoning
Dealing with Uncertainty
Class #21 – Monday, November 10
28th September 2005 Dr Bogdan L. Vrusias
Certainty Factor Model
Basic Probability Chapter Goal:
basic probability and bayes' rule
Presentation transcript:

Uncertainty Management in Rule-based Expert Systems What is uncertainty? Bayesian reasoning Bias of the Bayesian method Summary

What is uncertainty? Information can be incomplete, inconsistent, uncertain, or all three. In other words, information is often unsuitable for solving a problem. Uncertainty is defined as the lack of the exact knowledge that would enable us to reach a perfectly reliable conclusion. Classical logic permits only exact reasoning. IF A is true IF A is false THEN A is not false THEN A is not true

Sources of uncertain knowledge Weak implications. Domain experts and knowledge engineers have the painful task of establishing concrete correlations between IF (condition) and THEN (action) parts of the rules. Therefore, expert systems need to have the ability to handle vague associations, for example by accepting the degree of correlations as numerical certainty factors.

Imprecise language. Our natural language is ambiguous and imprecise Imprecise language. Our natural language is ambiguous and imprecise. We describe facts with such terms as often and sometimes, frequently and hardly ever. As a result, it can be difficult to express knowledge in the precise IF-THEN form of production rules. However, if the meaning of the facts is quantified, it can be used in expert systems.

Unknown data. When the data is incomplete or missing, the only solution is to accept the value “unknown” and proceed to an approximate reasoning with this value. Combining the views of different experts. Large expert systems usually combine the knowledge and expertise of a number of experts. Unfortunately, experts often have contradictory opinions and produce conflicting rules. To resolve the conflict, the knowledge engineer has to attach a weight to each expert and then calculate the composite conclusion. But no systematic method exists to obtain these weights.

Conditional Probability Let A be an event in the world and B be another event. Suppose that events A and B are not mutually exclusive, but occur conditionally on the occurrence of the other. The probability that event A will occur if event B occurs is called the conditional probability.

The number of times A and B can occur, or the probability that both A and B will occur, is called the joint probability of A and B. It is represented mathematically as p(AB). The number of ways B can occur is the probability of B, p(B), and thus Similarly, the conditional probability of event B occurring given that event A has occurred equals

Hence, and Substituting the last equation into the equation yields the Bayesian rule:

Bayesian Rule where: p(AB) is the conditional probability that event A occurs given that event B has occurred; p(BA) is the conditional probability of event B occurring given that event A has occurred; p(A) is the probability of event A occurring; p(B) is the probability of event B occurring.

Joint Probability

If the occurrence of event A depends on only two mutually exclusive events, B and NOT B, we obtain: where  is the logical function NOT. Similarly, Substituting this equation into the Bayesian rule yields:

Bayesian Reasoning Suppose all rules in the knowledge base are represented in the following form: IF E is true THEN H is true {with probability p} This rule implies that if event E occurs, then the probability that event H will occur is p.

The Bayesian rule expressed in terms of hypotheses and evidence looks like this: where: p(H) is the prior probability of hypothesis H being true; p(EH) is the probability that hypothesis H being true will result in evidence E; p(H) is the prior probability of hypothesis H being false; p(EH) is the probability of finding evidence E even when hypothesis H is false.

In expert systems, the probabilities required to solve a problem are provided by experts. An expert determines the prior probabilities for possible hypotheses p(H) and p(H), and also the conditional probabilities for observing evidence E if hypothesis H is true, p(EH), and if hypothesis H is false, p(EH). Users provide information about the evidence observed and the expert system computes p(HE) for hypothesis H in light of the user-supplied evidence E. Probability p(HE) is called the posterior probability of hypothesis H upon observing evidence E.

We can take into account both multiple hypotheses H1, H2, We can take into account both multiple hypotheses H1, H2,..., Hm and multiple evidences E1, E2,..., En. The hypotheses as well as the evidences must be mutually exclusive and exhaustive. Single evidence E and multiple hypotheses follow: Multiple evidences and multiple hypotheses follow:

This requires to obtain the conditional probabilities of all possible combinations of evidences for all hypotheses, and thus places an enormous burden on the expert. Therefore, in expert systems, conditional independence among different evidences assumed. Thus, instead of the unworkable equation, we attain:

Ranking potentially true hypotheses Suppose an expert, given three conditionally independent evidences E1, E2 and E3, creates three mutually exclusive and exhaustive hypotheses H1, H2 and H3, and provides prior probabilities for these hypotheses: p(H1), p(H2) and p(H3), respectively. The expert also determines the conditional probabilities of observing each evidence for all possible hypotheses.

The prior and conditional probabilities Assume that we first observe evidence E3. The expert system computes the posterior probabilities for all hypotheses as

Thus, After evidence E3 is observed, belief in hypothesis H2 increases and becomes equal to belief in hypothesis H1. Belief in hypothesis H3 also increases and even nearly reaches beliefs in hypotheses H1 and H2.

Suppose now that we observe evidence E1 Suppose now that we observe evidence E1. The posterior probabilities are calculated as Hence, Hypothesis H2 has now become the most likely one.

After observing evidence E2, the final posterior probabilities for all hypotheses are calculated: Although the initial ranking was H1, H2 and H3, only hypotheses H1 and H3 remain under consideration after all evidences (E1, E2 and E3) were observed.

Bias of the Bayesian method The framework for Bayesian reasoning requires probability values as primary inputs. The assessment of these values usually involves human judgement. However, psychological research shows that humans cannot elicit probability values consistent with the Bayesian rules. This suggests that the conditional probabilities may be inconsistent with the prior probabilities given by the expert.

THEN the starter is bad {with probability 0.7} Consider, for example, a car that does not start and makes odd noises when you press the starter. The conditional probability of the starter being faulty if the car makes odd noises may be expressed as: IF the symptom is “odd noises” THEN the starter is bad {with probability 0.7} p(starter is not badodd noises) = = p(starter is goododd noises) = 10.7 = 0.3

Therefore, we can obtain a companion rule that states IF the symptom is “odd noises” THEN the starter is good {with probability 0.3} Domain experts do not deal with conditional probabilities and often deny the very existence of the hidden implicit probability (0.3 in our example). We would also use available statistical information and empirical studies to derive the following rules: IF the starter is bad THEN the symptom is “odd noises” {probability 0.85} THEN the symptom is not “odd noises” {probability 0.15}

To use the Bayesian rule, we still need the prior probability, the probability that the starter is bad if the car does not start. Suppose, the expert supplies us the value of 5 per cent. Now we can apply the Bayesian rule to obtain: The number obtained is significantly lower than the expert’s estimate of 0.7 given at the beginning of this section. The reason for the inconsistency is that the expert made different assumptions when assessing the conditional and prior probabilities.

The Bayesian method is likely to be the most appropriate if reliable statistical data exists, the knowledge engineer is able to lead, and the expert is available for serious decision makings. In many areas of possible applications of expert systems, reliable statistical information is not available or we cannot assume the conditional independence of evidence. As a result, many researchers have found the Bayesian method unsuitable for their work. In the absence of any of the specified conditions, the Bayesian approach might be too arbitrary and even biased to produce meaningful results.