EE 5322: Intelligent Control Systems. Dempster Shafer Theory

Similar documents
Topic Dempster-Shafer Theory

COMPUTATIONAL METHODS FOR A MATHEMATICAL THEORY OF EVIDENCE

E3: PROBABILITY AND STATISTICS lecture notes

Multi-ultrasonic sensor fusion for autonomous mobile robots

So let us begin our quest to find the holy grail of real analysis.

5 Aleatory Variability and Epistemic Uncertainty

Lecture Note 1 Set and Probability Theory. MIT Spring 2006 Herman Bennett

abc Mark Scheme Statistics 6380 General Certificate of Education 2006 examination - January series SS02 Statistics 2

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

Software Verification and System Assurance

Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Math/Stats 425 Introduction to Probability. 1. Uncertainty and the axioms of probability

Using game-theoretic probability for probability judgment. Glenn Shafer

Science Stage 6 Skills Module 8.1 and 9.1 Mapping Grids

CONTENTS OF DAY 2. II. Why Random Sampling is Important 9 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE

3. The Junction Tree Algorithms

CHAPTER 2 Estimating Probabilities

99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, cm

Point and Interval Estimates

The evidential reasoning approach for MADA under both probabilistic and fuzzy uncertainties

3. Mathematical Induction

Review. Bayesianism and Reliability. Today s Class

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

Basic Probability Concepts

Notes on Continuous Random Variables

Design of an Intelligent Housing System Using Sensor Data Fusion Approaches

UNCERTAINTIES OF MATHEMATICAL MODELING

Removing Partial Inconsistency in Valuation- Based Systems*

8 Divisibility and prime numbers

Unit 19: Probability Models

1. True or False? A voltage level in the range 0 to 2 volts is interpreted as a binary 1.

A Few Basics of Probability

Sample Size Issues for Conjoint Analysis

WRITING PROOFS. Christopher Heil Georgia Institute of Technology

COMP 250 Fall 2012 lecture 2 binary representations Sept. 11, 2012

9. Momentum and Collisions in One Dimension*

University of Chicago Graduate School of Business. Business 41000: Business Statistics

Open World Target Identification Using the Transferable Belief Model

Fuzzy Probability Distributions in Bayesian Analysis

Knowledge Representation (II)

A Novel Solution on Alert Conflict Resolution Model in Network Management

ACH 1.1 : A Tool for Analyzing Competing Hypotheses Technical Description for Version 1.1

INDUCTIVE & DEDUCTIVE RESEARCH APPROACH

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

What Is Probability?

Normal distribution. ) 2 /2σ. 2π σ

CALCULATIONS & STATISTICS

BOOLEAN ALGEBRA & LOGIC GATES

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Basics of Statistical Machine Learning

8 Square matrices continued: Determinants

6.3 Conditional Probability and Independence

Prediction of future observations using belief functions

1 The Brownian bridge construction

Chapter 4. Probability and Probability Distributions

Generalized expressions of reliability of series-parallel and parallel-series systems using the Transferable Belief Model

Contradiction measures and specificity degrees of basic belief assignments

Likelihood: Frequentist vs Bayesian Reasoning

A FUZZY LOGIC APPROACH FOR SALES FORECASTING

LEARNING OBJECTIVES FOR THIS CHAPTER

Conditional probability of actually detecting a financial fraud a neutrosophic extension to Benford s law

CHAPTER 3. Methods of Proofs. 1. Logical Arguments and Formal Proofs

QUANTITATIVE METHODS BIOLOGY FINAL HONOUR SCHOOL NON-PARAMETRIC TESTS

MA 323 Geometric Modelling Course Notes: Day 02 Model Construction Problem

What is Statistic? OPRE 6301

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

AMS 5 CHANCE VARIABILITY

Decision and uncertainty management for human and human/agent teams. Dr. David Ullman President Robust Decisions, Inc.

IV. ALGEBRAIC CONCEPTS

MATHEMATICS FOR ENGINEERING INTEGRATION TUTORIAL 3 - NUMERICAL INTEGRATION METHODS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 5 9/17/2008 RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS INTRODUCTION

Hypothesis testing - Steps

Definition and Calculus of Probability

Bayesian Tutorial (Sheet Updated 20 March)

Analysis of Bayesian Dynamic Linear Models

1 o Semestre 2007/2008

1 Review of Newton Polynomials

Penalized regression: Introduction

Guide to the Uniform mark scale (UMS) Uniform marks in A-level and GCSE exams

Continued Fractions and the Euclidean Algorithm

Optimal Replacement of Underground Distribution Cables

Bayesian probability theory

Probability: Terminology and Examples Class 2, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

MATH 10: Elementary Statistics and Probability Chapter 5: Continuous Random Variables

Lemma 5.2. Let S be a set. (1) Let f and g be two permutations of S. Then the composition of f and g is a permutation of S.

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Section 6.2 Definition of Probability

CHAPTER 12 TESTING DIFFERENCES WITH ORDINAL DATA: MANN WHITNEY U

ROUGH SETS AND DATA MINING. Zdzisław Pawlak

2. Spin Chemistry and the Vector Model

Class 19: Two Way Tables, Conditional Distributions, Chi-Square (Text: Sections 2.5; 9.1)

DATA MINING IN FINANCE

Conditional Independence in Evidence Theory

Chi-square test Fisher s Exact test

Reliability Reliability Models. Chapter 26 Page 1

4. Continuous Random Variables, the Pareto and Normal Distributions

Review of Fundamental Mathematics

Transcription:

EE 5322: Intelligent Control Systems Dempster Shafer Theory Copyright F. L. Lewis 2004 All rights reserved. Prepared by: Prasanna Ballal Introduction: Recently, the scientific and engineering community has begun to recognize the importance of defining multiple types of uncertainty. The dual nature of uncertainty is described by the following concepts: Aleatory Uncertainty: The type of uncertainty which results from the fact that a system can behave in random ways (ex. Noise). Epistemic Uncertainty: The type of uncertainty which results from the lack of knowledge about a system and is a property of the analysts performing the analysis (Subective uncertainty). Probability theory has been traditionally used to characterize both types of uncertainty. E.g. aleatory uncertainty can be best dealt with using the frequentist approach associated with traditional probability theory. However, probability theory is not capable of completely capturing epistemic uncertainty. The traditional method was called Bayesian Probability. In this method, it is necessary to have information on the probability of all the events. When this is not available, uniform distribution function is often used, which means that all simple events for which a probability distribution is not known in a given sample space are equally likely. An additional assumption in the classical probability is the axiom of additivity where all probabilities that satisfy specific properties must add to 1. i.e. P [ A] + P[ A] = 1 The 3 axioms of Bayesian Theory are: 1. P[Ф]=0; 2. P[θ]=1; 3. if A B = 0 then P [ A B] = P[ A] + P[ B] Dempster Shafer Theory offers an alternative to traditional probabilistic theory for the mathematical representation of uncertainty. The significant innovation of this framework is that it allows for the allocation of probability mass to sets or intervals. It does not require an assumption regarding the probability of the individual element of the interval or set. An important aspect of this theory is the combination of evidence obtained by multiple sources and the modeling of conflict between them. It provides good results for evaluation of risk and reliability in engineering applications when it is not possible to obtain a correct and precise measurement from experiments. Dempster Shafer Theory ignores the 3 rd Bayesian Axiom and states that, i.e. P[ A B] or P[ A] + P[ B] P[ A B]. Definition of Terms: A set is represented by θ which contains all the possible elements of interest in each particular context and its elements are exhaustive and mutually exclusive events. This θ is called universal

set or the frame of discernment. E.g. in tossing a fair die, the frame of discernment is {1, 2, 3, 4, 5, 6}. This looks similar to the sample space in probability theory, but the difference is that in DS theory, the number of possible hypothesis is 2 θ while in probability theory it is θ where X is the cardinality of the set X. For simplicity we assume the number of elements of θ to be finite. Definition 1: If θ is a frame of discernment then a function m: 2 θ [0, 1] is called a basic probability assignment if m(φ) =0 and A θ m( A) = 1 The term m(a) is called A s basic probability number and m(a) is the measure of the belief that is committed exactly to A. 1. It is not required that m(θ)=1. 2. It is not required that m(a) m(b) when A B. 2 3. No relationship between m (A) and m (A) is required. 4. Also m ( A) + m( A) does not always have to be 1. Definition 2: Belief function Bel: 2 θ [0, 1] For any Bel(A)= m ( B) B A A θ. Bel(A) measures the total belief of all possible subsets of A. Properties of Belief Function: 1. Bel(Ф)=0. 2. Bel(θ)=1. n+ 1 3. Bel( A1... An ) Bel( ) Bel( A ) +.. + ( 1) Bel( A1... An ). i i< Definition 3: Plausibility function Pl: 2 θ [0, 1] Also, Pl(A)=1-Bel(A). Pl(A)= m ( B) B A 0 and, Bel(A) Pl(A). Properties of Plausibility Function: 1. Pl(Ф)=0. 2. Pl(θ)=1. 3. Pl A n+ 1 (... An ) Pl( ) Pl( A ) +.. + ( 1) Pl( A1 Definition 4: Doubt is defined as: Doubt(A) = Bel (A).... A ). 1 n i i<

Pl(A) measures the total belief mass that can move into A, whereas Bel(A) measures the total belief mass that is confined to A. Belief is also called lower probability and Plausibility is called upper probability. We can show a graphical representation the definition as follows: If A, AB, ABC, CD and BC are events and their probabilities are known, then we can represent probability, belief and plausibility of AB as: Rule of Combination: Suppose there are two distinct bodies of evidence and they are expressed by two basic assignments m1 and m2 on some focal points from power sets of the frame of discernment. The combined two belief functions can be represented by means of a oint basic assignment m1, 2. From the figure, m1 is a basic assignment over θ and its focal points are B1...Bn. The other basic assignment m2 has focal points on C1...Cn. Then the oint basic assignment for A is given by:

m ) 1,2 ( A) = m1 ( Bi ) m2 ( C. Bi C = A But, there is a problem in this formula when B i C = 0 for particular i and ( such as (i, )=(1,2), (2,n),,(m,2) as shown in the figure). In this case, the total area of rectangles which are now the focal elements for m1,2 is less than one, which violates second condition of basic probability assignment. To make the sum of m s to be equal to 1 we multiply them by the factor: (1 1 B = φ m ( B ) m i 2 ( C )). 1 Thus the rule of combination is expressed by: m1,2(a)= m ( Bi ) m2( C 1 κ 1 ) where к = m1 ( Bi ) m2( C ). B C= φ к is also called conflict. It indicates the degree of conflict between two distinct bodies of evidence. Larger к indicates that there is more conflict in evidence. This rule satisfies both associative and commutative rules; i.e. if m3(a) comes in,,then use m12 and m3. (m13, m2, m23, m1: Order does not matter.). Example 1: Suppose there are 100 cars in a parking lot consisting of type A and B. Two policemen check the type of cars in the lot. First policeman says that there are 30 A cars and 20 B cars among the 50 cars he counted. Second policeman says that there are 20 A cars and 20 cars that could A or B. Thus the basic probability assignment could be: m1(a)=0.3 m1(b)=0.2 m1(θ)=0.5 m2(a)=0.2 m2(ab)=0.2 m2(θ)=0.6 m1(a) 0.3 m1(b) 0.2 m1(θ) 0.5 m2(a) 0.2 0.06 0.04 (0 intersection) 0.1 m2(ab) 0.2 0.06 0.04 0.1 m2(θ) 0.6 0.18 0.12 0.3 к= 0.04. 1-к = 0.96. Thus, using the rule of combination,

m12(a)= (0.06+0.06+0.18+0.1)/(1-к)=0.42. m12(b)= (0.04+0.12)/(1-к)= 0.17. m12(ab)= 0.1/(1-к)= 0.1. m12(θ)=0.3/(1-к)=0.31. Bel(A)=m12(A)=0.42. (42 A cars) Bel(B)=m12(B)=0.17. (17 B cars) Bel(AB)= m12(a)+m12(b)+m12(ab)=0.69. (69 cars) Pl(A)= m12(a)+m12(ab)+m12(θ)=0.83. (83 A cars) Pl(B)= m12(b)+m12(ab)+m12(θ)=0.58. (58 B cars) Pl(θ)= 1.00. (100 cars) Thus by the rule of combination, we get a range of values such as 42-83 A cars and 17-58 B cars. Example 2: Comparison of Dempster Shafer Theory and Bayesian Theory. Suppose if m1(a)=0.6 and m1( A )=0.2 and m2(a)=0.8 and m2( A )=0.1. For the rest of the values, let m1(θ)=0.2 and m2(θ)=0.1. m1(a) m1( A ) m1(θ) 0.6 0.2 0.2 m2(a) 0.8 0.48 0.16 0.1 m2( A ) 0.1 0.06 0.02 0.02 m2(θ) 0.1 0.06 0.02 0.02 К=0.16+0.6=0.22. 1-к=0.78. Using the formulas from example (1), we can find the following: m12(a)=(0.48+0.06+0.16)/(1-к)=0.9. m12( A )=(0.02+0.02+0.02)/(1-к)=0.08. m12(θ)=0.02/(1-к)=0.03. Using the formula for belief and plausibility functions: Bel(A)=0.9. Bel( A )=0.08. Bel(θ)=1. Pl(A)=0.92. Pl( A )=0.1. Pl(θ)=1. If we represent it graphically:

If we solve this problem using Bayesian Theory: Covariance of A by m1 and by m2 is: P = ( P P 1 1 1 1 + 2 ) = (1/0.1 + 1/0.05)^(-1) = (10+20)^(-1)=0.033. Here, we are actually calculating the confidence in the value of A using two different sources. Now, mean, 1 1 m= P( P m1 + P 2 ) 1 2 m =0.033(10*0.7+20*0.85) =0.8. Comparing the range of uncertainties of Dempster Shafer (0.9-0.92) and Bayesian (0.8+orcovariance), we can conclude that Dempster Shafer provides the estimate with a smaller degree of uncertainty. Dempster Shafer combines more efficiently the two pieces of evidence than Bayesian. Example 3: If Doctor1 says that a patient has 99% chances of having cough and 1% chance of having flu and Doctor2 says that the same patient has 99% chances of having backache and 1% chance of having flu, then we can surely say that the patient has the lowest probability of having flu. Let us solve this using Dempster Shafer Theory. Thus, m1(a)=0.99 (cough) m1(b)=0.1 (flu)

m2(c)=0.99 (backache) m2(b)=0.01 (flu) Then. m2(c) 0.99 m2(b) 0.01 m2(θ) 0 m1(a) 0.99 0.98 0.099 0 m1(b) 0.01 0.0099 0.0001 0 m1(θ) 0 0 0 0 К=0.9999. 1-к=0.0001. Now using the rules of combination, m12(a)=0; m12(c)=0; m12(b)=1. From this result, combination of both the doctors prescription and using Dempster Shafer Theory, the chances of the patient having the flu is 100%. But this is wrong. Thus, Dempster Shafer Theory fails to give satisfactory results when a high degree of conflict exists among the sources of evidence. Applications of Dempster Shafer Theory: 1. Sensor Fusion: The main function of sensor fusion techniques is to extract more reliable and accurate information from an obect using two or more sensors. Since there are not only operating limits of a sensor but also noise corrupted environment, no single sensor can guarantee to deliver acceptably accurate information all the time. We can use Dempster Shafer Theory in sensor fusion. According to this scheme, sensor classifier produces two parameters for each hypothesis, supportability variable and plausibility variable. The difference is the ignorance of the hypothesis. Dempster Shafer s rule of combination provides a means of computing composite supportability/plausibility intervals for each hypothesis. As a result, a supportability/plausibility interval vector is obtained with a normalization vector. If one hypothesis is strongly supported by a group of sensor classifiers, then the credibility interval is reduced for that hypothesis. The normalization vector indicates the consistency of each sensor. Potential advantage of using Dempster Shafer theory in Sensor Fusion: 1. Different level of abstraction. 2. Representation of Ignorance. 3. Conflict Resolution. 2. Expert Systems: Uncertainty in rule based systems can be represented by associating a basic assignment with each rule. m E H We can have an expert system which aims at selecting a forecast technique or its combination based on the characteristics of a forecasting situation given by the user. The obective of forecasting is to predict the future. Since its decision is directly related with the future of the business, it is very important to forecasting or marketing managers in the field. This system has five modules, knowledge acquisition, knowledge base, belief base, Dempster Shafer module and Inference Engine. 3. Fault Detection and Identification: The FDI algorithm is developed for detecting and identifying a failure in a system. Suppose that there are more than one estimable parameters,

the Dempster Shafer Theory is useful method to detect and identify a failure from the signatures of the parameters. In the last stage of FDI algorithm, the decision is made with an index which measures the reliability of the decision, which is called the degree of certainty. Future of Dempster Shafer Theory: 1. Relationship between Fuzzy Logic, Neural Networks, Bayesian and Dempster Shafer Theory. 2. Modification of Dempster Shafer rule using weighted sources of evidence: In Dempster Shafer rule, the weights of independent sources of evidence are considered to be the same. Also, if belief functions m1 and m2 are identical, then their combination may not be equal to either of them. As a result, in repetitive combination of belief with same bpa, the resultant bpa either goes to zero or unity (Flu example). If we use weighted sources of evidence, we can solve this problem. Weighted Dempster Shafer allows taking into account the different reliabilities of the sensors. A high (low) weight is assigned to a more reliable (less reliable) source of evidence. Research is devoted to dynamically update these weights on-line.