probabilistic reasoning

44
Medical Decision-Support Systems Probabilistic Reasoning in Diagnostic Systems Yuval Shahar, M.D., Ph.D.

Upload: roger961

Post on 04-Jul-2015

329 views

Category:

Technology


0 download

TRANSCRIPT

Page 1: Probabilistic reasoning

Medical Decision-Support Systems

Probabilistic Reasoning in Diagnostic Systems

Yuval Shahar, M.D., Ph.D.

Page 2: Probabilistic reasoning

Reasoning Under Uncertainty in Medicine

• Uncertainty is inherent to medical reasoning– relation of diseases to clinical and laboratory

findings is probabilistic– Patient data itself is often uncertain with respect

to value and time – Patient preferences regarding outcomes vary– Cost of interventions and therapy can change

Page 3: Probabilistic reasoning

Probability: A Quick Introduction

• Probability function, range: [0, 1]• Prior probability of A, P(A): with no new

information (e.g., no patient information)• Posterior probability of A: P(A) given certain

information (e.g. laboratory tests)• Conditional probability: P(B|A)• Independence of A, B: P(B) = P(B|A)• Conditional independence of B,C, given A: P(B|

A) = P(B|A & C) – (e.g., two symptoms, given a specific disease)

Page 4: Probabilistic reasoning

Probabilistic Calculus

• P(not(A)) = 1-P(A)• In general:

– P(A & B) = P(A) * P(B|A)

• If A, B are independent: – P(A & B) = P(A) * P(B)

• If A, B are mutually exclusive:– P(A or B) = P(A) + P(B)

• If A,B not mutually exclusive , but independent:– P(A or B) = 1-P(not(A) & not(B)) = 1-(1-P(A))(1-P(B))

Page 5: Probabilistic reasoning

Test Characteristics

FP+TNTP+FN

FN+TNTrue negative (TN)

False negative (FN)

Negative

TP+FPFalse positive (FP)

True positive (TP)

Positive

TotalDisease absent

Disease present

Disease

Test result

Page 6: Probabilistic reasoning

Test Performace Measures

• The gold standard test: the procedure that defines presence or absence of a disease (often, very costly)

• The index test: The test whose performance is examined• True positive rate (TPR) = Sensitivity:

– P(Test is positive|patient has disease) = P(T+|D+)– Ratio of number of diseased patients with positive tests to total

number of patient: TP/(TP+FN)• True negative rate (TNR) = Specificity

– P(Test is negative|patient has no disease) = P(T-|D-)– Ratio of number of nondiseased patients with negative tests to

total number of patients: TN/(TN+FP)

Page 7: Probabilistic reasoning

Test Predictive Values

• Positive predictive value (PV+) = P(D|T+) = TP/(TP+FP)

• Negative predictive value (PV-) = P(D-|T-) = TN/(TN+FN)

Page 8: Probabilistic reasoning

Lab Tests: What is “Abnormal”?

Page 9: Probabilistic reasoning

The Cut-off Value Trade off

• Sensitivity and specificity depend on the cut off value between what we define as normal and abnormal

• Assume high test values are abnormal; then, moving the cut-off value to a higher one increases FN results and decreases FP results (i.e. more specific) and vice versa

• There is always a trade off in setting the cut-off point

Page 10: Probabilistic reasoning

Receiver Operating Characteristic (ROC) Curves: Examples

Page 11: Probabilistic reasoning

Receiver Operating Characteristic (ROC) Curves: Interpretation

• ROC curves summarize the trade-off between the TPR (sensitivity) and the false positive rate (FPR) (1-specificity) for a particular test, as we vary the cut-off treshold

• The greater the area under the ROC curve, the better (more sensitive, more specific)

Page 12: Probabilistic reasoning

Bayes Theorem

FPRDPySensitivitDP

DTPDP

DTPDPDTPDP

DTPDP

TP

DTPDP

TDPpositivetestdiseaseP

AP

BAPBPABP

*))(1(*)(

)|()(

)|()(()|()(

)|()(

)(

)|()(

)|():|(

)(

)|()()|(

B),|P(A P(B) A)|P(A)P(B B)&P(A

−++=

−+−+++=

++=

+=

==>

==

Page 13: Probabilistic reasoning

Odds-Likelihood (Odds Ratio) Form of Bayes Theorem

• Odds = P(A)/(1-P(A)); P = Odds/(1+Odds)

• Post-test odds = pretest odds * likehood ratio

TNR

FNR

DTP

DTPLRratiolikelihoodNegative

FPR

TPR

DTP

DTPLRratioLikelihood

DTP

DTP

DP

DP

TDP

TDP

=−−

−=−−−

=−+

+=+−

−++

−=

+−+

)|(

)|()(

)|(

)|()(

)|(

)|(*

)(

)(

)|(

)|(

Page 14: Probabilistic reasoning

Application of Bayes Theorem• Needs reliable pre-test probabilities• Needs reliable post-test likelihood ratios• Assumes one disease only (mutual exclusivity of

diseases)• Can be used in sequence for several tests, but only if

they are conditionally independent given the disease; then we use the post-test probability of Ti as the pre-test probability for Ti+1 (Simple, or Naïve, Bayes)

∏=−

=− ni

i

i

i

i

i

LRDP

DP

TDP

TDP

..1)(

)(

)|(

)|(

Page 15: Probabilistic reasoning

Relation of Pre-Test and Post-Test Probabilities

Page 16: Probabilistic reasoning

Example: Computing Predictive Values

• Assume P(Down Syndrom):– (A) 0.1% (age 30)– (B) 2% (age 45)

• Assume amniocentesis with Sensitivity of 99%, Specificity of 99% for Down Syndrom

• PV+ = P(DS|Amnio+)

• PV- = P(DS-|Amnio-) = 99.999%

Page 17: Probabilistic reasoning

Predictive Values: Down Syndrom

0.99979 -PV

66891.001.0*98.099.0*02.0

99.0*02.0

2%(B)P(DS)

99.999% Amnio-)|P(DS- -PV

0901.000999.000099.0

00099.0

01.0*999.099.0*001.0

99.0*001.0

%1.0)()(

=

=+

=+

=

==

=+

=+

=+

=

PV

PV

DSPA

Page 18: Probabilistic reasoning

Example: de Dombal’s System (1972)• Domain: Acute abdominal pain (7 possible diagnoses)• Input: Signs and symptoms of patient• Output: Probability distribution of diagnoses• Method: Naïve Bayesian classification• Evaluation: an eight-center study involving 250 physicians and

16,737 patients• Results:

– Diagnostic accuracy rose from 46 to 65%– The negative laparotomy rate fell by almost half– Perforation rate among patients with appendicitis fell by half– Mortality rate fell by 22%

• Results using survey data consistently better than the clinicians’ opinions and even the results using human probability estimates!

Page 19: Probabilistic reasoning

Decision Trees

• A convenient way to explicitly show the order and relationships of possible decisions, uncertain outcomes of decisions , and outcome utilities

• Enable computation of the decision that maximizes expected utility

Page 20: Probabilistic reasoning

Decision Trees Conventions

Decision node Chance node

Information link

Influence link

Page 21: Probabilistic reasoning

A Generic Decision Tree

Page 22: Probabilistic reasoning

Decision Trees: an HIV Example

Decision node

Chance node

Page 23: Probabilistic reasoning

Computation With Decision Trees

• Decision trees are “folded back” to the top most (leftmost, or initial) decision

• Computation is performed by averaging expected utility recursively over tree branches from right to left (bottom up), maximizing utility for every decision made and assuming that this is the expected utility for the subtree that follows the computed decision

Page 24: Probabilistic reasoning

Influence Diagrams: Node Conventions

Chance node

Decision node

Utility node

Page 25: Probabilistic reasoning

Link Semanticsin Influence Diagrams

Dependence link

Information link

Influence link

Page 26: Probabilistic reasoning

Influence Diagrams: An HIV Example

Page 27: Probabilistic reasoning

The Structure of Influence Diagram Links

Page 28: Probabilistic reasoning

Belief Networks (Bayesian/Causal Probabilistic/Probabilistic Networks, etc)

Disease

Fever Sinusitis

Runny nose

Headache

Influence diagrams without decision and utility nodes

Gender

Page 29: Probabilistic reasoning

Link Semantics in Belief Networks

Dependence

Independence

Conditional independence of B and C, given A

B

CA

Page 30: Probabilistic reasoning

Advantages of Influence Diagrams and Belief Networks

• Excellent modeling tool that supports acquisition from domain experts– Intuitive semantics (e.g., information and influence links)– Explicit representation of dependencies– very concise representation of large decision models

• “Anytime” algorithms available (using probability theory) to compute the distribution of values at any node given the values of any subset of the nodes (e.g., at any stage of information gathering)

• Explicit support for value of information computations

Page 31: Probabilistic reasoning

Disadvantages of Influence Diagrams and Belief Networks

• Explicit representation of dependencies often requires acquisition of joint probability distributions (P(A|B,C))

• Computation in general intractable (NP hard)

• Order of decisions and relations between decisions and available information might be obscured

Page 32: Probabilistic reasoning

Value of Information (VI)• We often need to decide what would be the next best

piece of information to gather (e.g., within a diagnostic process); that is, what is the best next question to ask (e.g., what would be the result of a urine culture?)

• The Value of Information (VI) of feature f is the marginal expected utility of an optimal decision made knowing f, compared to making it without knowing f

• The net value of information (NVI) of f = VI(f)-cost(f)• NVI is highly useful for a hypothetico-deductive

diagnostic approach to decide what would be the next information item, if any, to investigate

Page 33: Probabilistic reasoning

Examples of Successful Belief- Network Applications

• In clinical medicine:– Pathological diagnosis at the level of a

subspecialized medical expert (Pathfinder)– Endocrinological diagnosis (NESTOR)

• In bioinformatics:– Recognition of meaningful sites and features in

DNA sequences– Educated guess of tertiary structure of proteins

Page 34: Probabilistic reasoning

The Pathfinder Project(Heckerman, Horvitz, Nathwani 1992)

• Task and domain: Diagnosis of lymph node biopsy, an important medical problem– Large difference between expert and general pathologist

opinions (almost 65%!)

• Problems in the domain include– Misrecognition of features (information gathering)– Misintegration of evidence (information processing)

• The Pathfinder project focused mainly on assistance in information processing

• A Stanford/USC collaboration; eventually commercialized as Intellipath, marketed by the ACP, used as early as 1992 by at least 200 pathology sites

Page 35: Probabilistic reasoning

Pathfinder Domain

• More than 60 diseases• More than 130 findings, such as:

– Microscopic– immunological– molecular biology– Laboratory– Clinical

• Commercial product extended to at least 10 more medical domains

Page 36: Probabilistic reasoning

Pathfinder I/O behavior• Input: set of <Feature, Instance> (<Fi, Ii>) pairs

(e.g., <NECROSIS, ABSENT>– Instances are mutually exclusive values of each feature– Prior probability of each disease Dk is known

– P(F1I1, F2I2…FtIt | Dk,ξ) is in acquired knowledge base

• Output: P(Dk|F1I1, F2I2…FmIm,ξ)− ξ = background knowledge (context)

• User can ask what is the next best (cost-effective) feature to investigate or enter- Probabilistic (decision-theoretic) hypothethico-deductive

approach

• Distribution of each Dk is updated dynamically

Page 37: Probabilistic reasoning

Pathfinder Methodology:Probabilities and Utilities

• Decision-theoretic computation• Bayesian approach: Probabilities represent beliefs of

experts (data can update beliefs)• Utilities represented as a matrix of all diseases• A matrix entry pair < Dj Dk> encodes the (patient) utility

of diagnosing Dk when patient really has Dk

• Since no therapeutic recommendations are made, the model can use one representative patient (the expert), expressed in micromorts and willingness-to-pay to avoid risk of each outcome

Page 38: Probabilistic reasoning

Pathfinder Computation

• Normally we would use the general form of Bayes Theorem:

∑ ……=…

lD)|D(P),D|IFIF ,IP(F

)|D(P),D|IFIF ,IP(F ), IFIF ,IF|P(D

lltt2211

kktt2211tt2211k

ξξξξξ

• But that involves exponential number of probabilities to be acquired and represented

Page 39: Probabilistic reasoning

Pathfinder 1: The Simple Bayes Version

• Assuming conditional independence of features (Simple or Naïve Bayes):

∏==…

i

P

),D|IF P(

),D|IF()...,D|IF )P(,D|IP(F),D|IFIF ,IP(F

kii

kt)tk22k11ktt2211

ξξξξξ

• Assuming mutual exclusivity and exhaustiveness of diseases the overall computation is tractable:

∑∏∏

=

lD i

)|D(P),DIP(F

)|D(P),DIP(F),IF|P(D

ll|ii

i

kk|ii

i

iikξξ

ξξξ

Page 40: Probabilistic reasoning

Pathfinder 2: The Belief Network Version

• Mutual exclusivity and exhaustiveness of diseases is reasonable in lymphnode pathology– Single disease per examined lymph node– Large, exhaustive knowledge base

• Conditional independence is less reasonable and can lead to erroneous conclusions

• The simple Bayes representation of Pathfinder 1 was therefore enhanced to a belief network in Pathfinder 2 which included explicit dependencies between different features, still taking advantage of any explicit global and conditional independencies

Page 41: Probabilistic reasoning

Decision-Theoretic Diagnosis

• Using the utility matrix and given observations φ, the expected diagnostic utility using φ is averaged over all diagnoses:– EU(Dk(φ)) = ΣjP(Dj| φ)U(Dj,Dk)

• Thus, Dx(φ) = ARGMAXk [EU(Dk (φ))

• However, since the diagnosis is sensitive to the utility model, Pathfinder does not recommend it, only the probabilities P(Dk |φ)

Page 42: Probabilistic reasoning

Pathfinder: Gathering Information

• Next best feature to observe is recommended using a myopic approximation, which considers only up to one single feature to be observed

• The feature chosen maximizes EU given that a diagnosis would be made after observing it

• Feature f is chosen that maximizes NVI(f)• Although myopic approximation could backfire, in

practice it works well– especially when U(Dj,Dk) =is set to 0 if one of the diseases is

malignant and the other benign, and set to 1 if they are both malignant or both benign

Page 43: Probabilistic reasoning

Pathfinder 2: Knowledge Acquisition

• To facilitate acquisition of multiple probabilities, a Similarity Network model was developed

• Using similarity networks, an expert creates multiple small belief networks, representing 2 or more diseases that are difficult to distinguish

• The local belief networks are then unified into a global belief network, preserving soundness

• The graphical interface also allows partitioning of diseases into sets, relative to each set some feature is independent, thus further assisting in the construction

Page 44: Probabilistic reasoning

Pathfinder 1 and 2: Evaluation• Pathfinder 1 was compared to Pathfinder 2 using 53 cases,

a new user, and a thorough analysis of each case– Diagnostic accuracy of PF2 is greater than that of PF1 (gold

standard: the main domain expert’s distribution and his assessment on a scale of 1 to 10)

– Difference is due to better probabilistic representation (better acquisition and inference)

– Cost of constructing PF2 rather than PF1 is justified by the improvements, (measure: the utility of the diagnosis)

– PF2 is at least as good as the main domain expert, with respect to diagnostic accuracy