94
Abductive Plan Recognition By Extending Bayesian Logic Programs Sindhu V. Raghavan & Raymond J. Mooney The University of Texas at Austin 1

Abductive Plan Recognition By Extending Bayesian Logic Programs

  • Upload
    heinz

  • View
    28

  • Download
    0

Embed Size (px)

DESCRIPTION

Abductive Plan Recognition By Extending Bayesian Logic Programs. Sindhu V. Raghavan & Raymond J. Mooney The University of Texas at Austin. Plan Recognition. Predict an agent’s top-level plans based on the observed actions Abductive reasoning involving inference of cause from effect - PowerPoint PPT Presentation

Citation preview

Page 1: Abductive Plan Recognition By Extending Bayesian Logic Programs

Abductive Plan Recognition By Extending Bayesian Logic Programs

Sindhu V. Raghavan & Raymond J. Mooney

The University of Texas at Austin

1

Page 2: Abductive Plan Recognition By Extending Bayesian Logic Programs

Plan RecognitionPredict an agent’s top-level plans based on the

observed actions

Abductive reasoning involving inference of cause from effect

ApplicationsStory UnderstandingStrategic Planning Intelligent User Interfaces

2

Page 3: Abductive Plan Recognition By Extending Bayesian Logic Programs

$ cd test-dir$ cp test1.txt my-dir$ rm test1.txt

$ cd test-dir$ cp test1.txt my-dir$ rm test1.txtWhat task is the user performing?move-file

Which files and directories are involved?test1.txt and test-dir

Plan Recognition in Intelligent User Interfaces

3

Data is relational in nature - several files and directories and several relations between them

Page 4: Abductive Plan Recognition By Extending Bayesian Logic Programs

Related WorkFirst-order logic based approaches [Kautz and Allen, 1986; Ng

and Mooney, 1992]

Knowledge base of plans and actions Default reasoning or logical abduction to predict the best plan

based on the observed actions Unable to handle uncertainty in data or estimate likelihood of

alternative plans

Probabilistic graphical models [Charniak and Goldman, 1989; Huber et al., 1994; Pynadath and Wellman, 2000; Bui, 2003; Blaylock and Allen, 2005]

Encode the domain knowledge using Bayesian networks, abstract hidden Markov models, or statistical n-gram models

Unable to handle relational/structured data

Statistical Relational Learning based approaches Markov Logic Networks for plan recognition [Kate and Mooney, 2009;

Singla and Mooney, 2011]

4

Page 5: Abductive Plan Recognition By Extending Bayesian Logic Programs

Our Approach

Extend Bayesian Logic Programs (BLPs) [Kersting and

De Raedt, 2001] for plan recognitionBLPs integrate first-order logic and Bayesian

networks

Why BLPs?Efficient grounding mechanism that includes only those

variables that are relevant to the queryEasy to extend by incorporating any type of logical

inference to construct networksWell suited for capturing causal relations in data

5

Page 6: Abductive Plan Recognition By Extending Bayesian Logic Programs

Outline

MotivationBackground

Logical AbductionBayesian Logic Programs (BLPs)

Extending BLPs for Plan RecognitionExperimentsConclusions

6

Page 7: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical AbductionAbduction

Process of finding the best explanation for a set of observations

GivenBackground knowledge, B, in the form of a set of (Horn) clauses in

first-order logicObservations, O, in the form of atomic facts in first-order logic

FindA hypothesis, H, a set of assumptions (atomic facts) that logically

entail the observations given the theory:

B H OBest explanation is the one with the fewest assumptions

7

Page 8: Abductive Plan Recognition By Extending Bayesian Logic Programs

Bayesian Logic Programs (BLPs) [Kersting and De Raedt, 2001]

Set of Bayesian clauses a|a1,a2,....,an

Definite clauses that are universally quantifiedRange-restricted, i.e variables{head} variables{body}Associated conditional probability table (CPT)

o P(head|body)

Bayesian predicates a, a1, a2, …, an have finite domainsCombining rule like noisy-or for mapping multiple CPTs

into a single CPT.

8

Page 9: Abductive Plan Recognition By Extending Bayesian Logic Programs

Inference in BLPs[Kersting and De Raedt, 2001]

Logical inferenceGiven a BLP and a query, SLD resolution is used to construct

proofs for the query

Bayesian network constructionEach ground atom is a random variableEdges are added from ground atoms in the body to the ground

atom in headCPTs specified by the conditional probability distribution for the

corresponding clauseP(X) = P(Xi | Pa(Xi))

Probabilistic inferenceMarginal probability given evidenceMost Probable Explanation (MPE) given evidence

9€

i

Page 10: Abductive Plan Recognition By Extending Bayesian Logic Programs

BLPs for Plan Recognition

SLD resolution is deductive inference, used for predicting observations from top-level plans

Plan recognition is abductive in nature and involves predicting the top-level plan from observations

10

BLPs cannot be used as is for plan recognition

Page 11: Abductive Plan Recognition By Extending Bayesian Logic Programs

Extending BLPs for Plan Recognition

11

BLPsBLPs Logical Abduction

Logical Abduction

BALPsBALPs

BALPs – Bayesian Abductive Logic Programs

+

=

Page 12: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Abduction in BALPs

Given A set of observation literals O = {O1, O2,….On} and a

knowledge base KB

Compute a set abductive proofs of O using Stickel’s abduction algorithm [Stickel 1988]

Backchain on each Oi until it is proved or assumed

A literal is said to be proved if it unifies with a fact or the head of some rule in KB, otherwise it is said to be assumed

Construct a Bayesian network using the resulting set of proofs as in BLPs.

12

Page 13: Abductive Plan Recognition By Extending Bayesian Logic Programs

Example – Intelligent User InterfacesTop-level plan predicates

copy-file, move-file, remove-file

Action predicatescp, rm

Knowledge Base (KB)cp(Filename,Destdir) | copy-file(Filename,Destdir)cp(Filename,Destdir) | move-file(Filename,Destdir) rm(Filename) | move-file(Filename,Destdir) rm(Filename) | remove-file(Filename)

Observed actionscp(test1.txt, mydir) rm(test1.txt)

13

Page 14: Abductive Plan Recognition By Extending Bayesian Logic Programs

Abductive Inference

14

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

cp(Filename,Destdir) | copy-file(Filename,Destdir)

Assumed literalAssumed literal

Page 15: Abductive Plan Recognition By Extending Bayesian Logic Programs

Abductive Inference

15

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

cp(Filename,Destdir) | move-file(Filename,Destdir)

Assumed literalAssumed literal

Page 16: Abductive Plan Recognition By Extending Bayesian Logic Programs

Abductive Inference

16

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(Filename) | move-file(Filename,Destdir)

rm(test1.txt)

Match existing assumptionMatch existing assumption

Page 17: Abductive Plan Recognition By Extending Bayesian Logic Programs

Abductive Inference

17

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(Filename) | remove-file(Filename)

rm(test1.txt)

remove-file(test1)

Assumed literalAssumed literal

Page 18: Abductive Plan Recognition By Extending Bayesian Logic Programs

Structure of Bayesian network

18

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Page 19: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

Specifying probabilistic parametersNoisy-and

o Specify the CPT for combining the evidence from conjuncts in the body of the clause

Noisy-oro Specify the CPT for combining the evidence from

disjunctive contributions from different ground clauses with the same head

o Models “explaining away”Noisy-and and noisy-or models reduce the number of

parameters learned from data

19

Page 20: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

20

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Page 21: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

Most Probable Explanation (MPE)For multiple plans, compute MPE, the most likely

combination of truth values to all unknown literals given this evidence

Marginal ProbabilityFor single top level plan prediction, compute marginal

probability for all instances of plan predicate and pick the instance with maximum probability

When exact inference is intractable, SampleSearch [Gogate

and Dechter, 2007], an approximate inference algorithm for graphical models with deterministic constraints is used

21

Page 22: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

22

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Page 23: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

23

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Evidence

Page 24: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

24

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Evidence

Query variables

Page 25: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

25

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Evidence

Query variablesTRUE FALSEFALSE

MPE

Page 26: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

26

copy-file(test1.txt,mydir)

cp(test1.txt,mydir)

move-file(test1.txt,mydir)

rm(test1.txt)

remove-file(test1)

Noisy-or Noisy-or

Evidence

Query variablesTRUE FALSEFALSE

MPE

Page 27: Abductive Plan Recognition By Extending Bayesian Logic Programs

Parameter Learning

Learn noisy-or/noisy-and parameters using the EM algorithm adapted for BLPs [Kersting and De Raedt, 2008]

Partial observability In plan recognition domain, data is partially observableEvidence is present only for observed actions and top-level

plans; sub-goals, noisy-or, and noisy-and nodes are not observed

Simplify learning problemLearn noisy-or parameters onlyUsed logical-and instead of noisy-and to combine evidence

from conjuncts in the body of a clause

27

Page 28: Abductive Plan Recognition By Extending Bayesian Logic Programs

Experimental Evaluation

Monroe (Strategic planning)

Linux (Intelligent user interfaces)

Story Understanding (Story understanding)

28

Page 29: Abductive Plan Recognition By Extending Bayesian Logic Programs

Monroe and Linux [Blaylock and Allen, 2005]

TaskMonroe involves recognizing top level plans in an

emergency response domain (artificially generated using HTN planner)

Linux involves recognizing top-level plans based on linux commands

Single correct plan in each example

Data

29

No. examples

Avg. observations/ example

Total top-level plan predicates

Total observed action predicates

Monroe 1000 10.19 10 30

Linux 457 6.1 19 43

Page 30: Abductive Plan Recognition By Extending Bayesian Logic Programs

Monroe and Linux

MethodologyManually encoded the knowledge base Learned noisy-or parameters using EMComputed marginal probability for plan instances

Systems comparedBALPsMLN-HCAM [Singla and Mooney, 2011]

o MLN-PC and MLN-HC do not run on Monroe and Linux due to scaling issues

Blaylock and Allen’s system [Blaylock and Allen, 2005]

Performance metricConvergence score - measures the fraction of examples for

which the plan predicate was predicted correctly

30

Page 31: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Monroe

31

94.2 *

Co

nve

rgen

ce S

core

BALPs MLN-HCAM Blaylock & Allen

* - Differences are statistically significant wrt BALPs

Page 32: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Linux

32

Co

nve

rgen

ce S

core

BALPs MLN-HCAM Blaylock & Allen

36.1 *

* - Differences are statistically significant wrt BALPs

Page 33: Abductive Plan Recognition By Extending Bayesian Logic Programs

Experiments with partial observability Limitations of convergence score

Does not account for predicting the plan arguments correctly

Requires all the observations to be seen before plans can be predicted

Early plan recognition with partial set of observationsPerform plan recognition after observing the first 25%,

50%, 75%, and 100% of the observationsAccuracy – Assign partial credit for the predicting plan

predicate and a subset of the arguments correctly

Systems comparedBALPsMLN-HCAM [Singla and Mooney, 2011]

33

Page 34: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Monroe

34Percent observations seen

Ac

cu

rac

y

Page 35: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Linux

35

Ac

cu

rac

y

Percent observations seen

Page 36: Abductive Plan Recognition By Extending Bayesian Logic Programs

Story Understanding [Charniak and Goldman, 1991; Ng and Mooney, 1992]

TaskRecognize character’s top level plans based on actions

described in narrative textMultiple top-level plans in each example

Data25 examples in development set and 25 examples in test

set12.6 observations per example8 top-level plan predicates

36

Page 37: Abductive Plan Recognition By Extending Bayesian Logic Programs

Story UnderstandingMethodology

Knowledge base was created for ACCEL [Ng and Mooney, 1992]

Parameters set manuallyo Insufficient number of examples in the development set to learn

parameters

Computed MPE to get the best set of plans

Systems comparedBALPsMLN-HCAM [Singla and Mooney, 2011]

o Best performing MLN model

ACCEL-Simplicity [Ng and Mooney, 1992]

ACCEL-Coherence [Ng and Mooney, 1992]

o Specific for Story Understanding

37

Page 38: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Story Understanding

38* - Differences are statistically significant wrt BALPs

* *

Page 39: Abductive Plan Recognition By Extending Bayesian Logic Programs

Conclusion

BALPS – Extension of BLPs for plan recognition by employing logical abduction to construct Bayesian networks

Automatic learning of model parameters using EM

Empirical results on all benchmark datasets demonstrate advantages over existing methods

39

Page 40: Abductive Plan Recognition By Extending Bayesian Logic Programs

Future Work

Learn abductive knowledge base automatically from data

Compare BALPs with other probabilistic logics like ProbLog [De Raedt et. al, 2007], PRISM [Sato, 1995] and Poole’s Horn Abduction [Poole, 1993] on plan recognition

40

Page 41: Abductive Plan Recognition By Extending Bayesian Logic Programs

Questions

41

Page 42: Abductive Plan Recognition By Extending Bayesian Logic Programs

Backup

42

Page 43: Abductive Plan Recognition By Extending Bayesian Logic Programs

Completeness in First-order Logic

Completeness - If a sentence is entailed by a KB, then it is possible to find the proof that entails it

Entailment in first-order logic is semidecidable, i.e it is not possible to know if a sentence is entailed by a KB or not

Resolution is complete in first-order logic If a set of sentences is unsatisfiable, then it is possible to

find a contradiction

43

Page 44: Abductive Plan Recognition By Extending Bayesian Logic Programs

First-order LogicTerms

Constants – individual entities like anna, bobVariables – placeholders for objects like X, Y

Predicates Relations over entities like worksFor, capitalOf

Literal – predicate or its negation applied to terms Atom – Positive literal like worksFor(X,Y)Ground literal – literal with no variables like

worksFor(anna,bob)

Clause – disjunction of literalsHorn clause has at most one positive literalDefinite clause has exactly one positive literal

44

Page 45: Abductive Plan Recognition By Extending Bayesian Logic Programs

First-order Logic

QuantifiersUniversal quantification - true for all objects in the domain

Existential quantification - true for some objects in the domain

Logical InferenceForward Chaining– For every implication pq, if p is true,

then q is concluded to be trueBackward Chaining – For a query literal q, if an implication

pq is present and p is true, then q is concluded to be true, otherwise backward chaining tries to prove p

45

∀( )

∃( )

Page 46: Abductive Plan Recognition By Extending Bayesian Logic Programs

Forward chaining

For every implication pq, if p is true, then q is concluded to be true

Results in addition of a new fact to KBEfficient, but incompleteInference can explode and forward chaining may

never terminateAddition of new facts might result in rules being satisfied

It is data-driven, not goal-drivenMight result in irrelevant conclusions

46

Page 47: Abductive Plan Recognition By Extending Bayesian Logic Programs

Backward chaining

For a query literal q, if an implication pq is present and p is true, then q is concluded to be true, otherwise backward chaining tries to prove p

Efficient, but not completeMay never terminate, might get stuck in infinite

loopExponentialGoal-driven

47

Page 48: Abductive Plan Recognition By Extending Bayesian Logic Programs

Herbrand Model Semantics

Herbrand universeAll constants in the domain

Herbrand baseAll ground atoms atoms over Herbrand universe

Herbrand interpretationA set of ground atoms from Herbrand base that are true

Herbrand modelHerbrand interpretation that satisfies all clauses in the

knowledge base

48

Page 49: Abductive Plan Recognition By Extending Bayesian Logic Programs

Advantages of SRL models over vanilla probabilistic models

Compactly represent domain knowledge in first-order logic

Employ logical inference to construct ground networks

Enables parameter sharing

49

Page 50: Abductive Plan Recognition By Extending Bayesian Logic Programs

Parameter sharing in SRL

50

father(john)father(john)

parent(john)parent(john)

father(mary)father(mary)

parent(mary)parent(mary)

father(alice)father(alice)

parent(alice)parent(alice)

dummydummy

θ1θ1 θ2θ2 θ3θ3

Page 51: Abductive Plan Recognition By Extending Bayesian Logic Programs

Parameter sharing in SRL

father(X) parent(X)

51

father(john)father(john)

parent(john)parent(john)

father(mary)father(mary)

parent(mary)parent(mary)

father(alice)father(alice)

parent(alice)parent(alice)

dummydummy

θθ θθ θθ

θθ

Page 52: Abductive Plan Recognition By Extending Bayesian Logic Programs

Noisy-and Model

Several causes ci have to occur simultaneously if event e has to occur

ci fails to trigger e with probability pi

inh accounts for some unknown cause due to which e has failed to trigger

P(e) = (I – inh) Πi(1-pi)^(1-ci)

52

Page 53: Abductive Plan Recognition By Extending Bayesian Logic Programs

Noisy-or Model

Several causes ci cause event e has to occur

ci independently triggers e with probability pi

leak accounts for some unknown cause due to which e has triggered

P(e) = 1 – [(I – inh) Πi (1-pi)^(1-ci)]

Models explaining away If there are several causes of an event, and if there is

evidence for one of the causes, then the probability that the other causes have caused the event goes down

53

Page 54: Abductive Plan Recognition By Extending Bayesian Logic Programs

Noisy-and And Noisy-or Models

54

alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

Page 55: Abductive Plan Recognition By Extending Bayesian Logic Programs

Noisy-and And Noisy-or Models

55alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

dummy1 dummy2

Noisy/logical-and

Noisy/logical-andNoisy/logical-and

Noisy-or

Page 56: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

56

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

Example from Ngo and Haddawy, 1997

Page 57: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

57

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

Example from Ngo and Haddawy, 1997

Page 58: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

58

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

Example from Ngo and Haddawy, 1997

Page 59: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

59

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

Example from Ngo and Haddawy, 1997

Page 60: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

60

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

Example from Ngo and Haddawy, 1997

Page 61: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

61

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

Example from Ngo and Haddawy, 1997

Page 62: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

62

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

lives(james,Y) tornado(Y)

Example from Ngo and Haddawy, 1997

Page 63: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

63

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

lives(james,Y) tornado(Y)

lives(james,yorkshire)

Example from Ngo and Haddawy, 1997

Page 64: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

64

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

lives(james,Y) tornado(Y)

lives(james,yorkshire)

tornado(yorkshire)

Example from Ngo and Haddawy, 1997

Page 65: Abductive Plan Recognition By Extending Bayesian Logic Programs

Logical Inference in BLPs SLD Resolution

65

BLP

lives(james,yorkshire).lives(stefan,freiburg).neighborhood(james).tornado(yorkshire).

burglary(X) | neighborhood(X).alarm(X) | burglary(X).alarm(X) | lives(X,Y), tornado(Y).

Queryalarm(james)

Proof

alarm(james)

burglary(james)

neighborhood(james)

lives(james,Y) tornado(Y)

lives(james,yorkshire)

tornado(yorkshire)

Example from Ngo and Haddawy, 1997

Page 66: Abductive Plan Recognition By Extending Bayesian Logic Programs

Bayesian Network Construction

66

alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

Each ground atom becomes a node (random variable) in the Bayesian network

Edges are added from ground atoms in the body of a clause to the ground atom in the head

Specify probabilistic parameters using the CPTs associated with Bayesian clauses

Page 67: Abductive Plan Recognition By Extending Bayesian Logic Programs

Bayesian Network Construction

67

alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

Each ground atom becomes a node (random variable) in the Bayesian network

Edges are added from ground atoms in the body of a clause to the ground atom in the head

Specify probabilistic parameters using the CPTs associated with Bayesian clauses

Page 68: Abductive Plan Recognition By Extending Bayesian Logic Programs

Bayesian Network Construction

68

alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

Each ground atom becomes a node (random variable) in the Bayesian network

Edges are added from ground atoms in the body of a clause to the ground atom in the head

Specify probabilistic parameters using the CPTs associated with Bayesian clauses

Page 69: Abductive Plan Recognition By Extending Bayesian Logic Programs

Bayesian Network Construction

69

alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

Each ground atom becomes a node (random variable) in the Bayesian network

Edges are added from ground atoms in the body of a clause to the ground atom in the head

Specify probabilistic parameters using the CPTs associated with Bayesian clauses

Use combining rule to combine multiple CPTs into a single CPT

lives(stefan,freiburg)✖

Page 70: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

70

copy-file(Test1,txt,Mydir)

cp(Test1.txt,Mydir)

move-file(Test1,txt,Mydir)

rm(Test1.txt)

remove-file(Test1)

4 parameters

4 parameters

Page 71: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference

71

copy-file(Test1,txt,Mydir)

cp(Test1.txt,Mydir)

move-file(Test1,txt,Mydir)

rm(Test1.txt)

remove-file(Test1)

2 parameters 2 parameters

θ1 θ2 θ3θ4

Noisy models require parameters linear in the number of parents

Page 72: Abductive Plan Recognition By Extending Bayesian Logic Programs

Learning in BLPs[Kersting and De Raedt, 2008]

Parameter learningExpectation Maximization Gradient-ascent based learningBoth approaches optimize likelihood

Structure learningHill climbing search through the space of possible

structures Initial structure obtained from CLAUDIEN [De Raedt and

Dehaspe, 1997]

Learns from only positive examples

72

Page 73: Abductive Plan Recognition By Extending Bayesian Logic Programs

Probabilistic Inference and Learning

Probabilistic inferenceMarginal probability given evidenceMost Probable Explanation (MPE) given evidence

Learning [Kersting and De Raedt, 2008]

Parameterso Expectation Maximizationo Gradient-ascent based learning

Structureo Hill climbing search through the space of possible

structures

73

Page 74: Abductive Plan Recognition By Extending Bayesian Logic Programs

Expectation Maximization for BLPs/BALPs

• Perform logical inference to construct a ground Bayesian network for each example

• Let r denote rule, X denote a node, and Pa(X) denote parents of X

• E Step

• The inner sum is over all groundings of rule r

• M Step

74

*

*

* From SRL tutorial at ECML 07 74

Page 75: Abductive Plan Recognition By Extending Bayesian Logic Programs

Decomposable Combining Rules

Express different influences using separate nodes

These nodes can be combined using a deterministic function

75

Page 76: Abductive Plan Recognition By Extending Bayesian Logic Programs

Combining Rules

76alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

dummy1 dummy2

Logical-and

Logical-andLogical-and

Noisy-or

Page 77: Abductive Plan Recognition By Extending Bayesian Logic Programs

Decomposable Combining Rules

77alarm(james)

burglary(james)

neighborhood(james)

lives(james,yorkshire) tornado(yorkshire)

dummy1 dummy2

Logical-and

Logical-andLogical-and

Noisy-or Noisy-or

dummy1-new dummy2 -new

Logical-or

Page 78: Abductive Plan Recognition By Extending Bayesian Logic Programs

BLPs vs. PLPs

Differences in representation In BLPs, Bayesian atoms take finite set of values, but in

PLPs, each atom is logical in nature and it can take true or false

Instead of having neighborhood(x) = bad, in PLPs, we have neighborhood(x,bad)

To compute probability of a query alarm(james), PLPs have to construct one proof tree for all possible values for all predicates

Inference is cumbersome

BLPs subsume PLPs

78

Page 79: Abductive Plan Recognition By Extending Bayesian Logic Programs

BLPs vs. Poole's Horn Abduction

Differences in representationFor example, if P(x) and R(x) are two competing

hypothesis, then either P(x) could be true or R(x) could be true

Prior probabilities of P(x) and R(x) should sum to 1 Restrictions of these kind are not there in BLPsPLPs and hence BLPs are more flexible and have a richer

representation

79

Page 80: Abductive Plan Recognition By Extending Bayesian Logic Programs

BLPs vs. PRMs

BLPs subsume PRMsPRMs use entity-relationship models to represent

knowledge and they use KBMC-like construction to construct a ground Bayesian networkEach attribute becomes a random variable in the ground

network and relations over the entities are logical constraints In BLP, each attribute becomes a Bayesian atom and

relations become logical atoms Aggregator functions can be transformed into combining

rules

80

Page 81: Abductive Plan Recognition By Extending Bayesian Logic Programs

BLPs vs. RBNs

BLPs subsume RBNs In RBNs, each node in BN is a predicate and

probability formulae are used to specify probabilitiesCombining rules can be used to represent these

probability formulae in BLPs.

81

Page 82: Abductive Plan Recognition By Extending Bayesian Logic Programs

BALPs vs. BLPs

Like BLPs, BALPs use logic programs as templates for constructing Bayesian networks

Unlike BLPs, BALPs uses logical abduction instead of deduction to construct the network

82

Page 83: Abductive Plan Recognition By Extending Bayesian Logic Programs

Monroe [Blaylock and Allen, 2005]

TaskRecognize top level plans in an emergency response

domain (artificially generated using HTN planner)Plans include set-up-shelter, clear-road-wreck, provide-

medical-attentionSingle correct plan in each exampleDomain consists of several entities and sub-goalsTest the ability to scale to large domains

DataContains 1000 examples

83

Page 84: Abductive Plan Recognition By Extending Bayesian Logic Programs

Monroe

MethodologyKnowledge base constructed based on the domain knowledge

encoded in plannerLearn noisy-or parameters using EMCompute marginal probability for instances of top level plans

and pick the one with the highest marginal probabilitySystems compared

o BALPso MLN-HCAM [Singla and Mooney, 2011]

o Blaylock and Allen’s system [Blaylock and Allen, 2005]

Convergence score - measures the fraction of examples for which the plan schema was predicted correctly

84

Page 85: Abductive Plan Recognition By Extending Bayesian Logic Programs

Learning Results - Monroe

85

MW MW-Start Rand-Start

Conv Score 98.4 98.4 98.4

Acc-100 79.16 79.16 79.86

Acc-75 46.06 44.63 44.73

Acc-50 20.67 20.26 19.7

Acc-25 7.2 7.33 10.46

Page 86: Abductive Plan Recognition By Extending Bayesian Logic Programs

Linux [Blaylock and Allen, 2005]

TaskRecognize top level plans based on Linux commandsHuman users asked to perform tasks in Linux and commands

were recordedTop-level plans include find-file-by-ext, remove-file-by-ext,

copy-file, move-fileSingle correct plan in each exampleTests the ability to handle noise in data

o Users indicate success even when they have not achieved the task correctly

o Some top-level plans like find-file-by-ext and file-file-by-name have identical actions

DataContains 457 examples

86

Page 87: Abductive Plan Recognition By Extending Bayesian Logic Programs

Linux

MethodologyKnowledge base constructed based on the knowledge of Linux

commandsLearn noisy-or parameters using EMCompute marginal probability for instances of top level plans

and pick the one with the highest marginal probabilitySystems compared

o BALPso MLN-HCAM [Singla and Mooney, 2011]

o Blaylock and Allen’s system [Blaylock and Allen, 2005]

Convergence score - measures the fraction of examples for which the plan schema was predicted correctly o

87

Page 88: Abductive Plan Recognition By Extending Bayesian Logic Programs

Learning Results - Linux

88

Acc

ura

cy

Partial Observability

MW MW-Start Rand-Start

Conv Score 39.82 46.6 41.57

Page 89: Abductive Plan Recognition By Extending Bayesian Logic Programs

Story Understanding [Charniak and Goldman, 1991; Ng and Mooney, 1992]

TaskRecognize character’s top level plans based on actions

described in narrative textLogical representation of actions literals providedTop-level plans include shopping, robbing, restaurant

dining, partying Multiple top-level plans in each exampleTests the ability to predict multiple plans

Data25 development examples25 test examples

89

Page 90: Abductive Plan Recognition By Extending Bayesian Logic Programs

Story UnderstandingMethodology

Knowledge base constructed for ACCEL by Ng and Mooney [1992]

Insufficient number of examples to learn parameterso Noisy-or parameters set to 0.9o Noisy-and parameters set to 0.9o Priors tuned on development set

Compute MPE to get the best set of plansSystems compared

o BALPso MLN-HCAM [Singla and Mooney, 2011]

o ACCEL-Simplicity [Ng and Mooney, 1992]

o ACCEL-Coherence [Ng and Mooney, 1992]

– Specific for Story Understanding

90

Page 91: Abductive Plan Recognition By Extending Bayesian Logic Programs

Results on Story Understanding

91* - Differences are statistically significant wrt BALPs

Page 92: Abductive Plan Recognition By Extending Bayesian Logic Programs

Other Applications of BALPs

Medical diagnosisTextual entailmentComputational biology

Inferring gene relations based on the output of micro-array experiments

Any application that requires abductive reasoning

92

Page 93: Abductive Plan Recognition By Extending Bayesian Logic Programs

ACCEL[Ng and Mooney, 1992]

First-order logic based system for plan recognition

Simplicity metric selects explanations that have the least number of assumptions

Coherence metric selects explanations that connect maximum number of observationsMeasures explanatory coherenceSpecific to text interpretation

93

Page 94: Abductive Plan Recognition By Extending Bayesian Logic Programs

System by Blaylock and Allen[2005]

Statistical n-gram models to predict plans based on observed actions

Performs plan recognition in two phasesPredicts the plan schema firstPredicts arguments based on the predicted schema

94