Deakin University
Browse

File(s) under permanent embargo

On the recognition of abstract Markov policies

conference contribution
posted on 2000-01-01, 00:00 authored by H Bui, Svetha VenkateshSvetha Venkatesh, G West
Abstraction plays an essential role in the way the agents plan their behaviours, especially to reduce the computational complexity of planning in large domains. However, the effects of abstraction in the inverse process – plan recognition – are unclear. In this paper, we present a method for recognising the agent’s behaviour in noisy and uncertain domains, and across multiple levels of abstraction. We use the concept of abstract Markov policies in abstract probabilistic planning as the model of the agent’s behaviours and employ probabilistic inference in Dynamic Bayesian Networks (DBN) to infer the correct policy from a sequence of observations. When the states are fully observable, we show that for a broad and often-used class of abstract policies, the complexity of policy recognition scales well with the number of abstraction levels in the policy hierarchy. For the partially observable case, we derive an efficient hybrid inference scheme on the corresponding DBN to overcome the exponential complexity.

History

Event

National Conference on Artificial Intelligence (17th : 2000 : Austin, Tex.)

Pagination

524 - 530

Publisher

AAAI Press

Location

Austin, Tex.

Place of publication

Cambridge, Mass.

Start date

2000-07-30

End date

2000-08-03

ISBN-13

9780262511124

ISBN-10

0262511126

Language

eng

Publication classification

E1.1 Full written paper - refereed

Copyright notice

2000, AAAI

Title of proceedings

AAAI-2000 : Proceedings of the 17th National Conference on Artificial Intelligence