Deakin University
Browse

On the recognition of abstract Markov policies

conference contribution
posted on 2000-01-01, 00:00 authored by H Bui, Svetha VenkateshSvetha Venkatesh, G West
Abstraction plays an essential role in the way the agents plan their behaviours, especially to reduce the computational complexity of planning in large domains. However, the effects of abstraction in the inverse process – plan recognition – are unclear. In this paper, we present a method for recognising the agent’s behaviour in noisy and uncertain domains, and across multiple levels of abstraction. We use the concept of abstract Markov policies in abstract probabilistic planning as the model of the agent’s behaviours and employ probabilistic inference in Dynamic Bayesian Networks (DBN) to infer the correct policy from a sequence of observations. When the states are fully observable, we show that for a broad and often-used class of abstract policies, the complexity of policy recognition scales well with the number of abstraction levels in the policy hierarchy. For the partially observable case, we derive an efficient hybrid inference scheme on the corresponding DBN to overcome the exponential complexity.

History

Pagination

524-530

Location

Austin, Tex.

Start date

2000-07-30

End date

2000-08-03

ISBN-13

9780262511124

ISBN-10

0262511126

Language

eng

Publication classification

E1.1 Full written paper - refereed

Copyright notice

2000, AAAI

Title of proceedings

AAAI-2000 : Proceedings of the 17th National Conference on Artificial Intelligence

Event

National Conference on Artificial Intelligence (17th : 2000 : Austin, Tex.)

Publisher

AAAI Press

Place of publication

Cambridge, Mass.

Usage metrics

    Research Publications

    Categories

    No categories selected

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC