Deakin University
Browse

File(s) under permanent embargo

On the limitations of scalarisation for multi-objective reinforcement learning of pareto fronts

conference contribution
posted on 2008-12-01, 00:00 authored by P Vamplew, John YearwoodJohn Yearwood, Richard DazeleyRichard Dazeley, A Berry
Multiobjective reinforcement learning (MORL) extends RL to problems with multiple conflicting objectives. This paper argues for designing MORL systems to produce a set of solutions approximating the Pareto front, and shows that the common MORL technique of scalarisation has fundamental limitations when used to find Pareto-optimal policies. The work is supported by the presentation of three new MORL benchmarks with known Pareto fronts. © 2008 Springer Berlin Heidelberg.

History

Volume

5360 LNAI

Pagination

372-378

Location

Auckland, N.Z.

Start date

2008-12-01

End date

2008-12-05

ISSN

0302-9743

eISSN

1611-3349

ISBN-10

3540893776

Publication classification

EN.1 Other conference paper

Title of proceedings

21st Australasian Joint Conference on Artificial Intelligence

Publisher

Springer

Place of publication

Berlin, Germany

Usage metrics

    Research Publications

    Categories

    No categories selected

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC