File(s) under permanent embargo
On the limitations of scalarisation for multi-objective reinforcement learning of pareto fronts
conference contribution
posted on 2008-12-01, 00:00 authored by P Vamplew, John YearwoodJohn Yearwood, Richard DazeleyRichard Dazeley, A BerryMultiobjective reinforcement learning (MORL) extends RL to problems with multiple conflicting objectives. This paper argues for designing MORL systems to produce a set of solutions approximating the Pareto front, and shows that the common MORL technique of scalarisation has fundamental limitations when used to find Pareto-optimal policies. The work is supported by the presentation of three new MORL benchmarks with known Pareto fronts. © 2008 Springer Berlin Heidelberg.
History
Volume
5360 LNAIPagination
372-378Location
Auckland, N.Z.Publisher DOI
Start date
2008-12-01End date
2008-12-05ISSN
0302-9743eISSN
1611-3349ISBN-10
3540893776Publication classification
EN.1 Other conference paperTitle of proceedings
21st Australasian Joint Conference on Artificial IntelligencePublisher
SpringerPlace of publication
Berlin, GermanyUsage metrics
Categories
No categories selectedKeywords
Licence
Exports
RefWorksRefWorks
BibTeXBibTeX
Ref. managerRef. manager
EndnoteEndnote
DataCiteDataCite
NLMNLM
DCDC