File(s) under permanent embargo
Human feedback in continuous actor-critic reinforcement learning
conference contribution
posted on 2019-01-01, 00:00 authored by C Millán, B Fernandes, F Cruz Naranjo© 2019 ESANN (i6doc.com). All rights reserved. Reinforcement learning is utilized in contexts where an agent tries to learn from the environment. Using continuous actions, the performance may be improved in comparison to using discrete actions, however, this leads to excessive time to find a proper policy. In this work, we focus on including human feedback in reinforcement learning for a continuous action space. We unify the policy and the feedback to favor actions of low probability density. Furthermore, we compare the performance of the feedback for the continuous actor-critic algorithm and test our experiments in the cart-pole balancing task. The obtained results show that the proposed approach increases the accumulated reward in comparison to the autonomous learning method.
History
Pagination
661-666Location
Bruges, BelgiumStart date
2019-04-24End date
2019-04-26ISBN-13
9782875870650Language
engPublication classification
E1.1 Full written paper - refereedCopyright notice
2019, ESANNTitle of proceedings
ESANN 2019 : Proceedings, 27th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine LearningEvent
Artificial Neural Networks, Computational Intelligence and Machine Learning. European Symposium (27th : 2019 : Bruges, Belgium)Publisher
ESANNPlace of publication
[Bruges, Belgium]Usage metrics
Categories
No categories selectedLicence
Exports
RefWorksRefWorks
BibTeXBibTeX
Ref. managerRef. manager
EndnoteEndnote
DataCiteDataCite
NLMNLM
DCDC