Decentralized reinforcement learning applied to mobile robots
Artículo

Publication date
2017Metadata
Show full item record
Cómo citar
Leottau, David L.
Cómo citar
Decentralized reinforcement learning applied to mobile robots
Abstract
In this paper, decentralized reinforcement learning is applied to a control problem with a multidimensional action space. We propose a decentralized reinforcement learning architecture for a mobile robot, where the individual components of the commanded velocity vector are learned in parallel by separate agents. We empirically demonstrate that the decentralized architecture outperforms its centralized counterpart in terms of the learning time, while using less computational resources. The method is validated on two problems: an extended version of the 3-dimensional mountain car, and a ball-pushing behavior performed with a differential-drive robot, which is also tested on a physical setup.
Indexation
Artículo de publicación SCOPUS
Identifier
URI: https://repositorio.uchile.cl/handle/2250/169053
DOI: 10.1007/978-3-319-68792-6_31
ISSN: 16113349
03029743
Quote Item
Lecture Notes in Computer Science (LNCS, volume 9776), 2017
Collections