English   español  
Please use this identifier to cite or link to this item: http://hdl.handle.net/10261/30550
Share/Impact:
Statistics
logo share SHARE   Add this article to your Mendeley library MendeleyBASE
Visualizar otros formatos: MARC | Dublin Core | RDF | ORE | MODS | METS | DIDL
Exportar a otros formatos:
Title

Reinforcement learning for agents with many sensors and actuators acting in categorizable environments

AuthorsPorta, Josep M. ; Celaya, Enric
Issue Date2005
PublisherAssociation for the Advancement of Artificial Intelligence
CitationJournal of Artificial Intelligence Research 23: 79-122 (2005)
AbstractIn this paper, we confront the problem of applying reinforcement learning to agents that perceive the environment through many sensors and that can perform parallel actions using many actuators as is the case in complex autonomous robots. We argue that reinforcement learning can only be successfully applied to this case if strong assumptions are made on the characteristics of the environment in which the learning is performed, so that the relevant sensor readings and motor commands can be readily identified. The introduction of such assumptions leads to strongly-biased learning systems that can eventually lose the generality of traditional reinforcement-learning algorithms. In this line, we observe that, in realistic situations, the reward received by the robot depends only on a reduced subset of all the executed actions and that only a reduced subset of the sensor inputs (possibly different in each situation and for each action) are relevant to predict the reward. We formalize this property in the so called 'categorizability assumption' and we present an algorithm that takes advantage of the categorizability of the environment, allowing a decrease in the learning time with respect to existing reinforcement-learning algorithms. Results of the application of the algorithm to a couple of simulated realistic-robotic problems (landmark-based navigation and the six-legged robot gait generation) are reported to validate our approach and to compare it to existing flat and generalization-based reinforcement-learning approaches.
Publisher version (URL)https://www.jair.org/papers/paper1437.html
URIhttp://hdl.handle.net/10261/30550
ISSN1076-9757
Appears in Collections:(IRII) Artículos
Files in This Item:
File Description SizeFormat 
Reinforcement learning.pdf2,2 MBAdobe PDFThumbnail
View/Open
Show full item record
Review this work
 


WARNING: Items in Digital.CSIC are protected by copyright, with all rights reserved, unless otherwise indicated.