Por favor, use este identificador para citar o enlazar a este item: http://hdl.handle.net/10261/30153
COMPARTIR / EXPORTAR:
logo share SHARE BASE
Visualizar otros formatos: MARC | Dublin Core | RDF | ORE | MODS | METS | DIDL | DATACITE

Invitar a revisión por pares abierta
Campo DC Valor Lengua/Idioma
dc.contributor.authorCelaya, Enric-
dc.contributor.authorAgostini, Alejandro-
dc.date.accessioned2010-12-16T08:56:57Z-
dc.date.available2010-12-16T08:56:57Z-
dc.date.issued2010-
dc.identifier.citationProceedings of the 7th International Conference on Informatics in Control, Automation and Robotics: 160-168 (2010)-
dc.identifier.isbn978-989-8425-01-0-
dc.identifier.urihttp://hdl.handle.net/10261/30153-
dc.descriptionPresentado al ICINCO 2010 celebrado en Funchal (Portugal) del 15 al 18 de junio.-
dc.description.abstractThe successful application of Reinforcement Learning (RL) techniques to robot control is limited by the fact that, in most robotic tasks, the state and action spaces are continuous, multidimensional, and in essence, too large for conventional RL algorithms to work. The well known curse of dimensionality makes infeasible using a tabular representation of the value function, which is the classical approach that provides convergence guarantees. When a function approximation technique is used to generalize among similar states, the convergence of the algorithm is compromised, since updates unavoidably affect an extended region of the domain, that is, some situations are modified in a way that has not been really experienced, and the update may degrade the approximation. We propose a RL algorithm that uses a probability density estimation in the joint space of states, actions and Q-values as a means of function approximation. This allows us to devise an updating approach that, taking into account the local sampling density, avoids an excessive modification of the approximation far from the observed sample.-
dc.description.sponsorshipThis work was supported by the project 'CONSOLIDER-INGENIO 2010 Multimodal interaction in pattern recognition and computer vision' (V-00069). This research was partially supported by Consolider Ingenio 2010, project CSD2007-00018.-
dc.language.isoeng-
dc.publisherScience and technology publication-
dc.relation.isversionofPublisher's version-
dc.rightsopenAccess-
dc.subjectMachine learning-
dc.subjectReinforcement learning-
dc.titleReinforcement learning for robot control using probability density estimations-
dc.typecomunicación de congreso-
dc.description.peerreviewedPeer Reviewed-
dc.relation.publisherversionhttp://www.icinco.org/ICINCO2010/-
dc.type.coarhttp://purl.org/coar/resource_type/c_5794es_ES
item.languageiso639-1en-
item.fulltextWith Fulltext-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
item.cerifentitytypePublications-
item.grantfulltextopen-
item.openairetypecomunicación de congreso-
Aparece en las colecciones: (IRII) Comunicaciones congresos
Ficheros en este ítem:
Fichero Descripción Tamaño Formato
Reinforcement learning.pdf577,69 kBAdobe PDFVista previa
Visualizar/Abrir
Show simple item record

CORE Recommender

Page view(s)

371
checked on 24-abr-2024

Download(s)

297
checked on 24-abr-2024

Google ScholarTM

Check

Altmetric


NOTA: Los ítems de Digital.CSIC están protegidos por copyright, con todos los derechos reservados, a menos que se indique lo contrario.