Por favor, use este identificador para citar o enlazar a este item: http://hdl.handle.net/10261/96699
COMPARTIR / EXPORTAR:
logo share SHARE BASE
Visualizar otros formatos: MARC | Dublin Core | RDF | ORE | MODS | METS | DIDL | DATACITE

Invitar a revisión por pares abierta
Campo DC Valor Lengua/Idioma
dc.contributor.authorAgostini, Alejandro-
dc.contributor.authorCelaya, Enric-
dc.date.accessioned2014-05-14T11:08:20Z-
dc.date.available2014-05-14T11:08:20Z-
dc.date.issued2011-
dc.identifierisbn: 978-1-57735-514-4-
dc.identifier.citationInternational Joint Conference on Artificial Intelligence 2: 1146-1151 (2011)-
dc.identifier.urihttp://hdl.handle.net/10261/96699-
dc.descriptionTrabajo presentado al 22nd IJCAI celebrado en Barcelona del 16 al 22 de julio de 2011.-
dc.description.abstractIn this work we propose an approach for generalization in continuous domain Reinforcement Learning that, instead of using a single function approximator, tries many different function approximators in parallel, each one defined in a different region of the domain. Associated with each approximator is a relevance function that locally quantifies the quality of its approximation, so that, at each input point, the approximator with highest relevance can be selected. The relevance function is defined using parametric estimations of the variance of the q-values and the density of samples in the input space, which are used to quantify the accuracy and the confidence in the approximation, respectively. These parametric estimations are obtained from a probability density distribution represented as a Gaussian Mixture Model embedded in the input-output space of each approximator. In our experiments, the proposed approach required a lesser number of experiences for learning and produced more stable convergence profiles than when using a single function approximator.-
dc.description.sponsorshipThis research was partially supported by Consolider Ingenio 2010, project CSD2007-00018.-
dc.publisherAAAI Press-
dc.relation.isversionofPostprint-
dc.rightsopenAccess-
dc.titleA competitive strategy for function approximation in Q-learning-
dc.typecomunicación de congreso-
dc.relation.publisherversionhttp://ijcai.org/papers11/contents.php-
dc.date.updated2014-05-14T11:08:20Z-
dc.description.versionPeer Reviewed-
dc.language.rfc3066eng-
dc.type.coarhttp://purl.org/coar/resource_type/c_5794es_ES
item.openairetypecomunicación de congreso-
item.grantfulltextopen-
item.cerifentitytypePublications-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
item.fulltextWith Fulltext-
Aparece en las colecciones: (IRII) Libros y partes de libros
Ficheros en este ítem:
Fichero Descripción Tamaño Formato
A competitive strategy.pdf335,34 kBUnknownVisualizar/Abrir
Show simple item record

CORE Recommender

Page view(s)

255
checked on 23-abr-2024

Download(s)

93
checked on 23-abr-2024

Google ScholarTM

Check


NOTA: Los ítems de Digital.CSIC están protegidos por copyright, con todos los derechos reservados, a menos que se indique lo contrario.