Please use this identifier to cite or link to this item: http://dspace.mediu.edu.my:8181/xmlui/handle/1721.1/6718
Full metadata record
DC FieldValueLanguage
dc.creatorMartin, Martin C.-
dc.date2004-10-08T20:38:57Z-
dc.date2004-10-08T20:38:57Z-
dc.date2003-05-01-
dc.date.accessioned2013-10-09T02:46:33Z-
dc.date.available2013-10-09T02:46:33Z-
dc.date.issued2013-10-09-
dc.identifierAIM-2003-014-
dc.identifierhttp://hdl.handle.net/1721.1/6718-
dc.identifier.urihttp://koha.mediu.edu.my:8181/xmlui/handle/1721-
dc.descriptionThis paper presents a novel algorithm for learning in a class of stochastic Markov decision processes (MDPs) with continuous state and action spaces that trades speed for accuracy. A transform of the stochastic MDP into a deterministic one is presented which captures the essence of the original dynamics, in a sense made precise. In this transformed MDP, the calculation of values is greatly simplified. The online algorithm estimates the model of the transformed MDP and simultaneously does policy search against it. Bounds on the error of this approximation are proven, and experimental results in a bicycle riding domain are presented. The algorithm learns near optimal policies in orders of magnitude fewer interactions with the stochastic MDP, using less domain knowledge. All code used in the experiments is available on the project's web site.-
dc.format12 p.-
dc.format1085830 bytes-
dc.format303781 bytes-
dc.formatapplication/postscript-
dc.formatapplication/pdf-
dc.languageen_US-
dc.relationAIM-2003-014-
dc.subjectAI-
dc.subjectReinforcement learning-
dc.subjectbicycle-
dc.subjectpolicy search-
dc.subjectmarkov decision processes-
dc.titleThe Essential Dynamics Algorithm: Essential Results-
Appears in Collections:MIT Items

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.