Please use this identifier to cite or link to this item:
http://dspace2020.uniten.edu.my:8080/handle/123456789/6522
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Mohan, Y. | |
dc.contributor.author | Ponnambalam, S.G. | |
dc.contributor.author | Inayat-Hussain, J.I. | |
dc.date.accessioned | 2017-12-08T09:47:14Z | - |
dc.date.available | 2017-12-08T09:47:14Z | - |
dc.date.issued | 2009 | |
dc.identifier.uri | http://dspace.uniten.edu.my/jspui/handle/123456789/6522 | - |
dc.description.abstract | Q-learning is a machine learning technique that learns what to do and how to map states to actions to maximize rewards. Q-learning has been applied to various tasks such as foraging, soccer and prey-pursuing robots. In this paper, a simple foraging task has been considered to study the influences of the policies reported in the open literatures. A mobile robot is used to search and retrieve pucks back to a home location. The goal of this study is to identify an efficient policy for q-learning which maximizes the number of pucks collected and minimizes the number of collisions in the environment. Policies namely greedy, epsilon-greedy, Boltzmann distribution and random search are used to study their performances in the foraging task and the results are presented. ©2009 IEEE. | |
dc.title | A comparative study of policies in Q-learning for foraging tasks | |
item.fulltext | No Fulltext | - |
item.grantfulltext | none | - |
Appears in Collections: | COE Scholarly Publication |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.