File Download
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/IVS.2005.1505104
- Scopus: eid_2-s2.0-27944435493
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Double action Q-learning for obstacle avoidance in a dynamically changing environment
Title | Double action Q-learning for obstacle avoidance in a dynamically changing environment |
---|---|
Authors | |
Keywords | Obstacle avoidance Q-learning Reinforcement learning Temporal differences |
Issue Date | 2005 |
Publisher | IEEE. |
Citation | Ieee Intelligent Vehicles Symposium, Proceedings, 2005, v. 2005, p. 211-216 How to Cite? |
Abstract | In this paper, we propose a new method for solving the reinforcement learning problem in a dynamically changing environment, as in vehicle navigation, in which the Markov Decision Process used in traditional reinforcement learning is modified so that the response of the environment is taken into consideration for determining the agent's next state. This is achieved by changing the action-value function to handle three parameters at a time, namely, the current state, action taken by the agent, and action taken by the environment. As it considers the actions by the agent and environment, it is termed "Double Action". Based on the Q-learning method, the proposed method is implemented and the update rule is modified to handle all of the three parameters. Preliminary results show that the proposed method has the sum of rewards (negative) 89.5% less than that of the traditional method. Apart form that, our new method also has the total number of collisions and mean steps used in one episode 89.5% and 15.5% lower than that of the traditional method respectively. © 2005 IEEE. |
Persistent Identifier | http://hdl.handle.net/10722/45776 |
References |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ngai, DCK | en_HK |
dc.contributor.author | Yung, NHC | en_HK |
dc.date.accessioned | 2007-10-30T06:35:13Z | - |
dc.date.available | 2007-10-30T06:35:13Z | - |
dc.date.issued | 2005 | en_HK |
dc.identifier.citation | Ieee Intelligent Vehicles Symposium, Proceedings, 2005, v. 2005, p. 211-216 | en_HK |
dc.identifier.uri | http://hdl.handle.net/10722/45776 | - |
dc.description.abstract | In this paper, we propose a new method for solving the reinforcement learning problem in a dynamically changing environment, as in vehicle navigation, in which the Markov Decision Process used in traditional reinforcement learning is modified so that the response of the environment is taken into consideration for determining the agent's next state. This is achieved by changing the action-value function to handle three parameters at a time, namely, the current state, action taken by the agent, and action taken by the environment. As it considers the actions by the agent and environment, it is termed "Double Action". Based on the Q-learning method, the proposed method is implemented and the update rule is modified to handle all of the three parameters. Preliminary results show that the proposed method has the sum of rewards (negative) 89.5% less than that of the traditional method. Apart form that, our new method also has the total number of collisions and mean steps used in one episode 89.5% and 15.5% lower than that of the traditional method respectively. © 2005 IEEE. | en_HK |
dc.format.extent | 1037213 bytes | - |
dc.format.extent | 10863 bytes | - |
dc.format.mimetype | application/pdf | - |
dc.format.mimetype | text/plain | - |
dc.language | eng | en_HK |
dc.publisher | IEEE. | en_HK |
dc.relation.ispartof | IEEE Intelligent Vehicles Symposium, Proceedings | en_HK |
dc.rights | ©2005 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. | - |
dc.subject | Obstacle avoidance | en_HK |
dc.subject | Q-learning | en_HK |
dc.subject | Reinforcement learning | en_HK |
dc.subject | Temporal differences | en_HK |
dc.title | Double action Q-learning for obstacle avoidance in a dynamically changing environment | en_HK |
dc.type | Conference_Paper | en_HK |
dc.identifier.email | Yung, NHC:nyung@eee.hku.hk | en_HK |
dc.identifier.authority | Yung, NHC=rp00226 | en_HK |
dc.description.nature | published_or_final_version | en_HK |
dc.identifier.doi | 10.1109/IVS.2005.1505104 | en_HK |
dc.identifier.scopus | eid_2-s2.0-27944435493 | en_HK |
dc.identifier.hkuros | 102239 | - |
dc.relation.references | http://www.scopus.com/mlt/select.url?eid=2-s2.0-27944435493&selection=ref&src=s&origin=recordpage | en_HK |
dc.identifier.volume | 2005 | en_HK |
dc.identifier.spage | 211 | en_HK |
dc.identifier.epage | 216 | en_HK |
dc.identifier.scopusauthorid | Ngai, DCK=9332358900 | en_HK |
dc.identifier.scopusauthorid | Yung, NHC=7003473369 | en_HK |