File Download
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Adaptive vector quantization for reinforcement learning
Title | Adaptive vector quantization for reinforcement learning |
---|---|
Authors | |
Keywords | Learning algorithm intelligent control vector quantization robot navigation automated guided vehicles |
Issue Date | 2002 |
Publisher | International Federation of Automatic Control. |
Citation | The 15th World Congress of International Federation of Automatic Control (2002 IFAC), Barcelona, Spain, 21-26 July 2002 How to Cite? |
Abstract | Dynamic programming methods are capable of solving reinforcement learning
problems, in which an agent must improve its behavior through trial-and-error
interactions with a dynamic environment. However, these computational algorithms
suffer from the curse of dimensionality (Bellman, 1957) that the number of computational
operations increases exponentially with the cardinality of the state space. In practice, this
usually results in a very long training time and applications in continuous domain are far
from trivial. In order to ease this problem, we propose the use of vector quantization to
adaptively partition the state space based on the recent estimate of the action-value
function. In particular, this state-space partitioning operation is performed incrementally
to reflect the experience accumulated by the agent as it explores the underlying
environment. |
Persistent Identifier | http://hdl.handle.net/10722/100010 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lau, HYK | en_HK |
dc.contributor.author | Mak, KL | en_HK |
dc.contributor.author | Lee, SK | en_HK |
dc.date.accessioned | 2010-09-25T18:53:19Z | - |
dc.date.available | 2010-09-25T18:53:19Z | - |
dc.date.issued | 2002 | en_HK |
dc.identifier.citation | The 15th World Congress of International Federation of Automatic Control (2002 IFAC), Barcelona, Spain, 21-26 July 2002 | - |
dc.identifier.uri | http://hdl.handle.net/10722/100010 | - |
dc.description.abstract | Dynamic programming methods are capable of solving reinforcement learning problems, in which an agent must improve its behavior through trial-and-error interactions with a dynamic environment. However, these computational algorithms suffer from the curse of dimensionality (Bellman, 1957) that the number of computational operations increases exponentially with the cardinality of the state space. In practice, this usually results in a very long training time and applications in continuous domain are far from trivial. In order to ease this problem, we propose the use of vector quantization to adaptively partition the state space based on the recent estimate of the action-value function. In particular, this state-space partitioning operation is performed incrementally to reflect the experience accumulated by the agent as it explores the underlying environment. | - |
dc.language | eng | en_HK |
dc.publisher | International Federation of Automatic Control. | - |
dc.relation.ispartof | Proceedings of the 15th World Congress of International Federation of Automatic Control | en_HK |
dc.subject | Learning algorithm | - |
dc.subject | intelligent control | - |
dc.subject | vector quantization | - |
dc.subject | robot navigation | - |
dc.subject | automated guided vehicles | - |
dc.title | Adaptive vector quantization for reinforcement learning | en_HK |
dc.type | Conference_Paper | en_HK |
dc.identifier.email | Lau, HYK: hyklau@hkucc.hku.hk | en_HK |
dc.identifier.email | Mak, KL: makkl@hkucc.hku.hk | en_HK |
dc.identifier.authority | Lau, HYK=rp00137 | en_HK |
dc.identifier.authority | Mak, KL=rp00154 | en_HK |
dc.description.nature | link_to_OA_fulltext | - |
dc.identifier.hkuros | 81153 | en_HK |