File Download
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TSMCB.2003.808179
- Scopus: eid_2-s2.0-0037278069
- WOS: WOS:000180639100002
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance
Title | A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance |
---|---|
Authors | |
Keywords | Fuzzy system Obstacle avoidance Reinforcement learning Supervised learning Virtual environment (VE) |
Issue Date | 2003 |
Publisher | IEEE. |
Citation | Ieee Transactions On Systems, Man, And Cybernetics, Part B: Cybernetics, 2003, v. 33 n. 1, p. 17-27 How to Cite? |
Abstract | Fuzzy logic system promises an efficient way for obstacle avoidance. However, it is difficult to maintain the correctness, consistency, and completeness of a fuzzy rule base constructed and tuned by a human expert. Reinforcement learning method is capable of learning the fuzzy rules automatically. However, it incurs heavy learning phase and may result in an insufficiently learned rule base due to the curse of dimensionality. In this paper, we propose a neural fuzzy system with mixed coarse learning and fine learning phases. In the first phase, supervised learning method is used to determine the membership functions for the input and output variables simultaneously. After sufficient training, fine learning is applied which employs reinforcement learning algorithm to fine-tune the membership functions for the output variables. For sufficient learning, a new learning method using modified Sutton and Barto's model is proposed to strengthen the exploration. Through this two-step tuning approach, the mobile robot is able to perform collision-free navigation. To deal with the difficulty in acquiring large amount of training data with high consistency for the supervised learning, we develop a virtual environment (VE) simulator, which is able to provide desktop virtual environment (DVE) and immersive virtual environment (IVE) visualization. Through operating a mobile robot in the virtual environment (DVE/IVE) by a skilled human operator, the training data are readily obtained and used to train the neural fuzzy system. |
Persistent Identifier | http://hdl.handle.net/10722/42924 |
ISSN | 2014 Impact Factor: 6.220 |
ISI Accession Number ID | |
References |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ye, C | en_HK |
dc.contributor.author | Yung, NHC | en_HK |
dc.contributor.author | Wang, D | en_HK |
dc.date.accessioned | 2007-03-23T04:34:50Z | - |
dc.date.available | 2007-03-23T04:34:50Z | - |
dc.date.issued | 2003 | en_HK |
dc.identifier.citation | Ieee Transactions On Systems, Man, And Cybernetics, Part B: Cybernetics, 2003, v. 33 n. 1, p. 17-27 | en_HK |
dc.identifier.issn | 1083-4419 | en_HK |
dc.identifier.uri | http://hdl.handle.net/10722/42924 | - |
dc.description.abstract | Fuzzy logic system promises an efficient way for obstacle avoidance. However, it is difficult to maintain the correctness, consistency, and completeness of a fuzzy rule base constructed and tuned by a human expert. Reinforcement learning method is capable of learning the fuzzy rules automatically. However, it incurs heavy learning phase and may result in an insufficiently learned rule base due to the curse of dimensionality. In this paper, we propose a neural fuzzy system with mixed coarse learning and fine learning phases. In the first phase, supervised learning method is used to determine the membership functions for the input and output variables simultaneously. After sufficient training, fine learning is applied which employs reinforcement learning algorithm to fine-tune the membership functions for the output variables. For sufficient learning, a new learning method using modified Sutton and Barto's model is proposed to strengthen the exploration. Through this two-step tuning approach, the mobile robot is able to perform collision-free navigation. To deal with the difficulty in acquiring large amount of training data with high consistency for the supervised learning, we develop a virtual environment (VE) simulator, which is able to provide desktop virtual environment (DVE) and immersive virtual environment (IVE) visualization. Through operating a mobile robot in the virtual environment (DVE/IVE) by a skilled human operator, the training data are readily obtained and used to train the neural fuzzy system. | en_HK |
dc.format.extent | 1020341 bytes | - |
dc.format.extent | 5183 bytes | - |
dc.format.mimetype | application/pdf | - |
dc.format.mimetype | text/plain | - |
dc.language | eng | en_HK |
dc.publisher | IEEE. | en_HK |
dc.relation.ispartof | IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics | en_HK |
dc.rights | ©2003 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. | - |
dc.subject | Fuzzy system | en_HK |
dc.subject | Obstacle avoidance | en_HK |
dc.subject | Reinforcement learning | en_HK |
dc.subject | Supervised learning | en_HK |
dc.subject | Virtual environment (VE) | en_HK |
dc.title | A fuzzy controller with supervised learning assisted reinforcement learning algorithm for obstacle avoidance | en_HK |
dc.type | Article | en_HK |
dc.identifier.openurl | http://library.hku.hk:4550/resserv?sid=HKU:IR&issn=1083-4419&volume=33&spage=17&epage=27&date=2003&atitle=A+fuzzy+controller+with+supervised+learning+assisted+reinforcement+learning+algorithm+for+obstacle+avoidance | en_HK |
dc.identifier.email | Yung, NHC:nyung@eee.hku.hk | en_HK |
dc.identifier.authority | Yung, NHC=rp00226 | en_HK |
dc.description.nature | published_or_final_version | en_HK |
dc.identifier.doi | 10.1109/TSMCB.2003.808179 | en_HK |
dc.identifier.scopus | eid_2-s2.0-0037278069 | en_HK |
dc.identifier.hkuros | 81205 | - |
dc.relation.references | http://www.scopus.com/mlt/select.url?eid=2-s2.0-0037278069&selection=ref&src=s&origin=recordpage | en_HK |
dc.identifier.volume | 33 | en_HK |
dc.identifier.issue | 1 | en_HK |
dc.identifier.spage | 17 | en_HK |
dc.identifier.epage | 27 | en_HK |
dc.identifier.isi | WOS:000180639100002 | - |
dc.publisher.place | United States | en_HK |
dc.identifier.scopusauthorid | Ye, C=7202201245 | en_HK |
dc.identifier.scopusauthorid | Yung, NHC=7003473369 | en_HK |
dc.identifier.scopusauthorid | Wang, D=7407077210 | en_HK |
dc.identifier.issnl | 1083-4419 | - |