A Distributed Q-learning Classifier System for task decomposition in real robot learning problems
dc.contributor.author | Chapman, Kevin L. | en |
dc.contributor.committeechair | Bay, John S. | en |
dc.contributor.committeemember | Abbott, A. Lynn | en |
dc.contributor.committeemember | VanLandingham, Hugh F. | en |
dc.contributor.department | Electrical Engineering | en |
dc.date.accessioned | 2014-03-14T21:31:02Z | en |
dc.date.adate | 2009-03-04 | en |
dc.date.available | 2014-03-14T21:31:02Z | en |
dc.date.issued | 1996-09-05 | en |
dc.date.rdate | 2009-03-04 | en |
dc.date.sdate | 2009-03-04 | en |
dc.description.abstract | A distributed reinforcement-learning system is designed and implemented on a mobile robot for the study of complex task decomposition in real robot learning environments. The Distributed Q-learning Classifier System (DQLCS) is evolved from the standard Learning Classifier System (LCS) proposed by J.H. Holland. Two of the limitations of the standard LCS are its monolithic nature and its complex apportionment of credit scheme, the bucket brigade algorithm (BBA). The DQLCS addresses both of these problems as well as the inherent difficulties faced by learning systems operating in real environments. We introduce Q-learning as the apportionment of credit component of the DQLCS, and we develop a distributed learning architecture to facilitate complex task decomposition. Based upon dynamic programming, the Q-learning update equation is derived and its advantages over the complex BBA are discussed. The distributed architecture is implemented to provide for faster learning by allowing the system to effectively decrease the size of the problem space it must explore. Holistic and monolithic shaping approaches are used to distribute reward among the learning modules of the DQLCS in a variety of real robot learning experiments. The results of these experiments support the DQLCS as a useful reinforcement learning paradigm and suggest future areas of study in distributed learning systems. | en |
dc.description.degree | Master of Science | en |
dc.format.extent | ix, 110 leaves | en |
dc.format.medium | BTD | en |
dc.format.mimetype | application/pdf | en |
dc.identifier.other | etd-03042009-041449 | en |
dc.identifier.sourceurl | http://scholar.lib.vt.edu/theses/available/etd-03042009-041449/ | en |
dc.identifier.uri | http://hdl.handle.net/10919/41435 | en |
dc.language.iso | en | en |
dc.publisher | Virginia Tech | en |
dc.relation.haspart | LD5655.V855_1996.C437.pdf | en |
dc.relation.isformatof | OCLC# 36114106 | en |
dc.rights | In Copyright | en |
dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
dc.subject | Q-learning | en |
dc.subject | Learning Classifier Systems | en |
dc.subject | artificial intelligence | en |
dc.subject | mobile robots | en |
dc.subject | task decomposition | en |
dc.subject.lcc | LD5655.V855 1996.C437 | en |
dc.title | A Distributed Q-learning Classifier System for task decomposition in real robot learning problems | en |
dc.type | Thesis | en |
dc.type.dcmitype | Text | en |
thesis.degree.discipline | Electrical Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | masters | en |
thesis.degree.name | Master of Science | en |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- LD5655.V855_1996.C437.pdf
- Size:
- 12.21 MB
- Format:
- Adobe Portable Document Format
- Description: