Active exploratory Q-learning for large problems

Xianghai Wu, Jonathan Kofman, Hamid R. Tizhoosh

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Although reinforcement learning (RL) emerged more than a decade ago, it is still under extensive investigation in application to large problems, where the states and actions are multi-dimensional and continuous and result in the so-called curse of dimensionality. Conventional RL methods are still not efficient enough in huge state-action spaces, while value-function generalization-based approaches require a very large number of good training examples. This paper presents an active exploratory approach to address the challenge of RL in large problems. The core principle of this approach is that the agent does not rush to the next state. Instead, it attempts a number of actions at the current state first, and then selects the action that returns the greatest immediate reward. The state resulting from performing the action is considered as the next state. Four active exploration algorithms for good actions are proposed: random-based search, opposition-based random search, search by cyclical adjustment, and opposition-based cyclical adjustment of each action dimension. The efficiency of these algorithms is determined by a visual-servoing experiment with a 6-axis robot.

Original languageEnglish (US)
Title of host publication2007 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2007
Pages4040-4045
Number of pages6
DOIs
StatePublished - 2007
Event2007 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2007 - Montreal, QC, Canada
Duration: Oct 7 2007Oct 10 2007

Publication series

NameConference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
ISSN (Print)1062-922X

Conference

Conference2007 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2007
Country/TerritoryCanada
CityMontreal, QC
Period10/7/0710/10/07

ASJC Scopus subject areas

  • General Engineering

Fingerprint

Dive into the research topics of 'Active exploratory Q-learning for large problems'. Together they form a unique fingerprint.

Cite this