Abstract
In this paper, we present different opposition schemes for four reinforcement learning methods: Q-learning, Q(λ), Sarsa, and Sarsa(λ) under assumptions that are reasonable for many real-world problems where type-II opposites generally better reflect the nature of the problem at hand. It appears that the aggregation of opposition-based schemes with regular learning methods can significantly speed up the learning process, especially where the number of observations is small or the state space is large. We verify the performance of the proposed methods using two different applications: a grid-world problem and a single water reservoir management problem.
Original language | English (US) |
---|---|
Pages (from-to) | 101-114 |
Number of pages | 14 |
Journal | Information Sciences |
Volume | 275 |
DOIs | |
State | Published - Aug 10 2014 |
Keywords
- Grid world
- Opposition-based learning
- Q-learning
- Reinforcement learning
- Reservoir management
- Sarsa
ASJC Scopus subject areas
- Software
- Control and Systems Engineering
- Theoretical Computer Science
- Computer Science Applications
- Information Systems and Management
- Artificial Intelligence