JUCS - Journal of Universal Computer Science 13(9): 1246-1269, doi: 10.3217/jucs-013-09-1246
Focus of Attention in Reinforcement Learning
expand article infoLihong Li, Vadim Bulitko§, Russell Greiner§
‡ Rutgers University, Piscataway, United States of America§ University of Alberta, Edmonton, Canada
Open Access
Abstract
Classification-based reinforcement learning (RL) methods have recently been pro-posed as an alternative to the traditional value-function based methods. These methods use a classifier to represent a policy, where the input (features) to the classifier is the state and theoutput (class label) for that state is the desired action. The reinforcement-learning community knows that focusing on more important states can lead to improved performance. In this paper,we investigate the idea of focused learning in the context of classification-based RL. Specifically, we define a useful notation of state importance, which we use to prove rigorous bounds on policyloss. Furthermore, we show that a classification-based RL agent may behave arbitrarily poorly if it treats all states as equally important.
Keywords
reinforcement learning, function approximation, generalization, attention