Skip to main content
Article
Robust Reinforcement Learning in Motion Planning
Advances in Neural Information Processing Systems 6 (NIPS 1993) (1993)
  • Satinder P. Singh
  • Andrew G. Barto
  • Roderic Grupen, University of Massachusetts - Amherst
  • Christopher Connolly
Abstract
While exploring to find better solutions, an agent performing on-line reinforcement learning (RL) can perform worse than is acceptable. In some cases, exploration might have unsafe, or even catastrophic, results, often modeled in terms of reaching 'failure' states of the agent's environment. This paper presents a method that uses domain knowledge to reduce the number of failures during exploration. This method formulates the set of actions from which the RL agent composes a control policy to ensure that exploration is conducted in a policy space that excludes most of the unacceptable policies. The resulting action set has a more abstract relationship to the task being solved than is common in many applications of RL. Although the cost of this added safety is that learning may result in a suboptimal solution, we argue that this is an appropriate tradeoff in many problems. We illustrate this method in the domain of motion planning.
Publication Date
1993
Publisher Statement
Publisher's version: http://papers.nips.cc/paper/843-robust-reinforcement-learning-in-motion-planning Harvested from Citeseer.
Citation Information
Satinder P. Singh, Andrew G. Barto, Roderic Grupen and Christopher Connolly. "Robust Reinforcement Learning in Motion Planning" Advances in Neural Information Processing Systems 6 (NIPS 1993) (1993)
Available at: http://works.bepress.com/roderic_grupen/11/