Paper
18 February 2002 SAN-RL: combining spreading activation networks and reinforcement learning to learn configurable behaviors
Daniel M. Gaines, Don Mitchell Wilkes, Kanok Kusumalnukool, Siripun Thongchai, Kazuhiko Kawamura, John H. White
Author Affiliations +
Proceedings Volume 4573, Mobile Robots XVI; (2002) https://doi.org/10.1117/12.457458
Event: Intelligent Systems and Advanced Manufacturing, 2001, Boston, MA, United States
Abstract
Reinforcement learning techniques have been successful in allowing an agent to learn a policy for achieving tasks. The overall behavior of the agent can be controlled with an appropriate reward function. However, the policy that is learned will be fixed to this reward function. If the user wishes to change his or her preference about how the task is achieved the agent must be retrained with this new reward function. We address this challenge by combining Spreading Activation Networks and Reinforcement Learning in an approach we call SAN-RL. This approach provides the agent with a causal structure, the spreading activation network, relating goals to the actions that can achieve those goals. This enables the agent to select actions relative to the goal priorities. We combine this with reinforcement learning to enable the agent to learn a policy. Together, these approaches enable the learning of a configurable behaviors, a policy that can be adapted to meet the current preferences. We compare the approach with Q-learning on a robot navigation task. We demonstrate that SAN-RL exhibits goal-directed behavior before learning, exploits the causal structure of the network to focus its search during learning and results in configurable behaviors after learning.
© (2002) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Daniel M. Gaines, Don Mitchell Wilkes, Kanok Kusumalnukool, Siripun Thongchai, Kazuhiko Kawamura, and John H. White "SAN-RL: combining spreading activation networks and reinforcement learning to learn configurable behaviors", Proc. SPIE 4573, Mobile Robots XVI, (18 February 2002); https://doi.org/10.1117/12.457458
Lens.org Logo
CITATIONS
Cited by 2 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Curium

Sensors

Mobile robots

Defense and security

Detection and tracking algorithms

Robotics

Computer engineering

Back to Top