• Sorted by Date • Classified by Publication Type • Sorted by First Author Last Name • Classified by Research Category •
Matthew Adams, Robert Loftin, Matthew E. Taylor, Michael Littman,
and David Roberts. An Empirical Analysis of RL's Drift From Its Behaviorism Roots. In Proceedings of the Adaptive
and Learning Agents workshop (at AAMAS-12), June 2012.
ALA-12
We present an empirical survey of reinforcement learning techniques and relate these techniques to concepts from behaviorism, a field of psychology concerned with the learning process. Specifically, we examine two standard RL algorithms, model-free SARSA, and model-based R-MAX, when used with various shaping techniques. We consider multiple techniques for incorporating shaping into these algorithms, including the use of options and potentialbased shaping. Findings indicate any improvement in sample complexity that results from shaping is limited at best. We suggest that this is either due to reinforcement learning not modeling behaviorism well, or behaviorism not modeling animal learning well. We further suggest that a paradigm shift in reinforcement learning techniques is required before the kind of learning performance that techniques from behaviorism indicate are possible can be realized.
@inproceedings(ALA12-Adams, author="Matthew Adams and Robert Loftin and Matthew~E.\ Taylor and Michael Littman and David Roberts", title="An Empirical Analysis of RL's Drift From Its Behaviorism Roots", Booktitle="Proceedings of the Adaptive and Learning Agents workshop (at AAMAS-12)", month="June", year= "2012", wwwnote={<a href="http://como.vub.ac.be/ALA2012/">ALA-12</a>}, abstract="We present an empirical survey of reinforcement learning techniques and relate these techniques to concepts from behaviorism, a field of psychology concerned with the learning process. Specifically, we examine two standard RL algorithms, model-free SARSA, and model-based R-MAX, when used with various shaping techniques. We consider multiple techniques for incorporating shaping into these algorithms, including the use of options and potentialbased shaping. Findings indicate any improvement in sample complexity that results from shaping is limited at best. We suggest that this is either due to reinforcement learning not modeling behaviorism well, or behaviorism not modeling animal learning well. We further suggest that a paradigm shift in reinforcement learning techniques is required before the kind of learning performance that techniques from behaviorism indicate are possible can be realized.", )
Generated by bib2html.pl (written by Patrick Riley ) on Thu Jul 24, 2014 16:09:11