Matthew E. Taylor's Publications

Sorted by DateClassified by Publication TypeSorted by First Author Last NameClassified by Research Category

An Empirical Analysis of RL's Drift From Its Behaviorism Roots

Matthew Adams, Robert Loftin, Matthew E. Taylor, Michael Littman, and David Roberts. An Empirical Analysis of RL's Drift From Its Behaviorism Roots. In Proceedings of the Adaptive and Learning Agents workshop (at AAMAS-12), June 2012.
ALA-12

Download

[PDF]338.4kB  

Abstract

We present an empirical survey of reinforcement learning techniques and relate these techniques to concepts from behaviorism, a field of psychology concerned with the learning process. Specifically, we examine two standard RL algorithms, model-free SARSA, and model-based R-MAX, when used with various shaping techniques. We consider multiple techniques for incorporating shaping into these algorithms, including the use of options and potentialbased shaping. Findings indicate any improvement in sample complexity that results from shaping is limited at best. We suggest that this is either due to reinforcement learning not modeling behaviorism well, or behaviorism not modeling animal learning well. We further suggest that a paradigm shift in reinforcement learning techniques is required before the kind of learning performance that techniques from behaviorism indicate are possible can be realized.

BibTeX Entry

@inproceedings(ALA12-Adams,
  author="Matthew Adams and Robert Loftin and Matthew~E.\ Taylor and Michael Littman and David Roberts",
  title="An Empirical Analysis of RL's Drift From Its Behaviorism Roots",
  Booktitle="Proceedings of the Adaptive and Learning Agents workshop (at AAMAS-12)",
  month="June",
  year= "2012",
  wwwnote={<a href="http://como.vub.ac.be/ALA2012/">ALA-12</a>},
 abstract="We present an empirical survey of reinforcement learning techniques
 and relate these techniques to concepts from behaviorism,
 a field of psychology concerned with the learning process. Specifically,
 we examine two standard RL algorithms, model-free SARSA,
 and model-based R-MAX, when used with various shaping techniques.
 We consider multiple techniques for incorporating shaping
 into these algorithms, including the use of options and potentialbased
 shaping. Findings indicate any improvement in sample complexity
 that results from shaping is limited at best. We suggest
 that this is either due to reinforcement learning not modeling behaviorism
 well, or behaviorism not modeling animal learning well.
 We further suggest that a paradigm shift in reinforcement learning
 techniques is required before the kind of learning performance that
 techniques from behaviorism indicate are possible can be realized.",
)

Generated by bib2html.pl (written by Patrick Riley ) on Thu Jul 24, 2014 16:09:11