David Bourget (Western Ontario)
David Chalmers (ANU, NYU)
Rafael De Clercq
Jack Alan Reynolds
Learn more about PhilPapers
The paper presents an approach for developing multi-agent reinforcement learning systems that are made up of a coalition of modular agents. We focus on learning to segment sequences (sequential decision tasks) to create modular structures, through a bidding process that is based on reinforcements received during task execution. The approach segments sequences (and divides them up among agents) to facilitate the learning of the overall task. Notably, our approach does not rely on a priori knowledge or a priori structures. Initial experiments demonstrated the basic promise of the approach. This work shows how bidding and reinforcement learning can be usefully combined, thus pointing to a new research direction
|Keywords||No keywords specified (fix it)|
|Categories||categorize this paper)|
|Through your library||Only published papers are available at libraries|
Similar books and articles
Karl Tuyls, Ann Nowe, Tom Lenaerts & Bernard Manderick (2004). An Evolutionary Game Theoretic Perspective on Learning in Multi-Agent Systems. Synthese 139 (2):297 - 330.
Ron Sun, Todd Peterson & Edward Merrill, Bottom-Up Skill Learning in Reactive Sequential Decision Tasks.
Mikhail N. Zhadin (2000). LTP and Reinforcement: Possible Role of the Monoaminergic Systems. Behavioral and Brain Sciences 23 (2):287-288.
Roland Mühlenbernd (2011). Learning with Neighbours. Synthese 183 (S1):87-109.
Reiko Yakushijin & Robert A. Jacobs (2011). Are People Successful at Learning Sequences of Actions on a Perceptual Matching Task? Cognitive Science 35 (5):939-962.
Added to index2012-09-05
Total downloads6 ( #160,134 of 1,005,066 )
Recent downloads (6 months)1 ( #64,743 of 1,005,066 )
How can I increase my downloads?