Learning to plan probabilistically from neural networks

Abstract

Di erent from existing reinforcement learning algorithms that generate only reactive policies and existing probabilis tic planning algorithms that requires a substantial amount of a priori knowledge in order to plan we devise a two stage bottom up learning to plan process in which rst reinforce ment learning dynamic programming is applied without the use of a priori domain speci c knowledge to acquire a reactive policy and then explicit plans are extracted from the learned reactive policy Plan extraction is based on a beam search algorithm that performs temporal projection in a restricted fashion guided by the value functions re sulting from reinforcement learning dynamic programming Experiments and theoretical analysis are presented..

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 93,867

External links

  • This entry has no external links. Add one.
Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

  • Only published works are available at libraries.

Analytics

Added to PP
2009-06-13

Downloads
20 (#758,044)

6 months
20 (#173,532)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Ruowei Sun
University of Warwick

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references