Learning to plan probabilistically from neural networks

Di erent from existing reinforcement learning algorithms that generate only reactive policies and existing probabilis tic planning algorithms that requires a substantial amount of a priori knowledge in order to plan we devise a two stage bottom up learning to plan process in which rst reinforce ment learning dynamic programming is applied without the use of a priori domain speci c knowledge to acquire a reactive policy and then explicit plans are extracted from the learned reactive policy Plan extraction is based on a beam search algorithm that performs temporal projection in a restricted fashion guided by the value functions re sulting from reinforcement learning dynamic programming Experiments and theoretical analysis are presented..
Keywords No keywords specified (fix it)
Categories (categorize this paper)
 Save to my reading list
Follow the author(s)
My bibliography
Export citation
Find it on Scholar
Edit this record
Mark as duplicate
Revision history
Request removal from index
Translate to english
Download options
Our Archive

Upload a copy of this paper     Check publisher's policy     Papers currently archived: 29,478
External links

Setup an account with your affiliations in order to access resources via your University's proxy server
Configure custom proxy (use this if your affiliation does not provide a proxy)
Through your library
References found in this work BETA

No references found.

Add more references

Citations of this work BETA

No citations found.

Add more citations

Similar books and articles
Added to PP index

Total downloads
4 ( #651,872 of 2,180,551 )

Recent downloads (6 months)
1 ( #302,815 of 2,180,551 )

How can I increase my downloads?

Monthly downloads
My notes
Sign in to use this feature

There  are no threads in this forum
Nothing in this forum yet.

Other forums