|Abstract||Di erent from existing reinforcement learning algorithms that generate only reactive policies and existing probabilis tic planning algorithms that requires a substantial amount of a priori knowledge in order to plan we devise a two stage bottom up learning to plan process in which rst reinforce ment learning dynamic programming is applied without the use of a priori domain speci c knowledge to acquire a reactive policy and then explicit plans are extracted from the learned reactive policy Plan extraction is based on a beam search algorithm that performs temporal projection in a restricted fashion guided by the value functions re sulting from reinforcement learning dynamic programming Experiments and theoretical analysis are presented..|
|Keywords||No keywords specified (fix it)|
No categories specified
(categorize this paper)
|Through your library||Only published papers are available at libraries|
Similar books and articles
Ron Sun, Beyond Simple Rule Extraction: The Extraction of Planning Knowledge From Reinforcement Learners.
Enrico Blanzieri (1997). Dynamical Learning Algorithms for Neural Networks and Neural Constructivism. Behavioral and Brain Sciences 20 (4):559-559.
Edward Merrillb & Todd Petersonb (2001). From Implicit Skills to Explicit Knowledge: A Bottom‐Up Model of Skill Learning. Cognitive Science 25 (2):203-244.
Ron Sun, Todd Peterson & Edward Merrill, Bottom-Up Skill Learning in Reactive Sequential Decision Tasks.
Ron Sun (1997). Learning, Action, and Consciousness: A Hybrid Approach Toward Modeling Consciousness. Neural Networks 10:1317-33.
Sorry, there are not enough data points to plot this chart.
Added to index2009-06-13
Total downloads1 ( #291,771 of 722,867 )
Recent downloads (6 months)1 ( #60,917 of 722,867 )
How can I increase my downloads?