𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Constructive reinforcement learning

✍ Scribed by Jose Hernandez-Orallo


Publisher
John Wiley and Sons
Year
2000
Tongue
English
Weight
163 KB
Volume
15
Category
Article
ISSN
0884-8173

No coin nor oath required. For personal study only.

✦ Synopsis


This paper presents an operative measure of reinforcement for constructive learning Ε½ . methods, i.e., eager learning methods using highly expressible or universal representation languages. These evaluation tools allow a further insight in the study of the growth of knowledge, theory revision, and abduction. The final approach is based on an apportionment of credit wrt the ''course'' that the evidence makes through the learned theory. Our measure of reinforcement is shown to be justified by cross-validation and by the connection with other successful evaluation criteria, like the minimum description length principle. Finally, the relation with the classical view of reinforcement is studied, where the actions of an intelligent system can be rewarded or penalized, and we discuss whether this should affect our distribution of reinforcement. The most important result of this paper is that the way we distribute reinforcement into knowledge results in a rated ontology, instead of a single prior distribution. Therefore, this detailed information can be exploited for guiding the space search of inductive learning algorithms. Likewise, knowledge revision may be done to the part of the theory which is not justified by the evidence.


πŸ“œ SIMILAR VOLUMES


Distributed reinforcement learning
✍ Gerhard WeiΞ² πŸ“‚ Article πŸ“… 1995 πŸ› Elsevier Science 🌐 English βš– 623 KB
Networked reinforcement learning
✍ Makito Oku; Kazuyuki Aihara πŸ“‚ Article πŸ“… 2008 πŸ› Springer Japan 🌐 English βš– 433 KB
Two steps reinforcement learning
✍ Fernando FernΓ‘ndez; Daniel Borrajo πŸ“‚ Article πŸ“… 2008 πŸ› John Wiley and Sons 🌐 English βš– 977 KB

When applying reinforcement learning in domains with very large or continuous state spaces, the experience obtained by the learning agent in the interaction with the environment must be generalized. The generalization methods are usually based on the approximation of the value functions used to comp