Abstract
Reinforcement learning is an optimisation technique for applications like control or scheduling problems. It is used in learning situations, where success and failure of the system are the only training information. Unfortunately, we have to pay a price for this powerful ability: long training times and the instability of the learning process are not tolerable for industrial applications with large continuous state spaces. From our point of view, the integration of prior knowledge is a key mechanism for making autonomous learning practicable for industrial applications. The learning control architecture Fynesse provides a unified view onto the integration of prior control knowledge in the reinforcement learning framework. In this way, other approaches in this area can be embedded into Fynesse. The key features of Fynesse are (1) the integration of prior control knowledge like linear controllers, control characteristics or fuzzy controllers, (2) autonomous learning of control strategies and (3) the interpretation of learned strategies in terms of fuzzy control rules. The benefits and problems of different methods for the integration of a priori knowledge are demonstrated on empirical studies.
Similar content being viewed by others
Author information
Authors and Affiliations
Corresponding author
Additional information
The research project F ynesse was supported by the Deutsche Forschungsgemeinschaft (DFG).
Rights and permissions
About this article
Cite this article
Schoknecht, R., Spott, M. & Riedmiller, M. Fynesse: An architecture for integrating prior knowledge in autonomously learning agents. Soft Computing 8, 397–408 (2004). https://doi.org/10.1007/s00500-003-0295-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00500-003-0295-x