๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

Tuning fuzzy PD and PI controllers using reinforcement learning

โœ Scribed by Hamid Boubertakh; Mohamed Tadjine; Pierre-Yves Glorennec; Salim Labiod


Publisher
Elsevier Science
Year
2010
Tongue
English
Weight
725 KB
Volume
49
Category
Article
ISSN
0019-0578

No coin nor oath required. For personal study only.

โœฆ Synopsis


In this paper, we propose a new auto-tuning fuzzy PD and PI controllers using reinforcement Q-learning (QL) algorithm for SISO (single-input single-output) and TITO (two-input two-output) systems. We first, investigate the design parameters and settings of a typical class of Fuzzy PD (FPD) and Fuzzy PI (FPI) controllers: zero-order Takagi-Sugeno controllers with equidistant triangular membership functions for inputs, equidistant singleton membership functions for output, Larsen's implication method, and average sum defuzzification method. Secondly, the analytical structures of these typical fuzzy PD and PI controllers are compared to their classical counterpart PD and PI controllers. Finally, the effectiveness of the proposed method is proven through simulation examples.


๐Ÿ“œ SIMILAR VOLUMES


A self-learning and tuning fuzzy logic c
โœ Hung-Yuan Chung; Chih-Kuan Chiang ๐Ÿ“‚ Article ๐Ÿ“… 1997 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 237 KB ๐Ÿ‘ 2 views

This article presents a new method for learning and tuning a fuzzy logic controller automatically. A reinforcement learning and a genetic algorithm are used in conjunction with a multilayer neural network model of a fuzzy logic controller, which can automatically generate the fuzzy control rules and

Learning fuzzy control by evolutionary a
โœ Munir-ul M. Chowdhury; Yun Li ๐Ÿ“‚ Article ๐Ÿ“… 1998 ๐Ÿ› John Wiley and Sons ๐ŸŒ English โš– 401 KB

In this paper, evolutionary and dynamic programming-based reinforcement learning techniques are combined to form an unsupervised learning scheme for designing autonomous optimal fuzzy logic control systems. A ''messy genetic algorithm'' and an ''advantage learning'' scheme are first compared as rein