Article ID | Journal | Published Year | Pages | File Type |
---|---|---|---|---|
5000079 | Automatica | 2017 | 9 Pages |
Abstract
In this paper, a model-free solution to the Hâ control of linear discrete-time systems is presented. The proposed approach employs off-policy reinforcement learning (RL) to solve the game algebraic Riccati equation online using measured data along the system trajectories. Like existing model-free RL algorithms, no knowledge of the system dynamics is required. However, the proposed method has two main advantages. First, the disturbance input does not need to be adjusted in a specific manner. This makes it more practical as the disturbance cannot be specified in most real-world applications. Second, there is no bias as a result of adding a probing noise to the control input to maintain persistence of excitation (PE) condition. Consequently, the convergence of the proposed algorithm is not affected by probing noise. An example of the Hâ control for an F-16 aircraft is given. It is seen that the convergence of the new off-policy RL algorithm is insensitive to probing noise.
Keywords
Related Topics
Physical Sciences and Engineering
Engineering
Control and Systems Engineering
Authors
Bahare Kiumarsi, Frank L. Lewis, Zhong-Ping Jiang,