Optimal Robust Output Containment of Unknown Heterogeneous Multiagent System Using Off-Policy Reinforcement Learning | IEEE Journals & Magazine | IEEE Xplore

Optimal Robust Output Containment of Unknown Heterogeneous Multiagent System Using Off-Policy Reinforcement Learning


Abstract:

This paper investigates optimal robust output containment problem of general linear heterogeneous multiagent systems (MAS) with completely unknown dynamics. A modelbased ...Show More

Abstract:

This paper investigates optimal robust output containment problem of general linear heterogeneous multiagent systems (MAS) with completely unknown dynamics. A modelbased algorithm using offline policy iteration (PI) is first developed, where the p-copy internal model principle is utilized to address the system parameter variations. This offline PI algorithm requires the nominal model of each agent, which may not be available in most real-world applications. To address this issue, a discounted performance function is introduced to express the optimal robust output containment problem as an optimal output-feedback design problem with bounded L2-gain. To solve this problem online in real time, a Bellman equation is first developed to evaluate a certain control policy and find the updated control policies, simultaneously, using only the state/output information measured online. Then, using this Bellman equation, a model-free off-policy integral reinforcement learning algorithm is proposed to solve the optimal robust output containment problem of heterogeneous MAS, in realtime, without requiring any knowledge of the system dynamics. Simulation results are provided to verify the effectiveness of the proposed method.
Published in: IEEE Transactions on Cybernetics ( Volume: 48, Issue: 11, November 2018)
Page(s): 3197 - 3207
Date of Publication: 30 October 2017

ISSN Information:

PubMed ID: 29989978

Funding Agency:


References

References is not available for this document.