Loading web-font TeX/Math/Italic
Q-Learning for Feedback Nash Strategy of Finite-Horizon Nonzero-Sum Difference Games | IEEE Journals & Magazine | IEEE Xplore

Q-Learning for Feedback Nash Strategy of Finite-Horizon Nonzero-Sum Difference Games


Abstract:

In this article, we study the feedback Nash strategy of the model-free nonzero-sum difference game. The main contribution is to present the Q -learning algorithm for t...Show More

Abstract:

In this article, we study the feedback Nash strategy of the model-free nonzero-sum difference game. The main contribution is to present the Q -learning algorithm for the linear quadratic game without prior knowledge of the system model. It is noted that the studied game is in finite horizon which is novel to the learning algorithms in the literature which are mostly for the infinite-horizon Nash strategy. The key is to characterize the Q -factors in terms of the arbitrary control input and state information. A numerical example is given to verify the effectiveness of the proposed algorithm.
Published in: IEEE Transactions on Cybernetics ( Volume: 52, Issue: 9, September 2022)
Page(s): 9170 - 9178
Date of Publication: 12 March 2021

ISSN Information:

PubMed ID: 33710965

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.