Selector-Actor-Critic and Tuner-Actor-Critic Algorithms for Reinforcement Learning | IEEE Conference Publication | IEEE Xplore

Selector-Actor-Critic and Tuner-Actor-Critic Algorithms for Reinforcement Learning


Abstract:

This work presents two reinforcement learning (RL) architectures, which mimic rational humans in the way of analyzing the available information and making decisions. The ...Show More

Abstract:

This work presents two reinforcement learning (RL) architectures, which mimic rational humans in the way of analyzing the available information and making decisions. The proposed algorithms are called selector-actor-critic (SAC) and tuner-actor-critic (TAC). They are obtained by modifying the well known actor-critic (AC) algorithm. SAC is equipped with an actor, a critic, and a selector. The role of the selector is to determine the most promising action at the current state based on the last estimate from the critic. TAC is model based, and consists of a tuner, a model-learner, an actor, and a critic. After receiving the approximated value of the current state-action pair from the critic and the learned model from the model-learner, the tuner uses the Bellman equation to tune the value of the current state-action pair. Then, this tuned value is used by the actor to optimize the policy. We investigate the performance of the proposed algorithms, and compare with AC algorithm to show the advantages of the proposed algorithms using numerical simulations.
Date of Conference: 23-25 October 2019
Date Added to IEEE Xplore: 08 December 2019
ISBN Information:

ISSN Information:

Conference Location: Xi'an, China

References

References is not available for this document.