Averaged Soft Actor-Critic for Deep Reinforcement Learning

Complexity 2021:1-16 (2021)
  Copy   BIBTEX

Abstract

With the advent of the era of artificial intelligence, deep reinforcement learning has achieved unprecedented success in high-dimensional and large-scale artificial intelligence tasks. However, the insecurity and instability of the DRL algorithm have an important impact on its performance. The Soft Actor-Critic algorithm uses advanced functions to update the policy and value network to alleviate some of these problems. However, SAC still has some problems. In order to reduce the error caused by the overestimation of SAC, we propose a new SAC algorithm called Averaged-SAC. By averaging the previously learned action-state estimates, it reduces the overestimation problem of soft Q-learning, thereby contributing to a more stable training process and improving performance. We evaluate the performance of Averaged-SAC through some games in the MuJoCo environment. The experimental results show that the Averaged-SAC algorithm effectively improves the performance of the SAC algorithm and the stability of the training process.

Links

PhilArchive



    Upload a copy of this work     Papers currently archived: 91,349

External links

Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Analytics

Added to PP
2021-04-02

Downloads
7 (#1,351,854)

6 months
5 (#652,053)

Historical graph of downloads
How can I increase my downloads?

Author Profiles

Jing Gao
Lan Zhou University
Li Peng
Abilene Christian University

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references