Advanced Search
    Chen Hongming, Liu Quan, Yan Yan, He Bin, Jiang Yubin, Zhang Linlin. An Experience-Guided Deep Deterministic Actor-Critic Algorithm with Multi-Actor[J]. Journal of Computer Research and Development, 2019, 56(8): 1708-1720. DOI: 10.7544/issn1000-1239.2019.20190155
    Citation: Chen Hongming, Liu Quan, Yan Yan, He Bin, Jiang Yubin, Zhang Linlin. An Experience-Guided Deep Deterministic Actor-Critic Algorithm with Multi-Actor[J]. Journal of Computer Research and Development, 2019, 56(8): 1708-1720. DOI: 10.7544/issn1000-1239.2019.20190155

    An Experience-Guided Deep Deterministic Actor-Critic Algorithm with Multi-Actor

    • The continuous control task has always been an important research direction in reinforce-ment learning. In recent years, the development of deep learning (DL) and the advent of deterministic policy gradients algorithm (DPG), provide many good ideas for solving continuous control problems. The main difficulty faced by these methods is the exploration in the continuous action space. And some of them engage in exploratory behavior through external noise injection in the action space. However, this exploration method does not perform well in some continuous control tasks. This paper proposes an experience-guided deep deterministic actor-critic algorithm with multi-actor (EGDDAC-MA) without external noise, which learns a guiding network from excellent experiences to guide the updates of the actor network and the critic network. Besides, it uses a multi-actor actor-critic (AC) model which configures different actors for each phase in an episode. These actors are independent of each other and do not interfere with each other. Finally, the experimental results show that compared with DDPG, TRPO and PPO algorithms, the proposed algorithm has better performance in most continuous tasks in GYM simulation platform.
    • loading

    Catalog

      Turn off MathJax
      Article Contents

      /

      DownLoad:  Full-Size Img  PowerPoint
      Return
      Return