Weighted dueling double deep Q-network
Author:
Clc Number:

TP273

  • Article
  • | |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • | |
  • Comments
    Abstract:

    In deep reinforcement learning, the deep Q-network algorithm seriously overestimates the action value, which degrades the performance of agents.The double deep Q-network and dueling network structure can partially alleviate the impact of overestimation, sometimes the former one even underestimate the action value.Here, a Weighted Dueling Double Deep Q-Network (WD3QN) algorithm is proposed, in which the improved double estimators and dueling network structure are combined into the deep Q-network, and the learned possible action values are weighted to produce the final action value, which can effectively reduce the estimation error.Finally, the algorithm is applied to the classical CartPole control problem on the open AI Gym platform.The simulation results show that compared with other existing algorithms, the proposed algorithm has better learning effect, convergence and training speed.

    Reference
    Related
    Cited by
Get Citation

WANG Chenxi, ZHAO Xueyan, GUO Xin. Weighted dueling double deep Q-network[J]. Journal of Nanjing University of Information Science & Technology,2021,13(5):564-570

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:September 13,2021
  • Online: December 02,2021
Article QR Code

Address:No. 219, Ningliu Road, Nanjing, Jiangsu Province

Postcode:210044

Phone:025-58731025