Asynchronous Deep Q-network in Continuous Environment Based on Prioritized Experience Replay
Deep Q-network is a classical algorithm of reinforce learning,which is widely used and has many variants.The research content of this paper is to optimize and integrate some variant algorithms so that it has the advantage of running in the continuous environment,and improve the learning efficiency by Prioritized Experience Replay and multiple agents asynchronous parallel method,and establish the asynchronous Deep Q-network framework based on priority Experience Replay in the continuous environment.This paper uses some games in the Atari 2600 domain to test our algorithm framework,which achieved good results,improved stability,convergence speed and improved performance.
Deep Q-network Continuous Environment Prioritized Experience Replay Asynchronous
Hongda Liu Hanqi Zhang Linying Gong
College of Computer Science and Technology,Jilin University,Changchun 130012,China
国际会议
大连
英文
472-477
2019-03-29(万方平台首次上网日期,不代表论文的发表时间)