会议专题

Asynchronous Deep Q-network in Continuous Environment Based on Prioritized Experience Replay

  Deep Q-network is a classical algorithm of reinforce learning,which is widely used and has many variants.The research content of this paper is to optimize and integrate some variant algorithms so that it has the advantage of running in the continuous environment,and improve the learning efficiency by Prioritized Experience Replay and multiple agents asynchronous parallel method,and establish the asynchronous Deep Q-network framework based on priority Experience Replay in the continuous environment.This paper uses some games in the Atari 2600 domain to test our algorithm framework,which achieved good results,improved stability,convergence speed and improved performance.

Deep Q-network Continuous Environment Prioritized Experience Replay Asynchronous

Hongda Liu Hanqi Zhang Linying Gong

College of Computer Science and Technology,Jilin University,Changchun 130012,China

国际会议

2019 2nd International Conference on Mechanical Engineering, Industrial Materials and Industrial Electronics (MEIMIE 2019)2019年第二届机械工程、工业材料和工业电子国际会议(Meimie 2019)

大连

英文

472-477

2019-03-29(万方平台首次上网日期,不代表论文的发表时间)