Synthetic Experiences for Accelerating DQN Performance in Discrete Non-Deterministic Environments

State-of-the-art Deep Reinforcement Learning Algorithms such as DQN and DDPG use the concept of a replay buffer called Experience Replay. The default usage contains only the experiences that have been gathered over the runtime. We propose a method called Interpolated Experience Replay that uses stor...

Full description

Bibliographic Details
Main Authors: Wenzel Pilar von Pilchau, Anthony Stein, Jörg Hähner
Format: Article
Language:English
Published: MDPI AG 2021-07-01
Series:Algorithms
Subjects:
Online Access:https://www.mdpi.com/1999-4893/14/8/226