Enhanced Off-Policy Reinforcement Learning With Focused Experience Replay
Utilizing the collected experience tuples in the replay buffer (RB) is the primary way of exploiting the experiences in the off-policy reinforcement learning (RL) algorithms, and, therefore, the sampling scheme for the experience tuples in the RB can be critical for experience utilization. In this p...
Main Authors: | Seung-Hyun Kong, I. Made Aswin Nahrendra, Dong-Hee Paek |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9444458/ |
Similar Items
-
THE FUTURE, THE CRISIS, AND THE FUTURE OF REPLAY STORY
by: Eleonora Teresa Imbierowicz
Published: (2021-06-01) -
On Event Reproduction Ratio in Stateless and Stateful Replay of Real-World Traffic
by: Ying-Dar Lin, et al.
Published: (2013-09-01) -
Trajectory Based Prioritized Double Experience Buffer for Sample-Efficient Policy Optimization
by: Shengxiang Li, et al.
Published: (2021-01-01) -
Replay Debugger for Human Interactive Multiple Threaded Android Applications
Published: (2012) -
Experience Replay Using Transition Sequences
by: Thommen George Karimpanal, et al.
Published: (2018-06-01)