Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments
In this paper, we propose a deep deterministic policy gradient (DDPG)-based path-planning method for mobile robots by applying the hindsight experience replay (HER) technique to overcome the performance degradation resulting from sparse reward problems occurring in autonomous driving mobile robots....
| 出版年: | Sensors |
|---|---|
| 主要な著者: | , , , |
| フォーマット: | 論文 |
| 言語: | 英語 |
| 出版事項: |
MDPI AG
2022-12-01
|
| 主題: | |
| オンライン・アクセス: | https://www.mdpi.com/1424-8220/22/24/9574 |
| _version_ | 1851862593808891904 |
|---|---|
| author | Minjae Park Seok Young Lee Jin Seok Hong Nam Kyu Kwon |
| author_facet | Minjae Park Seok Young Lee Jin Seok Hong Nam Kyu Kwon |
| author_sort | Minjae Park |
| collection | DOAJ |
| container_title | Sensors |
| description | In this paper, we propose a deep deterministic policy gradient (DDPG)-based path-planning method for mobile robots by applying the hindsight experience replay (HER) technique to overcome the performance degradation resulting from sparse reward problems occurring in autonomous driving mobile robots. The mobile robot in our analysis was a robot operating system-based TurtleBot3, and the experimental environment was a virtual simulation based on Gazebo. A fully connected neural network was used as the DDPG network based on the actor–critic architecture. Noise was added to the actor network. The robot recognized an unknown environment by measuring distances using a laser sensor and determined the optimized policy to reach its destination. The HER technique improved the learning performance by generating three new episodes with normal experience from a failed episode. The proposed method demonstrated that the HER technique could help mitigate the sparse reward problem; this was further corroborated by the successful autonomous driving results obtained after applying the proposed method to two reward systems, as well as actual experimental results. |
| format | Article |
| id | doaj-art-6e8a731b9acf414a8cb2c0803ebfe0cf |
| institution | Directory of Open Access Journals |
| issn | 1424-8220 |
| language | English |
| publishDate | 2022-12-01 |
| publisher | MDPI AG |
| record_format | Article |
| spelling | doaj-art-6e8a731b9acf414a8cb2c0803ebfe0cf2025-08-19T22:20:03ZengMDPI AGSensors1424-82202022-12-012224957410.3390/s22249574Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward EnvironmentsMinjae Park0Seok Young Lee1Jin Seok Hong2Nam Kyu Kwon3Department of Electronic Engineering, Yeungnam University, Gyeongsan 38541, Republic of KoreaDepartment of Electronic Engineering, Soonchunhyang University, Asan 31538, Republic of KoreaDepartment of Electronic Engineering, Yeungnam University, Gyeongsan 38541, Republic of KoreaDepartment of Electronic Engineering, Yeungnam University, Gyeongsan 38541, Republic of KoreaIn this paper, we propose a deep deterministic policy gradient (DDPG)-based path-planning method for mobile robots by applying the hindsight experience replay (HER) technique to overcome the performance degradation resulting from sparse reward problems occurring in autonomous driving mobile robots. The mobile robot in our analysis was a robot operating system-based TurtleBot3, and the experimental environment was a virtual simulation based on Gazebo. A fully connected neural network was used as the DDPG network based on the actor–critic architecture. Noise was added to the actor network. The robot recognized an unknown environment by measuring distances using a laser sensor and determined the optimized policy to reach its destination. The HER technique improved the learning performance by generating three new episodes with normal experience from a failed episode. The proposed method demonstrated that the HER technique could help mitigate the sparse reward problem; this was further corroborated by the successful autonomous driving results obtained after applying the proposed method to two reward systems, as well as actual experimental results.https://www.mdpi.com/1424-8220/22/24/9574reinforcement learningdeep deterministic policy gradienthindsight experience replaymobile robotautonomous drivingsparse reward environments |
| spellingShingle | Minjae Park Seok Young Lee Jin Seok Hong Nam Kyu Kwon Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments reinforcement learning deep deterministic policy gradient hindsight experience replay mobile robot autonomous driving sparse reward environments |
| title | Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments |
| title_full | Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments |
| title_fullStr | Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments |
| title_full_unstemmed | Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments |
| title_short | Deep Deterministic Policy Gradient-Based Autonomous Driving for Mobile Robots in Sparse Reward Environments |
| title_sort | deep deterministic policy gradient based autonomous driving for mobile robots in sparse reward environments |
| topic | reinforcement learning deep deterministic policy gradient hindsight experience replay mobile robot autonomous driving sparse reward environments |
| url | https://www.mdpi.com/1424-8220/22/24/9574 |
| work_keys_str_mv | AT minjaepark deepdeterministicpolicygradientbasedautonomousdrivingformobilerobotsinsparserewardenvironments AT seokyounglee deepdeterministicpolicygradientbasedautonomousdrivingformobilerobotsinsparserewardenvironments AT jinseokhong deepdeterministicpolicygradientbasedautonomousdrivingformobilerobotsinsparserewardenvironments AT namkyukwon deepdeterministicpolicygradientbasedautonomousdrivingformobilerobotsinsparserewardenvironments |
