Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
Beyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9358136/ |
id |
doaj-33396cce226b4478a7b5bf346e679861 |
---|---|
record_format |
Article |
spelling |
doaj-33396cce226b4478a7b5bf346e6798612021-03-30T15:08:29ZengIEEEIEEE Access2169-35362021-01-019322823229710.1109/ACCESS.2021.30604269358136Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air CombatDongyuan Hu0https://orcid.org/0000-0002-2983-2471Rennong Yang1Jialiang Zuo2Ze Zhang3Jun Wu4Ying Wang5Air Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaBeyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose an intelligent maneuver planning method for BVR combat with using an improved deep Q network (DQN). First, a basic combat environment builds, which mainly includes flight motion model, relative motion model and missile attack model. Then, we create a maneuver decision framework for agent interaction with the environment. Basic perceptive variables are constructed for agents to form continuous state space. Also, considering the threat of each side missile and the constraint of airfield, the reward function is designed for agents to training. Later, we introduce a training algorithm and propose perceptional situation layers and value fitting layers to replace policy network in DQN. Based on long short-term memory (LSTM) cell, the perceptional situation layer can convert basic state to high-dimensional perception situation. The fitting layer does well in mapping action. Finally, three combat scenarios are designed for agent training and testing. Simulation result shows the agent can avoid the threat of enemy and gather own advantages to threat the target. It also proves the models and methods of agents are valid and intelligent air combat can be realized.https://ieeexplore.ieee.org/document/9358136/Beyond visual rangeintelligent air combatmaneuver planning and decisionmissile kill envelopedeep reinforcement learningLSTM-DQN |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Dongyuan Hu Rennong Yang Jialiang Zuo Ze Zhang Jun Wu Ying Wang |
spellingShingle |
Dongyuan Hu Rennong Yang Jialiang Zuo Ze Zhang Jun Wu Ying Wang Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat IEEE Access Beyond visual range intelligent air combat maneuver planning and decision missile kill envelope deep reinforcement learning LSTM-DQN |
author_facet |
Dongyuan Hu Rennong Yang Jialiang Zuo Ze Zhang Jun Wu Ying Wang |
author_sort |
Dongyuan Hu |
title |
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat |
title_short |
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat |
title_full |
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat |
title_fullStr |
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat |
title_full_unstemmed |
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat |
title_sort |
application of deep reinforcement learning in maneuver planning of beyond-visual-range air combat |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
Beyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose an intelligent maneuver planning method for BVR combat with using an improved deep Q network (DQN). First, a basic combat environment builds, which mainly includes flight motion model, relative motion model and missile attack model. Then, we create a maneuver decision framework for agent interaction with the environment. Basic perceptive variables are constructed for agents to form continuous state space. Also, considering the threat of each side missile and the constraint of airfield, the reward function is designed for agents to training. Later, we introduce a training algorithm and propose perceptional situation layers and value fitting layers to replace policy network in DQN. Based on long short-term memory (LSTM) cell, the perceptional situation layer can convert basic state to high-dimensional perception situation. The fitting layer does well in mapping action. Finally, three combat scenarios are designed for agent training and testing. Simulation result shows the agent can avoid the threat of enemy and gather own advantages to threat the target. It also proves the models and methods of agents are valid and intelligent air combat can be realized. |
topic |
Beyond visual range intelligent air combat maneuver planning and decision missile kill envelope deep reinforcement learning LSTM-DQN |
url |
https://ieeexplore.ieee.org/document/9358136/ |
work_keys_str_mv |
AT dongyuanhu applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat AT rennongyang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat AT jialiangzuo applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat AT zezhang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat AT junwu applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat AT yingwang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat |
_version_ |
1724179962723827712 |