Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat

Beyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose...

Full description

Bibliographic Details
Main Authors: Dongyuan Hu, Rennong Yang, Jialiang Zuo, Ze Zhang, Jun Wu, Ying Wang
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9358136/
id doaj-33396cce226b4478a7b5bf346e679861
record_format Article
spelling doaj-33396cce226b4478a7b5bf346e6798612021-03-30T15:08:29ZengIEEEIEEE Access2169-35362021-01-019322823229710.1109/ACCESS.2021.30604269358136Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air CombatDongyuan Hu0https://orcid.org/0000-0002-2983-2471Rennong Yang1Jialiang Zuo2Ze Zhang3Jun Wu4Ying Wang5Air Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaAir Force Engineering University, Xi’an, ChinaBeyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose an intelligent maneuver planning method for BVR combat with using an improved deep Q network (DQN). First, a basic combat environment builds, which mainly includes flight motion model, relative motion model and missile attack model. Then, we create a maneuver decision framework for agent interaction with the environment. Basic perceptive variables are constructed for agents to form continuous state space. Also, considering the threat of each side missile and the constraint of airfield, the reward function is designed for agents to training. Later, we introduce a training algorithm and propose perceptional situation layers and value fitting layers to replace policy network in DQN. Based on long short-term memory (LSTM) cell, the perceptional situation layer can convert basic state to high-dimensional perception situation. The fitting layer does well in mapping action. Finally, three combat scenarios are designed for agent training and testing. Simulation result shows the agent can avoid the threat of enemy and gather own advantages to threat the target. It also proves the models and methods of agents are valid and intelligent air combat can be realized.https://ieeexplore.ieee.org/document/9358136/Beyond visual rangeintelligent air combatmaneuver planning and decisionmissile kill envelopedeep reinforcement learningLSTM-DQN
collection DOAJ
language English
format Article
sources DOAJ
author Dongyuan Hu
Rennong Yang
Jialiang Zuo
Ze Zhang
Jun Wu
Ying Wang
spellingShingle Dongyuan Hu
Rennong Yang
Jialiang Zuo
Ze Zhang
Jun Wu
Ying Wang
Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
IEEE Access
Beyond visual range
intelligent air combat
maneuver planning and decision
missile kill envelope
deep reinforcement learning
LSTM-DQN
author_facet Dongyuan Hu
Rennong Yang
Jialiang Zuo
Ze Zhang
Jun Wu
Ying Wang
author_sort Dongyuan Hu
title Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
title_short Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
title_full Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
title_fullStr Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
title_full_unstemmed Application of Deep Reinforcement Learning in Maneuver Planning of Beyond-Visual-Range Air Combat
title_sort application of deep reinforcement learning in maneuver planning of beyond-visual-range air combat
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2021-01-01
description Beyond-visual-range (BVR) engagement becomes more and more popular in the modern air battlefield. The key and difficulty for pilots in the fight is maneuver planning, which reflects the tactical decision-making capacity of the both sides and determinates success or failure. In this paper, we propose an intelligent maneuver planning method for BVR combat with using an improved deep Q network (DQN). First, a basic combat environment builds, which mainly includes flight motion model, relative motion model and missile attack model. Then, we create a maneuver decision framework for agent interaction with the environment. Basic perceptive variables are constructed for agents to form continuous state space. Also, considering the threat of each side missile and the constraint of airfield, the reward function is designed for agents to training. Later, we introduce a training algorithm and propose perceptional situation layers and value fitting layers to replace policy network in DQN. Based on long short-term memory (LSTM) cell, the perceptional situation layer can convert basic state to high-dimensional perception situation. The fitting layer does well in mapping action. Finally, three combat scenarios are designed for agent training and testing. Simulation result shows the agent can avoid the threat of enemy and gather own advantages to threat the target. It also proves the models and methods of agents are valid and intelligent air combat can be realized.
topic Beyond visual range
intelligent air combat
maneuver planning and decision
missile kill envelope
deep reinforcement learning
LSTM-DQN
url https://ieeexplore.ieee.org/document/9358136/
work_keys_str_mv AT dongyuanhu applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
AT rennongyang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
AT jialiangzuo applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
AT zezhang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
AT junwu applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
AT yingwang applicationofdeepreinforcementlearninginmaneuverplanningofbeyondvisualrangeaircombat
_version_ 1724179962723827712