Air-Combat Strategy Using Approximate Dynamic Programming

Unmanned Aircraft Systems (UAS) have the potential to perform many of the dangerous missions currently own by manned aircraft. Yet, the complexity of some tasks, such as air combat, have precluded UAS from successfully carrying out these missions autonomously. This paper presents a formulation of a...

Full description

Bibliographic Details
Main Authors: McGrew, James S. (Contributor), How, Jonathan P. (Contributor), Bush, Lawrence (Contributor), Williams, Brian Charles (Contributor), Roy, Nicholas (Contributor)
Other Authors: Massachusetts Institute of Technology. Aerospace Controls Laboratory (Contributor), Massachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratory (Contributor), Massachusetts Institute of Technology. Department of Aeronautics and Astronautics (Contributor)
Format: Article
Language:English
Published: American Institute of Aeronautics and Astronautics, 2011-11-28T21:16:54Z.
Subjects:
Online Access:Get fulltext
Description
Summary:Unmanned Aircraft Systems (UAS) have the potential to perform many of the dangerous missions currently own by manned aircraft. Yet, the complexity of some tasks, such as air combat, have precluded UAS from successfully carrying out these missions autonomously. This paper presents a formulation of a level flight, fixed velocity, one-on-one air combat maneuvering problem and an approximate dynamic programming (ADP) approach for computing an efficient approximation of the optimal policy. In the version of the problem formulation considered, the aircraft learning the optimal policy is given a slight performance advantage. This ADP approach provides a fast response to a rapidly changing tactical situation, long planning horizons, and good performance without explicit coding of air combat tactics. The method's success is due to extensive feature development, reward shaping and trajectory sampling. An accompanying fast and e ffective rollout-based policy extraction method is used to accomplish on-line implementation. Simulation results are provided that demonstrate the robustness of the method against an opponent beginning from both off ensive and defensive situations. Flight results are also presented using micro-UAS own at MIT's Real-time indoor Autonomous Vehicle test ENvironment (RAVEN).
Defense University Research Instrumentation Program (U.S.) (grant number FA9550-07-1-0321)
United States. Air Force Office of Scientific Research (AFOSR # FA9550-08-1-0086)
American Society for Engineering Education (National Defense Science and Engineering Graduate Fellowship)