Application of deep reinforcement learning tomissile trajectory planning
CSTR:
Author:
Affiliation:

1.National Innovation Institute of Defense Technology, Academy of Military Sciences, Beijing 100071 , China ;2.Center for Control Theory and Guidance Technology, Harbin Institute of Technology, Harbin 150000 , China ;3.Laboratory for Space Environment and Physical Sciences, Harbin Institute of Technology, Harbin 150000 , China ;4.College of Aerospace Science and Engineering, National University of Defense Technology, Changsha 410073 , China

Clc Number:

TP18;TP27;V24

Fund Project:

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    Aiming for missile trajectory planning, an applicable Gym training evironment was established. An intelligent agent network structure and its reward functions were designed based on twin delayed deep deterministic policy gradient framework and according to terminal and process constraints, forming an intelligent trajectory planning method. Through deploying the algorithm on an embedded GPU computing acceleration platform, bias simulation and comparison tests were conducted. The results show that the method can reach the requirements of missile capability and process constraints under different range tasks and effectively overcome environmental disturbances with adaptability to distinct object models. Meanwhile, the method has an extremely fast calculation speed, far surpassing the popular GPOPS-Ⅱ toolbox. The computation time for single step trajectory command is less than a millisecond so that it can support real-time online trajectory generation, which provides an effective implementation path and technical support for engineering applications.

    Reference
    Related
    Cited by
Get Citation

张敬, 李彤, 李建锋, 等. 深度强化学习在导弹弹道规划中的应用[J]. 国防科技大学学报, 2025, 47(3): 109-118.

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:May 08,2023
  • Revised:
  • Adopted:
  • Online: June 03,2025
  • Published:
Article QR Code