|
|
Air Combat Decision-Making Method Based on Game Tree and Digital Parallel Simulation Battlefield |
ZHOU Wenjie1, FU Yulong1, GUO Xiangke2, QI Yutao1, ZHANG Haibin1 |
1. School of Cyber Engineering, Xidian University, Xi'an 710126, Shaanxi, China;
2. Air and Missile Defense College, Air Force Engineering University, Xi'an 710051, Shaanxi, China |
|
|
Abstract In air combat decision-making, effectively identifying the key states and improving the decision-making ability of intelligent bodies in these states is the key research direction of reinforcement learning algorithms. In this paper, a dynamic strategy switching framework built by deep reinforcement learning was proposed for the intelligent body problem in air combat decision-making, aiming at increasing the decision-making quality of the intelligent body in the complex environment. This study identified critical states using dimensionality reduction and classification of high-dimensional state space through representation learning and cluster analysis techniques in the non-critical state, a deep reinforcement learning algorithm was employed for decision-making; in the critical state, an inverse dynamics model was adopted to generates the target state's corresponding action sequence and a parallel simulation strategy was utilized to execute the action sequence in multiple simulation environments to approximate the target state rapidly. At the end of the simulation, the optimal decision path was determined by advantage value evaluation. The experimental results show that the method can improve the decision-making ability of the intelligent body in critical states, providing a new solution for intelligent decision-making in complex air combat environments.
|
Received: 29 January 2025
Published: 15 July 2025
|
|
|
|
|
[1] |
CHEN Shi, YANG Linsen, LIU Yihong, LUO Huan, ZANG Tianlei, ZHOU Buxiang. Optimal Scheduling Strategy of Newly-Built Microgrid in Small Sample Data-Driven Mode[J]. Journal of Shanghai Jiao Tong University, 2025, 59(6): 732-745. |
[2] |
LIU Fei, YU Yang, KUANG Yenian. Credibility Evaluation of Parallel Simulation: Current Status, Challenges, and Prospects[J]. Air & Space Defense, 2025, 8(3): 23-28. |
[3] |
YANG Yinghe, WEI Handi, FAN Dixia, LI Ang. Optimization Method of Underwater Flapping Foil Propulsion Performance Based on Gaussian Process Regression and Deep Reinforcement Learning[J]. Journal of Shanghai Jiao Tong University, 2025, 59(1): 70-78. |
[4] |
ZHOU Yi, ZHOU Liangcai, SHI Di, ZHAO Xiaoying, SHAN Xin. Coordinated Active Power-Frequency Control Based on Safe Deep Reinforcement Learning[J]. Journal of Shanghai Jiao Tong University, 2024, 58(5): 682-692. |
[5] |
DONG Yubo1 (董玉博), CUI Tao1 (崔涛), ZHOU Yufan1 (周禹帆),
SONG Xun2 (宋勋), ZHU Yue2 (祝月), DONG Peng1∗ (董鹏). Reward Function Design Method for Long Episode Pursuit Tasks Under Polar Coordinate in Multi-Agent Reinforcement Learning[J]. J Shanghai Jiaotong Univ Sci, 2024, 29(4): 646-655. |
[6] |
LI Shuyi (李舒逸), LI Minzhe (李旻哲), JING Zhongliang∗ (敬忠良). Multi-Agent Path Planning Method Based on Improved Deep Q-Network in Dynamic Environments[J]. J Shanghai Jiaotong Univ Sci, 2024, 29(4): 601-612. |
[7] |
MIAO Zhenhua(苗镇华), HUANG Wentao(黄文焘), ZHANG Yilian(张依恋), FAN Qinqin(范勤勤). Multi-Robot Task Allocation Using Multimodal Multi-Objective
Evolutionary Algorithm Based on Deep Reinforcement Learning[J]. J Shanghai Jiaotong Univ Sci, 2024, 29(3): 377-387. |
[8] |
QUAN Jiale, MA Xianlong, SHEN Yuheng. Multi-agent Formation Method Based on Dynamic Optimization of Proximal Policies[J]. Air & Space Defense, 2024, 7(2): 52-62. |
[9] |
ZHANG Weizhen, HE Zhen, TANG Zhangfan. Reinforcement Learning Control Design for Perching Maneuver of Unmanned Aerial Vehicles with Wind Disturbances[J]. Journal of Shanghai Jiao Tong University, 2024, 58(11): 1753-1761. |
[10] |
MA Chi, ZHANG Guoqun, SUN Junge, LYU Guangzhe, ZHANG Tao. Deep Reinforcement Learning-Based Reconfiguration Method for Integrated Electronic Systems[J]. Air & Space Defense, 2024, 7(1): 63-70. |
[11] |
JI Xiukun (冀秀坤), HAI Jintao (海金涛), LUO Wenguang (罗文广), LIN Cuixia (林翠霞), XIONG Yu(熊 禹), OU Zengkai (殴增开), WEN Jiayan(文家燕). Obstacle Avoidance in Multi-Agent Formation Process Based on Deep Reinforcement Learning[J]. J Shanghai Jiaotong Univ Sci, 2021, 26(5): 680-685. |
[12] |
LI Peng, RUAN Xiaogang, ZHU Xiaoqing, CHAI Jie, REN Dingqi, LIU Pengfei. A Regionalization Vision Navigation Method Based on Deep Reinforcement Learning[J]. Journal of Shanghai Jiao Tong University, 2021, 55(5): 575-585. |
[13] |
XIE De-Fu, TANG Min, MA Li-Zhuang, MAO Jun-Fa. A Parallel Simulation Technique of Interconnect Networks with Frequency-Dependent Parameters Based on Equivalent Circuit Models[J]. Journal of Shanghai Jiaotong University, 2012, 46(06): 921-925. |
|
|
|
|