•  
  •  
 

Journal of System Simulation

Abstract

Abstract: To solve the problem of strategy convergence caused by role homogenization in traditional self-play for imbalanced air combat, an intelligent decision-making method based on asymmetric selfplay was proposed. This method decoupled tactics from control by employing a hierarchical reinforcement learning framework and designed differentiated reward functions for advantaged and disadvantaged sides. Bidirectional independent policy pools were constructed to promote the co-evolution of strategies. The proximal policy optimization algorithm was utilized to train the model. Experiments in 1v1 weapon-imbalanced and 2v1 numerically-imbalanced scenarios demonstrate that compared to symmetric self-play, the proposed method increases the kill rate of the advantaged side by up to 12% and the survival rate of the disadvantaged side by up to 40%. The overall effectiveness in multi-agent combat is also significantly enhanced. The study verifies the effectiveness of the asymmetric design in enhancing the specialized combat capabilities and tactical diversity of intelligent agents for imbalanced air combat.

First Page

433

Last Page

446

CLC

TP391.9

Recommended Citation

Zheng Wei, Tang Jiahao, Xiong Xiaoping, et al. Intelligent Decision-making Method in Imbalanced Air Combat Based on Asymmetric Self-play[J]. Journal of System Simulation, 2026, 38(2): 433-446.

Corresponding Author

Tang Jiahao

DOI

10.16182/j.issn1004731x.joss.25-0621

Share

COinS