| 32 | 0 | 53 |
| 下载次数 | 被引频次 | 阅读次数 |
在军事模拟战略中,作战任务的执行往往依赖于团队的高效协同,多智能体强化学习方法正是为应对这一需求而被广泛引入的。然而,在实际应用过程中,由于多智能体对所处环境状态的观测构建不够充分,导致其获取的信息较为有限,难以满足复杂军事任务中战场感知的要求。针对这一问题,提出一种基于RACE的多模态信息融合技术改进方法(RACE-MFT)。该方法通过整合属性、文本和图像三方面信息,构建了更为丰富和全面的智能体状态表示,从而增加状态维度,增强智能体的信息感知能力,使其能够做出更优的决策。实验在即时战略游戏《星际争霸Ⅱ》和自建的“争夺要地任务”环境中开展。结果显示,在《星际争霸Ⅱ》中,使用RACE-MFT的智能体对阵游戏自带AI时,胜率提升了3%。在改进算法与原算法的对抗中,胜率稳定在80%。在“争夺要地”环境里,相比其他单一模块改进,RACE-MFT的收敛奖励达到最大值。证实了RACE-MFT在处理多智能体团队协同任务时的有效性。
Abstract:In military simulation strategies, the execution of combat tasks often relies on efficient teamwork, and multi-agent reinforcement learning methods have been widely introduced to meet this demand. However, in practical applications, due to insufficient observation and construction of the environmental state by multi-agent systems, the information they obtain is relatively limited, making it difficult to meet the requirements of battlefield perception in complex military tasks. A method for improving multi-modal information fusion technology based on RACE(RACE-MFT) is proposed to address this issue. This method integrates attribute, text, and image information to construct a richer and more comprehensive representation of the agent's state, thereby increasing the state dimension and enhancing the agent's information perception ability, enabling it to make better decisions. The experiment was conducted in the real-time strategy game StarCraft II and a self built "Battle for Key Tasks" environment. The results showed that in StarCraft II, when using RACE-MFT agents to compete against the game's built-in AI, the win rate increased by 3%. In the confrontation between the improved algorithm and the original algorithm, the winning rate remains stable at 80%. In the environment of "competing for important places", compared with other single module improvements, the convergence reward of RACE-MFT reaches the maximum. These all confirm the effectiveness of RACE-MFT in handling multi-agent team collaboration tasks.
[1]林艳,夏开元,张一晋.基于生成对抗网络辅助多智能体强化学习的边缘计算网络联邦切片资源管理[J/OL].电子与信息学报,1-12[2025-03-20]. http://kns. cnki. net/kcms/detail/11.4494. TN. 20250312.0924.018.html.
[2]唐翱,薛珮芸,强彦,等.多视角特征融合的孤立手语识别方法[J/OL].小型微型计算机系统,1-12[2025-03-20].http://kns.cnki.net/kcms/detail/21.1106.TP.20250221.1456.012.html.
[3]李安梦蝶.考虑驾驶人语言因素的行车愤怒情绪识别方法研究[D].山东:山东理工大学,2023.
[4]宫银凤,金久才,刘德庆,等.一种无人船载激光雷达-相机的融合感知方法[J/OL].海洋科学进展,1-12[2025-03-20].http://kns.cnki.net/kcms/detail/37.1387.P.20250228.1644.003.html.
[5]Cheng L,Nokhbatolfoghahai A,Groves M R, et al.Data level fusion of acoustic emission sensors using deep learning[J]. Journal of Intelligent Material Systems and Structures,2025,36(2):77-96.
[6]Jiang X,Fan X,Zhang T, et al.A Multi sensor Data Fusion Method for Smart Agriculture[J].World Scientific Research Journal,2024,10(8):24-30.
[7]王新奕,费泽松,周一青,等.面向物联网的通感算智融合:关键技术与未来展望[J/OL].电子与信息学报,1-21[2025-03-20].http://kns.cnki.net/kcms/detail/11.4494.TN.20250315.1835.004.html.
[8]王志伟,陆振宇.基于细粒度特征增强的多模态视觉问答研究[J/OL].南京信息工程大学学报, 1-17[2025-03-20]. https://doi. org/10.13878/j.cnki.jnuist.20250107001.
[9]Zhang S S,Wang P,Xiao B H, et al.Risk Prediction of Tunnel Water and Mud Inrush Based on Decision-Level Fusion of Multisource Data[J].Applied Geophysics,2025,(2):472-487.
[10]Lv C,Li B,Wang X, et al.CMS-net:Edge-aware multimodal MRI feature fusion for brain tumor segmentation[J].Image and Vision Computing,2025,156:105481.
[11]Zhao C,Liu Y,Chen W, et al.Multi-scale frequency feature fusion transformer for pediatric echocardiography analysis[J].Applied Soft Computing,2025,173:112950.
[12]丁祎.面向无人机光伏巡检的图像配准和目标检测算法研究[D].吉林:吉林大学,2024.
[13]吴鲲鹏.光学与SAR遥感图像配准及融合方法研究[D].吉林:吉林大学,2023.
[14]BAI Y,CHEN S,ZHANG J,et al. An adaptive activepower rolling dispatch strategy for high proportion ofrenewable energy based on distributed deepreinforcement learning[J]. Applied Energy, 2023, 330:120294.
[15]丁开源,艾斯卡尔·艾木都拉,朱斌,等.基于多模态深度强化学习的端到端无人车运动规划[J].系统仿真学报,2024,36(11):2631-2643.
[16]Moerland T M, Broekens J, Jonker C M. Learning multimodal transition dynamics for model-based reinforcement learning[J]. arXiv preprint arXiv:1705.00470, 2017.
[17]徐加开,陆奇,李祥云,等.基于改进型多模态信息融合深度强化学习的自主超声扫描方法[J/OL].计算机应用研究,1-8[2025-05-03].https://doi.org/10.19734/j.issn.1001-3695.2024.11.0476.
基本信息:
中图分类号:TP18
引用信息:
[1]陈亮,智鑫龙,王珺琳.基于RACE-MFT算法的多智能体战略状态融合模型研究[J].通信与信息技术,2026,No.279(01):1-6.
基金信息:
辽宁省教育厅高等学校基本科研项目青年项目(项目编号:1030040000668); 沈阳理工大学引进高层次人才项目(项目编号:1010147001228)