作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2026, Vol. 52 ›› Issue (5): 336-348. doi: 10.19678/j.issn.1000-3428.0070301

• 大模型与生成式人工智能 • 上一篇    下一篇

多智能体博弈环境下的大语言模型协同决策研究

余滔*(), 董军   

  1. 华北电力大学经济与管理学院, 北京 100096
  • 收稿日期:2024-08-29 修回日期:2024-11-11 出版日期:2026-05-15 发布日期:2025-01-10
  • 通讯作者: 余滔
  • 作者简介:

    余滔(CCF学生会员), 男, 博士研究生, 主研方向为复杂网络、人工智能、博弈论

    董军, 教授、博士

  • 基金资助:
    国家电网科学技术项目(1400-202256459A-2-0-ZN)

Research on Collaborative Decision-Making by Large Language Models in Multi-Agent Game Environments

YU Tao*(), DONG Jun   

  1. School of Economics and Management, North China Electric Power University, Beijing 100096, China
  • Received:2024-08-29 Revised:2024-11-11 Online:2026-05-15 Published:2025-01-10
  • Contact: YU Tao

摘要:

在多智能体博弈仿真中, 大语言模型(LLM)的性能已经被广泛研究, 但其在模糊任务目标或不确定性环境中引导多智能体合作的决策能力往往出现"失灵"现象。针对这一问题, 提出一种基于分布式贝叶斯推断的多层级协同决策框架。该框架集成了决策、互评、监管三大功能模块, 利用多个LLM进行协同决策, 并在空间囚徒困境博弈中进行了实验验证。实验结果表明, 该框架有效克服了LLM在模糊任务环境下的决策瓶颈, 成功促进了多智能体合作行为的涌现。此外, 通过对不同实验场景下模型决策能力的量化评估, 发现模型的决策误差与模型规模不呈线性关系。在模糊任务指令下, LLaMA3(70×109)模型的决策误差较LLaMA3(8×109)模型高出16.6%, 较LLaMA2(7×109)模型高出7.2%, 表明在更复杂的环境中, 单纯依赖模型规模的扩大未能显著提升决策性能。相反, LLM协同决策在提升决策一致性和有效性方面显示出显著优势。这些结果揭示了多模型协同在复杂决策环境中的关键作用, 并为未来在不确定性任务下的智能体系统设计提供了重要参考。

关键词: 大语言模型, 多智能体仿真, 空间囚徒困境博弈, 分布式贝叶斯推断, 多层级协同决策框架

Abstract:

In multi-agent game simulations, the performance of Large Language Model (LLM) has been widely studied; however, their decision-making ability to guide multi-agent cooperation in fuzzy task objectives or uncertain environments is often unreliable. To address this issue, a multi-level collaborative decision-making framework based on distributed Bayesian inferences is proposed. This framework integrates three major functional modules: decision making, peer evaluation, and supervision. It utilizes multiple LLM for collaborative decision making and has been experimentally validated in a spatial prisoner's dilemma game. The experimental results show that the framework effectively overcomes the decision-making bottleneck of LLM in fuzzy task environments and successfully promotes the emergence of multi-agent cooperative behavior. Additionally, a quantitative evaluation of the model's decision-making ability in different experimental scenarios reveals that the decision error of the model is not linearly related to the model size. Under fuzzy task instructions, the decision error of the LLaMA3 (70×109) model is 16.6% higher than that of the LLaMA3 (8×109) model and 7.2% higher than that of the LLaMA2 (7×109) model. This indicates that in more complex environments, relying solely on the expansion of the model size does not significantly improve the decision-making performance. By contrast, LLM collaborative decision making has shown significant advantages in improving decision consistency and effectiveness. These results reveal the crucial role of multi-model collaboration in complex decision-making environments and provide important references for the future design of intelligent agent systems for uncertain tasks.

Key words: Large Language Models (LLM), multi-agent simulation, spatial prisoner's dilemma game, distributed Bayesian inference, multi-level collaborative decision-making framework