具备可解释性的决策依据自编码多智能体强化学习方法
CSTR:
作者:
作者单位:

作者简介:

通讯作者:

中图分类号:

TP391.9

基金项目:

国家自然科学基金项目(62495090).


Interpretable decision-basis autoencoder for multi-agent reinforcement learning
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    深度强化学习已成为无人集群在复杂未知环境中实现自主决策的关键技术方案, 但是, 内部不可解释的“黑盒”结构使得人类难以理解、信任和监督智能体的自主决策, 严重阻碍其在高安全需求领域中的应用. 鉴于此, 提出一种具备可解释性的多智能体强化学习方法. 首先, 设计具备可解释性的狄利克雷变分自编码器, 从隐空间中编码匹配物理语义信息的决策依据概率分布; 然后, 使用门控网络线性混合决策依据生成动作决策; 最后, 在多智能体近端策略优化强化学习网络框架下完成可解释自编码器的集成训练. 所提出方法将智能体的决策表征为若干具备物理含义依据的混合概率分布, 使得人类可通过概率密度直观地理解智能体行为, 并可通过调整门控权重直接干预智能体决策. 仿真对比实验验证了所提出方法的决策性能, 所设计的可视化方法展示了智能体决策的可解释性以及人类干预决策的效果.

    Abstract:

    Deep reinforcement learning (DRL) has emerged as a pivotal technology for enabling autonomous decision-making in unmanned swarms operating within complex and unstructured environments. Nevertheless, the inherent “black-box” nature and lack of interpretability of DRL models impede human understanding, trust, and oversight of agents’ autonomous behaviors, thereby significantly restricting their deployment in safety-critical applications. To address this challenge, this paper proposes an interpretable multi-agent reinforcement learning approach. Firstly, an interpretable Dirichlet variational autoencoder is designed to encode decision rationales as probability distributions in a latent space aligned with physically meaningful semantics. Secondly, a gating network is employed to generate action decisions by linearly combining the encoded rationales. Lastly, the interpretable autoencoder is integrated into and jointly trained within a multi-agent proximal policy optimization framework. This method represents the agent’s decision-making process as a mixture of probability distributions, each corresponding to interpretable physical semantics, thereby enabling intuitive human understanding of agent behavior through probability densities and allowing direct human intervention by adjusting the gating weights. Comparative simulation experiments validate the decision-making performance of the proposed approach, while the developed visualization techniques demonstrate both the interpretability of agent decisions and the efficacy of human intervention.

    参考文献
    相似文献
    引证文献
引用本文

李佩璋,费庆,陈振,等.具备可解释性的决策依据自编码多智能体强化学习方法[J].控制与决策,2025,40(9):2748-2758

复制
相关视频

分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2025-01-07
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2025-08-08
  • 出版日期: 2025-09-20
文章二维码