Computer Science & AI
New GCM Model Unlocks Interpretable Cooperation and Performance in Multiagent AI
Original Authors: Wu, Zhu, Chen, Chen

Multiagent reinforcement learning (MARL) has been widely investigated, ranging from theoretical analysis to real-life applications. However, the utilization of existing non-transparent neural network architectures has resulted in opaque decision-making processes, making it difficult for humans to understand and trust the models being used. Fundamentally, all data is a topological structure, which provides reliable transparency for MARL tasks due to its powerful relational expression capability, scalability, and explicit structural relationships.
In this context, researchers propose a novel approach: Graph Cooperation Modeling (GCM). GCM is designed to explicitly capture and comprehend the complex dynamics of collaborative relationships among agents with the graph structure. It operates by learning a metric function to discern beneficial interactions among agents, which is then integrated into the agent aggregation strategy of a graph neural network (GNN) capable of modeling arbitrary-order interactions. Additionally, GCM utilizes identity semantics together with global state and individual value functions to estimate the credit of each agent, thereby enhancing each agent's distinct focus on task-related regions.
The effectiveness of GCM has been demonstrated through extensive experiments on a range of challenging MARL benchmarks. As lead author Wu notes in the paper, "Extensive experiments on a range of challenging MARL benchmarks demonstrate that GCM not only delivers up to 28.75% relative performance gains on super-hard maps but also offers clear interpretability that provides insights into the underlying cooperative patterns." This interpretability, offering insights into underlying cooperative patterns, is crucial for fostering human understanding and trust in multiagent AI systems, making them more effective in real-world applications.