Hierarchical Game Theory based Control for Large Scale Multi-Agent Systems: A Hybrid Reinforcement Learning Approach

Loading...
Thumbnail Image

Authors

Dey, Shawon

Issue Date

2025

Type

Dissertation

Language

en_US

Keywords

Research Projects

Organizational Units

Journal Issue

Alternative Title

Abstract

Distributed optimal control and coordination strategies for large-scale multi-agent systems (LS-MAS) operating under uncertainty have become increasingly critical due to the challenges of scalability, communication complexity, and computational intractability. Traditional control approaches face the curse of dimensionality as the number of interacting agents increases, making them unsuitable for real-time implementation in dense environments. To address these limitations, mean field game (MFG) theory has been employed as a scalable approximation tool, wherein the behavior of the overall population is represented through a probability density function (PDF). However, conventional MFG formulations often lack the ability to capture diverse coordination behaviors due to their reliance on a single global PDF, leading to suboptimal performance in heterogeneous or structured systems. To enhance flexibility and coordination efficiency, a hierarchical game-theoretic framework is introduced by decomposing the entire agent population into multiple interacting subgroups. Within this structure, inter-group and intra-group dynamics are governed through layered game formulations that incorporate both strategic leadership and population-level interactions. The framework enables scalable solutions to complex coordination problems such as flocking and formation, under uncertainty and in a distributed manner. To further improve coordination effectiveness, an extended formulation of MFG is developed to account for structural heterogeneity and asymmetric density constraints. A decomposition strategy is proposed to represent the global terminal behavior through a collection of subgroup-level distributions, allowing increased expressiveness and adaptability in control design. In addition, a reinforcement learning-based algorithm is designed to achieve distributed policy optimization within this hierarchical structure, enabling agents to adapt to changing environments with limited information exchange. Finally, a theoretical and empirical study is conducted to analyze the fundamental trade-off between coordination efficiency and computational complexity. Stability, convergence, and robustness of the proposed approach are validated through extensive simulations and comparative analysis.

Description

Citation

Publisher

License

Journal

Volume

Issue

PubMed ID

DOI

ISSN

EISSN