Centralized Training with Hybrid Execution in Multi-Agent Reinforcement LearningShow others and affiliations
2024 (English)In: AAMAS 2024 - Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems, International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS) , 2024, p. 2453-2455Conference paper, Published paper (Refereed)
Abstract [en]
We introduce hybrid execution in multi-agent reinforcement learning (MARL), a new paradigm in which agents aim to successfully complete cooperative tasks with arbitrary communication levels at execution time by taking advantage of information-sharing among the agents. Under hybrid execution, the communication level can range from a setting in which no communication is allowed between agents (fully decentralized), to a setting featuring full communication (fully centralized), but the agents do not know beforehand which communication level they will encounter at execution time. To formalize our setting, we define a new class of multi-agent partially observable Markov decision processes (POMDPs) that we name hybrid-POMDPs, which explicitly model a communication process between the agents.
Place, publisher, year, edition, pages
International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS) , 2024. p. 2453-2455
Keywords [en]
Machine Learning, Multi-Agent Reinforcement Learning, Multi-Agent Systems, Reinforcement Learning
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-348768Scopus ID: 2-s2.0-85196359245OAI: oai:DiVA.org:kth-348768DiVA, id: diva2:1878678
Conference
23rd International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2024, Auckland, New Zealand, May 6 2024 - May 10 2024
Note
QC 20240701
2024-06-272024-06-272024-07-01Bibliographically approved