|
1 | 1 | --- |
| 2 | +- title: "Equilibrium Policy Generalization: A Reinforcement Learning Framework for Cross-Graph Zero-Shot Generalization in Pursuit-Evasion Games" |
| 3 | + venue: To appear in Advances in Neural Information Processing Systems |
| 4 | + year: 2025 |
| 5 | + authors: |
| 6 | + - R Lu |
| 7 | + - P Zhang |
| 8 | + - R Shi |
| 9 | + - Y Zhu |
| 10 | + - D Zhao |
| 11 | + - D Wang |
| 12 | + - id:calippi |
| 13 | + keywords: |
| 14 | + - reinforcement learning |
| 15 | + abstract: Decentralized Multi-Agent Reinforcement Learning (MARL) methods allow for learning scalable multi-agent policies, but suffer from partial observability and induced non-stationarity. These challenges can be addressed by introducing mechanisms that facilitate coordination and high-level planning. Specifically, coordination and temporal abstraction can be achieved through communication (e.g., message passing) and Hierarchical Reinforcement Learning (HRL) approaches to decision-making. However, optimization issues limit the applicability of hierarchical policies to multi-agent systems. As such, the combination of these approaches has not been fully explored. To fill this void, we propose a novel and effective methodology for learning multi-agent hierarchies of message-passing policies. We adopt the feudal HRL framework and rely on a hierarchical graph structure for planning and coordination among agents. Agents at lower levels in the hierarchy receive goals from the upper levels and exchange messages with neighboring agents at the same level. To learn hierarchical multi-agent policies, we design a novel reward-assignment method based on training the lower-level policies to maximize the advantage function associated with the upper levels. Results on relevant benchmarks show that our method performs favorably compared to the state of the art. |
| 16 | + bibtex: > |
| 17 | + @misc{lu2025equilibrium, |
| 18 | + title={Equilibrium Policy Generalization: A Reinforcement Learning Framework for Cross-Graph Zero-Shot Generalization in Pursuit-Evasion Games}, |
| 19 | + author={Renyun Lu, Peng Zhang, Ruochan Shi, Yuanheng Zhu, Dongbin Zhao, Dong Wang, Cesare Alippi}, |
| 20 | + year={2025} |
| 21 | + } |
2 | 22 | - title: Hierarchical Message-Passing Policies for Multi-Agent Reinforcement Learning |
3 | 23 | links: |
4 | 24 | paper: https://arxiv.org/abs/2507.23604 |
|
23 | 43 | - title: "Over-squashing in Spatiotemporal Graph Neural Networks" |
24 | 44 | links: |
25 | 45 | paper: https://arxiv.org/abs/2506.15507 |
26 | | - venue: Preprint |
| 46 | + venue: To appear in Advances in Neural Information Processing Systems |
27 | 47 | year: 2025 |
28 | 48 | authors: |
29 | 49 | - id:imarisca |
|
50 | 70 | paper: https://arxiv.org/abs/2506.13652 |
51 | 71 | dataset: https://huggingface.co/datasets/MeteoSwiss/PeakWeather |
52 | 72 | code: https://github.com/Graph-Machine-Learning-Group/peakweather-wind-forecasting |
| 73 | + website: https://peakweather.readthedocs.io/ |
53 | 74 | venue: Preprint |
54 | 75 | year: 2025 |
55 | 76 | authors: |
|
107 | 128 | links: |
108 | 129 | paper: https://arxiv.org/abs/2405.19933 |
109 | 130 | github: https://github.com/allemanenti/Learning-Calibrated-Structures |
110 | | - venue: To appear in International Conference on Machine Learning |
| 131 | + venue: International Conference on Machine Learning |
111 | 132 | year: 2025 |
112 | 133 | authors: |
113 | 134 | - id:amanenti |
|
123 | 144 | title = {Learning {{Latent Graph Structures}} and Their {{Uncertainty}}}, |
124 | 145 | author = {Manenti, Alessandro and Zambon, Daniele and Alippi, Cesare}, |
125 | 146 | year = {2025}, |
126 | | - booktitle={To appear in Proceedings of the Forty-Second International Conference on Machine Learning (ICML)}, |
| 147 | + booktitle={Proceedings of the Forty-Second International Conference on Machine Learning (ICML)}, |
127 | 148 | } |
128 | 149 | - title: 'Temporal Graph ODEs for Irregularly-Sampled Time Series' |
129 | 150 | links: |
|
0 commit comments