Reinforcement Learning Journal, vol. TBD, 2025, pp. TBD.
Presented at the Reinforcement Learning Conference (RLC), Edmonton, Alberta, Canada, August 5–9, 2025.
Multi-agent reinforcement learning (MARL) has witnessed a remarkable surge in interest, fueled by the empirical success achieved in applications of single-agent reinforcement learning (RL). In this study, we consider a distributed Q-learning scenario, wherein a number of agents cooperatively solve a sequential decision making problem without access to the central reward function which is an average of the local rewards. In particular, we study finite-time analysis of a distributed Q-learning algorithm, and provide a new sample complexity result under tabular lookup setting for Markovian observation model.
Han-Dong Lim and Donghwan Lee. "A Finite-Time Analysis of Distributed Q-Learning." Reinforcement Learning Journal, vol. TBD, 2025, pp. TBD.
BibTeX:@article{lim2025finite,
title={A Finite-Time Analysis of Distributed {Q-Learning}},
author={Lim, Han-Dong and Lee, Donghwan},
journal={Reinforcement Learning Journal},
year={2025}
}