Zobrazeno 1 - 10
of 1 695
pro vyhledávání: '"LIM, HAN"'
Autor:
Lim, Han-Dong, Lee, Donghwan
Multi-agent reinforcement learning (MARL) has witnessed a remarkable surge in interest, fueled by the empirical success achieved in applications of single-agent reinforcement learning (RL). In this study, we consider a distributed Q-learning scenario
Externí odkaz:
http://arxiv.org/abs/2405.14078
Reinforcement learning has witnessed significant advancements, particularly with the emergence of model-based approaches. Among these, $Q$-learning has proven to be a powerful algorithm in model-free settings. However, the extension of $Q$-learning t
Externí odkaz:
http://arxiv.org/abs/2402.11877
Autor:
Lim, Han-Dong, Lee, Donghwan
The goal of this paper is to investigate distributed temporal difference (TD) learning for a networked multi-agent Markov decision process. The proposed approach is based on distributed optimization algorithms, which can be interpreted as primal-dual
Externí odkaz:
http://arxiv.org/abs/2310.00638
The main goal of this paper is to investigate continuous-time distributed dynamic programming (DP) algorithms for networked multi-agent Markov decision problems (MAMDPs). In our study, we adopt a distributed multi-agent framework where individual age
Externí odkaz:
http://arxiv.org/abs/2307.16706
Autor:
Lim, Han-Dong, Lee, Donghwan
Temporal-difference (TD) learning is widely regarded as one of the most popular algorithms in reinforcement learning (RL). Despite its widespread use, it has only been recently that researchers have begun to actively study its finite time behavior, i
Externí odkaz:
http://arxiv.org/abs/2306.09746
Autor:
Lim, Han-Dong, Lee, Donghwan
Off-policy learning ability is an important feature of reinforcement learning (RL) for practical applications. However, even one of the most elementary RL algorithms, temporal-difference (TD) learning, is known to suffer form divergence issue when th
Externí odkaz:
http://arxiv.org/abs/2302.09875
Autor:
Lim, Han-Dong, Lee, Donghwan
Q-learning has long been one of the most popular reinforcement learning algorithms, and theoretical analysis of Q-learning has been an active research topic for decades. Although researches on asymptotic convergence analysis of Q-learning have a long
Externí odkaz:
http://arxiv.org/abs/2207.12217
Autor:
Lim, Han-Dong, Lee, Donghwan
Q-learning is widely used algorithm in reinforcement learning community. Under the lookup table setting, its convergence is well established. However, its behavior is known to be unstable with the linear function approximation case. This paper develo
Externí odkaz:
http://arxiv.org/abs/2202.05404
Autor:
Lim, Han Yin, Dolzhenko, Anton V.
Publikováno v:
In European Journal of Medicinal Chemistry 5 October 2024 276
Publikováno v:
In Ophthalmology Science September-October 2024 4(5)