Zobrazeno 1 - 10
of 173
pro vyhledávání: '"Pham, Nhan"'
Autor:
Lobo, Elita, Hassanzadeh, Oktie, Pham, Nhan, Mihindukulasooriya, Nandana, Subramanian, Dharmashankar, Samulowitz, Horst
Enterprises often own large collections of structured data in the form of large databases or an enterprise data lake. Such data collections come with limited metadata and strict access policies that could limit access to the data contents and, theref
Externí odkaz:
http://arxiv.org/abs/2309.11506
In recent years, a proliferation of methods were developed for cooperative multi-agent reinforcement learning (c-MARL). However, the robustness of c-MARL agents against adversarial attacks has been rarely explored. In this paper, we propose to evalua
Externí odkaz:
http://arxiv.org/abs/2202.03558
Autor:
Alwood, Benjamin T., Meyer, Dawn M., Ionita, Chip, Snyder, Kenneth V., Santos, Roberta, Perrotta, Lindsey, Crooks, Ryan, Van Orden, Kimberlee, Torres, Dolores, Poynor, Briana, Pham, Nhan, Kelly, Sophie, Meyer, Brett C., Bolar, Divya S.
Publikováno v:
In Journal of Stroke and Cerebrovascular Diseases July 2024 33(7)
Autor:
van Thi, Khoa L., van Emmerik, Tim H.M., Vermeulen, Bart, Pham, Nhan Q., Hoitink, A.J.F. (Ton)
Publikováno v:
In Environmental Pollution 15 March 2024 345
Publikováno v:
NeurIPs 2021
We develop two new algorithms, called, FedDR and asyncFedDR, for solving a fundamental nonconvex composite optimization problem in federated learning. Our algorithms rely on a novel combination between a nonconvex Douglas-Rachford splitting method, r
Externí odkaz:
http://arxiv.org/abs/2103.03452
Motivated by broad applications in reinforcement learning and machine learning, this paper considers the popular stochastic gradient descent (SGD) when the gradients of the underlying objective function are sampled from Markov processes. This Markov
Externí odkaz:
http://arxiv.org/abs/2003.10973
Autor:
Pham, Nhan H., Nguyen, Lam M., Phan, Dzung T., Nguyen, Phuong Ha, van Dijk, Marten, Tran-Dinh, Quoc
Publikováno v:
Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR 108:374-385, 2020
We propose a novel hybrid stochastic policy gradient estimator by combining an unbiased policy gradient estimator, the REINFORCE estimator, with another biased one, an adapted SARAH estimator for policy optimization. The hybrid policy gradient estima
Externí odkaz:
http://arxiv.org/abs/2003.00430
Publikováno v:
ICML 2020
We develop two new stochastic Gauss-Newton algorithms for solving a class of non-convex stochastic compositional optimization problems frequently arising in practice. We consider both the expectation and finite-sum settings under standard assumptions
Externí odkaz:
http://arxiv.org/abs/2002.07290
Convergence Rates of Accelerated Markov Gradient Descent with Applications in Reinforcement Learning
Motivated by broad applications in machine learning, we study the popular accelerated stochastic gradient descent (ASGD) algorithm for solving (possibly nonconvex) optimization problems. We characterize the finite-time performance of this method when
Externí odkaz:
http://arxiv.org/abs/2002.02873
Akademický článek
Tento výsledek nelze pro nepřihlášené uživatele zobrazit.
K zobrazení výsledku je třeba se přihlásit.
K zobrazení výsledku je třeba se přihlásit.