no code implementations • 5 Mar 2024 • Trang H. Tran, Quoc Tran-Dinh, Lam M. Nguyen
The Stochastic Gradient Descent method (SGD) and its stochastic variants have become methods of choice for solving finite-sum optimization problems arising from machine learning and data science thanks to their ability to handle large-scale applications and big datasets.
no code implementations • 21 Nov 2023 • Trang H. Tran, Lam M. Nguyen, Kyongmin Yeo, Nam Nguyen, Roman Vaculin
Foundation models have recently gained attention within the field of machine learning thanks to its efficiency in broad data processing.
1 code implementation • 11 Jun 2023 • Anh Duy Nguyen, Trang H. Tran, Hieu H. Pham, Phi Le Nguyen, Lam M. Nguyen
Representation learning for time series has been an important research area for decades.
no code implementations • 1 Jun 2023 • Trang H. Tran, Lam M. Nguyen, Kyongmin Yeo, Nam Nguyen, Dzung Phan, Roman Vaculin, Jayant Kalagnanam
Time series forecasting using historical data has been an interesting and challenging topic, especially when the data is corrupted by missing values.
no code implementations • 21 Jun 2022 • Trang H. Tran, Lam M. Nguyen, Katya Scheinberg
In this work, we investigate the optimization aspects of the queueing model as a RL environment and provide insight to learn the optimal policy efficiently.
1 code implementation • 7 Feb 2022 • Trang H. Tran, Katya Scheinberg, Lam M. Nguyen
This rate is better than that of any other shuffling gradient methods in convex regime.
no code implementations • 7 Feb 2022 • Lam M. Nguyen, Trang H. Tran, Marten van Dijk
How and under what assumptions is guaranteed convergence to a \textit{global} minimum possible?
no code implementations • 29 Sep 2021 • Lam M. Nguyen, Trang H. Tran, Marten van Dijk
How and under what assumptions is guaranteed convergence to a \textit{global} minimum possible?
no code implementations • 24 Nov 2020 • Trang H. Tran, Lam M. Nguyen, Quoc Tran-Dinh
When the shuffling strategy is fixed, we develop another new algorithm that is similar to existing momentum methods, and prove the same convergence rates for this algorithm under the $L$-smoothness and bounded gradient assumptions.