Efficient and High-quality Recommendations via Momentum-incorporated Parallel Stochastic Gradient Descent-Based Learning
Document Type
Article
Publication Date
2-1-2021
Abstract
A recommender system (RS) relying on latent factor analysis usually adopts stochastic gradient descent (SGD) as its learning algorithm. However, owing to its serial mechanism, an SGD algorithm suffers from low efficiency and scalability when handling large-scale industrial problems. Aiming at addressing this issue, this study proposes a momentum-incorporated parallel stochastic gradient descent (MPSGD) algorithm, whose main idea is two-fold: a) implementing parallelization via a novel data-splitting strategy, and b) accelerating convergence rate by integrating momentum effects into its training process. With it, an MPSGD-based latent factor (MLF) model is achieved, which is capable of performing efficient and high-quality recommendations. Experimental results on four high-dimensional and sparse matrices generated by industrial RS indicate that owing to an MPSGD algorithm, an MLF model outperforms the existing state-of-the-art ones in both computational efficiency and scalability.
Identifier
85099413324 (Scopus)
Publication Title
IEEE Caa Journal of Automatica Sinica
External Full Text Location
https://doi.org/10.1109/JAS.2020.1003396
e-ISSN
23299274
ISSN
23299266
First Page
402
Last Page
411
Issue
2
Volume
8
Grant
61772493
Fund Ref
National Natural Science Foundation of China
Recommended Citation
Luo, Xin; Qin, Wen; Dong, Ani; Sedraoui, Khaled; and Zhou, Mengchu, "Efficient and High-quality Recommendations via Momentum-incorporated Parallel Stochastic Gradient Descent-Based Learning" (2021). Faculty Publications. 4369.
https://digitalcommons.njit.edu/fac_pubs/4369