An Asynchronously Alternative Stochastic Gradient Descent Algorithm for Efficiently Parallel Latent Feature Analysis on Shared-Memory

被引:2
|
作者
Qin, Wen [1 ,2 ,3 ]
Luo, Xin [4 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Comp Sci & Technol, Chongqing 400065, Peoples R China
[2] Chinese Acad Sci, Chongqing Inst Green & Intelligent Technol, Chongqing 400714, Peoples R China
[3] Univ Chinese Acad Sci, Chongqing Sch, Chongqing 400714, Peoples R China
[4] Southwest Univ, Coll Comp & Informat Sci, Chongqing 400715, Peoples R China
基金
中国国家自然科学基金;
关键词
Data Science; Parallel Stochastic Gradient Descent; Multicore; Latent Feature Analysis; Asynchronously Alternative Stochastic Gradient Descent; Convergence; Shared-Memory; High-Dimensional Incomplete data; SPARSE MATRICES; FACTOR MODEL; RECOMMENDER;
D O I
10.1109/ICKG55886.2022.00035
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A latent feature analysis (LFA) model is highly efficient in performing representation learning to high-dimensional incomplete (HDI) data like user-item interactions data from a recommender system. Stochastic gradient descent (SGD) is frequently adopted as the learning algorithm by an LFA model owing to its low computational complexity. However, a standard SGD algorithm is in nature a serial algorithm, which affects the resultant LFA model's scalability on mass HDI data. On the other hand, existing parallel SGD algorithms commonly suffer from low speedup when building an LFA model due to their frequent synchronizations during the training process. Motivated by this discovery, this paper proposes an Asynchronously Alternative Stochastic Gradient Descent (A(2)SGD) algorithm to achieve an efficiently parallelized LFA model on shared-memory with two fold-ideas: a) adopting the principle of an alternative stochastic gradient descent algorithm to decouple the LFA process, thereby achieving two parallelizable subtasks with minimal learning information loss; b) designing a novel parallelization scheme by eliminating synchronizations from both the subtask and the thread perspectives, i.e., both sub-tasks are taken simultaneously, and their affiliated threads also executes without synchronizations. Rigorously theoretical convergence proof illustrates that the newly-proposed parallelization scheme guarantees the convergence of a resultant LFA model. Detailed experimental results on four real HDI datasets indicate that an A(2)SGD-based LFA model outperforms several state-of-the-art parallel SGD-based LFA models in terms of both missing data estimation accuracy and parallelization speedup.
引用
收藏
页码:217 / 224
页数:8
相关论文
共 50 条
  • [1] Shared-memory and shared-nothing stochastic gradient descent algorithms for matrix completion
    Makari, Faraz
    Teflioudi, Christina
    Gemulla, Rainer
    Haas, Peter
    Sismanis, Yannis
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2015, 42 (03) : 493 - 523
  • [2] Shared-memory and shared-nothing stochastic gradient descent algorithms for matrix completion
    Faraz Makari
    Christina Teflioudi
    Rainer Gemulla
    Peter Haas
    Yannis Sismanis
    [J]. Knowledge and Information Systems, 2015, 42 : 493 - 523
  • [3] Fast brain tumor detection using adaptive stochastic gradient descent on shared-memory parallel environment
    Qin, Chuandong
    Li, Baosheng
    Han, Baole
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 120
  • [4] High Performance Parallel Stochastic Gradient Descent in Shared Memory
    Sallinen, Scott
    Satish, Nadathur
    Smelyanskiy, Mikhail
    Sury, Samantika S.
    Re, Christopher
    [J]. 2016 IEEE 30TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2016), 2016, : 873 - 882
  • [6] The Convergence of Stochastic Gradient Descent in Asynchronous Shared Memory
    Alistarh, Dan
    De Sa, Christopher
    Konstantinov, Nikola
    [J]. PODC'18: PROCEEDINGS OF THE 2018 ACM SYMPOSIUM ON PRINCIPLES OF DISTRIBUTED COMPUTING, 2018, : 169 - 177
  • [7] TUNING A PARALLEL DATABASE ALGORITHM ON A SHARED-MEMORY MULTIPROCESSOR
    GRAEFE, G
    THAKKAR, SS
    [J]. SOFTWARE-PRACTICE & EXPERIENCE, 1992, 22 (07): : 495 - 517
  • [8] Fast Convergence Stochastic Parallel Gradient Descent Algorithm
    Hu Dongting
    Shen Wen
    Ma Wenchao
    Liu Xinyu
    Su Zhouping
    Zhu Huaxin
    Zhang Xiumei
    Que Lizhi
    Zhu Zhuowei
    Zhang Yixin
    Chen Guoqing
    Hu Lifa
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2019, 56 (12)
  • [9] A Parallel Resampling Algorithm for Particle Filtering on Shared-Memory Architectures
    Gong, Peng
    Basciftci, Yuksel Ozan
    Ozguner, Fusun
    [J]. 2012 IEEE 26TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS & PHD FORUM (IPDPSW), 2012, : 1477 - 1483
  • [10] Shared-Memory Parallel Dynamic Louvain Algorithm for Community Detection
    Sahu, Subhajit
    Kothapalli, Kishore
    Banerjee, Dip Sankar
    [J]. 2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW 2024, 2024, : 1204 - 1205