Efficient and Scalable Multi-Task Regression on Massive Number of Tasks

被引:0
|
作者
He, Xiao [1 ]
Alesiani, Francesco [1 ]
Shaker, Ammar [1 ]
机构
[1] NEC Labs Europe, Heidelberg, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many real-world large-scale regression problems can be formulated as Multi-task Learning (MTL) problems with a massive number of tasks, as in retail and transportation domains. However, existing MTL methods still fail to offer both the generalization performance and the scalability for such problems. Scaling up MTL methods to problems with a tremendous number of tasks is a big challenge. Here, we propose a novel algorithm, named Convex Clustering Multi-Task regression Learning (CCMTL), which integrates with convex clustering on the k-nearest neighbor graph of the prediction models. Further, CCMTL efficiently solves the underlying convex problem with a newly proposed optimization method. CCMTL is accurate, efficient to train, and empirically scales linearly in the number of tasks. On both synthetic and real-world datasets, the proposed CCMTL outperforms seven state-of-the-art (SoA) multi-task learning methods in terms of prediction accuracy as well as computational efficiency. On a real-world retail dataset with 23; 812 tasks, CCMTL requires only around 30 seconds to train on a single thread, while the SoA methods need up to hours or even days.
引用
下载
收藏
页码:3763 / 3770
页数:8
相关论文
共 50 条
  • [1] Dominantly Truthful Multi-task Peer Prediction with a Constant Number of Tasks
    Kong, Yuqing
    PROCEEDINGS OF THE THIRTY-FIRST ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS (SODA'20), 2020, : 2398 - 2411
  • [2] Dominantly Truthful Multi-task Peer Prediction with a Constant Number of Tasks
    Kong, Yuqing
    PROCEEDINGS OF THE 2020 ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, SODA, 2020, : 2398 - 2411
  • [3] Massive Choice, Ample Tasks (MACHAMP): A Toolkit for Multi-task Learning in NLP
    van der Goot, Rob
    Ustun, Ahmet
    Ramponi, Alan
    Sharaf, Ibrahim
    Plank, Barbara
    EACL 2021: THE 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: PROCEEDINGS OF THE SYSTEM DEMONSTRATIONS, 2021, : 176 - 197
  • [4] Scalable efficient reproducible multi-task learning via data splitting
    Wen, Xin
    Li, Yang
    Zheng, Zemin
    STATISTICS & PROBABILITY LETTERS, 2024, 208
  • [5] Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks
    Liu, Yen-Cheng
    Ma, Chih-Yao
    Tian, Junjiao
    He, Zijian
    Kira, Zsolt
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [6] Scalable Privacy in Multi-Task Image Compression
    Alvar, Saeed Ranjbar
    Bajic, Ivan, V
    2021 INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP), 2021,
  • [7] Multi-task ordinal regression with task weight discovery
    Xiao, Yanshan
    Zeng, Mengyue
    Liu, Bo
    Zhao, Liang
    Kong, Xiangjun
    Hao, Zhifeng
    Knowledge-Based Systems, 2024, 305
  • [8] Multi-task Learning with Labeled and Unlabeled Tasks
    Pentina, Anastasia
    Lampert, Christoph H.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [9] Multi-task classification with sequential instances and tasks
    Xu, Wei
    Liu, Wei
    Chi, Haoyuan
    Huang, Xiaolin
    Yang, Jie
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2018, 64 : 59 - 67
  • [10] TASK AWARE MULTI-TASK LEARNING FOR SPEECH TO TEXT TASKS
    Indurthi, Sathish
    Zaidi, Mohd Abbas
    Lakumarapu, Nikhil Kumar
    Lee, Beomseok
    Han, Hyojung
    Ahn, Seokchan
    Kim, Sangha
    Kim, Chanwoo
    Hwang, Inchul
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7723 - 7727