SelfPAB: large-scale pre-training on accelerometer data for human activity recognition

被引:0
|
作者
Aleksej Logacjov
Sverre Herland
Astrid Ustad
Kerstin Bach
机构
[1] Norwegian University of Science and Technology,Department of Computer Science
[2] Norwegian University of Science and Technology,Department of Neuromedicine and Movement Science
来源
Applied Intelligence | 2024年 / 54卷
关键词
Accelerometer; Human activity recognition; Machine learning; Physical activity behavior; Self-supervised learning; Transformer;
D O I
暂无
中图分类号
学科分类号
摘要
Annotating accelerometer-based physical activity data remains a challenging task, limiting the creation of robust supervised machine learning models due to the scarcity of large, labeled, free-living human activity recognition (HAR) datasets. Researchers are exploring self-supervised learning (SSL) as an alternative to relying solely on labeled data approaches. However, there has been limited exploration of the impact of large-scale, unlabeled datasets for SSL pre-training on downstream HAR performance, particularly utilizing more than one accelerometer. To address this gap, a transformer encoder network is pre-trained on various amounts of unlabeled, dual-accelerometer data from the HUNT4 dataset: 10, 100, 1k, 10k, and 100k hours. The objective is to reconstruct masked segments of signal spectrograms. This pre-trained model, termed SelfPAB, serves as a feature extractor for downstream supervised HAR training across five datasets (HARTH, HAR70+, PAMAP2, Opportunity, and RealWorld). SelfPAB outperforms purely supervised baselines and other SSL methods, demonstrating notable enhancements, especially for activities with limited training data. Results show that more pre-training data improves downstream HAR performance, with the 100k-hour model exhibiting the highest performance. It surpasses purely supervised baselines by absolute F1-score improvements of 7.1% (HARTH), 14% (HAR70+), and an average of 11.26% across the PAMAP2, Opportunity, and RealWorld datasets. Compared to related SSL methods, SelfPAB displays absolute F1-score enhancements of 10.4% (HARTH), 18.8% (HAR70+), and 16% (average across PAMAP2, Opportunity, RealWorld).
引用
收藏
页码:4545 / 4563
页数:18
相关论文
共 50 条
  • [1] SelfPAB: large-scale pre-training on accelerometer data for human activity recognition
    Logacjov, Aleksej
    Herland, Sverre
    Ustad, Astrid
    Bach, Kerstin
    [J]. APPLIED INTELLIGENCE, 2024, 54 (06) : 4545 - 4563
  • [2] Pre-training on Large-Scale Heterogeneous Graph
    Jiang, Xunqiang
    Jia, Tianrui
    Fang, Yuan
    Shi, Chuan
    Lin, Zhe
    Wang, Hui
    [J]. KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 756 - 766
  • [3] Large-scale weakly-supervised pre-training for video action recognition
    Ghadiyaram, Deepti
    Du Tran
    Mahajan, Dhruv
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 12038 - 12047
  • [4] Synthetic Augmentation with Large-Scale Unconditional Pre-training
    Ye, Jiarong
    Ni, Haomiao
    Jin, Peng
    Huang, Sharon X.
    Xue, Yuan
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT II, 2023, 14221 : 754 - 764
  • [5] PreDet: Large-scale weakly supervised pre-training for detection
    Ramanathan, Vignesh
    Wang, Rui
    Mahajan, Dhruv
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2845 - 2855
  • [6] An Optimized Method for Large-Scale Pre-Training in Symbolic Music
    Liu, Shike
    Xu, Hongguang
    Xu, Ke
    [J]. Proceedings of the International Conference on Anti-Counterfeiting, Security and Identification, ASID, 2022, 2022-December : 105 - 109
  • [7] Automating Code Review Activities by Large-Scale Pre-training
    Li, Zhiyu
    Lu, Shuai
    Guo, Daya
    Duan, Nan
    Jannu, Shailesh
    Jenks, Grant
    Majumder, Deep
    Green, Jared
    Svyatkovskiy, Alexey
    Fu, Shengyu
    Sundaresan, Neel
    [J]. PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 1035 - 1047
  • [8] MusicBERT: Symbolic Music Understanding with Large-Scale Pre-Training
    Zeng, Mingliang
    Tan, Xu
    Wang, Rui
    Ju, Zeqian
    Qin, Tao
    Liu, Tie-Yan
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 791 - 800
  • [9] Editorial for Special Issue on Large-scale Pre-training: Data, Models, and Fine-tuning
    Wen, Ji-Rong
    Huang, Zi
    Zhang, Hanwang
    [J]. MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 145 - 146
  • [10] MuralDiff: Diffusion for Ancient Murals Restoration on Large-Scale Pre-Training
    Xu, Zishan
    Zhang, Xiaofeng
    Chen, Wei
    Liu, Jueting
    Xu, Tingting
    Wang, Zehua
    [J]. IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (03): : 2169 - 2181