Self-attention Deep Field-embedded Factorization Machine for Click-through Rate Prediction

被引:0
|
作者
Li, Guangli [1 ]
Ye, Yiyuan [1 ]
Xu, Guangxin [1 ]
Zhang, Hongbin [1 ]
Wu, Guangting [1 ]
Lyu, Jingqin [1 ]
机构
[1] School of Info. and Software Eng., East China Jiaotong Univ., Nanchang,330013, China
关键词
Deep neural networks;
D O I
10.12454/j.jsuese.202201373
中图分类号
学科分类号
摘要
Objective Click-through rate (CTR) prediction realizes accurate recommendation of digital advertisements by predicting the user's click probability on advertisements or commodities. However, current CTR prediction models have the following key issues. First, the raw embedding vectors have not been fully refined. Second, the corresponding feature interaction method is too simple. As a result, the performance of the models is heavily restricted. To alleviate these issues, a novel CTR model named self-attention deep field-embedded factorization machine (Self-AtDFEFM) is proposed. Methods First, a well-known multi-head self-attention mechanism is employed to capture the implicit information of the raw embedding vectors on different sub-spaces, and the corresponding weight is calculated to further refine the key low-level features. Second, a novel field-embedded factorization machine (FEFM) is designed to strengthen the interaction intensity between different feature fields by the field pair symmetric matrix. The key low-order feature combinations are fully optimized by the FEFM module for the subsequent high-order feature interaction. Third, a deep neural network (DNN) is built based on the low-order feature combinations to complete implicit high-order feature interaction. Finally, both the explicit and implicit feature interactions are combined together to implement CTR prediction. Results and Discussions Extensive experiments have been performed on the two public available datasets, namely Criteo and Avazu. First, the proposed Self-AtDFEFM is compared with numerous state-of-the-art baselines on the AUC (area under curve) and LogLoss metrics. Second, all parameters of Self-AtDFEFM was tuned, and the parameters included the number of the explicit high-order feature interaction layers, the number of the attention heads, the embedding dimension, and the number of the implicit high-order feature interaction layers. Further, ablation experiments of our model were completed. The results of the experiments showed that: the Self-AtDFEFM model outperformed mainstream baseline models on the AUC and LogLoss metrics; all parameters of Self-AtDFEFM have been adjusted to their optimal values; each module form a kind of joint force to improve the final CTR prediction performance. Notably, the explicit high-order feature interaction layer plays the most important role in Self-AtDFEFM. Conclusions Each module of Self-AtDFEFM is plug-and-play, that is, the Self-AtDFEFM is easier to build and deploy. Hence, Self-AtDFEFM achieves a good trade-off between prediction performance and model complexity, making it highly practical. © 2024 Sichuan University. All rights reserved.
引用
收藏
页码:287 / 296
相关论文
共 50 条
  • [1] Field-Aware Neural Factorization Machine for Click-Through Rate Prediction
    Zhang, Li
    Shen, Weichen
    Huang, Jianhang
    Li, Shijian
    Pan, Gang
    IEEE ACCESS, 2019, 7 : 75032 - 75040
  • [2] Click-Through Rate Prediction of Multi-Head Self-Attention in Hyperbolic Space
    Han Y.-L.
    Wang X.-Y.
    Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications, 2021, 44 (05): : 127 - 132
  • [3] Deep Field-Aware Interaction Machine for Click-Through Rate Prediction
    Qi, Gaofeng
    Li, Ping
    MOBILE INFORMATION SYSTEMS, 2021, 2021
  • [4] Disentangled self-attention neural network based on information sharing for click-through rate prediction
    Wang, Yingqi
    Ji, Huiqin
    He, Xin
    Yu, Junyang
    Han, Hongyu
    Zhai, Rui
    Wang, Longge
    PEERJ COMPUTER SCIENCE, 2024, 10 : 1 - 22
  • [5] Deep Interest with Hierarchical Attention Network for Click-Through Rate Prediction
    Xu, Weinan
    He, Hengxu
    Tan, Minshi
    Li, Yunming
    Lang, Jun
    Guo, Dongbai
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1905 - 1908
  • [6] Sparse Factorization Machines for Click-through Rate Prediction
    Pan, Zhen
    Chen, Enhong
    Liu, Qi
    Xu, Tong
    Ma, Haiping
    Lin, Hongjie
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2016, : 400 - 409
  • [7] Deep Spatio-Temporal Attention Network for Click-Through Rate Prediction
    Li, Xin-Lu
    Gao, Peng
    Lei, Yuan-Yuan
    Zhang, Le-Xuan
    Fang, Liang-Kuan
    INTELLIGENT COMPUTING METHODOLOGIES, PT III, 2022, 13395 : 626 - 638
  • [8] Deep Field Relation Neural Network for click-through rate prediction
    Zou, Dafang
    Wang, Zidong
    Zhang, Leimin
    Zou, Jinting
    Li, Qi
    Chen, Yun
    Sheng, Weiguo
    INFORMATION SCIENCES, 2021, 577 : 128 - 139
  • [9] Interpretable Click-Through Rate Prediction through Hierarchical Attention
    Li, Zeyu
    Cheng, Wei
    Chen, Yang
    Chen, Haifeng
    Wang, Wei
    PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM '20), 2020, : 313 - 321
  • [10] Field-weighted Factorization Machines for Click-Through Rate Prediction in Display Advertising
    Pan, Junwei
    Xu, Jian
    Ruiz, Alfonso Lobos
    Zhao, Wenliang
    Pan, Shengjun
    Sun, Yu
    Lu, Quan
    WEB CONFERENCE 2018: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW2018), 2018, : 1349 - 1357