共 50 条
- [1] U-BERT: Pre-training User Representations for Improved Recommendation [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4320 - 4327
- [3] Learning Transferable User Representations with Sequential Behaviors via Contrastive Pre-training [J]. 2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 51 - 60
- [4] Temporal Contrastive Pre-Training for Sequential Recommendation [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1925 - 1934
- [5] Advances in Pre-Training Distributed Word Representations [J]. PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 52 - 55
- [6] Pre-training Mention Representations in Coreference Models [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 8534 - 8540
- [7] Multi-Modal Contrastive Pre-training for Recommendation [J]. PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 99 - 108
- [8] Pre-training of Graph Augmented Transformers for Medication Recommendation [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5953 - 5959
- [10] Pre-Training Audio Representations With Self-Supervision [J]. IEEE SIGNAL PROCESSING LETTERS, 2020, 27 : 600 - 604