共 50 条
- [32] A survey of transformer-based multimodal pre-trained modals [J]. NEUROCOMPUTING, 2023, 515 : 89 - 106
- [34] Chemformer: a pre-trained transformer for computational chemistry [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01):
- [36] Integrally Pre-Trained Transformer Pyramid Networks [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18610 - 18620
- [37] On the Role of Pre-trained Embeddings in Binary Code Analysis [J]. PROCEEDINGS OF THE 19TH ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ACM ASIACCS 2024, 2024, : 795 - 810
- [38] Pre-trained Embeddings for Entity Resolution: An Experimental Analysis [J]. PROCEEDINGS OF THE VLDB ENDOWMENT, 2023, 16 (09): : 2225 - 2238
- [39] On the Sentence Embeddings from Pre-trained Language Models [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9119 - 9130
- [40] An integrated model based on deep learning classifiers and pre-trained transformer for phishing URL detection [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 269 - 285