共 15 条
- [1] Does Pre-training Induce Systematic Inference? How Masked Language Models Acquire Commonsense Knowledge NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 4550 - 4557
- [2] Evaluation of pre-training large language models on leadership-class supercomputers JOURNAL OF SUPERCOMPUTING, 2023, 79 (18): : 20747 - 20768
- [3] Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion APPLIED SCIENCES-BASEL, 2023, 13 (09):
- [4] Evaluation of pre-training large language models on leadership-class supercomputers The Journal of Supercomputing, 2023, 79 : 20747 - 20768
- [5] How Much Do Modifications to Transformer Language Models Affect Their Ability to Learn Linguistic Knowledge? PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 46 - 53
- [6] Pre-Wiring and pre-training: what does a neural network need to learn truly general identity rules? 1600, AI Access Foundation (61):
- [7] SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2134 - 2146
- [8] Pre-Wiring and Pre-Training: What Does a Neural Network Need to Learn Truly General Identity Rules? JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2018, 61 : 927 - 946
- [9] WuDaoCorpora: A super large-scale Chinese corpora for pre-training language models AI OPEN, 2021, 2 : 65 - 68
- [10] A Conceptual Framework for Subdomain Specific Pre-Training of Large Language Models for Green Claim Detection EUROPEAN JOURNAL OF SUSTAINABLE DEVELOPMENT, 2023, 12 (04): : 319 - 329