共 50 条
- [31] Do Syntax Trees Help Pre-trained Transformers Extract Information? 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 2647 - 2661
- [32] Unsupervised Out-of-Domain Detection via Pre-trained Transformers 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1052 - 1061
- [33] On Checking Robustness on Named Entity Recognition with Pre-trained Transformers Models BALTIC JOURNAL OF MODERN COMPUTING, 2023, 11 (04): : 591 - 606
- [34] Syntax-BERT: Improving Pre-trained Transformers with Syntax Trees 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3011 - 3020
- [36] Logical Transformers: Infusing Logical Structures into Pre-Trained Language Models FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1762 - 1773
- [37] Finding and Editing Multi-Modal Neurons in Pre-Trained Transformers FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1012 - 1037
- [39] Harnessing Pre-Trained Sentence Transformers for Offensive Language Detection in Indian Languages CEUR Workshop Proc., (427-434):