共 50 条
- [41] VLMO: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
- [42] Enhancing Vision-Language Pre-Training with Jointly Learned Questioner and Dense Captioner PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5120 - 5131
- [43] MAP: Multimodal Uncertainty-Aware Vision-Language Pre-training Model 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 23262 - 23271
- [44] Automated Bridge Inspection Image Interpretation Based on Vision-Language Pre-Training COMPUTING IN CIVIL ENGINEERING 2023-DATA, SENSING, AND ANALYTICS, 2024, : 1 - 8
- [46] Leveraging per Image-Token Consistency for Vision-Language Pre-training 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 19155 - 19164
- [47] GilBERT: Generative Vision-Language Pre-Training for Image-Text Retrieval SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 1379 - 1388
- [48] Enhancing Vision-Language Pre-Training with Jointly Learned Questioner and Dense Captioner MM 2023 - Proceedings of the 31st ACM International Conference on Multimedia, 2023, : 5120 - 5131
- [49] Multimodal detection of hateful memes by applying a vision-language pre-training model PLOS ONE, 2022, 17 (09):
- [50] Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-training 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2136 - 2148