共 50 条
- [1] COOKIE: Contrastive Cross-Modal Knowledge Sharing Pre-training for Vision-Language Representation [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 2188 - 2197
- [2] VLMixer: Unpaired Vision-Language Pre-training via Cross-Modal CutMix [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
- [3] Vision Language Pre-training by Contrastive Learning with Cross-Modal Similarity Regulation [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 14660 - 14679
- [4] CMAL: A Novel Cross-Modal Associative Learning Framework for Vision-Language Pre-Training [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 4515 - 4524
- [5] COTS: Collaborative Two-Stream Vision-Language Pre-Training Model for Cross-Modal Retrieval [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15671 - 15680
- [6] Contrastive Vision-Language Pre-training with Limited Resources [J]. COMPUTER VISION, ECCV 2022, PT XXXVI, 2022, 13696 : 236 - 253
- [7] Vision-Language Pre-Training with Triple Contrastive Learning [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15650 - 15659
- [9] PiTL: Cross-modal Retrieval with Weakly-supervised Vision-language Pre-training via Prompting [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2261 - 2265
- [10] Unifying Cross-Lingual and Cross-Modal Modeling Towards Weakly Supervised Multilingual Vision-Language Pre-training [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5939 - 5958