共 50 条
- [21] Twins: Revisiting the Design of Spatial Attention in Vision Transformers ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
- [23] From Multimodal to Unimodal Attention in Transformers using Knowledge Distillation 2021 17TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS 2021), 2021,
- [24] Are You Paying Attention? Multimodal Linear Attention Transformers for Affect Prediction in Video Conversations PROCEEDINGS OF THE 2ND INTERNATIONAL WORKSHOP ON MULTIMODAL AND RESPONSIBLE AFFECTIVE COMPUTING, MRAC 2024, 2024, : 15 - 23
- [25] Accumulated Trivial Attention Matters in Vision Transformers on Small Datasets 2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 3973 - 3981
- [26] An Attention-Based Token Pruning Method for Vision Transformers ROUGH SETS, IJCRS 2022, 2022, 13633 : 274 - 288
- [27] RAWAtten: Reconfigurable Accelerator for Window Attention in Hierarchical Vision Transformers 2023 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2023,
- [28] Focal Attention for Long-Range Interactions in Vision Transformers ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
- [29] Are Vision-Language Transformers Learning Multimodal Representations? A Probing Perspective THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11248 - 11257
- [30] Vision-and-Language or Vision-for-Language? On Cross-Modal Influence in Multimodal Transformers 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9847 - 9857