共 24 条
- [1] V e r y d e e p convolutional networks for large-scale image recognition, 3rd International Conference on Learning Representations, (2015)
- [2] HE Kaiming, ZHANG Xiangyu, REN Shaoqing, Et al., Deep residual learning for image recognition[C], 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, (2016)
- [3] SZEGEDY C, LIU Wei, JIA Yangqing, Et al., Going deeper with convolutions[C], 2015 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-9, (2015)
- [4] VASWANI A, SHAZEER N, PARMAR N, Et al., Attention is all you need[C], The 31st International Conference on Neural Information Processing Systems, pp. 6000-6010, (2017)
- [5] CARION N, MASSA F, SYNNAEVE G, Et al., End-to-end object detection with transformers[C], The 16th European Conference on Computer Vision, pp. 213-229, (2020)
- [6] CHEN Ying, KUANG Cheng, Pedestrian re-identification based on CNN and Transformer multi-scale learning[J], Journal of Electronics & Information Technology, 45, 6, pp. 2256-2263, (2023)
- [7] ZHAI Xiaohua, KOLESNIKOV A, HOULSBY N, Et al., Scaling vision transformers[C], 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1204-1213, (2022)
- [8] DOSOVITSKIY A, BEYER L, KOLESNIKOV A, Et al., An image is worth 16x16 words: transformers for image recognition at scale, 9th International Conference on Learning Representations, (2021)
- [9] WANG Teng, GONG Lei, WANG Chao, Et al., ViA: A novel vision-transformer accelerator based on FPGA[J], IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 41, 11, pp. 4088-4099, (2022)
- [10] NAG S, DATTA G, KUNDU S, Et al., ViTA: A vision transformer inference accelerator for edge applications[C], 2023 IEEE International Symposium on Circuits and Systems, pp. 1-5, (2023)