Pre-training transformer with dual-branch context content module for table detection in document images

被引:0
|
作者
Li, Yongzhi [1 ]
Zhang, Pengle [1 ]
Sun, Meng [2 ]
Huang, Jin [1 ,3 ]
He, Ruhan [1 ,3 ]
机构
[1] School of Computer Science and Artificial Intelligence, Wuhan Textile University, Wuhan,430064, China
[2] School of Computer Science, South-Central Minzu University, Wuhan,430064, China
[3] Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion, Wuhan Textile University, Wuhan,430064, China
来源
Virtual Reality and Intelligent Hardware | 2024年 / 6卷 / 05期
关键词
Deformable convolution - Dilated convolution - Document image analysis - Document images - Features extraction - Features fusions - Pre-training - Shape and size - Table detection - Transformer;
D O I
10.1016/j.vrih.2024.06.003
中图分类号
学科分类号
摘要
33
引用
收藏
页码:408 / 420
相关论文
共 47 条
  • [1] Pre-training transformer with dual-branch context content module for table detection in document images
    Yongzhi LI
    Pengle ZHANG
    Meng SUN
    Jin HUANG
    Ruhan HE
    虚拟现实与智能硬件(中英文), 2024, 6 (05) : 408 - 420
  • [2] Dual-branch dilated context convolutional for table detection transformer in the document images
    Ni, Ying
    Wang, Xiaoli
    Peng, Hanghang
    Li, Yonzhi
    Wang, Jinyang
    Li, Haoxuan
    Huang, Jin
    VISUAL COMPUTER, 2025, 41 (04): : 2709 - 2720
  • [3] MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance Detection
    Matero, Matthew
    Soni, Nikita
    Balasubramanian, Niranjan
    Schwartz, H. Andrew
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2959 - 2966
  • [4] Context-Interactive Pre-Training for Document Machine Translation
    Yang, Pengcheng
    Zhang, Pei
    Chen, Boxing
    Xie, Jun
    Luo, Weihua
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3589 - 3595
  • [5] DiT: Self-supervised Pre-training for Document Image Transformer
    Li, Junlong
    Xu, Yiheng
    Lv, Tengchao
    Cui, Lei
    Zhang, Cha
    Wei, Furu
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3530 - 3539
  • [6] Joint Classification of Hyperspectral Images and LiDAR Data Based on Dual-Branch Transformer
    Wang, Qingyan
    Zhou, Binbin
    Zhang, Junping
    Xie, Jinbao
    Wang, Yujing
    SENSORS, 2024, 24 (03)
  • [7] HTTD: A Hierarchical Transformer for Accurate Table Detection in Document Images
    Kasem, Mahmoud SalahEldin
    Mahmoud, Mohamed
    Yagoub, Bilel
    Senussi, Mostafa Farouk
    Abdalla, Mahmoud
    Kang, Hyun-Soo
    MATHEMATICS, 2025, 13 (02)
  • [8] Context-Aware Transformer Pre-Training for Answer Sentence Selection
    Di Liello, Luca
    Garg, Siddhant
    Moschitti, Alessandro
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 458 - 468
  • [9] Long Document Extractive Summarization Method Based on Pre-training Model and Transformer
    Zhou, Xinxin
    Guo, Yuechen
    Huang, Yuning
    Yan, Yuming
    Li, Maoyuan
    Journal of Network Intelligence, 2023, 8 (03): : 913 - 931
  • [10] Ship Recognition for Complex SAR Images via Dual-Branch Transformer Fusion Network
    Sun, Zhongzhen
    Leng, Xiangguang
    Zhang, Xianghui
    Xiong, Boli
    Ji, Kefeng
    Kuang, Gangyao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21