CAM-Vtrans: real-time sports training utilizing multi-modal robot data

被引:0
|
作者
Hong, LinLin [1 ]
Lee, Sangheang [1 ]
Song, GuanTing [2 ]
机构
[1] Jeonju Univ, Coll Phys Educ, Jeonju, Jeonrabug Do, South Korea
[2] Gongqing Inst Sci & Technol, Jiujiang, Jiangxi, Peoples R China
来源
关键词
assistive robotics; human-machine interaction; balance control; movement recovery; vision-transformer; CLIP; cross-attention; REPRESENTATION; CLASSIFICATION;
D O I
10.3389/fnbot.2024.1453571
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Introduction Assistive robots and human-robot interaction have become integral parts of sports training. However, existing methods often fail to provide real-time and accurate feedback, and they often lack integration of comprehensive multi-modal data.Methods To address these issues, we propose a groundbreaking and innovative approach: CAM-Vtrans-Cross-Attention Multi-modal Visual Transformer. By leveraging the strengths of state-of-the-art techniques such as Visual Transformers (ViT) and models like CLIP, along with cross-attention mechanisms, CAM-Vtrans harnesses the power of visual and textual information to provide athletes with highly accurate and timely feedback. Through the utilization of multi-modal robot data, CAM-Vtrans offers valuable assistance, enabling athletes to optimize their performance while minimizing potential injury risks. This novel approach represents a significant advancement in the field, offering an innovative solution to overcome the limitations of existing methods and enhance the precision and efficiency of sports training programs.
引用
收藏
页数:15
相关论文
共 50 条
  • [11] Multi-modal recognition control system for real-time robot welding penetration control and quality enhancement
    Ji, Tao
    Nor, Norzalilah Mohamad
    Abdullah, Ahmad Baharuddin
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2024, 135 (9-10): : 4359 - 4378
  • [12] Delivering real-time multi-modal materials analysis with enterprise beamlines
    Maffettone, Phillip M.
    Campbell, Stuart
    Hanwell, Marcus D.
    Wilkins, Stuart
    Olds, Daniel
    CELL REPORTS PHYSICAL SCIENCE, 2022, 3 (11):
  • [13] Real-Time Multi-Modal Sensing and Feedback for Catheterization in Porcine Tissue
    Heunis, Christoff M.
    Suligoj, Filip
    Fambuena Santos, Carlos
    Misra, Sarthak
    SENSORS, 2021, 21 (01) : 1 - 21
  • [14] Multi-modal Advanced Technology for Real-Time Information eXchange (MATRIX)
    Bridges, David
    Mostashfi, Shervin
    NAECON 2008 - IEEE NATIONAL AEROSPACE AND ELECTRONICS CONFERENCE, 2008, : 6 - 9
  • [15] Real-Time Multi-Modal Semantic Fusion on Unmanned Aerial Vehicles
    Bultmann, Simon
    Quenzel, Jan
    Behnke, Sven
    10TH EUROPEAN CONFERENCE ON MOBILE ROBOTS (ECMR 2021), 2021,
  • [16] Aucher: Multi-modal Queries on Live Audio Streams in Real-time
    Wen, Zeyi
    Liang, Mingyu
    He, Bingsheng
    Xia, Zexin
    Li, Bo
    2019 IEEE 35TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2019), 2019, : 1960 - 1963
  • [17] Impact of smartphone-delivered real-time multi-modal information
    Gan, Hongcheng
    Zhao, YaFei
    Wei, June
    INTERNATIONAL JOURNAL OF MOBILE COMMUNICATIONS, 2016, 14 (03) : 244 - 255
  • [18] Integrating multi-modal data into AFSA-LSTM model for real-time oil production prediction
    Jiang, Wei
    Wang, Xin
    Zhang, Shu
    ENERGY, 2023, 279
  • [19] Real-time Emotion Pre-Recognition in Conversations with Contrastive Multi-modal Dialogue Pre-training
    Ju, Xincheng
    Zhang, Dong
    Zhu, Suyang
    Li, Junhui
    Li, Shoushan
    Zhou, Guodong
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 1045 - 1055
  • [20] Real-time deformable registration of multi-modal whole slides for digital pathology
    Mueller, Dan
    Vossen, Dirk
    Hulsken, Bas
    COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2011, 35 (7-8) : 542 - 556