Learning shared embedding representation of motion and text using contrastive learning

被引:0
|
作者
Junpei Horie
Wataru Noguchi
Hiroyuki Iizuka
Masahito Yamamoto
机构
[1] Hokkaido University,Graduate School of Information Science and Technology
[2] Hokkaido University,Faculty of Information Science and Technology
[3] Hokkaido University,Center for Human Nature, Artificial Intelligence, and Neuroscience
来源
关键词
Multi-modal learning; Contrastive learning; Skeleton-based action recognition; Motion retrieval;
D O I
暂无
中图分类号
学科分类号
摘要
Multimodal learning of motion and text tries to find the correspondence between skeletal time-series data acquired by motion capture and the text that describes the motion. In this field, good associations can realize both motion-to-text and text-to-motion applications. However, the previous methods failed to associate motion with text, taking into account details of descriptions, for example, whether to move the left or right arm. In this paper, we propose a motion-text contrastive learning method for making correspondences between motion and text in a shared embedding space. We showed that our model outperforms the previous studies in the task of action recognition. We also qualitatively show that, by using a pre-trained text encoder, our model can perform motion retrieval with detailed correspondences between motion and text.
引用
收藏
页码:148 / 157
页数:9
相关论文
共 50 条
  • [1] Learning shared embedding representation of motion and text using contrastive learning
    Horie, Junpei
    Noguchi, Wataru
    Iizuka, Hiroyuki
    Yamamoto, Masahito
    ARTIFICIAL LIFE AND ROBOTICS, 2023, 28 (01) : 148 - 157
  • [2] Contrastive Learning with Transformer Initialization and Clustering Prior for Text Representation
    Liu, Chenjing
    Chen, Xiangru
    Hu, Peng
    Lin, Jie
    Wang, Junfeng
    Geng, Xue
    APPLIED SOFT COMPUTING, 2024, 166
  • [3] MoCoUTRL: a momentum contrastive framework for unsupervised text representation learning
    Zou, Ao
    Hao, Wenning
    Jin, Dawei
    Chen, Gang
    Sun, Feiyan
    CONNECTION SCIENCE, 2023, 35 (01)
  • [4] Description-Enhanced Label Embedding Contrastive Learning for Text Classification
    Zhang, Kun
    Wu, Le
    Lv, Guangyi
    Chen, Enhong
    Ruan, Shulan
    Liu, Jing
    Zhang, Zhiqiang
    Zhou, Jun
    Wang, Meng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 14889 - 14902
  • [5] Graph Embedding Contrastive Multi-Modal Representation Learning for Clustering
    Xia, Wei
    Wang, Tianxiu
    Gao, Quanxue
    Yang, Ming
    Gao, Xinbo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1170 - 1183
  • [6] Regularizing Visual Semantic Embedding With Contrastive Learning for Image-Text Matching
    Liu, Yang
    Liu, Hong
    Wang, Huaqiu
    Liu, Mengyuan
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1332 - 1336
  • [7] Mutual Contrastive Learning for Visual Representation Learning
    Yang, Chuanguang
    An, Zhulin
    Cai, Linhang
    Xu, Yongjun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3045 - 3053
  • [8] Masked Contrastive Representation Learning for Reinforcement Learning
    Zhu, Jinhua
    Xia, Yingce
    Wu, Lijun
    Deng, Jiajun
    Zhou, Wengang
    Qin, Tao
    Liu, Tie-Yan
    Li, Houqiang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3421 - 3433
  • [9] Molecular representation contrastive learning via transformer embedding to graph neural networks
    Liu, Yunwu
    Zhang, Ruisheng
    Li, Tongfeng
    Jiang, Jing
    Ma, Jun
    Yuan, Yongna
    Wang, Ping
    APPLIED SOFT COMPUTING, 2024, 164
  • [10] Contrastive Code Representation Learning
    Jain, Paras
    Jain, Ajay
    Zhang, Tianjun
    Abbeel, Pieter
    Gonzalez, Joseph E.
    Stoica, Ion
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5954 - 5971