Aligning Actions and Walking to LLM-Generated Textual Descriptions

被引:0
|
作者
Chivereanu, Radu [1 ]
Cosma, Adrian [1 ]
Catruna, Andy [1 ]
Rughinis, Razvan [1 ]
Radoi, Emilian [1 ]
机构
[1] Natl Univ Sci & Technol Politehn Bucharest, Bucharest, Romania
关键词
D O I
10.1109/FG59268.2024.10581994
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) have demonstrated remarkable capabilities in various domains, including data augmentation and synthetic data generation. This work explores the use of LLMs to generate rich textual descriptions for motion sequences, encompassing both actions and walking patterns. We leverage the expressive power of LLMs to align motion representations with high-level linguistic cues, addressing two distinct tasks: action recognition and retrieval of walking sequences based on appearance attributes. For action recognition, we employ LLMs to generate textual descriptions of actions in the BABEL-60 dataset, facilitating the alignment of motion sequences with linguistic representations. In the domain of gait analysis, we investigate the impact of appearance attributes on walking patterns by generating textual descriptions of motion sequences from the DenseGait dataset using LLMs. These descriptions capture subtle variations in walking styles influenced by factors such as clothing choices and footwear. Our approach demonstrates the potential of LLMs in augmenting structured motion attributes and aligning multi-modal representations. The findings contribute to the advancement of comprehensive motion understanding and open up new avenues for leveraging LLMs in multi-modal alignment and data augmentation for motion analysis. We make the code publicly available at https://github.com/Radu1999/WalkAndText
引用
收藏
页数:7
相关论文
共 41 条
  • [1] LLM-generated Explanations for Recommender Systems
    Lubos, Sebastian
    Tran, Thi Ngoc Trang
    Felfernig, Alexander
    Erdeniz, Seda Polat
    Le, Viet-Man
    ADJUNCT PROCEEDINGS OF THE 32ND ACM CONFERENCE ON USER MODELING, ADAPTATION AND PERSONALIZATION, UMAP 2024, 2024, : 276 - 285
  • [2] The Science of Detecting LLM-Generated Text
    Tang, Ruixiang
    Chuang, Yu-Neng
    Hu, Xia
    COMMUNICATIONS OF THE ACM, 2024, 67 (04) : 47 - 56
  • [3] Analyzing Students' Preferences for LLM-Generated Analogies
    Bernstein, Seth
    Denny, Paul
    Leinonen, Juho
    Littlefield, Matt
    Hellas, Arto
    MacNeil, Stephen
    PROCEEDINGS OF THE 2024 CONFERENCE INNOVATION AND TECHNOLOGY IN COMPUTER SCIENCE EDUCATION, VOL 2, ITICSE 2024, 2024, : 812 - 812
  • [4] A Performance Study of LLM-Generated Code on Leetcode
    Coignion, Tristan
    Quinton, Clement
    Rouvoy, Romain
    PROCEEDINGS OF 2024 28TH INTERNATION CONFERENCE ON EVALUATION AND ASSESSMENT IN SOFTWARE ENGINEERING, EASE 2024, 2024, : 79 - 89
  • [5] Meaning by Courtesy: LLM-Generated Texts and the Illusion of Content
    Ostertag, Gary
    AMERICAN JOURNAL OF BIOETHICS, 2023, 23 (10): : 91 - 93
  • [6] A Field Guide to Automatic Evaluation of LLM-Generated Summaries
    van Schaik, Tempest A.
    Pugh, Brittany
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2832 - 2836
  • [7] Lost in Transformation: Rediscovering LLM-Generated Campaigns in Social Media
    Grimme, Britta
    Pohl, Janina
    Winkelmann, Hendrik
    Stampe, Lucas
    Grimme, Christian
    DISINFORMATION IN OPEN ONLINE MEDIA, MISDOOM 2023, 2023, 14397 : 72 - 87
  • [8] Contrasting Linguistic Patterns in Human and LLM-Generated News Text
    Munoz-Ortiz, Alberto
    Gomez-Rodriguez, Carlos
    Vilares, David
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (09)
  • [9] AudiLens: Configurable LLM-Generated Audiences for Public Speech Practice
    Park, Jeongeon
    Choi, DaEun
    ADJUNCT PROCEEDINGS OF THE 36TH ANNUAL ACM SYMPOSIUM ON USER INTERFACE SOFTWARE & TECHNOLOGY, UIST 2023 ADJUNCT, 2023,
  • [10] Showing LLM-Generated Code Selectively Based on Confidence of LLMs
    School of Computer Science, Peking University, China
    不详
    arXiv,