A Cross-Modal Tactile Reproduction Utilizing Tactile and Visual Information Generated by Conditional Generative Adversarial Networks

被引:0
|
作者
Hatori, Koki [1 ]
Morikura, Takashi [2 ]
Funahashi, Akira [2 ]
Takemura, Kenjiro [3 ]
机构
[1] Keio Univ, Sch Sci Open & Environm Syst, Yokohama 2238522, Japan
[2] Keio Univ, Dept Biosci & Informat, Yokohama 2238522, Japan
[3] Keio Univ, Dept Mech Engn, Yokohama 2238522, Japan
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Generators; Visualization; Vectors; Training; Acoustics; Solid modeling; Generative adversarial networks; Fingers; Ultrasonic transducers; Tactile sensors; Tactile reproduction; cross-modal recognition; conditional generative adversarial networks;
D O I
10.1109/ACCESS.2025.3527946
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Tactile reproduction technology represents a promising advancement within the rapidly expanding field of virtual/augmented reality, necessitating the development of innovative methods specifically tailored to correspond with tactile sensory labels. Since human tactile perception is known to be influenced by visual information, this study has developed a cross-modal tactile sensory display using Conditional Generative Adversarial Networks, CGANs, to generate both mechanical and visual information. Initially, sensory evaluation experiments were conducted with 32 participants using twelve metal plate samples to collect tactile information. Subsequently, we prepared 320 images of variety of materials and conducted sensory evaluation experiments with 30 participants per image to gather tactile information evoked by viewing the images. Utilizing the collected tactile information, used as labels, and images as a dataset, we developed four types of visual information generation models using CGAN, each trained with weighted concatenated data of images and labels, in which image elements are amplified by factors of 1, 1,000, 5,000, and 10,000, respectively. Each of these four models was then used to generate twelve images corresponding to the sensory evaluation result of twelve different metal plate samples. We performed a cross-modal tactile reproduction experiment using the previously developed tactile information generation model to input signals to a tactile display, alongside the images generated by the visual information generation model. In this experiment, 20 subjects conducted sensory evaluations where tactile sensations were displayed concurrently with the visual display of the images. The results confirmed that the concurrent display of mechanical and visual information significantly reduced the mean absolute error between the displayed tactile information and that of the metal plate samples from 2.2 to 1.6 out of a 7-digit scale in sensory evaluation. These findings underscore the effectiveness of the visual information generation model and highlight the potential of integrating tactile and visual information for enhanced tactile reproduction systems.
引用
收藏
页码:9223 / 9229
页数:7
相关论文
共 50 条
  • [41] Visual-tactile processing in primary somatosensory cortex emerges before cross-modal experience
    Bieler, Malte
    Sieben, Kay
    Schildt, Sandra
    Roeder, Brigitte
    Hanganu-Opatz, Ileana L.
    SYNAPSE, 2017, 71 (06)
  • [42] Cross-modal responses in the primary visual cortex encode complex objects and correlate with tactile discrimination
    Vasconcelos, Nivaldo
    Pantoja, Janaina
    Belchior, Hindiael
    Caixeta, Fabio Viegas
    Faber, Jean
    Freire, Marco Aurelio M.
    Cota, Vinicius Rosa
    de Macedo, Edson Anibal
    Laplagne, Diego A.
    Gomes, Herman Martins
    Ribeiro, Sidarta
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2011, 108 (37) : 15408 - 15413
  • [43] When Visual Distractors Predict Tactile Search: The Temporal Profile of Cross-Modal Spatial Learning
    Chen, Siyi
    Shi, Zhuanghua
    Mueller, Hermann J.
    Geyer, Thomas
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-LEARNING MEMORY AND COGNITION, 2021, 47 (09) : 1453 - 1470
  • [44] Tactile-to-Visual Cross-Modal Transfer of Texture Categorisation Following Training: An fMRI Study
    O'Callaghan, Georgia
    O'Dowd, Alan
    Simoes-Franklin, Cristina
    Stapleton, John
    Newell, Fiona N.
    FRONTIERS IN INTEGRATIVE NEUROSCIENCE, 2018, 12
  • [45] Cross-Modal Semantic Matching Generative Adversarial Networks for Text-to-Image Synthesis
    Tan, Hongchen
    Liu, Xiuping
    Yin, Baocai
    Li, Xin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 832 - 845
  • [46] CM-GANs: Cross-modal Generative Adversarial Networks for Common Representation Learning
    Peng, Yuxin
    Qi, Jinwei
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2019, 15 (01)
  • [47] Review of Generative Adversarial Networks in mono- and cross-modal biomedical image registration
    Han, Tingting
    Wu, Jun
    Luo, Wenting
    Wang, Huiming
    Jin, Zhe
    Qu, Lei
    FRONTIERS IN NEUROINFORMATICS, 2022, 16
  • [48] Integration of visual and tactile information in reproduction of traveled distance
    Churan, Jan
    Paul, Johannes
    Klingenhoefer, Steffen
    Bremmer, Frank
    JOURNAL OF NEUROPHYSIOLOGY, 2017, 118 (03) : 1650 - 1663
  • [49] SEMANTIC PRESERVING GENERATIVE ADVERSARIAL NETWORK FOR CROSS-MODAL HASHING
    Wu, Fei
    Luo, Xiaokai
    Huang, Qinghua
    Wei, Pengfei
    Sun, Ying
    Dong, Xiwei
    Wu, Zhiyong
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 2743 - 2747
  • [50] Simultaneity learning in vision, audition, tactile sense and their cross-modal combinations
    Virsu, Veijo
    Oksanen-Hennah, Henna
    Vedenpaa, Anita
    Jaatinen, Pentti
    Lahti-Nuuttila, Pekka
    EXPERIMENTAL BRAIN RESEARCH, 2008, 186 (04) : 525 - 537