Bridging the gap between visual and auditory feature spaces for cross-media retrieval

被引:0
|
作者
Hong Zhang [1 ]
Fei Wu [1 ]
机构
[1] Zhejiang Univ, Inst Artificial Intelligence, Hangzhou 310027, Peoples R China
来源
基金
中国国家自然科学基金;
关键词
cross-media retrieval; canonical correlation; relevance feedback; dynamic cross-media ranking;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-media retrieval is an interesting research problem, which seeks to breakthrough the limitation of modality so that users can query multimedia objects by examples of different modalities. In this paper we present a novel approach to learn the underlying correlation between visual and auditory feature spaces for cross-media retrieval. A semi-supervised Correlation Preserving Mapping (SSCPM) is described to learn the isomorphic SSCPM subspace where canonical correlations between original visual and auditory features are furthest preserved. Based on user interactions of relevance feedback, local semantic clusters are formed for images and audios respectively. With the dynamic spread of ranking scores of positive and negative examples, crossmedia semantic correlations are refined, and cross-media distance is accurately estimated. Experiment results are encouraging and show that the performance of our approach is effective.
引用
收藏
页码:596 / 605
页数:10
相关论文
共 50 条
  • [1] Boosting Cross-media Retrieval via Visual-Auditory Feature Analysis and Relevance Feedback
    Zhang, Hong
    Yuan, Junsong
    Gao, Xingyu
    Chen, Zhenyu
    [J]. PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, : 953 - 956
  • [3] Understanding visual-auditory correlation from heterogeneous features for cross-media retrieval
    Zhang, Hong
    Wang, Yan-yun
    Pan, Hong
    Wu, Fei
    [J]. JOURNAL OF ZHEJIANG UNIVERSITY-SCIENCE A, 2008, 9 (02): : 241 - 249
  • [4] Understanding visual-auditory correlation from heterogeneous features for cross-media retrieval
    Hong Zhang
    Yan-yun Wang
    Hong Pan
    Fei Wu
    [J]. Journal of Zhejiang University SCIENCE A, 2008, 9 : 241 - 249
  • [5] CSRNCVA: A MODEL OF CROSS-MEDIA SEMANTIC RETRIEVAL BASED ON NEURAL COMPUTING OF VISUAL AND AUDITORY SENSATIONS
    Liu, Y.
    Cai, K.
    Liu, C.
    Zheng, F.
    [J]. NEURAL NETWORK WORLD, 2018, 28 (04) : 305 - 323
  • [6] Cross-media retrieval method based on feature subspace learning
    Zhang, Hong
    Wu, Fei
    Zhuang, Yue-Ting
    [J]. Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2008, 21 (06): : 739 - 745
  • [7] COUPLED FEATURE MAPPING AND CORRELATION MINING FOR CROSS-MEDIA RETRIEVAL
    Fan, Mengdi
    Wang, Wenmin
    Wang, Ronggang
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2016,
  • [8] COUPLED FEATURE SELECTION FOR MODALITY-DEPENDENT CROSS-MEDIA RETRIEVAL
    Yu, En
    Sun, Jiande
    Wang, Li
    Zhang, Huaxiang
    Li, Jing
    [J]. 2017 INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING AND COMMUNICATION SYSTEMS (ISPACS 2017), 2017, : 315 - 320
  • [9] LARGE-SCALE CROSS-MEDIA RETRIEVAL BY HETEROGENEOUS FEATURE AUGMENTATION
    Li, Qiang
    Han, Yahong
    Dang, Jianwu
    [J]. 2014 12TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP), 2014, : 977 - 980
  • [10] Nonnegative cross-media recoding of visual-auditory content for social media analysis
    Hong Zhang
    Xin Xu
    [J]. Multimedia Tools and Applications, 2015, 74 : 577 - 593