LEWIS: Latent Embeddings for Word Images and their Semantics

被引:10
|
作者
Gordo, Albert [1 ]
Almazan, Jon [1 ]
Murray, Naila [1 ]
Perronnin, Florent [2 ,3 ]
机构
[1] Xerox Res Ctr Europe, Meylan, France
[2] Facebook AI Res, Menlo Pk, CA USA
[3] Xerox Res Ctr Europe, Comp Vis Grp, Meylan, France
关键词
D O I
10.1109/ICCV.2015.147
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The goal of this work is to bring semantics into the tasks of text recognition and retrieval in natural images. Although text recognition and retrieval have received a lot of attention in recent years, previous works have focused on recognizing or retrieving exactly the same word used as a query, without taking the semantics into consideration. In this paper, we ask the following question: can we predict semantic concepts directly from a word image, without explicitly trying to transcribe the word image or its characters at any point? For this goal we propose a convolutional neural network (CNN) with a weighted ranking loss objective that ensures that the concepts relevant to the query image are ranked ahead of those that are not relevant. This can also be interpreted as learning a Euclidean space where word images and concepts are jointly embedded. This model is learned in an end-to-end manner, from image pixels to semantic concepts, using a dataset of synthetically generated word images and concepts mined from a lexical database (WordNet). Our results show that, despite the complexity of the task, word images and concepts can indeed be associated with a high degree of accuracy.
引用
收藏
页码:1242 / 1250
页数:9
相关论文
共 50 条
  • [1] Compressing and interpreting word embeddings with latent space regularization and interactive semantics probing
    Li, Haoyu
    Wang, Junpeng
    Zheng, Yan
    Wang, Liang
    Zhang, Wei
    Shen, Han-Wei
    [J]. INFORMATION VISUALIZATION, 2023, 22 (01) : 52 - 68
  • [2] Lexical semantics enhanced neural word embeddings
    Yang, Dongqiang
    Li, Ning
    Zou, Li
    Ma, Hongwei
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 252
  • [3] Decoupled Word Embeddings using Latent Topics
    Park, Heesoo
    Lee, Jongwuk
    [J]. PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 875 - 882
  • [4] Jointly Learning Word Embeddings and Latent Topics
    Shi, Bei
    Lam, Wai
    Jameel, Shoaib
    Schockaert, Steven
    Lai, Kwun Ping
    [J]. SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 375 - 384
  • [5] Further results on latent discourse models and word embeddings
    Khalife, Sammy
    Goncalves, Douglas
    Allouah, Youssef
    Liberti, Leo
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [6] Further results on latent discourse models and word embeddings
    Khalife, Sammy
    Gonçalves, Douglas
    Allouah, Youssef
    Liberti, Leo
    [J]. Journal of Machine Learning Research, 2021, 22
  • [7] Semantics-Driven Recognition of Collocations Using Word Embeddings
    Rodriguez-Fernandez, Sara
    Espinosa-Anke, Luis
    Carlini, Roberto
    Wanner, Leo
    [J]. PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, : 499 - 505
  • [8] Semantics-assisted Wasserstein Learning for Topic and Word Embeddings
    Li, Changchun
    Li, Ximing
    Ouyang, Jihong
    Wang, Yiming
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 292 - 301
  • [9] Distributional Semantics in Aphasia: An Exploratory Analysis Using Word Embeddings
    Bello Lepe, Sebastian
    [J]. REVISTA SIGNOS, 2022, 55 (110): : 1029 - 1044
  • [10] Incorporating Latent Meanings of Morphological Compositions to Enhance Word Embeddings
    Xu, Yang
    Liu, Jiawei
    Yang, Wei
    Huang, Liusheng
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 1232 - 1242