Long-Tail Hashing

被引:6
|
作者
Chen, Yong [1 ,2 ]
Hou, Yuqing [3 ]
Leng, Shu [4 ]
Zhang, Qing [3 ]
Lin, Zhouchen [1 ,2 ]
Zhang, Dell [5 ,6 ]
机构
[1] Peking Univ, Sch EECS, Key Lab Machine Percept MoE, Beijing, Peoples R China
[2] Pazhou Lab, Guangzhou, Peoples R China
[3] Meituan, Beijing, Peoples R China
[4] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
[5] Blue Prism AI Labs, London, England
[6] Birkbeck Univ London, London, England
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
learning to hash; long-tail datasets; memory network; large-scale; multimedia retrieval; ITERATIVE QUANTIZATION; PROCRUSTEAN APPROACH; DISTRIBUTIONS; PARETO; CODES; SMOTE;
D O I
10.1145/3404835.3462888
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Hashing, which represents data items as compact binary codes, has been becoming a more and more popular technique, e.g., for large-scale image retrieval, owing to its super fast search speed as well as its extremely economical memory consumption. However, existing hashing methods all try to learn binary codes from artificially balanced datasets which are not commonly available in real-world scenarios. In this paper, we propose Long-Tail Hashing Network (LTHNet), a novel two-stage deep hashing approach that addresses the problem of learning to hash for more realistic datasets where the data labels roughly exhibit a long-tail distribution. Specifically, the first stage is to learn relaxed embeddings of the given dataset with its long-tail characteristic taken into account via an end-to-end deep neural network; the second stage is to binarize those obtained embeddings. A critical part of LTHNet is its dynamic meta-embedding module extended with a determinantal point process which can adaptively realize visual knowledge transfer between head and tail classes, and thus enrich image representations for hashing. Our experiments have shown that LTHNet achieves dramatic performance improvements over all state-of-the-art competitors on long-tail datasets, with no or little sacrifice on balanced datasets. Further analyses reveal that while to our surprise directly manipulating class weights in the loss function has little effect, the extended dynamic meta-embedding module, the usage of cross-entropy loss instead of square loss, and the relatively small batch-size for training all contribute to LTHNet's success.
引用
收藏
页码:1328 / 1338
页数:11
相关论文
共 50 条
  • [31] Managing Long-Tail Processes Using FormSys
    Weber, Ingo
    Paik, Rye-Young
    Benatallah, Boualem
    Vorwerk, Corren
    Gong, Zifei
    Zheng, Liangliang
    Kim, Sung Wook
    SERVICE-ORIENTED COMPUTING - ICSOC 2010, PROCEEDINGS, 2010, 6470 : 702 - 703
  • [32] Retroactive liability and the insurability of long-tail risks
    Faure, M
    Fenn, P
    INTERNATIONAL REVIEW OF LAW AND ECONOMICS, 1999, 19 (04) : 487 - 500
  • [33] Open Knowledge Enrichment for Long-tail Entities
    Cao, Ermei
    Wang, Difeng
    Huang, Jiacheng
    Hu, Wei
    WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 384 - 394
  • [34] The long-tail of global engagement and international librarianship
    Witt, Steven W.
    IFLA JOURNAL-INTERNATIONAL FEDERATION OF LIBRARY ASSOCIATIONS, 2015, 41 (04): : 297 - 297
  • [35] SisGExp: Rethinking Long-Tail Agronomic Experiments
    Serra da Cruz, Sergio Manuel
    Pires do Nascimento, Jose Antonio
    PROVENANCE AND ANNOTATION OF DATA AND PROCESSES, IPAW 2016, 2016, 9672 : 214 - 217
  • [36] The Future of Academic Publishing: Application of the Long-Tail Theory
    Gould, Thomas H. P.
    PUBLISHING RESEARCH QUARTERLY, 2009, 25 (04) : 232 - 245
  • [37] The role of knowledge in determining identity of long-tail entities
    Ilievski, Filip
    Hovy, Eduard
    Vossen, Piek
    Schlobach, Stefan
    Xie, Qizhe
    JOURNAL OF WEB SEMANTICS, 2020, 61-62
  • [38] Long-tail Detection with Effective Class-Margins
    Cho, Jang Hyun
    Krahenbuhl, Philipp
    COMPUTER VISION, ECCV 2022, PT VIII, 2022, 13668 : 698 - 714
  • [39] Rumour Detection on Social Media with Long-Tail Strategy
    Zhang, Guixian
    Liang, Rongjiao
    Yu, Zhongyi
    Zhang, Shichao
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [40] Improving long-tail classification via decoupling and regularisation
    Gao, Shuzheng
    Wang, Chaozheng
    Gao, Cuiyun
    Luo, Wenjian
    Han, Peiyi
    Liao, Qing
    Xu, Guandong
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2025, 10 (01) : 62 - 71