A framework for benchmarking in CBIR

被引:16
|
作者
Müller, H [1 ]
Müller, W
Marchand-Maillet, S
Pun, T
Squire, DM
机构
[1] Univ Geneva, Vis Grp, CH-1211 Geneva 4, Switzerland
[2] Monash Univ, CSSE, Melbourne, Vic 3004, Australia
关键词
evaluation; content-based image retrieval; benchmarking; Benchathlon; TREC;
D O I
10.1023/A:1025034215859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Content-based image retrieval (CBIR) has been a very active research area for more than ten years. In the last few years the number of publications and retrieval systems produced has become larger and larger. Despite this, there is still no agreed objective way in which to compare the performance of any two of these systems. This fact is blocking the further development of the field since good or promising techniques can not be identified objectively, and the potential commercial success of CBIR systems is hindered because it is hard to establish the quality of an application. We are thus in the position in which other research areas, such as text retrieval or the database systems, found themselves several years ago. To have serious applications, as well as commercial success, objective proof of system quality is needed: in text retrieval the TREC benchmark is a widely accepted performance measure; in the transaction processing field for databases it is the TPC benchmark that has wide support. This paper describes a framework that enables the creation of a benchmark for CBIR. Parts of this framework have already been developed and systems can be evaluated against a small, freely-available database via a web interface. Much work remains to be done with respect to making available large, diverse image databases and obtaining relevance judgments for those large databases. We also need to establish an independent body, accepted by the entire community, that would organize a benchmarking event, give out official results and update the benchmark regularly. The Benchathlon could get this role if it manages to gain the confidence of the field. This should also prevent the negative effects, e.g., "benchmarketing", experienced with other benchmarks, such as the TPC predecessors. This paper sets out our ideas for an open framework for performance evaluation. We hope to stimulate discussion on evaluation in image retrieval so that systems can be compared on the same grounds. We also identify query paradigms beyond query by example (QBE) that may be integrated into a benchmarking framework, and we give examples of application-based benchmarking areas.
引用
收藏
页码:55 / 73
页数:19
相关论文
共 50 条
  • [41] A framework for benchmarking uncertainty in deep regression
    Schmaehling, Franko
    Martin, Joerg
    Elster, Clemens
    [J]. APPLIED INTELLIGENCE, 2023, 53 (08) : 9499 - 9512
  • [42] A CBIR-framework: using both syntactical and semantical information for image description
    Besson, L
    Da Costa, A
    [J]. SEVENTH INTERNATIONAL DATABASE ENGINEERING AND APPLICATIONS SYMPOSIUM, PROCEEDINGS, 2003, : 385 - 390
  • [43] Benchmarking with the BLASST Sessional Staff Standards Framework
    Luzia, Karina
    Harvey, Marina
    Parker, Nicola
    McCormack, Coralie
    Brown, Natalie R.
    McKenzie, Jo
    [J]. JOURNAL OF UNIVERSITY TEACHING AND LEARNING PRACTICE, 2013, 10 (03):
  • [44] A Framework for Benchmarking Open Government Data Efforts
    Sayogo, Djoko Sigit
    Pardo, Theresa A.
    Cook, Meghan
    [J]. 2014 47TH HAWAII INTERNATIONAL CONFERENCE ON SYSTEM SCIENCES (HICSS), 2014, : 1896 - 1905
  • [45] BenchFoundry: A Benchmarking Framework for Cloud Storage Services
    Bermbach, David
    Kuhlenkamp, Joern
    Dey, Akon
    Ramachandran, Arunmoezhi
    Fekete, Alan
    Tai, Stefan
    [J]. SERVICE-ORIENTED COMPUTING, ICSOC 2017, 2017, 10601 : 314 - 330
  • [46] ModelXGlue: a benchmarking framework for ML tools in MDE
    Lopez, Jose Antonio Hernandez
    Cuadrado, Jesus Sanchez
    Rubei, Riccardo
    Di Ruscio, Davide
    [J]. SOFTWARE AND SYSTEMS MODELING, 2024,
  • [47] A framework for benchmarking Arabic verb morphological tools
    Yagi, Sane M.
    Yaghi, Jim
    [J]. INNOVATIONS IN E-LEARNING, INSTRUCTION TECHNOLOGY, ASSESSMENT, AND ENGINEERING EDUCATION, 2007, : 181 - +
  • [48] Diverse Datasets and a Customizable Benchmarking Framework for Phishing
    Zeng, Victor
    Baki, Shahryar
    El Aassal, Ayman
    Verma, Rakesh
    De Moraes, Luis Felipe Teixeira
    Das, Avisha
    [J]. PROCEEDINGS OF THE SIXTH INTERNATIONAL WORKSHOP ON SECURITY AND PRIVACY ANALYTICS (IWSPA'20), 2020, : 35 - 41
  • [49] Benchmarking in the public sector: A critical conceptual framework
    Triantafillou, Peter
    [J]. PUBLIC ADMINISTRATION, 2007, 85 (03) : 829 - 846
  • [50] A Framework for Systematic Benchmarking of Monitoring and Diagnostic Systems
    Kurtoglu, Tolga
    Mengshoel, Ole J.
    Poll, Scott
    [J]. 2008 INTERNATIONAL CONFERENCE ON PROGNOSTICS AND HEALTH MANAGEMENT (PHM), 2008, : 339 - +