Reducing Query Latencies in Web Search Using Fine-Grained Parallelism

被引:0
|
作者
Eitan Frachtenberg
机构
[1] Microsoft,
来源
World Wide Web | 2009年 / 12卷
关键词
semantic web; search engines; performance evaluation; multi-core processors; parallel algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
Semantic Web search is a new application of recent advances in information retrieval (IR), natural language processing, artificial intelligence, and other fields. The Powerset group in Microsoft develops a semantic search engine that aims to answer queries not only by matching keywords, but by actually matching meaning in queries to meaning in Web documents. Compared to typical keyword search, semantic search can pose additional engineering challenges for the back-end and infrastructure designs. Of these, the main challenge addressed in this paper is how to lower query latencies to acceptable, interactive levels. Index-based semantic search requires more data processing, such as numerous synonyms, hypernyms, multiple linguistic readings, and other semantic information, both on queries and in the index. In addition, some of the algorithms can be super-linear, such as matching co-references across a document. Consequently, many semantic queries can run significantly slower than the same keyword query. Users, however, have grown to expect Web search engines to provide near-instantaneous results, and a slow search engine could be deemed unusable even if it provides highly relevant results. It is therefore imperative for any search engine to meet its users’ interactivity expectations, or risk losing them. Our approach to tackle this challenge is to exploit data parallelism in slow search queries to reduce their latency in multi-core systems. Although all search engines are designed to exploit parallelism, at the single-node level this usually translates to throughput-oriented task parallelism. This paper focuses on the engineering of two latency-oriented approaches (coarse- and fine-grained) and compares them to the task-parallel approach. We use Powerset’s deployed search engine to evaluate the various factors that affect parallel performance: workload, overhead, load balancing, and resource contention. We also discuss heuristics to selectively control the degree of parallelism and consequent overhead on a query-by-query level. Our experimental results show that using fine-grained parallelism with these dynamic heuristics can significantly reduce query latencies compared to fixed, coarse-granularity parallelization schemes. Although these results were obtained on, and optimized for, Powerset’s semantic search, they can be readily generalized to a wide class of inverted-index search engines.
引用
收藏
页码:441 / 460
页数:19
相关论文
共 50 条
  • [1] Reducing Query Latencies in Web Search Using Fine-Grained Parallelism
    Frachtenberg, Eitan
    [J]. WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2009, 12 (04): : 441 - 460
  • [2] FINE-GRAINED PARALLELISM IN ELLIE
    ANDERSEN, B
    [J]. JOURNAL OF OBJECT-ORIENTED PROGRAMMING, 1992, 5 (03): : 55 - 61
  • [3] Accelerating RSA with Fine-Grained Parallelism Using GPU
    Yang, Yang
    Guan, Zhi
    Sun, Huiping
    Chen, Zhong
    [J]. INFORMATION SECURITY PRACTICE AND EXPERIENCE, ISPEC 2015, 2015, 9065 : 454 - 468
  • [4] Fine-grained parallelism in computational mathematics
    Bandman, OL
    [J]. PROGRAMMING AND COMPUTER SOFTWARE, 2001, 27 (04) : 170 - 182
  • [5] Fine-Grained Parallelism in Computational Mathematics
    O. L. Bandman
    [J]. Programming and Computer Software, 2001, 27 : 170 - 182
  • [6] Fine-grained parallelism in dynamic web content generation: The parse and dispatch approach
    Papastavrou, S
    Samaras, G
    Evripidou, P
    Chrysanthis, PK
    [J]. ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS 2003: COOPIS, DOA, AND ODBASE, 2003, 2888 : 573 - 588
  • [7] Evaluation of Fine-grained Parallelism in AUTOSAR Applications
    Stegmeier, Alexander
    Kehr, Sebastian
    George, Dave
    Bradatsch, Christian
    Panic, Milos
    Bodekker, Bert
    Ungerer, Theo
    [J]. INTERNATIONAL CONFERENCE ON EMBEDDED COMPUTER SYSTEMS: ARCHITECTURES, MODELING, AND SIMULATION (SAMOS 2017), 2017, : 121 - 128
  • [8] A MATCHING APPROACH TO UTILIZING FINE-GRAINED PARALLELISM
    GUPTA, R
    SOFFA, ML
    [J]. PROCEEDINGS OF THE TWENTY-FIRST, ANNUAL HAWAII INTERNATIONAL CONFERENCE ON SYSTEM SCIENCES, VOLS 1-4: ARCHITECTURE TRACK, SOFTWARE TRACK, DECISION SUPPORT AND KNOWLEDGE BASED SYSTEMS TRACK, APPLICATIONS TRACK, 1988, : 148 - 156
  • [9] Exploiting Fine-Grained Parallelism on Cell Processors
    Hoffmann, Ralf
    Prell, Andreas
    Rauber, Thomas
    [J]. EURO-PAR 2010 - PARALLEL PROCESSING, PART II, 2010, 6272 : 175 - 186
  • [10] Graph Analytics Through Fine-Grained Parallelism
    Shang, Zechao
    Li, Feifei
    Yu, Jeffrey Xu
    Zhang, Zhiwei
    Cheng, Hong
    [J]. SIGMOD'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2016, : 463 - 478