Reducing Query Latencies in Web Search Using Fine-Grained Parallelism

被引:0
|
作者
Eitan Frachtenberg
机构
[1] Microsoft,
来源
World Wide Web | 2009年 / 12卷
关键词
semantic web; search engines; performance evaluation; multi-core processors; parallel algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
Semantic Web search is a new application of recent advances in information retrieval (IR), natural language processing, artificial intelligence, and other fields. The Powerset group in Microsoft develops a semantic search engine that aims to answer queries not only by matching keywords, but by actually matching meaning in queries to meaning in Web documents. Compared to typical keyword search, semantic search can pose additional engineering challenges for the back-end and infrastructure designs. Of these, the main challenge addressed in this paper is how to lower query latencies to acceptable, interactive levels. Index-based semantic search requires more data processing, such as numerous synonyms, hypernyms, multiple linguistic readings, and other semantic information, both on queries and in the index. In addition, some of the algorithms can be super-linear, such as matching co-references across a document. Consequently, many semantic queries can run significantly slower than the same keyword query. Users, however, have grown to expect Web search engines to provide near-instantaneous results, and a slow search engine could be deemed unusable even if it provides highly relevant results. It is therefore imperative for any search engine to meet its users’ interactivity expectations, or risk losing them. Our approach to tackle this challenge is to exploit data parallelism in slow search queries to reduce their latency in multi-core systems. Although all search engines are designed to exploit parallelism, at the single-node level this usually translates to throughput-oriented task parallelism. This paper focuses on the engineering of two latency-oriented approaches (coarse- and fine-grained) and compares them to the task-parallel approach. We use Powerset’s deployed search engine to evaluate the various factors that affect parallel performance: workload, overhead, load balancing, and resource contention. We also discuss heuristics to selectively control the degree of parallelism and consequent overhead on a query-by-query level. Our experimental results show that using fine-grained parallelism with these dynamic heuristics can significantly reduce query latencies compared to fixed, coarse-granularity parallelization schemes. Although these results were obtained on, and optimized for, Powerset’s semantic search, they can be readily generalized to a wide class of inverted-index search engines.
引用
收藏
页码:441 / 460
页数:19
相关论文
共 50 条
  • [41] Efficient Concurrent Search Trees Using Portable Fine-Grained Locality
    Phuong Hoai Ha
    Anshus, Otto J.
    Umar, Ibrahim
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (07) : 1580 - 1595
  • [42] Augmenting Strong Supervision Using Web Data for Fine-grained Categorization
    Xu, Zhe
    Huang, Shaoli
    Zhang, Ya
    Tao, Dacheng
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2524 - 2532
  • [43] Enabling Fine-Grained HTTP Caching of SPARQL Query Results
    Williams, Gregory Todd
    Weaver, Jesse
    SEMANTIC WEB - ISWC 2011, PT I, 2011, 7031 : 762 - 777
  • [44] Towards Fine-grained Parallelism in Parallel and Distributed Python']Python Libraries
    Kerney, Jamison
    Raicu, Joan
    Raicu, John
    Chard, Kyle
    2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW 2024, 2024, : 706 - 715
  • [45] A Fine-grained Asynchronous Bulk Synchronous parallelism model for PGAS applications
    Paul, Sri Raj
    Hayashi, Akihiro
    Chen, Kun
    Elmougy, Youssef
    Sarkar, Vivek
    JOURNAL OF COMPUTATIONAL SCIENCE, 2023, 69
  • [46] Fine-grained adaptive parallelism for automotive systems through AMALTHEA and OpenMP
    Munera, Adrian
    Royuela, Sara
    Pressler, Michael
    Mackamul, Harald
    Ziegenbein, Dirk
    Quinones, Eduardo
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 146
  • [48] Exploiting Fine-Grained Pipeline Parallelism for Wavefront Computations on Multicore Platforms
    Wu, Guiming
    Wang, Miao
    Dou, Yong
    Xia, Fei
    2009 INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING WORKSHOPS (ICPPW 2009), 2009, : 402 - 408
  • [49] Orchard: Heterogeneous Parallelism and Fine-grained Fusion for Complex Tree Traversals
    Singhal, Vidush
    Sakka, Laith
    Sundararajah, Kirshanthan
    Newton, Ryan R.
    Kulkarni, Milind
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2024, 21 (02)
  • [50] Exploiting fine-grained data parallelism with chip multiprocessors and fast barriers
    Sampson, Jack
    Gonzalez, Ruben
    Collard, Jean-Francois
    Jouppi, Norman P.
    Schlansker, Mike
    Calder, Brad
    MICRO-39: PROCEEDINGS OF THE 39TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, 2006, : 235 - +