Reducing Query Latencies in Web Search Using Fine-Grained Parallelism

被引:0
|
作者
Eitan Frachtenberg
机构
[1] Microsoft,
来源
World Wide Web | 2009年 / 12卷
关键词
semantic web; search engines; performance evaluation; multi-core processors; parallel algorithms;
D O I
暂无
中图分类号
学科分类号
摘要
Semantic Web search is a new application of recent advances in information retrieval (IR), natural language processing, artificial intelligence, and other fields. The Powerset group in Microsoft develops a semantic search engine that aims to answer queries not only by matching keywords, but by actually matching meaning in queries to meaning in Web documents. Compared to typical keyword search, semantic search can pose additional engineering challenges for the back-end and infrastructure designs. Of these, the main challenge addressed in this paper is how to lower query latencies to acceptable, interactive levels. Index-based semantic search requires more data processing, such as numerous synonyms, hypernyms, multiple linguistic readings, and other semantic information, both on queries and in the index. In addition, some of the algorithms can be super-linear, such as matching co-references across a document. Consequently, many semantic queries can run significantly slower than the same keyword query. Users, however, have grown to expect Web search engines to provide near-instantaneous results, and a slow search engine could be deemed unusable even if it provides highly relevant results. It is therefore imperative for any search engine to meet its users’ interactivity expectations, or risk losing them. Our approach to tackle this challenge is to exploit data parallelism in slow search queries to reduce their latency in multi-core systems. Although all search engines are designed to exploit parallelism, at the single-node level this usually translates to throughput-oriented task parallelism. This paper focuses on the engineering of two latency-oriented approaches (coarse- and fine-grained) and compares them to the task-parallel approach. We use Powerset’s deployed search engine to evaluate the various factors that affect parallel performance: workload, overhead, load balancing, and resource contention. We also discuss heuristics to selectively control the degree of parallelism and consequent overhead on a query-by-query level. Our experimental results show that using fine-grained parallelism with these dynamic heuristics can significantly reduce query latencies compared to fixed, coarse-granularity parallelization schemes. Although these results were obtained on, and optimized for, Powerset’s semantic search, they can be readily generalized to a wide class of inverted-index search engines.
引用
收藏
页码:441 / 460
页数:19
相关论文
共 50 条
  • [21] Exploiting fine grained parallelism for acceleration of web retrieval
    Yuan, JL
    Chi, CH
    Sun, QB
    WEB AND COMMUNICATION TECHNOLOGIES AND INTERNET -RELATED SOCIAL ISSUES - HSI 2005, 2005, 3597 : 125 - 134
  • [22] Temporal Query Networks for Fine-grained Video Understanding
    Zhang, Chuhan
    Gupta, Ankush
    Zisserman, Andrew
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4484 - 4494
  • [23] Construct and Query A Fine-Grained Geospatial Knowledge Graph
    Wei, Bo
    Guo, Xi
    Li, Xiaodi
    Wu, Ziyan
    Zhao, Jing
    Zou, Qiping
    DATA SCIENCE AND ENGINEERING, 2024, 9 (02) : 152 - 176
  • [24] Enhanced Query Classification with Millions of Fine-Grained Topics
    Ye, Qi
    Wang, Feng
    Li, Bo
    Liu, Zhimin
    WEB-AGE INFORMATION MANAGEMENT, PT II, 2016, 9659 : 120 - 131
  • [25] FiGO: Fine-Grained Query Optimization in Video Analytics
    Cao, Jiashen
    Sarkar, Karan
    Hadidi, Ramyad
    Arulraj, Joy
    Kim, Hyesoon
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 559 - 572
  • [26] Study of Fine-grained Nested Parallelism in CDCL SAT Solvers
    Edwards, James
    Vishkin, Uzi
    ACM TRANSACTIONS ON PARALLEL COMPUTING, 2021, 8 (03)
  • [27] Fine-grained parallelism in probabilistic parsing with Habanero Java']Java
    Francis-Landau, Matthew
    Xue, Bing
    Eisner, Jason
    Sarkar, Vivek
    PROCEEDINGS OF 2016 6TH WORKSHOP ON IRREGULAR APPLICATIONS: ARCHITECTURE AND ALGORITHMS (IA3), 2016, : 78 - 81
  • [28] Carbon: Architectural Support for Fine-Grained Parallelism on Chip Multiprocessors
    Kumar, Sanjeev
    Hughes, Christopher J.
    Nguyen, Anthony
    ISCA'07: 34TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE, CONFERENCE PROCEEDINGS, 2007, : 162 - 173
  • [29] FINGERS: Exploiting Fine-Grained Parallelism in Graph Mining Accelerators
    Chen, Qihang
    Tian, Boyu
    Gao, Mingyu
    ASPLOS '22: PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS, 2022, : 43 - 55
  • [30] An FPGA Overlay for CNN Inference with Fine-grained Flexible Parallelism
    Choudhury, Ziaul
    Shrivastava, Shashwat
    Ramapantulu, Lavanya
    Purini, Suresh
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2022, 19 (03)