Large Language Models are Built-in Autoregressive Search Engines

被引:0
|
作者
Ziems, Noah [1 ]
Yu, Wenhao [1 ]
Zhang, Zhihan [1 ]
Jiang, Meng [1 ]
机构
[1] Univ Notre Dame, Notre Dame, IN 46556 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Document retrieval is a key stage of standard Web search engines. Existing dual-encoder dense retrievers obtain representations for questions and documents independently, allowing for only shallow interactions between them. To overcome this limitation, recent autoregressive search engines replace the dual-encoder architecture by directly generating identifiers for relevant documents in the candidate pool. However, the training cost of such autoregressive search engines rises sharply as the number of candidate documents increases. In this paper, we find that large language models (LLMs) can follow human instructions to directly generate URLs for document retrieval. Surprisingly, when providing a few Query-URL pairs as in-context demonstrations, LLMs can generate Web URLs where nearly 90% of the corresponding documents contain correct answers to open-domain questions. In this way, LLMs can be thought of as built-in search engines, since they have not been explicitly trained to map questions to document identifiers. Experiments demonstrate that our method can consistently achieve better retrieval performance than existing retrieval approaches by a significant margin on three open-domain question answering benchmarks, under both zero and few-shot settings. The code for this work can be found at https://github.com/Ziems/llm-url.
引用
收藏
页码:2666 / 2678
页数:13
相关论文
共 50 条
  • [31] CoSearchAgent: A Lightweight Collaborative Search Agent with Large Language Models
    Gong, Peiyuan
    Li, Jiamian
    Mao, Jiaxin
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2729 - 2733
  • [32] Mathematical discoveries from program search with large language models
    Romera-Paredes, Bernardino
    Barekatain, Mohammadamin
    Novikov, Alexander
    Balog, Matej
    Kumar, M. Pawan
    Dupont, Emilien
    Ruiz, Francisco J. R.
    Ellenberg, Jordan S.
    Wang, Pengming
    Fawzi, Omar
    Kohli, Pushmeet
    Fawzi, Alhussein
    NATURE, 2024, 625 (7995) : 468 - 475
  • [33] Mathematical discoveries from program search with large language models
    Bernardino Romera-Paredes
    Mohammadamin Barekatain
    Alexander Novikov
    Matej Balog
    M. Pawan Kumar
    Emilien Dupont
    Francisco J. R. Ruiz
    Jordan S. Ellenberg
    Pengming Wang
    Omar Fawzi
    Pushmeet Kohli
    Alhussein Fawzi
    Nature, 2024, 625 : 468 - 475
  • [34] FRESHLLMS: Refreshing Large Language Models with Search Engine Augmentation
    Vu, Tu
    Iyyer, Mohit
    Wang, Xuezhi
    Constant, Noah
    Wei, Jerry
    Wei, Jason
    Tar, Chris
    Sung, Yun-Hsuan
    Zhou, Denny
    Le, Quoc
    Luong, Thang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 13697 - 13720
  • [35] Large language models should be used as scientific reasoning engines, not knowledge databases
    Truhn, Daniel
    Reis-Filho, Jorge S.
    Kather, Jakob Nikolas
    NATURE MEDICINE, 2023, 29 (12) : 2983 - 2984
  • [36] Cellular dynamic models with built-in causal relation functions
    Vachkov, GL
    Christova, NG
    2002 FIRST INTERNATIONAL IEEE SYMPOSIUM INTELLIGENT SYSTEMS, VOL II, EUNITE INVITED SESSION, PROCEEDINGS, 2002, : 34 - 39
  • [37] Comparing the Built-In Application Architecture Models in the Web Browser
    Taivalsaari, Antero
    Mikkonen, Tommi
    Pautasso, Cesare
    Systa, Kari
    2017 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE ARCHITECTURE (ICSA 2017), 2017, : 51 - 54
  • [38] BUILT-IN STABILIZATION AND VARIANCES IN LINEAR DYNAMIC-MODELS
    PETTERSEN, O
    MANCHESTER SCHOOL OF ECONOMIC AND SOCIAL STUDIES, 1978, 46 (02): : 155 - 163
  • [39] LEGALBENCH: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Language Models
    Guha, Neel
    Nyarko, Julian
    Ho, Daniel E.
    Re, Christopher
    Chilton, Adam
    Narayana, Aditya
    Chohlas-Wood, Alex
    Peters, Austin
    Waldon, Brandon
    Rockmore, Daniel N.
    Zambrano, Diego
    Talisman, Dmitry
    Hoque, Enam
    Surani, Faiz
    Fagan, Frank
    Sarfaty, Galit
    Dickinson, Gregory M.
    Porat, Haggai
    Hegland, Jason
    Wu, Jessica
    Nudell, Joe
    Niklaus, Joel
    Nay, John
    Choi, Jonathan H.
    Tobia, Kevin
    Hagan, Margaret
    Ma, Megan
    Livermore, Michael
    Rasumov-Rahe, Nikon
    Holzenberger, Nils
    Kolt, Noam
    Henderson, Peter
    Rehaag, Sean
    Goel, Sharad
    Gao, Shang
    Williams, Spencer
    Gandhi, Sunny
    Zur, Tom
    Iyer, Varun
    Li, Zehua
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [40] Search strategies for identifying a composite plate delamination using built-in transducers
    Keilers, CH
    STRUCTURAL HEALT H MONITORING: CURRENT STATUS AND PERSPECTIVES, 1997, : 466 - 477