Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3539618.3591715acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

Lexically-Accelerated Dense Retrieval

Published: 18 July 2023 Publication History
  • Get Citation Alerts
  • Abstract

    Retrieval approaches that score documents based on learned dense vectors (i.e., dense retrieval) rather than lexical signals (i.e., conventional retrieval) are increasingly popular. Their ability to identify related documents that do not necessarily contain the same terms as those appearing in the user's query (thereby improving recall) is one of their key advantages. However, to actually achieve these gains, dense retrieval approaches typically require an exhaustive search over the document collection, making them considerably more expensive at query-time than conventional lexical approaches. Several techniques aim to reduce this computational overhead by approximating the results of a full dense retriever. Although these approaches reasonably approximate the top results, they suffer in terms of recall -- one of the key advantages of dense retrieval. We introduce 'LADR' (Lexically-Accelerated Dense Retrieval), a simple-yet-effective approach that improves the efficiency of existing dense retrieval models without compromising on retrieval effectiveness. LADR uses lexical retrieval techniques to seed a dense retrieval exploration that uses a document proximity graph. Through extensive experiments, we find that LADR establishes a new dense retrieval effectiveness-efficiency Pareto frontier among approximate k nearest neighbor techniques. When tuned to take around 8ms per query in retrieval latency on our hardware, LADR consistently achieves both precision and recall that are on par with an exhaustive search on standard benchmarks. Importantly, LADR accomplishes this using only a single CPU -- no hardware accelerators such as GPUs -- which reduces the deployment cost of dense retrieval systems.

    Supplementary Material

    MP4 File (SIGIR23-frp1854.mp4)
    Traditionally, lexical methods have been used to obtain retrieval results efficiently. Retrieval approaches scoring documents based on learned dense vectors outperform lexical methods by eliminating the term overlap dependency. This comes at the cost of latency as an exhaustive search over the document collection is required. Several techniques aim to reduce this computational overhead by approximating the results of a full dense retriever. Even though reasonable results are achieved, approximation methods suffer in terms of recall. We introduce LADR (Lexically-Accelerated Dense Retrieval) which improves efficiency of dense retrieval methods without compromising on retrieval effectiveness. Further, through extensive experiments we infer that LADR establishes a Pareto frontier among approximate k nearest neighbor techniques.

    References

    [1]
    Negar Arabzadeh, Xinyi Yan, and Charles L. A. Clarke. 2021. Predicting Efficiency/Effectiveness Trade-Offs for Dense vs. Sparse Retrieval Strategy Selection. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (Virtual Event, Queensland, Australia) (CIKM '21). Association for Computing Machinery, New York, NY, USA, 2862--2866. https://doi.org/10.1145/3459637.3482159
    [2]
    Arian Askari and Suzan Verberne. 2021. Combining Lexical and Neural Retrieval with Longformer-based Summarization for Effective Case Law Retrieval. In DESIRES.
    [3]
    Nick Craswell, Bhaskar Mitra, Emine Yilmaz, and Daniel Campos. 2021. Overview of the TREC 2020 deep learning track. https://doi.org/10.48550/ARXIV.2102.07662
    [4]
    Nick Craswell, Bhaskar Mitra, Emine Yilmaz, Daniel Campos, and Ellen M. Voorhees. 2020. Overview of the TREC 2019 deep learning track. https://doi.org/10.48550/ARXIV.2003.07820
    [5]
    Zhuyun Dai and Jamie Callan. 2020. Context-Aware Term Weighting For First Stage Passage Retrieval. Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (2020).
    [6]
    Shuai Ding and Torsten Suel. 2011. Faster Top-k Document Retrieval Using Block-Max Indexes. In Proceedings of the 34th International ACM SIGIR Conference on Research and Development in Information Retrieval (Beijing, China) (SIGIR '11). Association for Computing Machinery, New York, NY, USA, 993--1002. https://doi.org/10.1145/2009916.2010048
    [7]
    Yixing Fan, Xiaohui Xie, Yinqiong Cai, Jia Chen, Xinyu Ma, Xiangsheng Li, Ruqing Zhang, and Jiafeng Guo. 2021. Pre-training Methods in Information Retrieval. https://doi.org/10.48550/ARXIV.2111.13853
    [8]
    Thibault Formal, Benjamin Piwowarski, and Stéphane Clinchant. 2021. SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (Virtual Event, Canada) (SIGIR '21). Association for Computing Machinery, New York, NY, USA, 2288--2292. https://doi.org/10.1145/3404835.3463098
    [9]
    Luyu Gao, Zhuyun Dai, Tongfei Chen, Zhen Fan, Benjamin Van Durme, and Jamie Callan. 2021. Complement Lexical Retrieval Model with Semantic Residual Embeddings. In Advances in Information Retrieval: 43rd European Conference on IR Research, ECIR 2021, Virtual Event, March 28 -- April 1, 2021, Proceedings, Part I. Springer-Verlag, Berlin, Heidelberg, 146--160. https://doi.org/10.1007/978-3-030-72113-8_10
    [10]
    Jiafeng Guo, Yixing Fan, Qingyao Ai, and W. Bruce Croft. 2016. A Deep Relevance Matching Model for Ad-Hoc Retrieval. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management (Indianapolis, Indiana, USA) (CIKM '16). Association for Computing Machinery, New York, NY, USA, 55--64. https://doi.org/10.1145/2983323.2983769
    [11]
    Ruiqi Guo, Philip Sun, Erik Lindgren, Quan Geng, David Simcha, Felix Chern, and Sanjiv Kumar. 2020. Accelerating Large-Scale Inference with Anisotropic Vector Quantization. In International Conference on Machine Learning. https://arxiv.org/abs/1908.10396
    [12]
    Charles R. Harris, K. Jarrod Millman, Sté fan J. van der Walt, Ralf Gommers, Pauli Virtanen, David Cournapeau, Eric Wieser, Julian Taylor, Sebastian Berg, Nathaniel J. Smith, Robert Kern, Matti Picus, Stephan Hoyer, Marten H. van Kerkwijk, Matthew Brett, Allan Haldane, Jaime Ferná ndez del Río, Mark Wiebe, Pearu Peterson, Pierre Gérard-Marchant, Kevin Sheppard, Tyler Reddy, Warren Weckesser, Hameer Abbasi, Christoph Gohlke, and Travis E. Oliphant. 2020. Array programming with NumPy. Nature, Vol. 585, 7825 (Sept. 2020), 357--362. https://doi.org/10.1038/s41586-020-2649-2
    [13]
    Sebastian Hofstätter, Sheng-Chieh Lin, Jheng-Hong Yang, Jimmy Lin, and Allan Hanbury. 2021. Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (Virtual Event, Canada) (SIGIR '21). Association for Computing Machinery, New York, NY, USA, 113--122. https://doi.org/10.1145/3404835.3462891
    [14]
    N. Jardine and C. J. van Rijsbergen. 1971. The use of hierarchic clustering in information retrieval. Inf. Storage Retr., Vol. 7 (1971), 217--240.
    [15]
    Jeff Johnson, Matthijs Douze, and Hervé Jégou. 2019. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, Vol. 7, 3 (2019), 535--547.
    [16]
    Herve Jégou, Matthijs Douze, and Cordelia Schmid. 2011. Product Quantization for Nearest Neighbor Search. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 33, 1 (2011), 117--128. https://doi.org/10.1109/TPAMI.2010.57
    [17]
    Kyungho Kim, Kyungjae Lee, Seung-won Hwang, Young-In Song, and Seungwook Lee. 2021. Query Generation for Multimodal Documents. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume. Association for Computational Linguistics, Online, 659--668. https://doi.org/10.18653/v1/2021.eacl-main.54
    [18]
    Carlos Lassance, Thibault Formal, and Stéphane Clinchant. 2021. Composite code sparse autoencoders for first stage retrieval. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2136--2140.
    [19]
    Fabian Lechtenberg, Javier Farreres, Aldwin-Lois Galvan-Cara, Ana Somoza-Tornos, Antonio Espuña, and Moisès Graells. 2022. Information retrieval from scientific abstract and citation databases: A query-by-documents approach based on Monte-Carlo sampling. Expert Systems with Applications, Vol. 199 (2022), 116967. https://doi.org/10.1016/j.eswa.2022.116967
    [20]
    Jurek Leonhardt, Koustav Rudra, Megha Khosla, Abhijit Anand, and Avishek Anand. 2022. Efficient Neural Ranking Using Forward Indexes. In Proceedings of the ACM Web Conference 2022 (Virtual Event, Lyon, France) (WWW '22). Association for Computing Machinery, New York, NY, USA, 266--276. https://doi.org/10.1145/3485447.3511955
    [21]
    Jimmy Lin. 2019. The Neural Hype and Comparisons Against Weak Baselines. SIGIR Forum, Vol. 52, 2 (jan 2019), 40--51. https://doi.org/10.1145/3308774.3308781
    [22]
    Sheng-Chieh Lin, Jheng-Hong Yang, and Jimmy Lin. 2021. In-Batch Negatives for Knowledge Distillation with Tightly-Coupled Teachers for Dense Retrieval. In Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021). Association for Computational Linguistics, Online, 163--173. https://doi.org/10.18653/v1/2021.repl4nlp-1.17
    [23]
    Bing Liu. 2007. Web usage mining. Web Data Mining: Exploring Hyperlinks, Contents, and Usage Data (2007), 449--483.
    [24]
    H. P. Luhn. 1958. The Automatic Creation of Literature Abstracts. IBM Journal of Research and Development, Vol. 2, 2 (1958), 159--165. https://doi.org/10.1147/rd.22.0159
    [25]
    Sean MacAvaney, Franco Maria Nardini, Raffaele Perego, Nicola Tonellotto, Nazli Goharian, and Ophir Frieder. 2020. Expansion via Prediction of Importance with Contextualization. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 1573--1576. https://doi.org/10.1145/3397271.3401262
    [26]
    Sean MacAvaney, Nicola Tonellotto, and Craig Macdonald. 2022. Adaptive Re-Ranking with a Corpus Graph. In 31st ACM International Conference on Information and Knowledge Management. https://doi.org/10.1145/3511808.3557231
    [27]
    Yu A. Malkov and D. A. Yashunin. 2020. Efficient and Robust Approximate Nearest Neighbor Search Using Hierarchical Navigable Small World Graphs. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 42, 4 (2020), 824--836. https://doi.org/10.1109/TPAMI.2018.2889473
    [28]
    Antonio Mallia, Joel Mackenzie, Torsten Suel, and Nicola Tonellotto. 2022. Faster Learned Sparse Retrieval with Guided Traversal. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (Madrid, Spain) (SIGIR '22). Association for Computing Machinery, New York, NY, USA, 1901--1905. https://doi.org/10.1145/3477495.3531774
    [29]
    Antonio Mallia, Michal Siedlaczek, Joel M. Mackenzie, and Torsten Suel. 2019. PISA: Performant Indexes and Search for Academia. In OSIRRC@SIGIR.
    [30]
    Tomas Mikolov, Kai Chen, Gregory S. Corrado, and Jeffrey Dean. 2013. Efficient Estimation of Word Representations in Vector Space. In ICLR.
    [31]
    Bhaskar Mitra and Nick Craswell. 2018. An Introduction to Neural Information Retrieval. Foundations and Trends® in Information Retrieval, Vol. 13, 1 (December 2018), 1--126. https://www.microsoft.com/en-us/research/publication/introduction-neural-information-retrieval/
    [32]
    Bhaskar Miutra and Nick Craswell. 2018. An Introduction to Neural Information Retrieval. Found. Trends Inf. Retr., Vol. 13, 1 (dec 2018), 1--126. https://doi.org/10.1561/1500000061
    [33]
    Liang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu, and Xueqi Cheng. 2016. A Study of MatchPyramid Models on Ad-hoc Retrieval. https://doi.org/10.48550/ARXIV.1606.04648
    [34]
    Samarth Rawal and Chitta Baral. 2020. Multi-Perspective Semantic Information Retrieval. https://doi.org/10.48550/ARXIV.2009.01938
    [35]
    Stephen Robertson and Hugo Zaragoza. 2009. The Probabilistic Relevance Framework: BM25 and Beyond. Foundations and Trends® in Information Retrieval, Vol. 3, 4 (2009), 333--389. https://doi.org/10.1561/1500000019
    [36]
    Saar Kuzi, Mingyang Zhang 001, Cheng Li 012, Michael Bendersky, and Marc Najork. 2020. Leveraging Semantic and Lexical Matching to Improve the Recall of Document Retrieval Systems: A Hybrid Approach. CoRR (2020).
    [37]
    Sheikh Muhammad Sarwar, John Foley, Liu Yang, and James Allan. 2019. Sentence Retrieval for Entity List Extraction with a Seed, Context, and Topic. In Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval (Santa Clara, CA, USA) (ICTIR '19). Association for Computing Machinery, New York, NY, USA, 209--212. https://doi.org/10.1145/3341981.3344250
    [38]
    Yingxia Shao Zhao Cao Shitao Xiao, Zheng Liu. 2022. RetroMAE: Pre-Training Retrieval-oriented Language Models Via Masked Auto-Encoder. In EMNLP. https://arxiv.org/abs/2205.12035
    [39]
    Sivic and Zisserman. 2003. Video Google: a text retrieval approach to object matching in videos. In Proceedings Ninth IEEE International Conference on Computer Vision. 1470--1477 vol.2. https://doi.org/10.1109/ICCV.2003.1238663
    [40]
    Nandan Thakur, Nils Reimers, Andreas Rücklé, Abhishek Srivastava, and Iryna Gurevych. 2021. BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2). https://openreview.net/forum?id=wCu6T5xFjeJ
    [41]
    Peter D. Turney and Patrick Pantel. 2010. From Frequency to Meaning: Vector Space Models of Semantics. J. Artif. Int. Res., Vol. 37, 1 (jan 2010), 141--188.
    [42]
    Xiao Wang, Sean MacAvaney, Craig Macdonald, and Iadh Ounis. 2022. An Inspection of the Reproducibility and Replicability of TCT-ColBERT. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (Madrid, Spain) (SIGIR '22). Association for Computing Machinery, New York, NY, USA, 2790--2800. https://doi.org/10.1145/3477495.3531721
    [43]
    William Webber, Alistair Moffat, and Justin Zobel. 2010. A similarity measure for indefinite rankings. ACM Trans. Inf. Syst., Vol. 28 (2010), 20:1--20:38.
    [44]
    Lee Xiong, Chenyan Xiong, Ye Li, Kwok-Fung Tang, Jialin Liu, Paul Bennett, Junaid Ahmed, and Arnold Overwijk. 2021. Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text Retrieval. ArXiv, Vol. abs/2007.00808 (2021).
    [45]
    Xiaobing Xue, Van Dang, and W. Bruce Croft. 2009. Query Substitution based on N-gram Analysis. In SIGIR.
    [46]
    Andrew Yates, Rodrigo Nogueira, and Jimmy Lin. 2021. Pretrained Transformers for Text Ranking: BERT and Beyond. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Tutorials. Association for Computational Linguistics, Online, 1--4. https://doi.org/10.18653/v1/2021.naacl-tutorials.1
    [47]
    Jiangbo Yuan, Guillaume Gravier, Sébastien Campion, Xiuwen Liu, and Hervé Jégou. 2012. Efficient Mining of Repetitions in Large-Scale TV Streams with Product Quantization Hashing. In Computer Vision - ECCV 2012. Workshops and Demonstrations - Florence, Italy, October 7--13, 2012, Proceedings, Part I (Lecture Notes in Computer Science, Vol. 7583), Andrea Fusiello, Vittorio Murino, and Rita Cucchiara (Eds.). Springer, 271--280. https://doi.org/10.1007/978-3-642-33863-2_27
    [48]
    Han Zhang, Hongwei Shen, Yiming Qiu, Yunjiang Jiang, Songlin Wang, Sulong Xu, Yun Xiao, Bo Long, and Wen-Yun Yang. 2021. Joint Learning of Deep Retrieval Model and Product Quantization Based Embedding Index. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval (Virtual Event, Canada) (SIGIR '21). Association for Computing Machinery, New York, NY, USA, 1718--1722. https://doi.org/10.1145/3404835.3462988
    [49]
    Jingwei Zhuo, Ziru Xu, Wei Dai, Han Zhu, Han Li, Jian Xu, and Kun Gai. 2020. Learning Optimal Tree Models under Beam Search. In Proceedings of the 37th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 119), Hal Daumé III and Aarti Singh (Eds.). PMLR, 11650--11659. https://proceedings.mlr.press/v119/zhuo20a.html

    Cited By

    View all
    • (2024)Two-Step SPLADE: Simple, Efficient and Effective Approximation of SPLADEAdvances in Information Retrieval10.1007/978-3-031-56060-6_23(349-363)Online publication date: 16-Mar-2024
    • (2023)Genetic Generative Information RetrievalProceedings of the ACM Symposium on Document Engineering 202310.1145/3573128.3609340(1-4)Online publication date: 22-Aug-2023

    Index Terms

    1. Lexically-Accelerated Dense Retrieval

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      SIGIR '23: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval
      July 2023
      3567 pages
      ISBN:9781450394086
      DOI:10.1145/3539618
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 18 July 2023

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. adaptive re-ranking
      2. approximate k nearest neighbor
      3. dense retrieval

      Qualifiers

      • Research-article

      Conference

      SIGIR '23
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 792 of 3,983 submissions, 20%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)441
      • Downloads (Last 6 weeks)21

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Two-Step SPLADE: Simple, Efficient and Effective Approximation of SPLADEAdvances in Information Retrieval10.1007/978-3-031-56060-6_23(349-363)Online publication date: 16-Mar-2024
      • (2023)Genetic Generative Information RetrievalProceedings of the ACM Symposium on Document Engineering 202310.1145/3573128.3609340(1-4)Online publication date: 22-Aug-2023

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media