Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1007/978-3-031-39847-6_31guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

MIRS: [MASK] Insertion Based Retrieval Stabilizer for Query Variations

Published: 28 August 2023 Publication History

Abstract

Pre-trained Language Models (PLMs) have greatly pushed the frontier of document retrieval tasks. Recent studies, however, show that PLMs are vulnerable to query variations, i.e., queries containing misspellings or word re-ordering of original queries, and etc. Despite the increasing interest to robustify the retriever performance, the impact of the query variations is not fully exploited. To effectively address this problem, this paper revisits the Masked-Language Modeling (MLM) and proposes a robust fine-tuning algorithm, termed [MASK] Insertion based Retrieval Stabilizer (MIRS). The proposed algorithm differs from existing methods via the injection of [MASK] tokens into query variations and further encouraging the representation similarity between the pair of original queries and their variations. In comparison to MLM, the traditional [MASK] substitution-then-prediction is less emphasized in MIRS. Additionally, an in-depth analysis of our algorithm is also provided to reveal: (1) the latent representation (or semantic) of the original query forms a convex hull, while the impact of the query variation is then quantified as a “distortion” to this hull via deviating the hull vertices; and (2) inserted [MASK] tokens play a significant role in enlarging the intersection between the newly-formed hull (after variations) and the original one, thereby preserving more semantic from original queries. With the proposed [MASK] injection, MIRS exhibits a relative 1.8 MRR@10 absolute point enhancement on average in the retrieval accuracy, verified using 5 baselines across 3 public datasets with 4 types of query variations. We also provide intensive ablation studies to investigate the hyperparameter sensitiveness, to breakdown the model into individual components to manifest their efficacy, and further, to evaluate the out-of-domain model generalizability.

References

[1]
Chen, X., Luo, J., He, B., Sun, L., Sun, Y.: Towards robust dense retrieval via local ranking alignment. In: Raedt, L.D. (ed.) Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, pp. 1980–1986. International Joint Conferences on Artificial Intelligence Organization (7 2022)
[2]
Craswell, N., Mitra, B., Yilmaz, E., Campos, D., Voorhees, E.M.: Overview of the TREC 2019 deep learning track. In: Proceedings of the Twenty-Ninth Text REtrieval Conference (NIST Special Publication). National Institute of Standards and Technology (NIST) (2020)
[3]
El Boukkouri, H., Ferret, O., Lavergne, T., Noji, H., Zweigenbaum, P., Tsujii, J.: CharacterBERT: reconciling ELMo and BERT for word-level open-vocabulary representations from characters. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 6903–6915. Barcelona, Spain (Online) (2020)
[4]
Jeong, S., Baek, J., Cho, S., Hwang, S.J., Park, J.: Augmenting document representations for dense retrieval with interpolation and perturbation. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 442–452. Association for Computational Linguistics, Dublin, Ireland (2022)
[5]
Karpukhin, V., et al.: Dense passage retrieval for open-domain question answering. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 6769–6781. Association for Computational Linguistics, Online (Nov 2020)
[6]
Khattab, O., Zaharia, M.: ColBERT: efficient and effective passage search via contextualized late interaction over BERT, pp. 39–48. Association for Computing Machinery, New York, NY, USA (2020)
[7]
Ma, X., Nogueira dos Santos, C., Arnold, A.O.: Contrastive fine-tuning improves robustness for neural rankers. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 570–582. Association for Computational Linguistics, Online (Aug 2021)
[8]
Mao, Y., et al.: Generation-augmented retrieval for open-domain question answering. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 4089–4100. Association for Computational Linguistics, Online (2021)
[9]
Nguyen, T., Rosenberg, M., Song, X., Gao, J., Tiwary, S., Majumder, R., Deng, L.: Ms marco: A human generated machine reading comprehension dataset. In: Proceedings of the Workshop on Cognitive Computation: Integrating neural and symbolic approaches 2016 co-located with the 30th Annual Conference on Neural Information Processing Systems (CEUR Workshop Proceedings, Vol. 1773). CEUR-WS.org (2016)
[10]
Nogueira, R., Cho, K.: Passage re-ranking with bert. arXiv preprint arXiv:1901.04085 (2019)
[11]
Parkin L, Chardin B, Jean S, Hadjali A, and Baron M Strauss C, Kotsis G, Tjoa AM, and Khalil I Dealing with plethoric answers of SPARQL queries Database and Expert Systems Applications 2021 Cham Springer 292-304
[12]
Penha G, Câmara A, Hauff C, et al. Hagen M et al. Evaluating the robustness of retrieval pipelines with query variation generators Advances in Information Retrieval 2022 Cham Springer 397-412
[13]
Raman, N., Shah, S., Veloso, M.: Structure and semantics preserving document representations. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 780–790. SIGIR 2022, Association for Computing Machinery, New York, NY, USA (2022)
[14]
Robertson, S.E., Walker, S.: Some simple effective approximations to the 2-poisson model for probabilistic weighted retrieval. In: Croft, B.W., van Rijsbergen, C.J. (eds.) SIGIR 1994, pp. 232–241. SIGIR ’94, Springer, London (1994).
[15]
Sidiropoulos, G., Kanoulas, E.: Analysing the robustness of dual encoders for dense retrieval against misspellings. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2132–2136. SIGIR 2022, Association for Computing Machinery, New York, NY, USA (2022)
[16]
Tang, H., Sun, X., Jin, B., Wang, J., Zhang, F., Wu, W.: Improving document representations by generating pseudo query embeddings for dense retrieval. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 5054–5064. Association for Computational Linguistics (2021)
[17]
Xiong, L., et al.: Approximate nearest neighbor negative contrastive learning for dense text retrieval. In: International Conference on Learning Representations (2021)
[18]
Zhu, X., Hao, T., Cheng, S., Wang, F.L., Liu, H.: A self-supervised joint training framework for document reranking. In: Findings of the Association for Computational Linguistics: NAACL 2022, pp. 1056–1065. Association for Computational Linguistics, Seattle, United States (2022)
[19]
Zhuang, S., Zuccon, G.: Dealing with typos for BERT-based passage retrieval and ranking. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 2836–2842. Association for Computational Linguistics, Online and Punta Cana, Dominican Republic (Nov 2021)
[20]
Zhuang, S., Zuccon, G.: Characterbert and self-teaching for improving the robustness of dense retrievers on queries with typos. In: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 38–45. SIGIR ’22, Association for Computing Machinery, New York, NY, USA (2022)

Cited By

View all
  • (2024)Masking the unknownProceedings of the Fortieth Conference on Uncertainty in Artificial Intelligence10.5555/3702676.3702863(3997-4010)Online publication date: 15-Jul-2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Guide Proceedings
Database and Expert Systems Applications: 34th International Conference, DEXA 2023, Penang, Malaysia, August 28–30, 2023, Proceedings, Part I
Aug 2023
560 pages
ISBN:978-3-031-39846-9
DOI:10.1007/978-3-031-39847-6

Publisher

Springer-Verlag

Berlin, Heidelberg

Publication History

Published: 28 August 2023

Author Tags

  1. Document Retrieval
  2. Masked-Language Modeling
  3. Model Robustness
  4. Query Variations
  5. Query Representation

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 14 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Masking the unknownProceedings of the Fortieth Conference on Uncertainty in Artificial Intelligence10.5555/3702676.3702863(3997-4010)Online publication date: 15-Jul-2024

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media