Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3397271.3401094acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

Training Curricula for Open Domain Answer Re-Ranking

Published: 25 July 2020 Publication History
  • Get Citation Alerts
  • Abstract

    In precision-oriented tasks like answer ranking, it is more important to rank many relevant answers highly than to retrieve all relevant answers. It follows that a good ranking strategy would be to learn how to identify the easiest correct answers first (i.e., assign a high ranking score to answers that have characteristics that usually indicate relevance, and a low ranking score to those with characteristics that do not), before incorporating more complex logic to handle difficult cases (e.g., semantic matching or reasoning). In this work, we apply this idea to the training of neural answer rankers using curriculum learning. We propose several heuristics to estimate the difficulty of a given training sample. We show that the proposed heuristics can be used to build a training curriculum that down-weights difficult samples early in the training process. As the training process progresses, our approach gradually shifts to weighting all samples equally, regardless of difficulty. We present a comprehensive evaluation of our proposed idea on three answer ranking datasets. Results show that our approach leads to superior performance of two leading neural ranking architectures, namely BERT and ConvKNRM, using both pointwise and pairwise losses. When applied to a BERT-based ranker, our method yields up to a 4% improvement in MRR and a 9% improvement in P@1 (compared to the model trained without a curriculum). This results in models that can achieve comparable performance to more expensive state-of-the-art techniques.

    References

    [1]
    Y. Bengio, J. Louradour, R. Collobert, and J. Weston. 2009. Curriculum Learning. In ICML. 41--48.
    [2]
    Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching Word Vectors with Subword Information. TACL, Vol. 5 (2017), 135--146.
    [3]
    X. Chen and A. Gupta. 2015. Weakly Supervised Learning of Convolutional Networks. In International Conference on Computer Vision. 1431--1439.
    [4]
    T. F. Coleman and Z. Wu. 1996. Parallel Continuation-based Global Optimization for Molecular Conformation and Protein Folding. Journal of Global Optimization, Vol. 8, 1 (January 1996), 49--65.
    [5]
    R. Collobert, J. Weston, L. Bottou, M. Karlen, K. Kavukcuoglu, and P. Kuksa. 2011. Natural Language Processing (Almost) from Scratch. The Journal of Machine Learning Research, Vol. 12 (August 2011), 2493--2537.
    [6]
    Nick Craswell, Bhaskar Mitra, and Daniel Campos. 2019. Overview of the TREC 2019 Deep Learning Track. In TREC.
    [7]
    Zhuyun Dai, Chenyan Xiong, Jamie Callan, and Zhiyuan Liu. 2018. Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search. In WSDM. 126--134.
    [8]
    Mostafa Dehghani, Arash Mehrjou, Stephan Gouws, Jaap Kamps, and Bernhard Schölkopf. 2017. Fidelity-Weighted Learning. In ICLR.
    [9]
    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL.
    [10]
    Laura Dietz, Ben Gamari, Jeff Dalton, and Nick Craswell. 2017. TREC Complex Answer Retrieval Overview. In TREC.
    [11]
    N. Ferro, C. Lucchese, M. Maistro, and R. Perego. 2018. Continuation Methods and Curriculum Learning for Learning to Rank. In CIKM. 1523--1526.
    [12]
    Jiafeng Guo, Yixing Fan, Qingyao Ai, and W. Bruce Croft. 2016. A Deep Relevance Matching Model for Ad-hoc Retrieval. In CIKM. 55--64. http://arxiv.org/abs/1711.08611 arXiv: 1711.08611.
    [13]
    Helia Hashemi, Mohammad Aliannejadi, Hamed Zamani, and W. Bruce Croft. 2019. ANTIQUE: A Non-Factoid Question Answering Benchmark. ArXiv, Vol. abs/1905.08957 (2019).
    [14]
    Baotian Hu, Zhengdong Lu, Hang Li, and Qingcai Chen. 2014a. Convolutional Neural Network Architectures for Matching Natural Language Sentences. In NIPS.
    [15]
    B. Hu, Z. Lu, H. Li, and Q. Chen. 2014b. Convolutional Neural Network Architectures for Matching Natural Language Sentences. In NIPS. 2042--2050.
    [16]
    Po-Sen Huang, Xiaodong He, Jianfeng Gao, Li Deng, Alex Acero, and Larry P. Heck. 2013. Learning deep structured semantic models for web search using clickthrough data. In CIKM.
    [17]
    Kai Hui, Andrew Yates, Klaus Berberich, and Gerard de Melo. 2018. Co-PACRR: A Context-Aware Neural IR Model for Ad-hoc Retrieval. In WSDM. ACM, 279--287.
    [18]
    Shiyu Ji, Jinjin Shao, and Tao Yang. 2019. Efficient Interaction-based Neural Ranking with Locality Sensitive Hashing. In WWW.
    [19]
    Lu Jiang, Deyu Meng, Shoou-I Yu, Zhen-Zhong Lan, Shiguang Shan, and Alexander G. Hauptmann. 2014. Self-Paced Learning with Diversity. In NIPS.
    [20]
    Lu Jiang, Deyu Meng, Qian Zhao, Shiguang Shan, and Alexander G. Hauptmann. 2015. Self-Paced Curriculum Learning. In AAAI.
    [21]
    Jimmy Lin. 2018. The Neural Hype and Comparisons Against Weak Baselines. SIGIR Forum, Vol. 52 (2018), 40--51.
    [22]
    Jimmy Lin and Peilin Yang. 2019. The Impact of Score Ties on Repeatability in Document Ranking. In SIGIR. http://arxiv.org/abs/1807.05798 arXiv: 1807.05798.
    [23]
    Sean MacAvaney, Andrew Yates, Arman Cohan, and Nazli Goharian. 2019. CEDR: Contextualized Embeddings for Document Ranking. In SIGIR 2019.
    [24]
    Sean MacAvaney, Andrew Yates, Arman Cohan, Luca Soldaini, Kai Hui, Nazli Goharian, and Ophir Frieder. 2018. Overcoming Low-Utility Facets for Complex Answer Retrieval. Information Retrieval Journal (2018).
    [25]
    Donald Metzler and W. Bruce Croft. 2005. A Markov random field model for term dependencies. In SIGIR.
    [26]
    Bhaskar Mitra and Nick Craswell. 2017. Neural Models for Information Retrieval. ArXiv, Vol. abs/1705.01509 (2017).
    [27]
    Tri Nguyen, Mir Rosenberg, Xia Song, Jianfeng Gao, Saurabh Tiwary, Rangan Majumder, and Li Deng. 2016. MS MARCO: A Human Generated MAchine Reading COmprehension Dataset. In CoCo@NIPS.
    [28]
    Rodrigo Nogueira and Kyunghyun Cho. 2019. Passage Re-ranking with BERT. arXiv:1901.04085 (2019).
    [29]
    Rodrigo Nogueira, Wei Yang, Jimmy Lin, and Kyunghyun Cho. 2019. Document Expansion by Query Prediction. ArXiv, Vol. abs/1904.08375 (2019).
    [30]
    Liang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu, and Xueqi Cheng. 2016. A Study of MatchPyramid Models on Ad-hoc Retrieval. In NeuIR @ SIGIR.
    [31]
    Liang Pang, Yanyan Lan, Jiafeng Guo, Jun Xu, Jingfang Xu, and Xueqi Cheng. 2017. DeepRank: A New Deep Architecture for Relevance Ranking in Information Retrieval. In CIKM. 257--266.
    [32]
    Gustavo Penha and Claudia Hauff. 2020. Curriculum Learning Strategies for IR: An Empirical Study on Conversation Response Ranking. In ECIR.
    [33]
    M. Qu, J. Tang, and J. Han. 2018. Curriculum Learning for Heterogeneous Star Network Embedding via Deep Reinforcement Learning. In WSDM. 468--476.
    [34]
    Mrinmaya Sachan and Eric P. Xing. 2016. Easy Questions First? A Case Study on Curriculum Learning for Question Answering. In ACL.
    [35]
    David W Scott. 2015. Multivariate density estimation: theory, practice, and visualization. John Wiley & Sons.
    [36]
    Yelong Shen, Xiaodong He, Jianfeng Gao, Li Deng, and Grégoire Mesnil. 2014. Learning semantic representations using convolutional neural networks for web search. In WWW.
    [37]
    Chenyan Xiong, Zhuyun Dai, Jamie Callan, Zhiyuan Liu, and Russell Power. 2017. End-to-End Neural Ad-hoc Ranking with Kernel Pooling. In SIGIR. 55--64. http://arxiv.org/abs/1706.06613 arXiv: 1706.06613.
    [38]
    Peilin Yang, Hui Fang, and Jimmy Lin. 2018. Anserini: Reproducible Ranking Baselines Using Lucene. J. Data and Information Quality, Vol. 10 (2018), 16:1--16:20.
    [39]
    Wei Yang, Kuang Lu, Peilin Yang, and Jimmy Lin. 2019. Critically Examining the "Neural Hype": Weak Baselines and the Additivity of Effectiveness Gains from Neural Ranking Models. In SIGIR.

    Cited By

    View all
    • (2024)Semantic Correlation Model of Socio-Formative Data for Curricular Planning EvaluationEuropean Journal of Educational Research10.12973/eu-jer.13.1.69volume-13-2024:volume-13-issue-1-january-2024(69-87)Online publication date: 15-Jan-2024
    • (2024)Unbiased, Effective, and Efficient Distillation from Heterogeneous Models for Recommender SystemsACM Transactions on Recommender Systems10.1145/3649443Online publication date: 23-Feb-2024
    • (2024)Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and ChallengesACM Computing Surveys10.1145/364847156:7(1-33)Online publication date: 14-Feb-2024
    • Show More Cited By

    Index Terms

    1. Training Curricula for Open Domain Answer Re-Ranking

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      SIGIR '20: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval
      July 2020
      2548 pages
      ISBN:9781450380164
      DOI:10.1145/3397271
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 25 July 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. curriculum learning
      2. neural re-ranking
      3. open domain question answering

      Qualifiers

      • Research-article

      Funding Sources

      • EU Horizon 2020 research and innovation programme
      • Italian Ministry of Education and Research (MIUR)

      Conference

      SIGIR '20
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 792 of 3,983 submissions, 20%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)8
      • Downloads (Last 6 weeks)1
      Reflects downloads up to 09 Aug 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Semantic Correlation Model of Socio-Formative Data for Curricular Planning EvaluationEuropean Journal of Educational Research10.12973/eu-jer.13.1.69volume-13-2024:volume-13-issue-1-january-2024(69-87)Online publication date: 15-Jan-2024
      • (2024)Unbiased, Effective, and Efficient Distillation from Heterogeneous Models for Recommender SystemsACM Transactions on Recommender Systems10.1145/3649443Online publication date: 23-Feb-2024
      • (2024)Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and ChallengesACM Computing Surveys10.1145/364847156:7(1-33)Online publication date: 14-Feb-2024
      • (2024)Scalable and Effective Generative Information RetrievalProceedings of the ACM on Web Conference 202410.1145/3589334.3645477(1441-1452)Online publication date: 13-May-2024
      • (2023)Achieving Human Parity on Visual Question AnsweringACM Transactions on Information Systems10.1145/357283341:3(1-40)Online publication date: 4-Apr-2023
      • (2022)From Easy to HardProceedings of the 31st ACM International Conference on Information & Knowledge Management10.1145/3511808.3557328(2784-2794)Online publication date: 17-Oct-2022
      • (2022)Curriculum Learning for Dense Retrieval DistillationProceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3477495.3531791(1979-1983)Online publication date: 6-Jul-2022
      • (2021)Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware SamplingProceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3404835.3462891(113-122)Online publication date: 11-Jul-2021

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media