Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Auxiliary Task Guided Interactive Attention Model for Question Difficulty Prediction

  • Conference paper
  • First Online:
Artificial Intelligence in Education (AIED 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13355))

Included in the following conference series:

  • 4662 Accesses

Abstract

Online learning platforms conduct exams to evaluate the learners in a monotonous way, where the questions in the database may be classified into Bloom’s Taxonomy as varying levels in complexity from basic knowledge to advanced evaluation. The questions asked in these exams to all learners are very much static. It becomes important to ask new questions with different difficulty levels to each learner to provide a personalized learning experience. In this paper, we propose a multi-task method with an interactive attention mechanism, Qdiff, for jointly predicting Bloom’s Taxonomy and difficulty levels of academic questions. We model the interaction between the predicted bloom taxonomy representations and the input representations using an attention mechanism to aid in difficulty prediction. The proposed learning method would help learn representations that capture the relationship between Bloom’s taxonomy and difficulty labels. The proposed multi-task method learns a good input representation by leveraging the relationship between the related tasks and can be used in similar settings where the tasks are related. The results demonstrate that the proposed method performs better than training only on difficulty prediction. However, Bloom’s labels may not always be given for some datasets. Hence we soft label another dataset with a model fine-tuned to predict Bloom’s labels to demonstrate the applicability of our method to datasets with only difficulty labels.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://www.prepscholar.com/gre/blog/how-is-the-gre-scored/.

  2. 2.

    We don’t disclose the identity of the source due to the anonymity requirement.

  3. 3.

    Kinder-garden to grade-12.

References

  1. Alsubait, T., Parsia, B., Sattler, U.: A similarity-based theory of controlling MCQ difficulty. In: ICEEE 2013, pp. 283–288 (2013)

    Google Scholar 

  2. Benedetto, L., Aradelli, G., Cremonesi, P., Cappelli, A., Giussani, A., Turrin, R.: On the application of transformers for estimating the difficulty of multiple-choice questions from text. In: Proceedings of the 16th Workshop on Innovative Use of NLP for Building Educational Applications, Online, pp. 147–157. ACL (April 2021)

    Google Scholar 

  3. Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. CoRR abs/1810.04805 (2018)

    Google Scholar 

  4. François, T., Miltsakaki, E.: Do NLP and machine learning improve traditional readability formulas? In: Proceedings of the 1st Workshop on Predicting and Improving Text Readability for Target Reader Populations, Montréal, Canada. ACL (June 2012)

    Google Scholar 

  5. Gogus, A.: Bloom’s taxonomy of learning objectives, pp. 469–473 (2012)

    Google Scholar 

  6. Ha, L.A., Yaneva, V.: Automatic distractor suggestion for multiple-choice tests using concept embeddings and information retrieval. In: BEA (June 2018)

    Google Scholar 

  7. Howard, J., Ruder, S.: Universal language model fine-tuning for text classification. arXiv preprint arXiv:1801.06146 (2018)

  8. Liu, J., Wang, Q., Lin, C.Y., Hon, H.W.: Question difficulty estimation in community question answering services. In: Proceedings of the 2013 EMNLP, Seattle, Washington, USA. ACL (October 2013)

    Google Scholar 

  9. Nadeem, F., Ostendorf, M.: Language based mapping of science assessment items to skills. In: Proceedings of the 12th Workshop on Innovative Use of NLP for Building Educational Applications, Copenhagen, Denmark. ACL (September 2017)

    Google Scholar 

  10. Padó, U.: Question difficulty - how to estimate without norming, how to use for automated grading. In: BEA, Copenhagen, Denmark. ACL (September 2017)

    Google Scholar 

  11. Peters, M., et al.: Deep contextualized word representations. In: Proceedings of the 2018 Conference of the NAACL, Volume 1 (Long Papers). ACL (June 2018)

    Google Scholar 

  12. Ruder, S.: An overview of multi-task learning in deep neural networks. arXiv preprint arXiv:1706.05098 (2017)

  13. Smith, N.A., Heilman, M., Hwa, R.: Question generation as a competitive undergraduate course project. In: Proceedings of the NSF Workshop on the Question Generation Shared Task and Evaluation Challenge, pp. 4–6 (2008)

    Google Scholar 

  14. Supraja, S., Hartman, K., Tatinati, S., Khong, A.W.H.: Toward the automatic labeling of course questions for ensuring their alignment with learning outcomes. In: EDM (2017)

    Google Scholar 

  15. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)

    Google Scholar 

  16. Wang, Q., Liu, J., Wang, B., Guo, L.: A regularized competition model for question difficulty estimation in community question answering services. In: EMNLP 2014, pp. 1115–1126 (2014)

    Google Scholar 

  17. Xue, K., Yaneva, V., Runyon, C., Baldwin, P.: Predicting the difficulty and response time of multiple choice questions using transfer learning. In: BEA. ACL (July 2020)

    Google Scholar 

  18. Yaneva, V., Orăsan, C., Evans, R., Rohanian, O.: Combining multiple corpora for readability assessment for people with cognitive disabilities. In: BEA (September 2017)

    Google Scholar 

Download references

Acknowledgements

The authors acknowledge the support of Extramarks Education India Pvt. Ltd., SERB, FICCI (PM fellowship), Infosys Centre for AI and TiH Anubhuti (IIITD).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to V. Venktesh .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Venktesh, V., Akhtar, M.S., Mohania, M., Goyal, V. (2022). Auxiliary Task Guided Interactive Attention Model for Question Difficulty Prediction. In: Rodrigo, M.M., Matsuda, N., Cristea, A.I., Dimitrova, V. (eds) Artificial Intelligence in Education. AIED 2022. Lecture Notes in Computer Science, vol 13355. Springer, Cham. https://doi.org/10.1007/978-3-031-11644-5_39

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-11644-5_39

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-11643-8

  • Online ISBN: 978-3-031-11644-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics