Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2556195.2556220acmconferencesArticle/Chapter ViewAbstractPublication PageswsdmConference Proceedingsconference-collections
research-article

Modeling dwell time to predict click-level satisfaction

Published: 24 February 2014 Publication History

Abstract

Clicks on search results are the most widely used behavioral signals for predicting search satisfaction. Even though clicks are correlated with satisfaction, they can also be noisy. Previous work has shown that clicks are affected by position bias, caption bias, and other factors. A popular heuristic for reducing this noise is to only consider clicks with long dwell time, usually equaling or exceeding 30 seconds. The rationale is that the more time a searcher spends on a page, the more likely they are to be satisfied with its contents. However, having a single threshold value assumes that users need a fixed amount of time to be satisfied with any result click, irrespective of the page chosen. In reality, clicked pages can differ significantly. Pages have different topics, readability levels, content lengths, etc. All of these factors may affect the amount of time spent by the user on the page. In this paper, we study the effect of different page characteristics on the time needed to achieve search satisfaction. We show that the topic of the page, its length and its readability level are critical in determining the amount of dwell time needed to predict whether any click is associated with satisfaction. We propose a method to model and provide a better understanding of click dwell time. We estimate click dwell time distributions for SAT (satisfied) or DSAT (dissatisfied) clicks for different click segments and use them to derive features to train a click-level satisfaction model. We compare the proposed model to baseline methods that use dwell time and other search performance predictors as features, and demonstrate that the proposed model achieves significant improvements.

References

[1]
Ageev, M., Guo, Q., Lagun, D., and Agichtein, E. (2011). Find it if you can: a game for modeling different types of web search success using interaction data. Proc. SIGIR, 345--354.
[2]
Agichtein, E., Brill, E., and Dumais, S. (2006). Improving web search ranking by incorporating user behavior information. Proc. SIGIR, 19--26.
[3]
Agichtein, E., Brill, E., Dumais, S.T., and Ragno, R. (2006). Learning user interaction models for predicting web search result preferences. Proc. SIGIR, 3--10.
[4]
Bennett, P.N., Svore, K., and Dumais, S.T. (2010). Classification-enhanced ranking. Proc. WWW, 111--120.
[5]
Boldi, P., Bonchi, F., Castillo, C., Donato, D., Gionis, A., and Vigna, S. (2008). The query-flow graph: model and applications. Proc. CIKM, 609--618.
[6]
Buscher, G., van Elst, L., and Dengel, A. (2009). Segment-level display time as implicit feedback: a comparison to eye tracking. Proc. SIGIR, 67--74.
[7]
Cameron, A.C. and Trivedi, P.K. (1998). Regression Analysis of Count Data. Cambridge University Press.
[8]
Chang, C.-C. and Lin, C.-J. (2011). LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2(27): 1--27.
[9]
Choi, S.C. and Whette, R. (1969). Maximum likelihood estimation of the parameters of the gamma distribution and their bias. Technometrics, 11(4): 683--690.
[10]
Claypool, M., Le, P., Wased, M., and Brown, D. (2001). Implicit interest indicators. Proc. IUI, 33--40.
[11]
Collins-Thompson, K. and Callan, J. (2004). A language modeling approach to predicting reading difficulty. Proc. HLT, 193--200.
[12]
Cronen-Townsend, S., Zhou, Y., and Croft, W.B. (2002). Predicting query performance. Proc. SIGIR, 299--306.
[13]
Feild, H., Allan, J., and Jones, R. (2010). Predicting searcher frustration. Proc. SIGIR, 34--4
[14]
Fox, S., Karnawat, K., Mydland, M., Dumais, S., and White, T. (2005). Evaluating implicit measures to improve web search. ACM TOIS, 23(2): 147--168.
[15]
Friedman, J.H. (2001). Greedy function approximation: a gradient boosting machine. The Annals of Statistics, 29(5): 1189--1232.
[16]
Guo, Q. and Agichtein, E. (2012). Beyond dwell time: estimating document relevance from cursor movements and other post-click searcher behavior. Proc. WWW, 569--578.
[17]
Guo, Q., White, R.W., Dumais, S.T., Wang, J., and Anderson, B. (2010). Predicting query performance using query, result, and user interaction features. Proc. RIAO, 198--201.
[18]
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., and Witten, I.H. (2009). The WEKA data mining software: an update. SIGKDD Explorations, 11(1): 10--18.
[19]
Hassan, A. (2012). A semi-supervised approach to modeling web search satisfaction. Proc. SIGIR, 275--284.
[20]
Hassan, A., Jones, R., and Klinkner, K.L. (2010). Beyond DCG: user behavior as a predictor of a successful search. Proc. WSDM, 221--230.
[21]
Hassan, A., Shi, X., Craswell, N., and Ramsey, B. (2013). Beyond clicks: query reformulation as a predictor of search satisfaction. Proc. CIKM, 2019--2028.
[22]
Hassan, A., Song, Y., and He, L. (2011). A task level user satisfaction model and its application on improving relevance estimation. Proc. CIKM, 125--134.
[23]
He, B. and Ounis, I. (2006). Query performance prediction. Information System, 31(7): 585--594.
[24]
Hogg, R. V., McKean, J., and Craig, A.T. (2012). Introduction to Mathematical Statistics. Pearson, 7th Ed.
[25]
Huffman, S. and Hochster, M. (2007). How well does result relevance predict session satisfaction? Proc. SIGIR, 567--574.
[26]
T. Joachims. (2002). Optimizing search engines using click-through data. Proc. SIGKDD, 132--142.
[27]
Justel, A., Peña, D., and Zamar, R. (1997). A multivariate Kolmogorov-Smirnov test of goodness of fit. Statistics & Probability Letters, 35(3): 251--259.
[28]
Kang, I.-H. and Kim, G. (2003). Query type classification for web document retrieval. Proc. SIGIR, 64--71.
[29]
Kelly, D. and Teevan, J. (2003). Implicit feedback for infer-ring user preference: a bibliography. SIGIR Forum: 37(2).
[30]
Kidwell, P., Lebanon, G., and Collins-Thompson, K. (2009). Statistical estimation of word acquisition with application to readability prediction. Proc. EMNLP, 900--909.
[31]
Kelly, D. and Belkin, N.J. (2001). Reading time, scrolling, and interaction: exploring implicit sources of user preferences for relevance feedback. Proc. SIGIR, 408--409
[32]
Kelly, D. and Belkin, N.J. (2004). Display time as implicit feedback: understanding task effects. Proc. SIGIR, 377--384.
[33]
Kim, Y., Hassan, A., White, R.W., and Wang, Y.-M. (2013). Playing by the rules: mining query associations to predict search performance. Proc. WSDM, 133--142.
[34]
Leskovec, J., Dumais, S., and Horvitz, E. (2007). Web projections: learning from contextual sub graphs of the web. Proc. WWW, 471--480.
[35]
Liu C., White, R.W., and Dumais, S. (2010). Understanding web browsing behaviors through Weibull analysis of dwell time. Proc. SIGIR. 379--386.
[36]
Lloyd, S.P. (1982). Least squares quantization in PCM. IEEE Transactions on Information Theory, 28(2): 129--137.
[37]
Minka, T.P. (2002). Estimating a Gamma Distribution. Microsoft Research Technical Report.
[38]
Quinlan, J.R. (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers.
[39]
Radlinski, F., Kurup, M. and Joachims, T. (2008). How does clickthrough data reflect retrieval quality? Proc. CIKM, 43--52.
[40]
Wang, K., Thrasher, C., Viegas, E., Li, X. and Hsu, P. (2010). An overview of Microsoft web n-gram corpus and applications. NAACL HLT Demo Session, 45--48.
[41]
White, R.W. and Huang, J. (2010). Assessing the scenic route: measuring the value of search trails in web logs. Proc. Proc. SIGIR, 587--594.
[42]
White, R. W. and Kelly, D. (2006). A study on the effects of personalization and task information on implicit feedback performance. Proc. CIKM. 297--306.
[43]
Xue, G.-R., Xing, D., Yang, Q., and Yu, Y. (2008). Deep classification in large-scale text hierarchies. Proc. SIGIR, 619--626.
[44]
Yin, P., Luo, P., Lee, W.-C., and Wang, M. (2013). Silence is also evidence: interpreting dwell time for recommendation from psychological perspective. Proc. KDD, 989--997.
[45]
Zhou, Y., and Croft, W.B. (2006). Ranking robustness: a novel framework to predict query performance. Proc. CIKM, 567--574.
[46]
Zhou, Y. and Croft, W.B. (2007). Query performance prediction in web search environments. Proc. SIGIR, 543--550.

Cited By

View all
  • (2024)Interpretable Triplet Importance for Personalized RankingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679536(809-818)Online publication date: 21-Oct-2024
  • (2024)CWRCzech: 100M Query-Document Czech Click Dataset and Its Application to Web Relevance RankingProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3626772.3657851(1221-1231)Online publication date: 10-Jul-2024
  • (2024)Rethinking the Evaluation of Dialogue Systems: Effects of User Feedback on Crowdworkers and LLMsProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3626772.3657712(1952-1962)Online publication date: 10-Jul-2024
  • Show More Cited By

Index Terms

  1. Modeling dwell time to predict click-level satisfaction

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    WSDM '14: Proceedings of the 7th ACM international conference on Web search and data mining
    February 2014
    712 pages
    ISBN:9781450323512
    DOI:10.1145/2556195
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 February 2014

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. click satisfaction.
    2. dwell time analysis
    3. user behavior

    Qualifiers

    • Research-article

    Conference

    WSDM 2014

    Acceptance Rates

    WSDM '14 Paper Acceptance Rate 64 of 355 submissions, 18%;
    Overall Acceptance Rate 498 of 2,863 submissions, 17%

    Upcoming Conference

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)95
    • Downloads (Last 6 weeks)13
    Reflects downloads up to 09 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Interpretable Triplet Importance for Personalized RankingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679536(809-818)Online publication date: 21-Oct-2024
    • (2024)CWRCzech: 100M Query-Document Czech Click Dataset and Its Application to Web Relevance RankingProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3626772.3657851(1221-1231)Online publication date: 10-Jul-2024
    • (2024)Rethinking the Evaluation of Dialogue Systems: Effects of User Feedback on Crowdworkers and LLMsProceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval10.1145/3626772.3657712(1952-1962)Online publication date: 10-Jul-2024
    • (2024)Inverse Learning with Extremely Sparse Feedback for RecommendationProceedings of the 17th ACM International Conference on Web Search and Data Mining10.1145/3616855.3635797(396-404)Online publication date: 4-Mar-2024
    • (2024)Understanding Documentation Use Through Log Analysis: A Case Study of Four Cloud ServicesProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642721(1-17)Online publication date: 11-May-2024
    • (2024)A Self-Learning Framework for Large-Scale Conversational AI SystemsIEEE Computational Intelligence Magazine10.1109/MCI.2024.336397119:2(34-48)Online publication date: 5-Apr-2024
    • (2024)Individual Persistence Adaptation for User-Centric Evaluation of User Satisfaction in Recommender SystemsIEEE Access10.1109/ACCESS.2024.336069312(23626-23635)Online publication date: 2024
    • (2024)Dual perspective denoising model for session-based recommendationExpert Systems with Applications10.1016/j.eswa.2024.123845249(123845)Online publication date: Sep-2024
    • (2024)Robust enhanced collaborative filtering without explicit noise filteringThe Journal of Supercomputing10.1007/s11227-024-06086-w80:11(15763-15782)Online publication date: 6-Apr-2024
    • (2024)Modeling the impact of out-of-schema questions in task-oriented dialog systemsData Mining and Knowledge Discovery10.1007/s10618-024-01039-638:4(2466-2494)Online publication date: 4-Jun-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media