Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3394486.3403371acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Time-Aware User Embeddings as a Service

Published: 20 August 2020 Publication History

Abstract

Digital media companies typically collect rich data in the form of sequences of online user activities. Such data is used in various applications, involving tasks ranging from click or conversion prediction to recommendation or user segmentation. Nonetheless, each application depends upon specialized feature engineering that requires a lot of effort and typically disregards the time-varying nature of the online user behavior. Learning time-preserving vector representations of users (user embeddings), irrespective of a specific task, would save redundant effort and potentially lead to higher embedding quality. To that end, we address the limitations of the current state-of-the-art self-supervised methods for task-independent (unsupervised) sequence embedding, and propose a novel Time-Aware Sequential Autoencoder (TASA) that accounts for the temporal aspects of sequences of activities. The generated embeddings are intended to be readily accessible for many problem formulations and seamlessly applicable to desired tasks, thus sidestepping the burden of task-driven feature engineering. The proposed TASA shows improvements over alternative self-supervised models in terms of sequence reconstruction. Moreover, the embeddings generated by TASA yield increases in predictive performance on both proprietary and public data. It also achieves comparable results to supervised approaches that are trained on individual tasks separately and require substantially more computational effort. TASA has been incorporated within a pipeline designed to provide time-aware user embeddings as a service, and the use of its embeddings exhibited lifts in conversion prediction AUC on four audiences.

Supplementary Material

MP4 File (3394486.3403371.mp4)
In digital media companies, user activity histories are utilized by individual teams that work on different user-related tasks. However, each task depends on specialized feature engineering that requires redundant preprocessing and typically disregards the time-varying nature of online user behavior. Therefore, we propose the Time-Aware Sequential Autoencoder (TASA) to learn time-aware embeddings, intended to be seamlessly applicable to different user-related tasks, thus sidestepping the burden of task-driven feature engineering. TASA manifests sequence reconstruction improvements over alternative self-supervised models. Moreover, the embeddings generated by TASA yield increases in predictive performance on both proprietary and public data, and achieve comparable results to task-specific supervised models. Finally, TASA has been incorporated within a pipeline designed to provide time-aware user embeddings as a service, and the use of its embeddings exhibited conversion prediction lifts on four audiences.

References

[1]
Shahin Amiriparian, Michael Freitag, Nicholas Cummins, and Björn Schuller. 2017. Sequence to sequence autoencoders for unsupervised representation learning from audio. In Proc. of the DCASE 2017 Workshop.
[2]
Sarath Chandar AP, Stanislas Lauly, Hugo Larochelle, Mitesh Khapra, Balaraman Ravindran, Vikas C Raykar, and Amrita Saha. 2014. An autoencoder approach to learning bilingual word representations. In Advances in Neural Information Processing Systems. 1853--1861.
[3]
Tian Bai, Shanshan Zhang, Brian L Egleston, and Slobodan Vucetic. 2018. Interpretable representation learning for healthcare via capturing disease progression through time. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM, 43--51.
[4]
Dana H Ballard. 1987. Modular Learning in Neural Networks. In AAAI. 279--284.
[5]
Inci M Baytas, Cao Xiao, Xi Zhang, Fei Wang, Anil K Jain, and Jiayu Zhou. 2017. Patient subtyping via time-aware LSTM networks. In Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining. ACM, 65--74.
[6]
Sean Billings. 2018. Gradient Augmented Information Retrieval with Autoencoders and Semantic Hashing. arXiv preprint arXiv:1803.04494 (2018).
[7]
Hervé Bourlard and Yves Kamp. 1988. Auto-association by multilayer perceptrons and singular value decomposition. Biological cybernetics, Vol. 59, 4--5 (1988), 291--294.
[8]
Tianqi Chen and Carlos Guestrin. 2016. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 785--794.
[9]
Yu-An Chung, Chao-Chung Wu, Chia-Hao Shen, Hung-Yi Lee, and Lin-Shan Lee. 2016. Audio word2vec: Unsupervised learning of audio segment representations using sequence-to-sequence autoencoder. arXiv preprint arXiv:1603.00982 (2016).
[10]
Andrew M Dai and Quoc V Le. 2015. Semi-supervised sequence learning. In Advances in neural information processing systems. 3079--3087.
[11]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
[12]
Khoa D Doan, Pranjul Yadav, and Chandan K Reddy. 2019. Adversarial Factorization Autoencoder for Look-alike Modeling. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management. 2803--2812.
[13]
Djordje Gligorijevic, Jelena Gligorijevic, and Aaron Flores. 2019. Time-Aware Prospective Modeling of Users for Online Display Advertising. AdKDD 2019 workshop at the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2019).
[14]
Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep learning. MIT press.
[15]
Zhicheng He, Jie Liu, Na Li, and Yalou Huang. 2019. Learning Network-to-Network Model for Content-rich Network Embedding. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM, 1037--1045.
[16]
Geoffrey E Hinton and Sam T Roweis. 2003. Stochastic neighbor embedding. In Advances in neural information processing systems. 857--864.
[17]
Geoffrey E Hinton and Richard S Zemel. 1994. Autoencoders, minimum description length and Helmholtz free energy. In Advances in neural information processing systems. 3--10.
[18]
Sébastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. 2014. On using very large target vocabulary for neural machine translation. arXiv preprint arXiv:1412.2007 (2014).
[19]
Ian Jolliffe. 2011. Principal component analysis. Springer.
[20]
Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
[21]
Alex Krizhevsky and Geoffrey E Hinton. 2011. Using very deep autoencoders for content-based image retrieval. In ESANN, Vol. 1. 2.
[22]
Yann Le Cun. 1986. Modèles connexionnistes de l'apprentissage. These de Doctorat. Universite Paris (1986).
[23]
Daniel D Lee and H Sebastian Seung. 1999. Learning the parts of objects by non-negative matrix factorization. Nature, Vol. 401, 6755 (1999), 788.
[24]
Jiwei Li, Minh-Thang Luong, and Dan Jurafsky. 2015. A hierarchical neural autoencoder for paragraphs and documents. arXiv preprint arXiv:1506.01057 (2015).
[25]
Chin-Yew Lin. 2004. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out. 74--81.
[26]
Cheng-Yuan Liou, Wei-Chen Cheng, Jiun-Wei Liou, and Daw-Ran Liou. 2014. Autoencoder for words. Neurocomputing, Vol. 139 (2014), 84--96.
[27]
Minh-Thang Luong, Quoc V Le, Ilya Sutskever, Oriol Vinyals, and Lukasz Kaiser. 2015. Multi-task sequence to sequence learning. arXiv preprint arXiv:1511.06114 (2015).
[28]
Geoffrey McLachlan. 2004. Discriminant analysis and statistical pattern recognition. Vol. 544. John Wiley & Sons.
[29]
Al Mead. 1992. Review of the development of multidimensional scaling methods. Journal of the Royal Statistical Society: Series D (The Statistician), Vol. 41, 1 (1992), 27--39.
[30]
Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics. Association for Computational Linguistics, 311--318.
[31]
Viorica Patraucean, Ankur Handa, and Roberto Cipolla. 2015. Spatio-temporal video autoencoder with differentiable memory. arXiv preprint arXiv:1511.06309 (2015).
[32]
Wenjie Pei and David MJ Tax. 2018. Unsupervised Learning of Sequence Representations by Autoencoders. arXiv preprint arXiv:1804.00946 (2018).
[33]
Jonas Pfeiffer, Samuel Broscheit, Rainer Gemulla, and Mathias Göschl. 2018. A neural autoencoder approach for document ranking and query refinement in pharmacogenomic information retrieval. Association for Computational Linguistics.
[34]
Trang Pham, Truyen Tran, Dinh Phung, and Svetha Venkatesh. 2016. Deepcare: A deep dynamic memory model for predictive medicine. In Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer, 30--41.
[35]
Sam T Roweis and Lawrence K Saul. 2000. Nonlinear dimensionality reduction by locally linear embedding. science, Vol. 290, 5500 (2000), 2323--2326.
[36]
Ruslan Salakhutdinov and Geoffrey Hinton. 2009. Semantic hashing. International Journal of Approximate Reasoning, Vol. 50, 7 (2009), 969--978.
[37]
Nitish Srivastava, Elman Mansimov, and Ruslan Salakhudinov. 2015. Unsupervised learning of video representations using lstms. In International conference on machine learning. 843--852.
[38]
Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to sequence learning with neural networks. In Advances in neural information processing systems. 3104--3112.
[39]
Joshua B Tenenbaum, Vin De Silva, and John C Langford. 2000. A global geometric framework for nonlinear dimensionality reduction. science, Vol. 290, 5500 (2000), 2319--2323.
[40]
Vincent Wan, Yannis Agiomyrgiannakis, Hanna Silen, and Jakub Vit. 2017. Google's Next-Generation Real-Time Unit-Selection Synthesizer Using Sequence-to-Sequence LSTM-Based Autoencoders. In INTERSPEECH. 1143--1147.
[41]
Yuan Zhang, Xi Yang, Julie Ivy, and Min Chi. 2019. ATTAIN: Attention-based Time-Aware LSTM Networks for Disease Progression Modeling. IJCAI.
[42]
Yichao Zhou, Shaunak Mishra, Jelena Gligorijevic, Tarun Bhatia, and Narayan Bhamidipati. 2019. Understanding consumer journey using attention based recurrent neural networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. ACM, 3102--3111.

Cited By

View all
  • (2025)On the Use of Embedding Techniques for Modeling User Navigational Behavior in Intelligent Prefetching StrategiesConcurrency and Computation: Practice and Experience10.1002/cpe.835637:3Online publication date: 16-Jan-2025
  • (2024)Deep Journey Hierarchical Attention Networks for Conversion Predictions in Digital MarketingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3680066(4358-4365)Online publication date: 21-Oct-2024
  • (2022)How Is the Stroke? Inferring Shot Influence in Badminton Matches via Long Short-term DependenciesACM Transactions on Intelligent Systems and Technology10.1145/355139114:1(1-22)Online publication date: 9-Nov-2022
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '20: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
August 2020
3664 pages
ISBN:9781450379984
DOI:10.1145/3394486
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 20 August 2020

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. neural embeddings
  2. sequential models
  3. user representation

Qualifiers

  • Research-article

Funding Sources

  • Yahoo! Research "Faculty Research and Engagement Program"
  • National Science Foundation (NSF)

Conference

KDD '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Upcoming Conference

KDD '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)36
  • Downloads (Last 6 weeks)4
Reflects downloads up to 08 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2025)On the Use of Embedding Techniques for Modeling User Navigational Behavior in Intelligent Prefetching StrategiesConcurrency and Computation: Practice and Experience10.1002/cpe.835637:3Online publication date: 16-Jan-2025
  • (2024)Deep Journey Hierarchical Attention Networks for Conversion Predictions in Digital MarketingProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3680066(4358-4365)Online publication date: 21-Oct-2024
  • (2022)How Is the Stroke? Inferring Shot Influence in Badminton Matches via Long Short-term DependenciesACM Transactions on Intelligent Systems and Technology10.1145/355139114:1(1-22)Online publication date: 9-Nov-2022
  • (2022)Unsupervised Representation Learning of Player Behavioral Data with Confidence Guided MaskingProceedings of the ACM Web Conference 202210.1145/3485447.3512275(3396-3406)Online publication date: 25-Apr-2022
  • (2022)Optimizing Modality Weights in Topic Models of Transactional DataAutomation and Remote Control10.1134/S0005117922012005083:12(1908-1922)Online publication date: 1-Dec-2022
  • (2022)Field-aware Variational Autoencoders for Billion-scale User Representation Learning2022 IEEE 38th International Conference on Data Engineering (ICDE)10.1109/ICDE53745.2022.00321(3413-3425)Online publication date: May-2022
  • (2021)Time-Sequential Variational Conditional Auto-encoders for Recommendation行動時刻を考慮した条件付き変分オートエンコーダによる推薦システムTransactions of the Japanese Society for Artificial Intelligence10.1527/tjsai.36-3_C-KB736:3(C-KB7_1-10)Online publication date: 1-May-2021
  • (2021)CoPEProceedings of the 30th ACM International Conference on Information & Knowledge Management10.1145/3459637.3482419(2627-2636)Online publication date: 26-Oct-2021
  • (2021)Variable Interval Time Sequence Modeling for Career Trajectory Prediction: Deep Collaborative PerspectiveProceedings of the Web Conference 202110.1145/3442381.3449959(612-623)Online publication date: 19-Apr-2021
  • (2021)UIAE: Collaborative Filtering for User and Item based on Auto-Encoder2021 7th Annual International Conference on Network and Information Systems for Computers (ICNISC)10.1109/ICNISC54316.2021.00008(1-6)Online publication date: Jul-2021
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media