Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleOctober 2024
Towards Completeness-Oriented Tool Retrieval for Large Language Models
CIKM '24: Proceedings of the 33rd ACM International Conference on Information and Knowledge ManagementPages 1930–1940https://doi.org/10.1145/3627673.3679847Recently, integrating external tools with Large Language Models (LLMs) has gained significant attention as an effective strategy to mitigate the limitations inherent in their pre-training data. However, real-world systems often incorporate a wide array ...
- research-articleOctober 2024
Hyperbolic Contrastive Learning for Cross-Domain Recommendation
CIKM '24: Proceedings of the 33rd ACM International Conference on Information and Knowledge ManagementPages 2920–2929https://doi.org/10.1145/3627673.3679572Cross-Domain Recommendation (CDR) seeks to utilize knowledge from different domains to alleviate the problem of data sparsity in the target recommendation domain, and has been gaining more attention in recent years. Although there have been notable ...
- ArticleSeptember 2024
HetCAN: A Heterogeneous Graph Cascade Attention Network with Dual-Level Awareness
Machine Learning and Knowledge Discovery in Databases. Research TrackPages 57–73https://doi.org/10.1007/978-3-031-70365-2_4AbstractHeterogeneous graph neural networks (HGNNs) have recently shown impressive capability in modeling heterogeneous graphs that are ubiquitous in real-world applications. Most existing methods for heterogeneous graphs mainly learn node embeddings by ...
- ArticleSeptember 2024
PSP: Pre-training and Structure Prompt Tuning for Graph Neural Networks
Machine Learning and Knowledge Discovery in Databases. Research TrackPages 423–439https://doi.org/10.1007/978-3-031-70362-1_25AbstractGraph Neural Networks (GNNs) are powerful in learning semantics of graph data. Recently, a new paradigm “pre-train & prompt” has shown promising results in adapting GNNs to various tasks with less supervised data. The success of such paradigm can ...
- research-articleAugust 2024
MPGraf: a modular and pre-trained graphformer for learning to rank at web-scale (extended abstract)
IJCAI '24: Proceedings of the Thirty-Third International Joint Conference on Artificial IntelligenceArticle No.: 937, Pages 8439–8443https://doi.org/10.24963/ijcai.2024/937Both Transformer and Graph Neural Networks (GNNs) have been used in learning to rank (LTR), however, they adhere to two distinct yet complementary problem formulations, i.e., ranking score regression based on query-webpage pairs and link prediction ...
-
- research-articleAugust 2024
GS2P: a generative pre-trained learning to rank model with over-parameterization for web-scale search (extended abstract)
IJCAI '24: Proceedings of the Thirty-Third International Joint Conference on Artificial IntelligenceArticle No.: 936, Pages 8433–8438https://doi.org/10.24963/ijcai.2024/936While learning to rank (LTR) is widely employed in web searches to prioritize pertinent webpages from the retrieved contents based on input queries, traditional LTR models stumble over two principal stumbling blocks leading to subpar performance: 1) the ...
- research-articleJuly 2024
Unsupervised Large Language Model Alignment for Information Retrieval via Contrastive Feedback
- Qian Dong,
- Yiding Liu,
- Qingyao Ai,
- Zhijing Wu,
- Haitao Li,
- Yiqun Liu,
- Shuaiqiang Wang,
- Dawei Yin,
- Shaoping Ma
SIGIR '24: Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information RetrievalPages 48–58https://doi.org/10.1145/3626772.3657689Large language models (LLMs) have demonstrated remarkable capabilities across various research domains, including the field of Information Retrieval (IR). However, the responses generated by off-the-shelf LLMs tend to be generic, i.e., cannot capture the ...
- research-articleJune 2024
Toward Bias-Agnostic Recommender Systems: A Universal Generative Framework
ACM Transactions on Information Systems (TOIS), Volume 42, Issue 6Article No.: 142, Pages 1–30https://doi.org/10.1145/3655617User behavior data, such as ratings and clicks, has been widely used to build personalizing models for recommender systems. However, many unflattering factors (e.g., popularity, ranking position, users’ selection) significantly affect the performance of ...
- research-articleApril 2024
Towards Flexible and Adaptive Neural Process for Cold-Start Recommendation
IEEE Transactions on Knowledge and Data Engineering (IEEECS_TKDE), Volume 36, Issue 4Pages 1815–1828https://doi.org/10.1109/TKDE.2023.3304839Recommender systems have been widely adopted in various online personal e-commerce applications for improving user experience. A long-standing challenge in recommender systems is how to provide accurate recommendation to users in cold-start situations ...
- articleMarch 2024
Exploring the Potential of Large Language Models (LLMs)in Learning on Graphs
- Zhikai Chen,
- Haitao Mao,
- Hang Li,
- Wei Jin,
- Hongzhi Wen,
- Xiaochi Wei,
- Shuaiqiang Wang,
- Dawei Yin,
- Wenqi Fan,
- Hui Liu,
- Jiliang Tang
ACM SIGKDD Explorations Newsletter (SIGKDD), Volume 25, Issue 2Pages 42–61https://doi.org/10.1145/3655103.3655110Learning on Graphs has attracted immense attention due to its wide real-world applications. The most popular pipeline for learning on graphs with textual node attributes primarily relies on Graph Neural Networks (GNNs), and utilizes shallow text ...
- surveyFebruary 2024
Explainability for Large Language Models: A Survey
- Haiyan Zhao,
- Hanjie Chen,
- Fan Yang,
- Ninghao Liu,
- Huiqi Deng,
- Hengyi Cai,
- Shuaiqiang Wang,
- Dawei Yin,
- Mengnan Du
ACM Transactions on Intelligent Systems and Technology (TIST), Volume 15, Issue 2Article No.: 20, Pages 1–38https://doi.org/10.1145/3639372Large language models (LLMs) have demonstrated impressive capabilities in natural language processing. However, their internal mechanisms are still unclear and this lack of transparency poses unwanted risks for downstream applications. Therefore, ...
- research-articleDecember 2023
Learning to tokenize for generative retrieval
- Weiwei Sun,
- Lingyong Yan,
- Zheng Chen,
- Shuaiqiang Wang,
- Haichao Zhu,
- Pengjie Ren,
- Zhumin Chen,
- Dawei Yin,
- Maarten de Rijke,
- Zhaochun Ren
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 2010, Pages 46345–46361As a new paradigm in information retrieval, generative retrieval directly generates a ranked list of document identifiers (docids) for a given query using generative language models (LMs). How to assign each document a unique docid (denoted as document ...
- research-articleDecember 2023
<bold>COLTR</bold>: Semi-Supervised Learning to Rank With Co-Training and Over-Parameterization for Web Search
- Yuchen Li,
- Haoyi Xiong,
- Qingzhong Wang,
- Linghe Kong,
- Hao Liu,
- Haifang Li,
- Jiang Bian,
- Shuaiqiang Wang,
- Guihai Chen,
- Dejing Dou,
- Dawei Yin
IEEE Transactions on Knowledge and Data Engineering (IEEECS_TKDE), Volume 35, Issue 12Pages 12542–12555https://doi.org/10.1109/TKDE.2023.3270750While <italic>learning to rank</italic> (LTR) has been widely used in web search to prioritize most relevant webpages among the retrieved contents subject to the input queries, the traditional LTR models fail to deliver decent performance due to two main ...
- research-articleOctober 2023
I3 Retriever: Incorporating Implicit Interaction in Pre-trained Language Models for Passage Retrieval
CIKM '23: Proceedings of the 32nd ACM International Conference on Information and Knowledge ManagementPages 441–451https://doi.org/10.1145/3583780.3614923Passage retrieval is a fundamental task in many information systems, such as web search and question answering, where both efficiency and effectiveness are critical concerns. In recent years, neural retrievers based on pre-trained language models (PLM), ...
- ArticleSeptember 2023
LtrGCN: Large-Scale Graph Convolutional Networks-Based Learning to Rank for Web Search
Machine Learning and Knowledge Discovery in Databases: Applied Data Science and Demo TrackPages 635–651https://doi.org/10.1007/978-3-031-43427-3_38AbstractWhile traditional Learning to Rank (LTR) models use query-webpage pairs to perform regression tasks to predict the ranking scores, they usually fail to capture the structure of interactions between queries and webpages over an extremely large ...
- research-articleAugust 2023
S2phere: Semi-Supervised Pre-training for Web Search over Heterogeneous Learning to Rank Data
KDD '23: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data MiningPages 4437–4448https://doi.org/10.1145/3580305.3599935While Learning to Rank (LTR) models on top of transformers have been widely adopted to achieve decent performance, it is still challenging to train the model with sufficient data as only an extremely small number of query-webpage pairs could be annotated ...
- research-articleAugust 2023
Semantic-Enhanced Differentiable Search Index Inspired by Learning Strategies
KDD '23: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data MiningPages 4904–4913https://doi.org/10.1145/3580305.3599903Recently, a new paradigm called Differentiable Search Index (DSI) has been proposed for document retrieval, wherein a sequence-to-sequence model is learned to directly map queries to relevant document identifiers. The key idea behind DSI is to fully ...
- short-paperJuly 2023
Graph Enhanced BERT for Query Understanding
SIGIR '23: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information RetrievalPages 3315–3319https://doi.org/10.1145/3539618.3591845Query understanding plays a key role in exploring users' search intents and facilitating users to locate their most desired information. However, it is inherently challenging since it needs to capture semantic information from short and ambiguous queries ...
- research-articleDecember 2022
Pre-trained Language Model-based Retrieval and Ranking for Web Search
- Lixin Zou,
- Weixue Lu,
- Yiding Liu,
- Hengyi Cai,
- Xiaokai Chu,
- Dehong Ma,
- Daiting Shi,
- Yu Sun,
- Zhicong Cheng,
- Simiu Gu,
- Shuaiqiang Wang,
- Dawei Yin
ACM Transactions on the Web (TWEB), Volume 17, Issue 1Article No.: 4, Pages 1–36https://doi.org/10.1145/3568681Pre-trained language representation models (PLMs) such as BERT and Enhanced Representation through kNowledge IntEgration (ERNIE) have been integral to achieving recent improvements on various downstream tasks, including information retrieval. However, it ...