Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models

  • Yuhao Wang
  • , Junwei Pan
  • , Pengyue Jia
  • , Wanyu Wang
  • , Maolin Wang
  • , Zhixiang Feng
  • , Xiaotian Li
  • , Jie Jiang
  • , Xiangyu Zhao*
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Sequential Recommendation (SR) aims to leverage the sequential patterns in users’ historical interactions to accurately track their preferences. However, the primary reliance of existing SR methods on collaborative data results in challenges such as the cold-start problem and sub-optimal performance. Concurrently, despite the proven effectiveness of large language models (LLMs), their integration into commercial recommender systems is impeded by issues such as high inference latency, incomplete capture of all distribution statistics, and catastrophic forgetting. To address these issues, we introduce a novel Pre-train, Align, and Disentangle (PAD) framework to enhance SR models with LLMs. In particular, we initially pre-train both the SR and LLM models to obtain collaborative and textual embeddings. Subsequently, we propose a characteristic recommendation-anchored alignment loss using multi-kernel maximum mean discrepancy with Gaussian kernels. Lastly, a triple-experts architecture, comprising aligned and modality-specific experts with disentangled embeddings, is fine-tuned in a frequency-aware manner. Experimental results on three public datasets validate the efficacy of PAD, indicating substantial enhancements and compatibility with various SR backbone models, particularly for cold items. The code and datasets are accessible for reproduction.

Original languageEnglish
Title of host publicationSIGIR 2025 - Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval
PublisherAssociation for Computing Machinery, Inc
Pages1455-1465
Number of pages11
ISBN (Electronic)9798400715921
DOIs
StatePublished - 13 Jul 2025
Externally publishedYes
Event48th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2025 - Padua, Italy
Duration: 13 Jul 202518 Jul 2025

Publication series

NameSIGIR 2025 - Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval

Conference

Conference48th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2025
Country/TerritoryItaly
CityPadua
Period13/07/2518/07/25

Keywords

  • Large Language Model
  • Recommender System
  • Reproducing Kernel Hilbert Space
  • Sequential Recommendation

Fingerprint

Dive into the research topics of 'Pre-train, Align, and Disentangle: Empowering Sequential Recommendation with Large Language Models'. Together they form a unique fingerprint.

Cite this