Sequential recommender systems predict items that may interest users by
modeling their preferences based on historical interactions. Traditional
sequential recommendation methods rely on capturing implicit collaborative
filtering signals among items. Recent relation-aware sequential recommendation
models have achieved promising performance by explicitly incorporating item
relations into the modeling of user historical sequences, where most relations
are extracted from knowledge graphs. However, existing methods rely on manually
predefined relations and suffer the sparsity issue, limiting the generalization
ability in diverse scenarios with varied item relations. In this paper, we
propose a novel relation-aware sequential recommendation framework with Latent
Relation Discovery (LRD). Different from previous relation-aware models that
rely on predefined rules, we propose to leverage the Large Language Model (LLM)
to provide new types of relations and connections between items. The motivation
is that LLM contains abundant world knowledge, which can be adopted to mine
latent relations of items for recommendation. Specifically, inspired by that
humans can describe relations between items using natural language, LRD
harnesses the LLM that has demonstrated human-like knowledge to obtain language
knowledge representations of items. These representations are fed into a latent
relation discovery module based on the discrete state variational autoencoder
(DVAE). Then the self-supervised relation discovery tasks and recommendation
tasks are jointly optimized. Experimental results on multiple public datasets
demonstrate our proposed latent relations discovery method can be incorporated
with existing relation-aware sequential recommendation models and significantly
improve the performance. Further analysis experiments indicate the
effectiveness and reliability of the discovered latent relations.Comment: Accepted by SIGIR 202