UniEmbedding: Learning Universal Multi-Modal Multi-Domain Item Embeddings via User-View Contrastive Learning
Citations Over TimeTop 10% of 2024 papers
Abstract
Learning high-quality item embeddings is crucial for recommendation tasks such as matching and ranking. However, existing methods often rely on ID-based item embeddings learned end-to-end with downstream recommendation models, which may suffer from overfitting and limited generalizability. In this paper, we aim to learn universal item embeddings (dubbed UniEmbedding) that capture multi-modal semantics, generalize across multiple domains, and serve different downstream tasks. To achieve this goal, we introduce the UniEmbedding pretraining framework, which includes three modules: a domain-aware multi-modal adapter, a user-view projection module, and contrastive learning objectives across domains. Compared to naive ID embeddings, UniEmbedding provides rich semantic information that generalizes more effectively across domains. Unlike multi-modal embeddings directly extracted from off-the-shelf pretrained models, UniEmbedding achieves better alignment between content semantics and behaviors. We evaluated UniEmbedding on both public and industrial datasets, demonstrating its effectiveness in matching and ranking tasks. Furthermore, UniEmbedding has been deployed in multiple recommendation applications at Huawei, resulting in significant gains in user engagement metrics.
Related Papers
- → How Many Cars in the City Are Too Many? Towards Finding the Optimal Modal Split for a Multi-Modal Urban Road Network(2021)14 cited
- Identification and validation of test modal parameters of gearbox housing in a certain tracked vehicle(2011)
- MODAL DIVERSION. IN: HANDBOOK OF TRANSPORT SYSTEMS AND TRAFFIC CONTROL(2001)
- → An Examination about the Modal Shift to the Marine Transportation by the Modal Split Model(2001)1 cited
- MODAL PARTICLES IN THE YUEYANG DIALECT(2002)