Quick Links: 📋Blog | 🛠️Code | 🗃️Resource | 📭Citation |
This repository contains the source code for our Paper: ''Exploring Adapter-based Transfer Learning for Recommender Systems: Empirical Studies and Practical Insights''.
【blog】 (1)https://zhuanlan.zhihu.com/p/684389843/ (2)https://medium.com/@lifengyi_6964/parameter-efficient-adapter-transfer-for-multi-domain-recommendation-329bf3c32a60
🤗 Resources: four Large-scale datasets for evaluating foundation / transferable / multi-modal / LLM recommendaiton models.
-
NineRec(TPAMI): https://github.com/westlake-repl/NineRec
-
Tenrec(NeurIPS): https://github.com/yuangh-x/2022-NIPS-Tenrec
-
MicroLens(DeepMind Talk): https://github.com/westlake-repl/MicroLens
Adapters, a plug-in neural network module with some tunable parameters, have emerged as a parameter efficient transfer learning technique for adapting pre-trained models to downstream tasks, especially for natural language processing (NLP) and computer vision (CV) fields. Meanwhile, learning recommender system (RS) models directly from raw item modality features --- e.g., texts of NLP and images of CV --- can enable effective and transferable recommendations (called TransRec). In view of this, a natural question arises: can adapter-based learning techniques achieve parameter-efficient TransRec with good performance?
To this end, we perform empirical studies to address several key sub-questions. First, we ask whether the adapter-based TransRec performs comparably to TransRec based on standard full-parameter fine-tuning? does it hold for recommendation with different item modalities, e.g., textual RS and visual RS. If yes, we benchmark these existing adapters, which have been shown to be effective in NLP and CV tasks, in item recommendation tasks. Third, we carefully study several key factors for the adapter-based TransRec in terms of where and how to insert these adapters? Finally, we look at the effects of adapter-based TransRec by either scaling up its source training data or scaling down its target training data. Our paper provides key insights and practical guidance on unified & transferable recommendation --- a less studied recommendation scenario.
- torch == 1.7.1+cu110
- torchvision==0.8.2+cu110
- transformers==4.20.1
- loralib==0.1.1
The complete textual recommendation dataset (MIND & Adressa) is available under the Dataset directory.
Download the image files:
"hm_images.zip" for HM dataset from this link
"am_image.zip" for Amazon Clothes&Shoes dataset from this link
Then put them under Dataset/HM and Dataset/Amazon, respectively. Finally, please use the "build_lmdb_hm.py" scripts from IDvs.MoRec to generate the LMDB file.
We report details of the pre-trained Item ME we used in Table. Download the pytorch-version of them, and put the checkpoint pytorch_model.bin
into the corresponding path under pretrained_models/
Pre-trained model | #Param. | URL |
---|---|---|
BERTbase | 109M | https://huggingface.co/bert-base-uncased |
RoBERTabase | 125M | https://huggingface.co/roberta-base |
ViTbase | 86M | https://huggingface.co/google/vit-base-patch16-224 |
MAEbase | 86M | https://huggingface.co/facebook/vit-mae-base |
We provide the following checkpoints pre-trained on the source datasets:
SASRec+BERT, CPC+BERT, SASRec+ViT, and CPC+ViT
The SASRec/CPC+BERT pretrained checkpoint should be placed under Downstream/Text/pretrained_models and SASRec/CPC+ViT under Downstream/CV/pretrained_models.
An example of Source: For pretraining text TransRec:
cd Pretraining/Text/script
python sm_base_sasrec.py
We select the checkpoint with the best validation result as the pre-trained checkpoint for the downstream tasks. Please note that the pretraining process may take a considerable amount of time. Our pre-trained image recommendation model was trained on the source domain using four A40 GPUs for approximately 3 to 4 days.
An example of Downstream: For training text TransRec with the classic Houlsby adapter:
cd Downstream/Text/script
python adapter_houlsby.py
We report the test results of the checkpoint with the best validation result.
If you find our Adapter4Rec useful in your work, please cite our paper as:
@inproceedings{fu2024exploring,
title={Exploring adapter-based transfer learning for recommender systems: Empirical studies and practical insights},
author={Fu, Junchen and Yuan, Fajie and Song, Yu and Yuan, Zheng and Cheng, Mingyue and Cheng, Shenghui and Zhang, Jiaqi and Wang, Jie and Pan, Yunzhu},
booktitle={Proceedings of the 17th ACM International Conference on Web Search and Data Mining},
pages={208--217},
year={2024}
}