VisualLens: Personalization Through Task-Agnostic Visual History
Abstract
Existing recommendation systems either rely on user interaction logs, such as online shopping history for shopping recommendations, or focus on text signals. However, item-based histories are not always accessible and generalizable for multimodal recommendation. We hypothesize that a user's visual history --- comprising images from daily life --- can offer rich, task-agnostic insights into their interests and preferences, and thus be leveraged for effective personalization. To this end, we propose VisualLens, a novel framework that leverages multimodal large language models (MLLMs) to enable personalization using task-agnostic visual history. VisualLens extracts, filters, and refines a spectrum user profile from the visual history to support personalized recommendation. We created two new benchmarks, Google-Review-V and Yelp-V, with task-agnostic visual histories, and show that VisualLens improves over state-of-the-art item-based multimodal recommendations by 5-10\% on Hit@3, and outperforms GPT-4o by 2-5\%. Further analysis shows that VisualLens is robust across varying history lengths and excels at adapting to both longer histories and unseen content categories.
Cite
Text
Zhu et al. "VisualLens: Personalization Through Task-Agnostic Visual History." Advances in Neural Information Processing Systems, 2025.Markdown
[Zhu et al. "VisualLens: Personalization Through Task-Agnostic Visual History." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhu2025neurips-visuallens/)BibTeX
@inproceedings{zhu2025neurips-visuallens,
title = {{VisualLens: Personalization Through Task-Agnostic Visual History}},
author = {Zhu, Wang Bill and Fu, Deqing and Sun, Kai and Lu, Yi and Lin, Zhaojiang and Moon, Seungwhan and Narang, Kanika and Canim, Mustafa and Liu, Yue and Kumar, Anuj and Dong, Xin Luna},
booktitle = {Advances in Neural Information Processing Systems},
year = {2025},
url = {https://mlanthology.org/neurips/2025/zhu2025neurips-visuallens/}
}