Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings
Abstract
In this paper, we explore Latent Dirichlet Allocation (LDA) [1] and Polylingual Latent Dirichlet Allocation (PolyLDA) [6], as a means to discover trending styles in Overstock1 from deep visual semantic features transferred from a pretrained convolutional neural network and text-based item attributes. To utilize deep visual semantic features in conjunction with LDA, we develop a method for creating a bag of words representation of unrolled image vectors. By viewing the channels within the convolutional layers of a Resnet-50 [2] as being representative of a word, we can index these activations to create visual documents. We then train LDA over these documents to discover the latent style in the images. We also incorporate text-based data with PolyLDA, where each representation is viewed as an independent language attempting to describe the same style. The resulting topics are shown to be excellent indicators of visual style across our platform.
Cite
Text
Iqbal et al. "Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018. doi:10.1109/CVPRW.2018.00253Markdown
[Iqbal et al. "Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/iqbal2018cvprw-discovering/) doi:10.1109/CVPRW.2018.00253BibTeX
@inproceedings{iqbal2018cvprw-discovering,
title = {{Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings}},
author = {Iqbal, Murium and Kovac, Adair and Aryafar, Kamelia},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2018},
pages = {1899-1901},
doi = {10.1109/CVPRW.2018.00253},
url = {https://mlanthology.org/cvprw/2018/iqbal2018cvprw-discovering/}
}