Cross-Modal Image Clustering via Canonical Correlation Analysis
Abstract
A new algorithm via Canonical Correlation Analysis (CCA) is developed in this paper to support more effective cross-modal image clustering for large-scale annotated image collections. It can be treated as a bi-media multimodal mapping problem and modeled as a correlation distribution over multimodal feature representations. It integrates the multimodal feature generation with the Locality Linear Coding (LLC) and co-occurrence association network, multimodal feature fusion with CCA, and accelerated hierarchical k-means clustering, which aims to characterize the correlations between the inter-related visual features in images and semantic features in captions, and measure their association degree more precisely. Very positive results were obtained in our experiments using a large quantity of public data.
Cite
Text
Jin et al. "Cross-Modal Image Clustering via Canonical Correlation Analysis." AAAI Conference on Artificial Intelligence, 2015. doi:10.1609/AAAI.V29I1.9181Markdown
[Jin et al. "Cross-Modal Image Clustering via Canonical Correlation Analysis." AAAI Conference on Artificial Intelligence, 2015.](https://mlanthology.org/aaai/2015/jin2015aaai-cross/) doi:10.1609/AAAI.V29I1.9181BibTeX
@inproceedings{jin2015aaai-cross,
title = {{Cross-Modal Image Clustering via Canonical Correlation Analysis}},
author = {Jin, Cheng and Mao, Wenhui and Zhang, Ruiqi and Zhang, Yuejie and Xue, Xiangyang},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2015},
pages = {151-159},
doi = {10.1609/AAAI.V29I1.9181},
url = {https://mlanthology.org/aaai/2015/jin2015aaai-cross/}
}