Learning Compact Features via In-Training Representation Alignment
Abstract
Deep neural networks (DNNs) for supervised learning can be viewed as a pipeline of the feature extractor (i.e., last hidden layer) and a linear classifier (i.e., output layer) that are trained jointly with stochastic gradient descent (SGD) on the loss function (e.g., cross-entropy). In each epoch, the true gradient of the loss function is estimated using a mini-batch sampled from the training set and model parameters are then updated with the mini-batch gradients. Although the latter provides an unbiased estimation of the former, they are subject to substantial variances derived from the size and number of sampled mini-batches, leading to noisy and jumpy updates. To stabilize such undesirable variance in estimating the true gradients, we propose In-Training Representation Alignment (ITRA) that explicitly aligns feature distributions of two different mini-batches with a matching loss in the SGD training process. We also provide a rigorous analysis of the desirable effects of the matching loss on feature representation learning: (1) extracting compact feature representation; (2) reducing over-adaption on mini-batches via an adaptively weighting mechanism; and (3) accommodating to multi-modalities. Finally, we conduct large-scale experiments on both image and text classifications to demonstrate its superior performance to the strong baselines.
Cite
Text
Li et al. "Learning Compact Features via In-Training Representation Alignment." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I7.26044Markdown
[Li et al. "Learning Compact Features via In-Training Representation Alignment." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/li2023aaai-learning-b/) doi:10.1609/AAAI.V37I7.26044BibTeX
@inproceedings{li2023aaai-learning-b,
title = {{Learning Compact Features via In-Training Representation Alignment}},
author = {Li, Xin and Li, Xiangrui and Pan, Deng and Qiang, Yao and Zhu, Dongxiao},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {8675-8683},
doi = {10.1609/AAAI.V37I7.26044},
url = {https://mlanthology.org/aaai/2023/li2023aaai-learning-b/}
}