Cross-Task Affinity Learning for Multitask Dense Scene Predictions
Abstract
Multitask learning (MTL) has become prominent for its ability to predict multiple tasks jointly achieving better per-task performance with fewer parameters than single-task learning. Recently decoder-focused architectures have significantly improved multitask performance by refining task predictions using features from related tasks. However most refinement methods struggle to efficiently capture both local and long-range dependencies between task-specific representations and cross-task patterns. In this paper we introduce the Cross-Task Affinity Learning (CTAL) module a lightweight framework that enhances task refinement in multitask networks. CTAL effectively captures local and long-range cross-task interactions by optimizing task affinity matrices for parameter-efficient grouped convolutions without concern for information loss. Our results demonstrate state-of-the-art MTL performance for both CNN and transformer backbones using significantly fewer parameters than single-task learning.
Cite
Text
Sinodinos and Armanfard. "Cross-Task Affinity Learning for Multitask Dense Scene Predictions." Winter Conference on Applications of Computer Vision, 2025.Markdown
[Sinodinos and Armanfard. "Cross-Task Affinity Learning for Multitask Dense Scene Predictions." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/sinodinos2025wacv-crosstask/)BibTeX
@inproceedings{sinodinos2025wacv-crosstask,
title = {{Cross-Task Affinity Learning for Multitask Dense Scene Predictions}},
author = {Sinodinos, Dimitrios and Armanfard, Narges},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2025},
pages = {1546-1555},
url = {https://mlanthology.org/wacv/2025/sinodinos2025wacv-crosstask/}
}