Sequence Complementor: Complementing Transformers for Time Series Forecasting with Learnable Sequences
Abstract
Since its introduction, the transformer has shifted the development trajectory away from traditional models (e.g., RNN, MLP) in time series forecasting, which is attributed to its ability to capture global dependencies within temporal tokens. Follow-up studies have largely involved altering the tokenization and self-attention modules to better adapt Transformers for addressing special challenges like non-stationarity, channel-wise dependency, and variable correlation in time series. However, we found that the expressive capability of sequence representation is a key factor influencing Transformer performance in time forecasting after investigating several representative methods, where there is an almost linear relationship between sequence representation entropy and mean square error, with more diverse representations performing better. In this paper, we propose a novel attention mechanism with Sequence Complementors and prove feasible from an information theory perspective, where these learnable sequences are able to provide complementary information beyond current input to feed attention. We further enhance the Sequence Complementors via a diversification loss that is theoretically covered. The empirical evaluation of both long-term and short-term forecasting has confirmed its superiority over the recent state-of-the-art methods.
Cite
Text
Chen et al. "Sequence Complementor: Complementing Transformers for Time Series Forecasting with Learnable Sequences." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I15.33747Markdown
[Chen et al. "Sequence Complementor: Complementing Transformers for Time Series Forecasting with Learnable Sequences." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/chen2025aaai-sequence/) doi:10.1609/AAAI.V39I15.33747BibTeX
@inproceedings{chen2025aaai-sequence,
title = {{Sequence Complementor: Complementing Transformers for Time Series Forecasting with Learnable Sequences}},
author = {Chen, Xiwen and Qiu, Peijie and Zhu, Wenhui and Li, Huayu and Wang, Hao and Sotiras, Aristeidis and Wang, Yalin and Razi, Abolfazl},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {15913-15921},
doi = {10.1609/AAAI.V39I15.33747},
url = {https://mlanthology.org/aaai/2025/chen2025aaai-sequence/}
}