Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models
Abstract
In Multimodal Sentiment Analysis (MSA), data noise arising from various sources can lead to uncertainty in Aleatoric Uncertainty (AU), significantly impacting model performance. Current efforts to address AU have insufficiently explored its sources. They primarily focus on modeling noise rather than implementing targeted modeling based on its origin. Consequently, these approaches struggle to effectively mitigate the influence of AU, resulting in sustained limitations in model performance. Our research identifies that the AU primarily stems from two problems: subjective bias in the annotation process and the complex set relationships of sentiment features. To specifically address them, we propose DFMU, a Distribution-based Framework for Modeling Aleatoric Uncertainty, which incorporates an uncertainty modeling block capable of encoding uncertainty distributions and adaptively adjusting optimization objectives. Furthermore, we introduce distribution-based contrastive learning with sentiment words replacement to better capture the complex relationships among features. Extensive experiments on three public MSA datasets, i.e., MOSI, MOSEI, and SIMS, demonstrate that the proposed model maintains robust performance even under high noise conditions and achieves state-of-the-art results on these popular datasets.
Cite
Text
Wang et al. "Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models." International Joint Conference on Artificial Intelligence, 2024. doi:10.24963/ijcai.2024/917Markdown
[Wang et al. "Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/wang2024ijcai-beyond/) doi:10.24963/ijcai.2024/917BibTeX
@inproceedings{wang2024ijcai-beyond,
title = {{Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models}},
author = {Wang, Xindi and Salmani, Mahsa and Omidi, Parsa and Ren, Xiangyu and Rezagholizadeh, Mehdi and Eshaghi, Armaghan},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2024},
pages = {8299-8307},
doi = {10.24963/ijcai.2024/917},
url = {https://mlanthology.org/ijcai/2024/wang2024ijcai-beyond/}
}