Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models

Abstract

In Multimodal Sentiment Analysis (MSA), data noise arising from various sources can lead to uncertainty in Aleatoric Uncertainty (AU), significantly impacting model performance. Current efforts to address AU have insufficiently explored its sources. They primarily focus on modeling noise rather than implementing targeted modeling based on its origin. Consequently, these approaches struggle to effectively mitigate the influence of AU, resulting in sustained limitations in model performance. Our research identifies that the AU primarily stems from two problems: subjective bias in the annotation process and the complex set relationships of sentiment features. To specifically address them, we propose DFMU, a Distribution-based Framework for Modeling Aleatoric Uncertainty, which incorporates an uncertainty modeling block capable of encoding uncertainty distributions and adaptively adjusting optimization objectives. Furthermore, we introduce distribution-based contrastive learning with sentiment words replacement to better capture the complex relationships among features. Extensive experiments on three public MSA datasets, i.e., MOSI, MOSEI, and SIMS, demonstrate that the proposed model maintains robust performance even under high noise conditions and achieves state-of-the-art results on these popular datasets.

Cite

Text

Wang et al. "Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models." International Joint Conference on Artificial Intelligence, 2024. doi:10.24963/ijcai.2024/917

Markdown

[Wang et al. "Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models." International Joint Conference on Artificial Intelligence, 2024.](https://mlanthology.org/ijcai/2024/wang2024ijcai-beyond/) doi:10.24963/ijcai.2024/917

BibTeX

@inproceedings{wang2024ijcai-beyond,
  title     = {{Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models}},
  author    = {Wang, Xindi and Salmani, Mahsa and Omidi, Parsa and Ren, Xiangyu and Rezagholizadeh, Mehdi and Eshaghi, Armaghan},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {8299-8307},
  doi       = {10.24963/ijcai.2024/917},
  url       = {https://mlanthology.org/ijcai/2024/wang2024ijcai-beyond/}
}