Abnar, Samira

10 publications

ICLRW 2025 From Dense to Dynamic: Token-Difficulty Driven MoEfication of Pre-Trained LLMs Kumari Nishu, Sachin Mehta, Samira Abnar, Mehrdad Farajtabar, Maxwell Horton, Mahyar Najibi, Moin Nabi, Minsik Cho, Devang Naik
ICML 2025 Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models Samira Abnar, Harshay Shah, Dan Busbridge, Alaaeldin El-Nouby, Joshua M. Susskind, Vimal Thilak
ICLRW 2025 Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models Samira Abnar, Harshay Shah, Dan Busbridge, Alaaeldin El-Nouby, Joshua M. Susskind, Vimal Thilak
NeurIPS 2024 Aggregate-and-Adapt Natural Language Prompts for Downstream Generalization of CLIP Chen Huang, Skyler Seto, Samira Abnar, David Grangier, Navdeep Jaitly, Josh Susskind
ICLR 2023 Diffusion Probabilistic Fields Peiye Zhuang, Samira Abnar, Jiatao Gu, Alex Schwing, Joshua M. Susskind, Miguel Ángel Bautista
ICLR 2022 Exploring the Limits of Large Scale Pre-Training Samira Abnar, Mostafa Dehghani, Behnam Neyshabur, Hanie Sedghi
NeurIPS 2022 GAUDI: A Neural Architect for Immersive 3D Scene Generation Miguel Angel Bautista, Pengsheng Guo, Samira Abnar, Walter Talbott, Alexander Toshev, Zhuoyuan Chen, Laurent Dinh, Shuangfei Zhai, Hanlin Goh, Daniel Ulbricht, Afshin Dehghan, Joshua Susskind
ICLR 2022 Scale Efficiently: Insights from Pretraining and Finetuning Transformers Yi Tay, Mostafa Dehghani, Jinfeng Rao, William Fedus, Samira Abnar, Hyung Won Chung, Sharan Narang, Dani Yogatama, Ashish Vaswani, Donald Metzler
ICLR 2021 Long Range Arena : A Benchmark for Efficient Transformers Yi Tay, Mostafa Dehghani, Samira Abnar, Yikang Shen, Dara Bahri, Philip Pham, Jinfeng Rao, Liu Yang, Sebastian Ruder, Donald Metzler
AAAI 2020 A Comparison of Architectures and Pretraining Methods for Contextualized Multilingual Word Embeddings Niels van der Heijden, Samira Abnar, Ekaterina Shutova