Bahri, Dara

22 publications

TMLR 2026 Improving Detection of Watermarked Language Models Dara Bahri, John Frederick Wieting
ICLRW 2025 A Watermark for Black-Box Language Models Dara Bahri, John Frederick Wieting
TMLR 2025 Decoding-Based Regression Xingyou Song, Dara Bahri
ICML 2024 A Universal Class of Sharpness-Aware Minimization Algorithms Behrooz Tahmasebi, Ashkan Soleymani, Dara Bahri, Stefanie Jegelka, Patrick Jaillet
ICMLW 2024 A Universal Class of Sharpness-Aware Minimization Algorithms Behrooz Tahmasebi, Ashkan Soleymani, Dara Bahri, Stefanie Jegelka, Patrick Jaillet
NeurIPS 2023 Sharpness-Aware Minimization Leads to Low-Rank Features Maksym Andriushchenko, Dara Bahri, Hossein Mobahi, Nicolas Flammarion
ICLR 2023 UL2: Unifying Language Learning Paradigms Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Dara Bahri, Tal Schuster, Steven Zheng, Denny Zhou, Neil Houlsby, Donald Metzler
ICLR 2022 Charformer: Fast Character Transformers via Gradient-Based Subword Tokenization Yi Tay, Vinh Q. Tran, Sebastian Ruder, Jai Gupta, Hyung Won Chung, Dara Bahri, Zhen Qin, Simon Baumgartner, Cong Yu, Donald Metzler
ICLR 2022 Churn Reduction via Distillation Heinrich Jiang, Harikrishna Narasimhan, Dara Bahri, Andrew Cotter, Afshin Rostamizadeh
NeurIPS 2022 Confident Adaptive Language Modeling Tal Schuster, Adam Fisch, Jai Gupta, Mostafa Dehghani, Dara Bahri, Vinh Tran, Yi Tay, Donald Metzler
ICLR 2022 ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning Vamsi Aribandi, Yi Tay, Tal Schuster, Jinfeng Rao, Huaixiu Steven Zheng, Sanket Vaibhav Mehta, Honglei Zhuang, Vinh Q. Tran, Dara Bahri, Jianmo Ni, Jai Gupta, Kai Hui, Sebastian Ruder, Donald Metzler
ICLR 2022 Scarf: Self-Supervised Contrastive Learning Using Random Feature Corruption Dara Bahri, Heinrich Jiang, Yi Tay, Donald Metzler
NeurIPS 2022 Transformer Memory as a Differentiable Search Index Yi Tay, Vinh Tran, Mostafa Dehghani, Jianmo Ni, Dara Bahri, Harsh Mehta, Zhen Qin, Kai Hui, Zhe Zhao, Jai Gupta, Tal Schuster, William W. Cohen, Donald Metzler
NeurIPSW 2021 An Empirical Study of Pre-Trained Vision Models on Out-of-Distribution Generalization Yaodong Yu, Heinrich Jiang, Dara Bahri, Hossein Mobahi, Seungyeon Kim, Ankit Singh Rawat, Andreas Veit, Yi Ma
ICLR 2021 HyperGrid Transformers: Towards a Single Model for Multiple Tasks Yi Tay, Zhe Zhao, Dara Bahri, Donald Metzler, Da-Cheng Juan
ICML 2021 Locally Adaptive Label Smoothing Improves Predictive Churn Dara Bahri, Heinrich Jiang
ICLR 2021 Long Range Arena : A Benchmark for Efficient Transformers Yi Tay, Mostafa Dehghani, Samira Abnar, Yikang Shen, Dara Bahri, Philip Pham, Jinfeng Rao, Liu Yang, Sebastian Ruder, Donald Metzler
ICML 2021 OmniNet: Omnidirectional Representations from Transformers Yi Tay, Mostafa Dehghani, Vamsi Aribandi, Jai Gupta, Philip M Pham, Zhen Qin, Dara Bahri, Da-Cheng Juan, Donald Metzler
ICML 2021 Synthesizer: Rethinking Self-Attention for Transformer Models Yi Tay, Dara Bahri, Donald Metzler, Da-Cheng Juan, Zhe Zhao, Che Zheng
ICML 2020 Deep k-NN for Noisy Labels Dara Bahri, Heinrich Jiang, Maya Gupta
ICML 2020 Sparse Sinkhorn Attention Yi Tay, Dara Bahri, Liu Yang, Donald Metzler, Da-Cheng Juan
NeurIPS 2018 Diminishing Returns Shape Constraints for Interpretability and Regularization Maya Gupta, Dara Bahri, Andrew Cotter, Kevin Canini