Lewis, Mike

29 publications

NeurIPS 2025 FlexOLMo: Open Language Models for Flexible Data Use Weijia Shi, Akshita Bhagia, Kevin Farhat, Niklas Muennighoff, Jacob Morrison, Evan Pete Walsh, Dustin Schwenk, Shayne Longpre, Jake Poznanski, Allyson Ettinger, Daogao Liu, Margaret Li, Mike Lewis, Wen-tau Yih, Dirk Groeneveld, Luca Soldaini, Kyle Lo, Noah A. Smith, Luke Zettlemoyer, Pang Wei Koh, Hannaneh Hajishirzi, Ali Farhadi, Sewon Min
ICLR 2025 Law of the Weakest Link: Cross Capabilities of Large Language Models Ming Zhong, Aston Zhang, Xuewei Wang, Rui Hou, Wenhan Xiong, Chenguang Zhu, Zhengxing Chen, Liang Tan, Chloe Bi, Mike Lewis, Sravya Popuri, Sharan Narang, Melanie Kambadur, Dhruv Mahajan, Sergey Edunov, Jiawei Han, Laurens van der Maaten
TMLR 2025 Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models Weixin Liang, Lili Yu, Liang Luo, Srini Iyer, Ning Dong, Chunting Zhou, Gargi Ghosh, Mike Lewis, Wen-tau Yih, Luke Zettlemoyer, Xi Victoria Lin
ICLRW 2025 Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models Weixin Liang, Lili Yu, Liang Luo, Srini Iyer, Ning Dong, Chunting Zhou, Gargi Ghosh, Mike Lewis, Wen-tau Yih, Luke Zettlemoyer, Xi Victoria Lin
ICLRW 2025 Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models Weixin Liang, Lili Yu, Liang Luo, Srini Iyer, Ning Dong, Chunting Zhou, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Xi Victoria Lin
ICLR 2024 Efficient Streaming Language Models with Attention Sinks Guangxuan Xiao, Yuandong Tian, Beidi Chen, Song Han, Mike Lewis
ICLR 2024 In-Context Pretraining: Language Modeling Beyond Document Boundaries Weijia Shi, Sewon Min, Maria Lomeli, Chunting Zhou, Margaret Li, Xi Victoria Lin, Noah A. Smith, Luke Zettlemoyer, Wen-tau Yih, Mike Lewis
ICLR 2024 RA-DIT: Retrieval-Augmented Dual Instruction Tuning Xi Victoria Lin, Xilun Chen, Mingda Chen, Weijia Shi, Maria Lomeli, Richard James, Pedro Rodriguez, Jacob Kahn, Gergely Szilvasy, Mike Lewis, Luke Zettlemoyer, Wen-tau Yih
ICLR 2024 Self-Alignment with Instruction Backtranslation Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Omer Levy, Luke Zettlemoyer, Jason E Weston, Mike Lewis
ICML 2023 Coder Reviewer Reranking for Code Generation Tianyi Zhang, Tao Yu, Tatsunori Hashimoto, Mike Lewis, Wen-Tau Yih, Daniel Fried, Sida Wang
NeurIPSW 2023 FActScore: Fine-Grained Atomic Evaluation of Factual Precision in Long Form Text Generation Sewon Min, Kalpesh Krishna, Xinxi Lyu, Mike Lewis, Wen-tau Yih, Pang Wei Koh, Mohit Iyyer, Luke Zettlemoyer, Hannaneh Hajishirzi
ICLR 2023 InCoder: A Generative Model for Code Infilling and Synthesis Daniel Fried, Armen Aghajanyan, Jessy Lin, Sida Wang, Eric Wallace, Freda Shi, Ruiqi Zhong, Scott Yih, Luke Zettlemoyer, Mike Lewis
NeurIPS 2023 LIMA: Less Is More for Alignment Chunting Zhou, Pengfei Liu, Puxin Xu, Srinivasan Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu, Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Omer Levy
NeurIPS 2023 MEGABYTE: Predicting Million-Byte Sequences with Multiscale Transformers Lili Yu, Daniel Simig, Colin Flaherty, Armen Aghajanyan, Luke Zettlemoyer, Mike Lewis
ICLR 2023 Progressive Prompts: Continual Learning for Language Models Anastasia Razdaibiedina, Yuning Mao, Rui Hou, Madian Khabsa, Mike Lewis, Amjad Almahairi
ICML 2023 Retrieval-Augmented Multimodal Language Modeling Michihiro Yasunaga, Armen Aghajanyan, Weijia Shi, Richard James, Jure Leskovec, Percy Liang, Mike Lewis, Luke Zettlemoyer, Wen-Tau Yih
ICLR 2022 8-Bit Optimizers via Block-Wise Quantization Tim Dettmers, Mike Lewis, Sam Shleifer, Luke Zettlemoyer
NeurIPSW 2022 Branch-Train-Merge: Embarrassingly Parallel Training of Expert Language Models Margaret Li, Suchin Gururangan, Tim Dettmers, Mike Lewis, Tim Althoff, Noah A. Smith, Luke Zettlemoyer
NeurIPS 2022 GPT3.int8(): 8-Bit Matrix Multiplication for Transformers at Scale Tim Dettmers, Mike Lewis, Younes Belkada, Luke Zettlemoyer
ICLR 2022 HTLM: Hyper-Text Pre-Training and Prompting of Language Models Armen Aghajanyan, Dmytro Okhonko, Mike Lewis, Mandar Joshi, Hu Xu, Gargi Ghosh, Luke Zettlemoyer
ICLR 2022 Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation Ofir Press, Noah Smith, Mike Lewis
ICML 2021 BASE Layers: Simplifying Training of Large, Sparse Models Mike Lewis, Shruti Bhosale, Tim Dettmers, Naman Goyal, Luke Zettlemoyer
ICLR 2021 Nearest Neighbor Machine Translation Urvashi Khandelwal, Angela Fan, Dan Jurafsky, Luke Zettlemoyer, Mike Lewis
ICLR 2020 Generalization Through Memorization: Nearest Neighbor Language Models Urvashi Khandelwal, Omer Levy, Dan Jurafsky, Luke Zettlemoyer, Mike Lewis
NeurIPS 2020 Pre-Training via Paraphrasing Mike Lewis, Marjan Ghazvininejad, Gargi Ghosh, Armen Aghajanyan, Sida Wang, Luke Zettlemoyer
NeurIPS 2020 Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela
ICLR 2019 Generative Question Answering: Learning to Answer the Whole Question Mike Lewis, Angela Fan
NeurIPS 2019 Hierarchical Decision Making by Generating and Following Natural Language Instructions Hengyuan Hu, Denis Yarats, Qucheng Gong, Yuandong Tian, Mike Lewis
ICML 2018 Hierarchical Text Generation and Planning for Strategic Dialogue Denis Yarats, Mike Lewis