Rudra, Atri

25 publications

ICLR 2025 Towards Learning High-Precision Least Squares Algorithms with Sequence Models Jerry Weihong Liu, Jessica Grogan, Owen M Dugan, Ashish Rao, Simran Arora, Atri Rudra, Christopher Re
ICMLW 2024 Can Transformers Solve Least Squares to High Precision? Jerry Weihong Liu, Jessica Grogan, Owen M Dugan, Simran Arora, Atri Rudra, Christopher Re
ICMLW 2024 Can Transformers Solve Least Squares to High Precision? Jerry Weihong Liu, Jessica Grogan, Owen M Dugan, Simran Arora, Atri Rudra, Christopher Re
ICMLW 2024 Just Read Twice: Closing the Recall Gap for Recurrent Language Models Simran Arora, Aman Timalsina, Aaryan Singhal, Sabri Eyuboglu, Xinyi Zhao, Ashish Rao, Atri Rudra, Christopher Re
ICML 2024 Simple Linear Attention Language Models Balance the Recall-Throughput Tradeoff Simran Arora, Sabri Eyuboglu, Michael Zhang, Aman Timalsina, Silas Alberti, James Zou, Atri Rudra, Christopher Re
ICMLW 2024 Simple Linear Attention Language Models Balance the Recall-Throughput Tradeoff Simran Arora, Sabri Eyuboglu, Michael Zhang, Aman Timalsina, Silas Alberti, Dylan Zinsley, James Zou, Atri Rudra, Christopher Re
ICMLW 2024 Towards Smaller Language Models via Layer Looping Sabri Eyuboglu, Dylan Zinsley, Jon Saad-Falcon, Simran Arora, Atri Rudra, James Zou, Christopher Re
ICLR 2024 Zoology: Measuring and Improving Recall in Efficient Language Models Simran Arora, Sabri Eyuboglu, Aman Timalsina, Isys Johnson, Michael Poli, James Zou, Atri Rudra, Christopher Re
ICLR 2023 How to Train Your HIPPO: State Space Models with Generalized Orthogonal Basis Projections Albert Gu, Isys Johnson, Aman Timalsina, Atri Rudra, Christopher Re
ICLR 2023 Hungry Hungry Hippos: Towards Language Modeling with State Space Models Daniel Y Fu, Tri Dao, Khaled Kamal Saab, Armin W Thomas, Atri Rudra, Christopher Re
NeurIPS 2023 Laughing Hyena Distillery: Extracting Compact Recurrences from Convolutions Stefano Massaroli, Michael Poli, Dan Fu, Hermann Kumbong, Rom Parnichkun, David Romero, Aman Timalsina, Quinn McIntyre, Beidi Chen, Atri Rudra, Ce Zhang, Christopher Ré, Stefano Ermon, Yoshua Bengio
NeurIPS 2023 Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture Dan Fu, Simran Arora, Jessica Grogan, Isys Johnson, Evan Sabri Eyuboglu, Armin Thomas, Benjamin Spector, Michael Poli, Atri Rudra, Christopher Ré
ICML 2023 Simple Hardware-Efficient Long Convolutions for Sequence Modeling Daniel Y Fu, Elliot L Epstein, Eric Nguyen, Armin W Thomas, Michael Zhang, Tri Dao, Atri Rudra, Christopher Re
ICLRW 2023 Simple Hardware-Efficient Long Convolutions for Sequence Modeling Daniel Y Fu, Elliot L Epstein, Eric Nguyen, Armin W Thomas, Michael Zhang, Tri Dao, Atri Rudra, Christopher Re
NeurIPS 2022 FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, Christopher Ré
ICML 2022 Monarch: Expressive Structured Matrices for Efficient and Accurate Training Tri Dao, Beidi Chen, Nimit S Sohoni, Arjun Desai, Michael Poli, Jessica Grogan, Alexander Liu, Aniruddh Rao, Atri Rudra, Christopher Re
ICLR 2022 Pixelated Butterfly: Simple and Efficient Sparse Training for Neural Network Models Beidi Chen, Tri Dao, Kaizhao Liang, Jiaming Yang, Zhao Song, Atri Rudra, Christopher Re
NeurIPS 2021 Combining Recurrent, Convolutional, and Continuous-Time Models with Linear State Space Layers Albert Gu, Isys Johnson, Karan Goel, Khaled Saab, Tri Dao, Atri Rudra, Christopher Ré
NeurIPS 2021 Scatterbrain: Unifying Sparse and Low-Rank Attention Beidi Chen, Tri Dao, Eric Winsor, Zhao Song, Atri Rudra, Christopher Ré
NeurIPS 2020 HiPPO: Recurrent Memory with Optimal Polynomial Projections Albert Gu, Tri Dao, Stefano Ermon, Atri Rudra, Christopher Ré
ICLR 2020 Kaleidoscope: An Efficient, Learnable Representation for All Structured Linear Maps Tri Dao, Nimit Sohoni, Albert Gu, Matthew Eichhorn, Amit Blonder, Megan Leszczynski, Atri Rudra, Christopher Ré
ICML 2019 Learning Fast Algorithms for Linear Transforms Using Butterfly Factorizations Tri Dao, Albert Gu, Matthew Eichhorn, Atri Rudra, Christopher Re
NeurIPS 2018 Learning Compressed Transforms with Low Displacement Rank Anna Thomas, Albert Gu, Tri Dao, Atri Rudra, Christopher Ré
CVPRW 2015 A Multiple Server Scheme for Fingerprint Fuzzy Vaults Jesse Hartloff, Matthew Morse, Bingsheng Zhang, Thomas Effland, Jennifer Cordaro, Jim Schuler, Sergey Tulyakov, Atri Rudra, Venu Govindaraju
CVPRW 2014 Secure Fingerprint Matching with Generic Local Structures Matthew Morse, Jesse Hartloff, Thomas Effland, Jim Schuler, Jennifer Cordaro, Sergey Tulyakov, Atri Rudra, Venu Govindaraju