Transform Once: Efficient Operator Learning in Frequency Domain
Abstract
Spectrum analysis provides one of the most effective paradigms for information-preserving dimensionality reduction in data: often, a simple description of naturally occurring signals can be obtained via few terms of periodic basis functions. Neural operators designed for frequency domain learning are based on complex-valued transforms i.e. Fourier Transforms (FT), and layers that perform computation on the spectrum and input data separately. This design introduces considerable computational overhead: for each layer, a forward and inverse FT. Instead, this work introduces a blueprint for frequency domain learning through a single transform: transform once (T1). Our results significantly streamline the design process of neural operators, pruning redundant transforms, and leading to speedups of 3 x to 30 that increase with data resolution and model size. We perform extensive experiments on learning to solve partial differential equations, including incompressible Navier-Stokes, turbulent flows around airfoils, and high-resolution video of smoke dynamics. T1 models improve on the test performance of SOTA neural operators while requiring significantly less computation, with over $30\%$ reduction in predictive error across tasks.
Cite
Text
Poli et al. "Transform Once: Efficient Operator Learning in Frequency Domain." ICML 2022 Workshops: AI4Science, 2022.Markdown
[Poli et al. "Transform Once: Efficient Operator Learning in Frequency Domain." ICML 2022 Workshops: AI4Science, 2022.](https://mlanthology.org/icmlw/2022/poli2022icmlw-transform/)BibTeX
@inproceedings{poli2022icmlw-transform,
title = {{Transform Once: Efficient Operator Learning in Frequency Domain}},
author = {Poli, Michael and Massaroli, Stefano and Berto, Federico and Park, Jinkyoo and Dao, Tri and Re, Christopher and Ermon, Stefano},
booktitle = {ICML 2022 Workshops: AI4Science},
year = {2022},
url = {https://mlanthology.org/icmlw/2022/poli2022icmlw-transform/}
}