Maven: A Multimodal Foundation Model for Supernova Science

Abstract

We present Maven, a foundation model for supernova science. Maven is trained using self-supervised contrastive learning to align photometric and spectroscopic time-series observations in a shared embedding space. The model is first pre-trained on 0.5M synthetic supernovae, and then fine-tuned on 4,702 real observations from the Zwicky Transient Facility. Maven achieves state-of-the-art performance in supernova classification and redshift estimation, demonstrating the effectiveness of its learned embeddings for multiple downstream tasks. We find that pre-training with synthetic data significantly improves model performance. Maven has been designed to address the common challenge in astrophysics of consolidating sparse information-dense data with abundant lower-quality or synthetic data. Our approach offers a scalable solution for large, unlabeled, and multimodal astronomical datasets, and paves the way for upcoming projects like the Vera C. Rubin Observatory.

Cite

Text

Zhang et al. "Maven: A Multimodal Foundation Model for Supernova Science." NeurIPS 2024 Workshops: TSALM, 2024.

Markdown

[Zhang et al. "Maven: A Multimodal Foundation Model for Supernova Science." NeurIPS 2024 Workshops: TSALM, 2024.](https://mlanthology.org/neuripsw/2024/zhang2024neuripsw-maven-b/)

BibTeX

@inproceedings{zhang2024neuripsw-maven-b,
  title     = {{Maven: A Multimodal Foundation Model for Supernova Science}},
  author    = {Zhang, Gemma and Helfer, Thomas and Gagliano, Alexander Thomas and Mishra-Sharma, Siddharth and Villar, V Ashley},
  booktitle = {NeurIPS 2024 Workshops: TSALM},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/zhang2024neuripsw-maven-b/}
}