Understanding the Limitations of B-Spline KANs: Convergence Dynamics and Computational Efficiency

Abstract

Kolmogorov-Arnold Networks (KANs) have recently emerged as a potential alternative to multi-layer perceptrons (MLPs), leveraging the Kolmogorov Representation Theorem to introduce learnable activation functions on each edge rather than fixed activations at the nodes. While KANs have demonstrated promise in small-scale problems by achieving similar or better performance with fewer parameters, our empirical investigations reveal significant limitations when scaling to real-world tasks. Specifically, KANs suffer from increased computational costs and reduced performance, rendering them unsuitable for deep learning applications. Our study explores these limitations through extensive testing across diverse tasks, including computer vision and scientific machine learning, and provides a detailed comparison with MLPs.

Cite

Text

Pal and Das. "Understanding the Limitations of B-Spline KANs: Convergence Dynamics and Computational Efficiency." NeurIPS 2024 Workshops: SciForDL, 2024.

Markdown

[Pal and Das. "Understanding the Limitations of B-Spline KANs: Convergence Dynamics and Computational Efficiency." NeurIPS 2024 Workshops: SciForDL, 2024.](https://mlanthology.org/neuripsw/2024/pal2024neuripsw-understanding/)

BibTeX

@inproceedings{pal2024neuripsw-understanding,
  title     = {{Understanding the Limitations of B-Spline KANs: Convergence Dynamics and Computational Efficiency}},
  author    = {Pal, Avik and Das, Dipankar},
  booktitle = {NeurIPS 2024 Workshops: SciForDL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/pal2024neuripsw-understanding/}
}