Machine Learning: Science and Technology (Jan 2025)

Multi-exit Kolmogorov–Arnold networks: enhancing accuracy and parsimony

  • James Bagrow,
  • Josh Bongard

DOI
https://doi.org/10.1088/2632-2153/adf9bd
Journal volume & issue
Vol. 6, no. 3
p. 035037

Abstract

Read online

Kolmogorov–Arnold networks (KANs) uniquely combine high accuracy with interpretability, making them valuable for scientific modeling. However, it is unclear a priori how deep a network needs to be for any given task, and deeper KANs can be difficult to optimize and interpret. Here we introduce multi-exit KANs, where each layer includes its own prediction branch, enabling the network to make accurate predictions at multiple depths simultaneously. This architecture provides deep supervision that improves training while discovering the right level of model complexity for each task. Multi-exit KANs consistently outperform standard, single-exit versions on synthetic functions, dynamical systems, and real-world datasets. Remarkably, the best predictions often come from earlier, simpler exits, revealing that these networks naturally identify smaller, more parsimonious and interpretable models without sacrificing accuracy. To automate this discovery, we develop a differentiable ‘learning-to-exit’ algorithm that balances contributions from exits during training. Our approach offers scientists a practical way to achieve both high performance and interpretability, addressing a fundamental challenge in machine learning for scientific discovery.

Keywords