KAN Architecture
Kolmogorov-Arnold Networks (KANs) introduce a novel edge-based approach where learnable spline functions replace traditional fixed activation functions. Instead of applying non-linearities at nodes (like ReLU), each edge in a KAN applies a parameterized transformation , typically modeled using B-splines.
This design enables adaptive, interpretable, and compositional learning, inspired by the Kolmogorov–Arnold representation theorem. It enhances both model flexibility and explainability by learning customized activation behavior along each connection.
Figure 1: Schematic representation of the KAN architecture. Each layer applies learnable univariate spline functions along the edges (between nodes), enabling edge-based transformations rather than traditional node-based activations. While the classical Kolmogorov-Arnold formulation applied outer functions only at the final stage, modern deep KAN architectures generalize this by applying spline-based transformations at every layer, ensuring compositional and interpretable modeling across the network.
🌱 Key Variants of KAN
Since its introduction, the Kolmogorov-Arnold Network (KAN) has inspired several powerful architectural variants. Each is tailored to specific tasks or computational challenges, while preserving the core idea of edge-based, learnable spline activations.
- FastKAN: An optimized version of KAN that replaces B-splines with Gaussian RBFs for faster computation and improved scalability.
- T-KAN: A time-aware extension of KAN that incorporates temporal patterns, designed for sequence and time-series modeling.
- PDE-KAN: Combines KAN with physics-informed learning to solve partial differential equations, enforcing domain constraints through regularization.
- Wav-KAN: Integrates wavelet transformations into KAN, enhancing signal decomposition and localized feature extraction.
- SigKAN: Uses signature transforms to model temporal dynamics in continuous streams, improving interpretability in time-series data.
- GKAN: Adapts KAN to graph-structured data, supporting node-level and graph-level tasks via spline-transformed edge weights.
These variants demonstrate KAN's adaptability across domains such as physics, time-series analysis, graph learning, and signal processing.
📈 Research Impact & Community Engagement
This comprehensive survey has been officially published in ACM Computing Surveys, one of the most prestigious journals in computer science, reflecting its significance and contribution to the field. The work has garnered significant attention within the machine learning community, contributing to the growing understanding and adoption of Kolmogorov-Arnold Networks.
Beyond the academic publication, our research has fostered a vibrant community ecosystem. We maintain an awesome KAN repository - a curated collection of KAN (Kolmogorov-Arnold Network) resources including libraries, projects, tutorials, papers, and more, specifically designed for researchers and developers working in this exciting field. This repository serves as a comprehensive hub for anyone looking to explore, implement, or contribute to KAN-related research.
The work has been featured in various academic platforms and has sparked discussions about the future of interpretable neural architectures. To learn more about the broader impact of this research and its implications for the field, you can explore our publication story on Kudos, which provides additional insights into the research journey, community reception, and potential applications of KAN architectures.