CFP last date
20 February 2025
Reseach Article

Enhancing the Performance of Kolmogorov-Arnold Networks (KAN) using Residual Activations and Xavier Initialization

by Subhasis Mitra
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
Volume 186 - Number 48
Year of Publication: 2024
Authors: Subhasis Mitra
10.5120/ijca2024924146

Subhasis Mitra . Enhancing the Performance of Kolmogorov-Arnold Networks (KAN) using Residual Activations and Xavier Initialization. International Journal of Computer Applications. 186, 48 ( Nov 2024), 52-54. DOI=10.5120/ijca2024924146

@article{ 10.5120/ijca2024924146,
author = { Subhasis Mitra },
title = { Enhancing the Performance of Kolmogorov-Arnold Networks (KAN) using Residual Activations and Xavier Initialization },
journal = { International Journal of Computer Applications },
issue_date = { Nov 2024 },
volume = { 186 },
number = { 48 },
month = { Nov },
year = { 2024 },
issn = { 0975-8887 },
pages = { 52-54 },
numpages = {9},
url = { https://ijcaonline.org/archives/volume186/number48/enhancing-the-performance-of-kolmogorov-arnold-networks-kan-using-residual-activations-and-xavier-initialization/ },
doi = { 10.5120/ijca2024924146 },
publisher = {Foundation of Computer Science (FCS), NY, USA},
address = {New York, USA}
}
%0 Journal Article
%1 2024-11-27T00:39:23.341625+05:30
%A Subhasis Mitra
%T Enhancing the Performance of Kolmogorov-Arnold Networks (KAN) using Residual Activations and Xavier Initialization
%J International Journal of Computer Applications
%@ 0975-8887
%V 186
%N 48
%P 52-54
%D 2024
%I Foundation of Computer Science (FCS), NY, USA
Abstract

Kolmogorov-Arnold Networks (KANs), inspired by the Kolmogorov-Arnold representation theorem [4], are a novel class of neural networks characterized by learnable activation functions on edges instead of fixed activation functions on nodes. While KANs have demonstrated superior performance over traditional Multi-Layer Perceptrons (MLPs) in tasks requiring high-dimensional function approximation, their performance can be further optimized through effective initialization strategies and the introduction of residual connections. In this paper, an enhancement of KANs is proposed by combining Xavier initialization with residual activations. Xavier initialization ensures proper weight scaling, preventing vanishing or exploding gradients during training, while residual activations enable faster convergence and more efficient training of complex models. Through experimental evaluation on synthetic function approximation tasks, it demonstrates that these improvements yield faster convergence, better generalization, and increased robustness in training KANs.

References
  1. Glorot, X., & Bengio, Y. (2010). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (pp. 249-256).
  2. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778).
  3. Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Soljačić, M., Hou, T. Y., & Tegmark, M. (2024). Kolmogorov-Arnold Networks: Improving neural scaling laws for AI and science.
  4. Liu et al., 2024] Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Soljaˇci´c, M., Hou, T. Y., and Tegmark, M. (2024). Kan: Kolmogorov-arnold networks. arXiv preprint arXiv:2404.19756.
Index Terms

Computer Science
Information Sciences
Machine Learning
Neural Networks
Initialization Techniques
Function Approximation

Keywords

Kolmogorov-Arnold Networks (KANs) Xavier Initialization Residual Activations Deep Learning