International Journal of Computer Applications |
Foundation of Computer Science (FCS), NY, USA |
Volume 186 - Number 48 |
Year of Publication: 2024 |
Authors: Subhasis Mitra |
10.5120/ijca2024924146 |
Subhasis Mitra . Enhancing the Performance of Kolmogorov-Arnold Networks (KAN) using Residual Activations and Xavier Initialization. International Journal of Computer Applications. 186, 48 ( Nov 2024), 52-54. DOI=10.5120/ijca2024924146
Kolmogorov-Arnold Networks (KANs), inspired by the Kolmogorov-Arnold representation theorem [4], are a novel class of neural networks characterized by learnable activation functions on edges instead of fixed activation functions on nodes. While KANs have demonstrated superior performance over traditional Multi-Layer Perceptrons (MLPs) in tasks requiring high-dimensional function approximation, their performance can be further optimized through effective initialization strategies and the introduction of residual connections. In this paper, an enhancement of KANs is proposed by combining Xavier initialization with residual activations. Xavier initialization ensures proper weight scaling, preventing vanishing or exploding gradients during training, while residual activations enable faster convergence and more efficient training of complex models. Through experimental evaluation on synthetic function approximation tasks, it demonstrates that these improvements yield faster convergence, better generalization, and increased robustness in training KANs.