DropKAN: Dropout Kolmogorov–Arnold Networks

We propose DropKAN (Dropout Kolmogorov—Arnold Networks), a regularization method that introduces dropout masks at the edge level within Kolmogorov—Arnold Networks (KANs) layers, randomly masking a subset of activation outputs in the computation graph. Forward pass analysis reve...

Full description

Saved in:
Bibliographic Details
Main Author: Mohammed Ghaith Altarabichi
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11121822/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We propose DropKAN (Dropout Kolmogorov&#x2014;Arnold Networks), a regularization method that introduces dropout masks at the edge level within Kolmogorov&#x2014;Arnold Networks (KANs) layers, randomly masking a subset of activation outputs in the computation graph. Forward pass analysis reveals that DropKAN, when combined with scaling, accurately preserves the expected output signal magnitude in line with theoretical expectations. In contrast, conventional neuron-level Dropout&#x2014;with its scaling&#x2014;distorts signal propagation due to the nonlinear nature of KANs activations. Through extensive experiments on several classical benchmark datasets, DropKAN consistently achieves superior test accuracy compared to both unregularized KANs and KANs employing standard Dropout. Sensitivity analysis across dropout rates reveals DropKAN&#x2019;s robustness, effectively mitigating underfitting at high dropout levels by preserving partial neuron activity via edge-level masking. Additionally, DropKAN demonstrates enhanced sample efficiency under limited training data conditions, outperforming baselines on multiple large-scale datasets. In computer vision benchmarks (MNIST, Fashion MNIST, EMNIST, CIFAR-10), DropKAN further validates its regularization efficacy by consistently improving generalization over standard KAN and Dropout configurations. These results establish DropKAN as a principled and practical regularization technique for KANs architectures. Our implementation of DropKAN is available at (<uri>https://github.com/ghaith81/dropkan</uri>).
ISSN:2169-3536