DropKAN: Dropout Kolmogorov–Arnold Networks
We propose DropKAN (Dropout Kolmogorov—Arnold Networks), a regularization method that introduces dropout masks at the edge level within Kolmogorov—Arnold Networks (KANs) layers, randomly masking a subset of activation outputs in the computation graph. Forward pass analysis reve...
Saved in:
| Main Author: | |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/11121822/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| Summary: | We propose DropKAN (Dropout Kolmogorov—Arnold Networks), a regularization method that introduces dropout masks at the edge level within Kolmogorov—Arnold Networks (KANs) layers, randomly masking a subset of activation outputs in the computation graph. Forward pass analysis reveals that DropKAN, when combined with scaling, accurately preserves the expected output signal magnitude in line with theoretical expectations. In contrast, conventional neuron-level Dropout—with its scaling—distorts signal propagation due to the nonlinear nature of KANs activations. Through extensive experiments on several classical benchmark datasets, DropKAN consistently achieves superior test accuracy compared to both unregularized KANs and KANs employing standard Dropout. Sensitivity analysis across dropout rates reveals DropKAN’s robustness, effectively mitigating underfitting at high dropout levels by preserving partial neuron activity via edge-level masking. Additionally, DropKAN demonstrates enhanced sample efficiency under limited training data conditions, outperforming baselines on multiple large-scale datasets. In computer vision benchmarks (MNIST, Fashion MNIST, EMNIST, CIFAR-10), DropKAN further validates its regularization efficacy by consistently improving generalization over standard KAN and Dropout configurations. These results establish DropKAN as a principled and practical regularization technique for KANs architectures. Our implementation of DropKAN is available at (<uri>https://github.com/ghaith81/dropkan</uri>). |
|---|---|
| ISSN: | 2169-3536 |