Spline function (of order k)
B-spline of order k, which is basis function of splines of order k:
For Interpretability
For Accuracy: finer grid
Neural scaling laws are the phenomenon where test loss decreases with more model parameters:
where
For
Consider data from
for which
is a true solution.
Consider training loss:
Two main results:
They use network attribution methods to find that the signature
Human scientists later identified that
KANs not only rediscover these results with much smaller networks and much more automation, but also present some interesting new results and insights.
To investigate 1. , they treat 17 knot invariants as inputs and signature as outputs:
KANs have less parameters(2e2
v.s. 3e5
), but behave better on accuracy(.816
v.s. .78
).
To investigate 2. ,they formulate the problem as a regression task.
Due to time constraints, we'll skip this part. TO MUCH PHYSICS!
KANs are usually 10x slower(in calculation) than MLPs given same num of parameters.
Are KANs just RBF Networks? GitHub Issue#162
THANKS.
This Saturday, I will present an article about KAN. The author, inspired by the Kolmogorov-Arnold representation theorem, generalized the representation form and constructed a novel type of neural network. Compared to MLP, this network demonstrates stronger accuracy and interpretability. It holds great promise in scientific applications such as solving PDEs and related fields.