Skip to content

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What is the mathematic equation behind the model? #16

Closed
hoangthangta opened this issue May 10, 2024 · 4 comments
Closed

What is the mathematic equation behind the model? #16

hoangthangta opened this issue May 10, 2024 · 4 comments

Comments

@hoangthangta
Copy link

hoangthangta commented May 10, 2024

I wonder what the equation used in the KAN model, anybody knows?

@karlo-con-k
Copy link

From what I understand, conventional neural network models are based on the Universal Approximation Theorem. However, KAN is based on the Kolmogorov-Arnol Representation Theorem. I'm not sure if that answers your question :D

@hoangthangta
Copy link
Author

hoangthangta commented May 10, 2024

We may know that KAN uses math equations (instead of variables) to fit the data points. However, what type of math equation (abs, sin, cos, etc) does this package use? It may use curve2coeff (compute the coefficients of the curve that interpolates the given points) to solve the equation (numpy.linalg.lstsq, a @ x = b). I am not sure but it is correct since the authors does not give much information about math equations.

@minh-nguyenhoang
Copy link

minh-nguyenhoang commented May 11, 2024

I think the author still use b-spline, but just compute them in a more efficient way.

However, all activation functions are linear combination of a fixed set of basis functions which are B-splines; given that, we can reformulate the computation as activate the input with different basis functions and then combine them linearly.

@Blealtan
Copy link
Owner

Please refer to the original paper. I'm just reproducing it in a more computationally efficient way. The only change is in the sparsification loss term which seems not really critical to applications other than symbolic regression; if I get the time I will write on that one.

Repository owner locked and limited conversation to collaborators May 20, 2024
@Blealtan Blealtan converted this issue into discussion #34 May 20, 2024

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

Projects
None yet
Development

No branches or pull requests

4 participants