Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add kan and hyperparams #182

Closed
wants to merge 3 commits into from

Conversation

gkielian
Copy link
Collaborator

@gkielian gkielian commented Jun 7, 2024

These combine the latest PRs towards adding ability to experiment with KAN as a replacement for linear layers and for the MLP, while retaining backwards compatibility with prior features.

A small change was needed for the current KAN implementation, requiring a clamp to prevent div by zero.

After implementing the above, inference with the sample.py file became much more stable.

SenmiaoORZ and others added 3 commits June 4, 2024 21:52
This change prevents division by zero.

This appears to be the reason for the initial instability of inference,
so now we can explore various KAN applications.

Note:
We should convert this clamp into an argparse param and config for
experimentation.
@gkielian gkielian mentioned this pull request Jun 7, 2024
@gkielian gkielian requested a review from klei22 June 7, 2024 17:11
@klei22
Copy link
Collaborator

klei22 commented Jun 17, 2024

The wrapped nn.linear uses a different initialization, would it be possible to have this match the previous nn.Linear initialization values?

@gkielian
Copy link
Collaborator Author

Collaborator

Sounds good, will look into whether we can forward the initialization values from model.py

@klei22
Copy link
Collaborator

klei22 commented Jun 18, 2024

image

I tested too, and seems actually like it should be fine, will make some direct edits and merge in.

@klei22
Copy link
Collaborator

klei22 commented Jun 18, 2024

Merged in the latest repo changes and created a new PR with suggested adjustments:

#189

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants