Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Additional parameters to mlx_lm lora? r, lora_alpha, lora_dropout, scale? #454

Closed
ivanfioravanti opened this issue Feb 18, 2024 · 8 comments
Labels
enhancement New feature or request

Comments

@ivanfioravanti
Copy link
Contributor

When playing with fine-tuning sometimes I change from_linear in lora.py to play with them.
Should we add command line args for these?

@awni
Copy link
Member

awni commented Feb 18, 2024

Yes that's been on our list to add for a while.

Though, do you use mlx_lm.lora or just the original lora.py script? My preference is to update the package since that one has many more features.

@ivanfioravanti
Copy link
Contributor Author

I always us mlx_lm now, it's becoming more powerful at each release 💪

@ivanfioravanti
Copy link
Contributor Author

It would be great adding the new LR Scheduler or even the optimizer, but parameters become too complex then.
Probably something like PR #235 to read from a config file would be better.

@awni
Copy link
Member

awni commented Feb 18, 2024

Agree I think we may need to start using a yaml config

@chimezie
Copy link
Contributor

#235 is dated. I can rebase it to mlx-examples/main and update it (to support the parameters that have been added since I last worked on that PR) if there is interest. I have found it useful to pull parameters from yaml and override them from what is provided via the command line.

@awni awni added the enhancement New feature or request label Feb 25, 2024
@Solido
Copy link

Solido commented Feb 27, 2024

I too migrated to mlx_lm.lora and end up using a shell script generator.
#235 would simplify the whole process as use cases for MLX grow and simplify
the tuning by having different models/confs/benchmark in parallel.

@ivanfioravanti
Copy link
Contributor Author

@awni I think we can close this issue, in recent versions (thx @chimezie) -c parameter has been added to mlx_lm.generate and I added dropout in this pr #599

@awni
Copy link
Member

awni commented Mar 20, 2024

Yes indeed.. safe to close, thank you!

@awni awni closed this as completed Mar 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

4 participants