-
-
Notifications
You must be signed in to change notification settings - Fork 719
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any plans to support phi-2 finetuning? #85
Comments
@asmith26 On it! :) |
I tried, fine tuning of phi-2. Which is not good. |
@imrankh46 Ye smaller models can sometimes not follow instructions well. Could you approx give an example maybe to show Phi not following instructions? |
here is the code. https://colab.research.google.com/drive/1a7rL3UzWfo5I7OPyVmTEnR6_tRqIOblg?usp=sharing |
@danielhanchen I'm also interesting in contributing this, let me know if you have space on your PR for another helping hand. |
@cm2435 Oh more than happy to collab if you're into it!! I actually took a look at Phi the other day:
In general Phi is possible, just there are a few blockers - esp (2) and (3). But again - if you want to take a crack at this @cm2435 - I'll be super grateful + I'll collab with you! :) Possibly taking a stab at some steps like (2) finding the derivative might be better as a first step :) |
@danielhanchen Started a fork and opened a staging PR to work off. I can start by adding some unit test coverage around the kernels if we need to test their accuracy, something simple like assert torch.allclose(triton_out, torch_out) or I can try to contribute some of the other steps you mentioned- what's your preference? |
@cm2435 Oh lmao just noticed i never responded on this thread WHOOPS! Well I responded on your other threads already I guess - sorry!!! |
Hi there,
Just thought I'd ask if there were any plans to support phi-2 finetuning, https://huggingface.co/microsoft/phi-2?
Many thanks for any help, and this amazing lib!
The text was updated successfully, but these errors were encountered: