Skip to content

Conversation

@bendavidsteel
Copy link
Contributor

Specifically I need this to allow setting the torch dtype so I can load models that only just fit in my VRAM. Let me know if any extra tweaks are required!

@EngSalem
Copy link

Is this commit merged to the main branch?
I am looking into loading the HF model with float16 quantization, but it seems that this part isn't merged yet

@arnavsinghvi11
Copy link
Collaborator

Hi @bendavidsteel just following up on old PRs. can you resolve merge conflicts and lmk when ready to merge? Thanks!

@bendavidsteel
Copy link
Contributor Author

@arnavsinghvi11 resolved!:)

@arnavsinghvi11 arnavsinghvi11 merged commit 38e7e54 into stanfordnlp:main Apr 15, 2024
@arnavsinghvi11
Copy link
Collaborator

Thanks @bendavidsteel !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants