Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inference on Windows #10

Closed
devilismyfriend opened this issue Mar 15, 2023 · 5 comments
Closed

Inference on Windows #10

devilismyfriend opened this issue Mar 15, 2023 · 5 comments

Comments

@devilismyfriend
Copy link

Hi, would love to test the LORA model, did anyone try and do it on windows?

@devilismyfriend
Copy link
Author

devilismyfriend commented Mar 15, 2023

Update:

I managed to get it to work, here are some instructions:
1.created a new env with anaconda with python 3.8 (no particular reason for the version, just wanted to make sure everything will be compatible)
2.Installed Torch 13 with CUDA 11.6
3.Ran the lines provided here up to the bitsandbytes portion
4.Ran pip install bitsandbytes==0.35.0
5.Copied over the files here into the bitsandbytes folder https://github.com/devilismyfriend/StableTuner/tree/main/resources/bitsandbytes_windows

I didn't try training, but inference works.
EDIT:
Training does work but it requires upgrading to bitsandbytes 0.37.0 and editing the files differently.

@devilismyfriend
Copy link
Author

devilismyfriend commented Mar 15, 2023

Side note on performance, creative writing seems to be better on the Stanford one, I'm getting a lot of noinput results using the LORA.

@devilismyfriend
Copy link
Author

On another note, I seem to be getting better performance when using the inference kwargs from the Stanford alpaca (tatsu-lab/stanford_alpaca#35)

Here's the code snippet for anyone interested
generation_config = GenerationConfig( temperature=0.7, top_p=0.9, num_beams=1, max_new_tokens=600, do_sample=True, **kwargs, ) with torch.no_grad(): generation_output = model.generate( input_ids=input_ids, generation_config=generation_config, return_dict_in_generate=True, output_scores=True, )

@tloen tloen closed this as completed Mar 16, 2023
@tloen
Copy link
Owner

tloen commented Mar 16, 2023

@devilismyfriend are you noting this just for creative writing, or also for instruction-following? I've set the temperature really low to get "precise" outputs, which should come at the cost of creativity.

@devilismyfriend
Copy link
Author

devilismyfriend commented Mar 16, 2023

Well I'm giving it an instruction but it basically just repeated it to me most times (the instruction is to write a sentence based on another sentence), using the args from the Standford alpaca I'm getting less noinputs, it follows the instruction better and provides more creative results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants