Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tuning the ethical guidelines of ExLlamaV2 #335

Closed
redshiva opened this issue Feb 12, 2024 · 6 comments
Closed

Tuning the ethical guidelines of ExLlamaV2 #335

redshiva opened this issue Feb 12, 2024 · 6 comments

Comments

@redshiva
Copy link

I have used the raw FB Llama2 models in developing my application. When interacting with the model, I did not encounter any ethical constraints. As far as I experienced, I could ask any questions and get an answer, which can be problematic for an application that is user facing. However, after converting the Meta model to ExLlamaV2, I ask questions and I am hitting: "As a responsible AI language model, I cannot fulfill that request..."

I want ethical constraints, but I want to tune them. How can I do this? In looking through the code, I do not see where this is being set.

Thank you!

@DocShotgun
Copy link

If I'm understanding correctly, you're suggesting that by quantizing the model to exl2... it somehow became censored? And you want that?

@CrossPr0duct
Copy link

You keep asking the model to be ethical and you beg, I think that works because it's AGI

@Kimiko-AI
Copy link

FEEL THE AGI

@KaraKaraWitch
Copy link

I have used the raw FB Llama2 models in developing my application. When interacting with the model, I did not encounter any ethical constraints. As far as I experienced, I could ask any questions and get an answer, which can be problematic for an application that is user facing. However, after converting the Meta model to ExLlamaV2, I ask questions and I am hitting: "As a responsible AI language model, I cannot fulfill that request..."

I want ethical constraints, but I want to tune them. How can I do this? In looking through the code, I do not see where this is being set.

Thank you!

You should probably look into LoRA's or actual finetuning the model. Quantization is not the way to do model tuning. Consider reading up on how to finetune/LoRA a LLM.

Also... in what context you want to add ethical constraints? Corporate?

@turboderp
Copy link
Owner

ExLlamaV2 doesn't do anything to make inference more or less ethical, it just runs the model. Quantization introduces some level of inaccuracy which means the response from a quantized model is never going to be exactly the same as the original model, for a given prompt. Likewise sampling options will affect the output in various unpredictable ways.

If you're running the chat example in llama mode, you can try adjusting the system prompt. The default prompt is the one originally provided by Meta and it's extremely "aligned", to the point of being ridiculous. Try it with something like `-sp "Just answer the questions." instead, or with a blank string.

As for tuning alignment in general, that's a a whole science. LoRAs are an option, or you can pick from any of the thousands of compatible models on HF finetuned for various purposes.

@redshiva
Copy link
Author

redshiva commented Feb 12, 2024 via email

@turboderp turboderp closed this as not planned Won't fix, can't repro, duplicate, stale Mar 27, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants