-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Different Outputs Streaming (More Coherent) vs Single Message (Writes Transcript) #69
Comments
Awesome, thanks for the update on this! It's interesting that there's a difference... I wonder if you switch back to "single message" mode whether you would continue to get the improved behavior or if it would regress. The main difference might be that any of the settings which split the message into parts might give the bot more context around what a chat transcript should look like, so it's easier for it to get the gist of what we want it to generate. Let me know if the other investigation turns up anything, and thanks for the help! |
How would I try this? |
I get different responses even though I have the parameters identical in config.yml for oobabot and in text-gen-webui. I tried adding the new min_p parameter to config.yml and it loads but I don't know if its working. The params are In text-gen-webui I ask "how can I increase my power level past 9000?" and it gives me a list of things to do but in discord it just says "become one with the force" no matter what settings I set? |
could be a difference in prompting. The webui's selection and oobabot's preset system prompt are both very different. |
I am wondering that as well, does oobabot inherit the chat-instruction template from text-gen? the character context for text-gen and the oobabot personality are both "The following is a conversation with an AI Large Language Model. The AI has been trained to answer questions, provide recommendations, and help with decision making. The AI follows user requests. The AI thinks outside the box." |
Don't forget the instruction format, including tags like [INST] and <> if using a llama2-chat model. These aren't included by default in oobabot. |
I'm using dolphin-mixtral and the model card says this about the prompt format <|im_start|>system So in the persona I have <|im_start|>system You are a discord AI bot. The AI follows instructions and is helpful. <|im_end|> The model is still is wonky. right now it spams emojis with every reply. The temp is at 1 in the config but it doesn't seem to change anything? |
I am having similar issues, did you ever find anything out? |
Hey! I've been modifying and playing around with oobabot formatting and even playing with the code a bit.
When I first started using Llama 2 with the "single message" option I almost always receive messages that try to continue on the conversation itself (talking to itsellf, predicting user responses).
When I switched over to Streaming I'm finally getting coherent and single message replies.
Haven't looked into the code because I've been working on this most of the day yesterday and this morning, but if I do I'll post more here if I find any differences that could be causing this.
Thank you for your time!
The text was updated successfully, but these errors were encountered: