Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

categorize internal jobs and expose a mechanism for the user to specify/override what model/api gets used for different categories #37

Closed
dmarx opened this issue Apr 5, 2023 · 3 comments

Comments

@dmarx
Copy link

dmarx commented Apr 5, 2023

followup thought from #36

I haven't looked into this yet to see if it's already a configuration option, but it would be useful if the user could specify different models for different components of this system. concretely, gpt-3.5-turbo would probably be sufficient for the validation steps. If the action could be configured such that the more expensive API (gpt4) is used only for the main generations and a cheaper API could be used for stuff like the validation steps, that would potentially have a lot of benefits (cheaper, faster, less likely to get rate limited).

@irgolic
Copy link
Owner

irgolic commented Apr 6, 2023

Currently the model config option applies to all LLM calls. All agents should have their own model, as well as other config options. I'll open a PR implementing the refactor soon.

@irgolic
Copy link
Owner

irgolic commented May 5, 2023

Sorry for the long pause, I've implemented a refactor that allows agents to have their own config options, but haven't yet implemented picking different models for different actions. This will come with a refactor into an "action choice engine".

I did some testing, and I experienced a pretty bad performance drop with gpt-3.5-turbo, even for parsing steps. Although, earlier this week we changed how we were specifying instructions in guardrails, and it seems to greatly improve gpt-3.5-turbo performance; I'll try again to see if we can offload parsing steps to it.

@dmarx
Copy link
Author

dmarx commented May 5, 2023

yeah forget gpt-3.5-turbo. I think it's just too unreliable for code generation. may as well focus on using whatever tools are most likely to work rather than wrestling with 3.5

@irgolic irgolic closed this as completed Nov 5, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants