-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Architectural thoughts/discussion #14
Comments
First of all, thanks for all the great pull requests and feedback! This is super helpful! Good article! This is also worth a read: https://huyenchip.com/2023/04/11/llm-engineering.html Here are some of my thoughts on the topic:
Overall it's still early and I'm trying to get the project into a good shape before OpenAI gives GPT-4 access to everyone. |
I tend to agree with most points. What I found however, is that the speed of GPT 3.5 turbo makes a huge difference for me as a developer and prompt engineer because it allows for much faster experimentation. So even though GPT 4 WILL be the default and fast standard, right now it does come with a hefty penalty making debugging and iterating quite slow. So this could have a big impact for us as developers of these things. Also some things are fundamental and model independent, e.g. how can we prompt the model and run tools to save steps (showing available files in case of a file not found is one example) and thus enable us to iterate faster on the tooling and the users to get results more effectively. |
100% on the last part. And GPT 3.5 is supported (plus I'll add support for the OpenAI text models soon). Personally I have the same sense of rapid feedback and fast experimentation when I use GPT-3.5, but when I reflect on it, I tend to realize that even though it might feel fast and good, it probably is a case of walking faster in the wrong direction and solving soon-to-be-obsolete problems. Therefore I want to make GPT-4 the primary target and would like to avoid any GPT-3.5 specific optimizations. |
I've added FlexibleJsonActionFormat, which should help a lot with GPT-3.5-turbo. GPT-3.5-turbo agents should work a lot better now. |
I've been playing with this right now and we're definitely getting somewhere.
This blog post is a good summary, I'm guessing you know it: https://jina.ai/news/auto-gpt-unmasked-hype-hard-truths-production-pitfalls/
Now, autogpt allows subtasks to be executed at GPT 3.5 turbo, which is a lot faster and cheaper to execute. I believe that with a bit of prompt engineering that is possible at least for solving some subtasks here. (Prompt engineering in this case model specific, as 3.5 sometimes doesn't want to format responses nicely and machine readable. I also believe autogpt has a "json autofix" thing which adds missing parenthesis before parsing.)
Apart from that it would be interesting to explore, how we could reduce the number of needed steps to achieve small, autonomous tasks more quickly and cost effectively.
The text was updated successfully, but these errors were encountered: