-
Notifications
You must be signed in to change notification settings - Fork 48
dynamic model list #107
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
dynamic model list #107
Conversation
Deploying with
|
Latest commit: |
5554b61
|
Status: | ✅ Deploy successful! |
Preview URL: | https://287f8767.console-overthinker-dev.pages.dev |
Branch Preview URL: | https://taras-gpt-models.console-overthinker-dev.pages.dev |
This is awesome. I've pulled your branch locally, and started doing some review, fixes, and additions on top of it. I'll push up the changes in the next day or so. |
I've added a new commit to this. It does a number of things:
I think this needs a bunch of testing, and I want to see make sure i won't break current databases with this--I'm 95% sure I won't, but let's not land this yet. One cool thing about this PR is that it helps guide the user to see exactly which models their API key supports. For example, my OpenAI key does not support GPT-4, but @tarasglek's does, and when I switch keys, I get a different list of models everywhere we show them. That's very cool. |
I think both of these could be follow-ups. For 1. we'll have to figure out a good way to do this via Chakra-UI, which doesn't natively support submenus (first time it's let me down so far). Thinking about this connected to other discussions, I wonder if we should have a For 2., would a hash really help here? I find that responses from different models are often very, very similar but usually not identical, so a hash would basically always be different and hide the similarity. Maybe there is a way to do similarity scoring? Not sure how you do that without setting one of them as a baseline (maybe the current version is baseline?). Also, is time a useful indicator, since it's dependent on a bunch of things unrelated to the model (e.g., network and API congestion, speed of things in user's browser, etc). When I query these models, I find that GPT-4 can be faster than GPT-3.5 on some days, which is clearly not due to the speed of the model. |
OK, I've spent some more time reading this change, and I think we're safe to land it. From a data-loss perspective, this change is widening the type of |
pushed a fix to "retry with". it was using model from settings, not from menu. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice catch
About time we hit some limits :)
I was getting identical results due to above bug :) |
This explains so much, I couldn't understand how the models could be so close or how GPT-4 got so fast! Glad you caught this. |
Prod seems happy with this change, loading my current stuff. Let me know if you see anything breaking. |
Addresses #106 by querying openai for available models, also gets us ready for claude support.
new chatgpt model is using date suffix, so mark that as chatgpt and mark old one with date. Will change that after 27th.
Left for followup: