Description
Is your feature request related to a problem? Please describe.
As mentioned above, the issue is that Autogen has issues with new models. As all the model pricings are hardcoded. Curently you need to get the latest version to fix this. But you might not always want to get the latest version.
Describe the solution you'd like
Solution Idea A) Propose to allow setting prompt/completion
prices via llm_settings. See here:
{
"model": "gpt-3.5-turbo-1106",
"api_key": "sk-xxx",
"max_tokens": 4000,
"timeout": 100,
"max_retries": 2,
"input_cost" : 10, # USD per mil; alt naming prompt_tokens_cost
"output_cost" : 30 # USD per mil; alt naming completion_tokens_cost
}
It might make sense to also add vision pricing this way then:
Solution Idea B) It seems that curently Autogen is hardcoding the price for each model type
separately. But if it's not present, then you should simply compare the
prefix. So in this case, I believe the prefix "gpt-4-turbo" is enough to
understand what price range it is.
Additional context
No response