feat: add support for copilot models#3273
Conversation
|
I think for the claude models, we should keep haiku as the weak models. Having sonnet-3.5 as the weak model for sonnet-3.7 is expensive and overkill. |
|
@paul-gauthier Since the changes in litellm side were merged at BerriAI/litellm#8577, could you have a look at this to see if I'm missing anythings? |
Please change the weak models to haiku. sonnet-3.5 costs the same as sonnet-3.7 - there's no point in making 3.5 the weaker model. |
Sorry to intrude here, are the sonnet costs relevant when used via GH Copilot? Isn't it a flat fee regardless of the model used? |
Fair point. One could even argue that the weak model could be the same as the strong model in that case - why have them be different? My concern is twofold. The first is simple: Using such a heavy model is overkill for the weak model. The second is that (some) people who have used proxies to reroute Aider to copilot have gotten warning emails from Microsoft. It's not clear what triggers it - perhaps too much |
| use_repo_map: true | ||
| examples_as_sys_msg: true | ||
| extra_params: | ||
| max_tokens: 8192 |
There was a problem hiding this comment.
why is the number of tokens so small? As far as i see both 3.7 and 3.5 accept 200k context: https://docs.anthropic.com/en/docs/about-claude/models/all-models.
do I misunderstand how max_tokens works?
There was a problem hiding this comment.
@TheBestPessimist github has limits on input/output tokens. i think input is limited to ~80k, and output to ~8k
There was a problem hiding this comment.
Interesting, thanks for mentioning. In that case, does max_tokens represent the output value? i'd have expected that to be the max input size, so it's missing at least a 0: 8192 -> 81920
There was a problem hiding this comment.
the max_token parameter is equal to the max_output_tokens parameter everywhere in the aider docs website so i'd assume it's not the total amount of tokens
copilot doesn't provide haiku model. the list of copilot models: the weak model should be set to either 4o-mini or gemini-2.0-flash-001 |
|
It would be nice if you could add entries for the model-metadata.json file too. |
|
I merged this on my local to resolve conflicts. Thanks! Users will have a better experience if you are able to add the metadata json for these models too. |
solve #2227
Wait for BerriAI/litellm#8577 to be merged