Closed
Description
Describe the issue
I'd like to use llama3.1 70b or 405b with function calling. Any recommendations for what is supported in autogen?
I have tried but could not get either of these to work:
-
Google Cloud's Vertex AI solution to call llama3.1 (using autogen's Vertex AI supported code following https://microsoft.github.io/autogen/docs/topics/non-openai-models/cloud-gemini_vertexai worked for Gemini models but not llama, because vertexai is still trying to call
publishers/google/models/llama3_1
when it should bepublishers/llama/models/llama3_1
, issue at https://github.com/microsoft/autogen/blob/main/autogen/oai/gemini.py#L202-L207) -
Google Cloud's Vertex AI API solution (cloud-based proxy server)
-
Ollama llama 3.1 70b hosted locally
Steps to reproduce
No response
Screenshots and logs
No response
Additional Information
No response