Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add gpt-4o metadata #3613

Merged
merged 2 commits into from
May 13, 2024
Merged

Add gpt-4o metadata #3613

merged 2 commits into from
May 13, 2024

Conversation

ConnorDoyle
Copy link
Contributor

#3612

Title

Relevant issues

Type

πŸ†• New Feature
πŸ› Bug Fix
🧹 Refactoring
πŸ“– Documentation
πŸ’» Development Environment
πŸš„ Infrastructure
βœ… Test

Changes

Testing

Notes

Pre-Submission Checklist (optional but appreciated):

  • I have included relevant documentation updates (stored in /docs/my-website)

OS Tests (optional but appreciated):

  • Tested on Windows
  • Tested on MacOS
  • Tested on Linux

Copy link

vercel bot commented May 13, 2024

The latest updates on your projects. Learn more about Vercel for Git β†—οΈŽ

Name Status Preview Comments Updated (UTC)
litellm βœ… Ready (Inspect) Visit Preview πŸ’¬ Add feedback May 13, 2024 6:52pm

@@ -9,6 +9,18 @@
"mode": "chat",
"supports_function_calling": true
},
"gpt-4o": {

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we add an alias for gpt-4o-2024-05-13 too?

@mmmaia
Copy link

mmmaia commented May 13, 2024

Thanks for working on this! A couple observations:

  • max_output_tokens for GPT-4o is 2048 (at least it's what I see in the playgronud)
  • we should also add gpt-4o-2024-05-13

@HowieG
Copy link

HowieG commented May 13, 2024

How is mode handled with multimodal input?
"mode": ["chat", "audio_transcription", "image_generation"],

@ConnorDoyle
Copy link
Contributor Author

Thanks for working on this! A couple observations:

  • max_output_tokens for GPT-4o is 2048 (at least it's what I see in the playgronud)
  • we should also add gpt-4o-2024-05-13

I couldn't find docs supporting either explicit API limit, and I was unable to convince the model to produce more than 2048 tokens via some manual API testing. Updated this to 2048 since the limit is at least 2048 via the API (but possibly higher.)

@mmmaia
Copy link

mmmaia commented May 13, 2024

I couldn't find docs supporting either explicit API limit, and I was unable to convince the model to produce more than 2048 tokens via some manual API testing. Updated this to 2048 since the limit is at least 2048 via the API (but possibly higher.)

Now my playground is showing 4096. Unfortunately, there's no mention about it in the docs.

@krrishdholakia
Copy link
Contributor

LGTM - merging in, we'll take care of updating the max token limit based on testing

@Undertone0809
Copy link
Contributor

How is mode handled with multimodal input? "mode": ["chat", "audio_transcription", "image_generation"],

#3629 So the question is, is it because of this?

@krrishdholakia
Copy link
Contributor

@Undertone0809 mode is primarily used for vertex ai for routing between their different sdk's.

We also started using it for health checks, as a way to know whether it's a /completion, /embedding or /chat/completion call

No impact on multimodal.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

6 participants