You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We use LiteLLM's Proxy PromptCachingDeploymentCheck to route requests from the same conversation to the same provider, with the goal of increasing cache hits.
That routing step checks the length of the messages since anything under 1024 tokens won't be cached.
"message": "Error in is_prompt_caching_valid_prompt: Unsupported type <class 'dict'> for key cache_control in message {'content': 'You are presented with the following task:
...our truncated message...
tools.\\n', 'role': 'user', 'cache_control': {'type': 'ephemeral'}}",