diff --git a/README.md b/README.md index b179649..8f87046 100644 --- a/README.md +++ b/README.md @@ -20,7 +20,7 @@ edgee = Edgee("your-api-key") # Send a simple request response = edgee.send( - model="gpt-4o", + model="gpt-5.2", input="What is the capital of France?" ) @@ -34,7 +34,7 @@ The `send()` method makes non-streaming chat completion requests: ```python response = edgee.send( - model="gpt-4o", + model="gpt-5.2", input="Hello, world!" ) @@ -58,7 +58,7 @@ if response.compression: The `stream()` method enables real-time streaming responses: ```python -for chunk in edgee.stream("gpt-4o", "Tell me a story"): +for chunk in edgee.stream("gpt-5.2", "Tell me a story"): if chunk.text: print(chunk.text, end="", flush=True) diff --git a/example/compression.py b/example/compression.py index d446279..4dce59e 100644 --- a/example/compression.py +++ b/example/compression.py @@ -87,7 +87,7 @@ Based on this context, summarize the key milestones in AI development in 3 bullet points.""" response = edgee.send( - model="gpt-4o", + model="gpt-5.2", input={ "messages": [ {"role": "user", "content": user_message}, diff --git a/example/test.py b/example/test.py index 2351002..d1f3ee5 100644 --- a/example/test.py +++ b/example/test.py @@ -37,7 +37,7 @@ # Test 3: With tools print("Test 3: With tools") response3 = edgee.send( - model="gpt-4o", + model="gpt-5.2", input={ "messages": [{"role": "user", "content": "What is the weather in Paris?"}], "tools": [