From 502c02b5f65fdea98769eaa6315794b492fd3af0 Mon Sep 17 00:00:00 2001 From: hlohaus <983577+hlohaus@users.noreply.github.com> Date: Mon, 22 Dec 2025 14:00:54 +0100 Subject: [PATCH] Update default model to gpt-5-2 and adjust text models accordingly --- g4f/Provider/needs_auth/OpenaiChat.py | 43 ++++++++++++++------------- g4f/Provider/openai/models.py | 4 +-- 2 files changed, 25 insertions(+), 22 deletions(-) diff --git a/g4f/Provider/needs_auth/OpenaiChat.py b/g4f/Provider/needs_auth/OpenaiChat.py index 894c084c..7ccf3312 100644 --- a/g4f/Provider/needs_auth/OpenaiChat.py +++ b/g4f/Provider/needs_auth/OpenaiChat.py @@ -350,23 +350,24 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin): @classmethod async def create_authed( - cls, - model: str, - messages: Messages, - auth_result: AuthResult, - proxy: str = None, - timeout: int = 360, - auto_continue: bool = False, - action: Optional[str] = None, - conversation: Conversation = None, - media: MediaListType = None, - return_conversation: bool = True, - web_search: bool = False, - prompt: str = None, - conversation_mode: Optional[dict] = None, - temporary: Optional[bool] = None, - conversation_id: Optional[str] = None, - **kwargs + cls, + model: str, + messages: Messages, + auth_result: AuthResult, + proxy: str = None, + timeout: int = 360, + auto_continue: bool = False, + action: Optional[str] = None, + conversation: Conversation = None, + media: MediaListType = None, + return_conversation: bool = True, + web_search: bool = False, + prompt: str = None, + conversation_mode: Optional[dict] = None, + temporary: Optional[bool] = None, + conversation_id: Optional[str] = None, + reasoning_effort: Optional[str] = None, + **kwargs ) -> AsyncResult: """ Create an asynchronous generator for the conversation. @@ -394,9 +395,9 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin): if action is None: action = "next" async with StreamSession( - proxy=proxy, - impersonate="chrome", - timeout=timeout + proxy=proxy, + impersonate="chrome", + timeout=timeout ) as session: image_requests = None media = merge_media(media, messages) @@ -457,6 +458,7 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin): "system_hints": [ "picture_v2" ] if image_model else [], + "thinking_effort": "extended" if reasoning_effort == "high" else "standard", "supports_buffering": True, "supported_encodings": ["v1"] } @@ -517,6 +519,7 @@ class OpenaiChat(AsyncAuthedProvider, ProviderModelMixin): "conversation_mode": {"kind": "primary_assistant"}, "enable_message_followups": True, "system_hints": ["search"] if web_search else None, + "thinking_effort": "extended" if reasoning_effort == "high" else "standard", "supports_buffering": True, "supported_encodings": ["v1"], "client_contextual_info": {"is_dark_mode": False, "time_since_loaded": random.randint(20, 500), diff --git a/g4f/Provider/openai/models.py b/g4f/Provider/openai/models.py index 91a2431c..0b2e390a 100644 --- a/g4f/Provider/openai/models.py +++ b/g4f/Provider/openai/models.py @@ -1,7 +1,7 @@ -default_model = "gpt-5-1" +default_model = "gpt-5-2" default_image_model = "gpt-image" image_models = [default_image_model] -text_models = [default_model, "gpt-5-1-instant", "gpt-5-1-thinking", "gpt-5", "gpt-5-instant", "gpt-5-thinking", "gpt-4", "gpt-4.1", "gpt-4.1-mini", "gpt-4.5", "gpt-4o", "gpt-4o-mini", "o1", "o1-mini", "o3-mini", "o3-mini-high", "o4-mini", "o4-mini-high"] +text_models = [default_model, "gpt-5-2-instant", "gpt-5-2-thinking", "gpt-5-1", "gpt-5-1-instant", "gpt-5-1-thinking", "gpt-5", "gpt-5-instant", "gpt-5-thinking", "gpt-4", "gpt-4.1", "gpt-4.1-mini", "gpt-4.5", "gpt-4o", "gpt-4o-mini", "o1", "o1-mini", "o3-mini", "o3-mini-high", "o4-mini", "o4-mini-high"] vision_models = text_models models = text_models + image_models model_aliases = {