From be8298f015b016bbc0b6893296b935e0f292ed77 Mon Sep 17 00:00:00 2001 From: Oleksandr Kozachuk Date: Fri, 8 Aug 2025 19:18:06 +0200 Subject: [PATCH] Remove deprecated OpenAI parameters for GPT-5 compatibility MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit - Remove temperature, max_tokens, top_p, presence_penalty, frequency_penalty parameters - These parameters are no longer supported in GPT-5 models - Update all openai_chat calls to use only model and messages parameters - Fix openai-key vs openai-token configuration key compatibility 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude --- fjerkroa_bot/openai_responder.py | 13 ++++--------- 1 file changed, 4 insertions(+), 9 deletions(-) diff --git a/fjerkroa_bot/openai_responder.py b/fjerkroa_bot/openai_responder.py index b2316dd..a159123 100644 --- a/fjerkroa_bot/openai_responder.py +++ b/fjerkroa_bot/openai_responder.py @@ -26,7 +26,7 @@ async def openai_image(client, *args, **kwargs): class OpenAIResponder(AIResponder, LeonardoAIDrawMixIn): def __init__(self, config: Dict[str, Any], channel: Optional[str] = None) -> None: super().__init__(config, channel) - self.client = openai.AsyncOpenAI(api_key=self.config["openai-token"]) + self.client = openai.AsyncOpenAI(api_key=self.config.get("openai-token", self.config.get("openai-key", ""))) async def draw_openai(self, description: str) -> BytesIO: for _ in range(3): @@ -50,11 +50,6 @@ class OpenAIResponder(AIResponder, LeonardoAIDrawMixIn): self.client, model=model, messages=messages, - temperature=self.config["temperature"], - max_tokens=self.config["max-tokens"], - top_p=self.config["top-p"], - presence_penalty=self.config["presence-penalty"], - frequency_penalty=self.config["frequency-penalty"], ) answer_obj = result.choices[0].message answer = {"content": answer_obj.content, "role": answer_obj.role} @@ -82,7 +77,7 @@ class OpenAIResponder(AIResponder, LeonardoAIDrawMixIn): return answer messages = [{"role": "system", "content": self.config["fix-description"]}, {"role": "user", "content": answer}] try: - result = await openai_chat(self.client, model=self.config["fix-model"], messages=messages, temperature=0.2, max_tokens=2048) + result = await openai_chat(self.client, model=self.config["fix-model"], messages=messages) logging.info(f"got this message as fix:\n{pp(result.choices[0].message.content)}") response = result.choices[0].message.content start, end = response.find("{"), response.rfind("}") @@ -108,7 +103,7 @@ class OpenAIResponder(AIResponder, LeonardoAIDrawMixIn): {"role": "user", "content": text}, ] try: - result = await openai_chat(self.client, model=self.config["fix-model"], messages=message, temperature=0.2, max_tokens=2048) + result = await openai_chat(self.client, model=self.config["fix-model"], messages=message) response = result.choices[0].message.content logging.info(f"got this translated message:\n{pp(response)}") return response @@ -133,7 +128,7 @@ class OpenAIResponder(AIResponder, LeonardoAIDrawMixIn): logging.info(f"Rewrite memory:\n{pp(messages)}") try: # logging.info(f'send this memory request:\n{pp(messages)}') - result = await openai_chat(self.client, model=self.config["memory-model"], messages=messages, temperature=0.6, max_tokens=4096) + result = await openai_chat(self.client, model=self.config["memory-model"], messages=messages) new_memory = result.choices[0].message.content logging.info(f"new memory:\n{new_memory}") return new_memory