diff --git a/reginaldCog/openai_completion.py b/reginaldCog/openai_completion.py index b63b580..04b31df 100644 --- a/reginaldCog/openai_completion.py +++ b/reginaldCog/openai_completion.py @@ -34,9 +34,6 @@ class Completion: "model": self.__model, "messages": self.__messages, "max_completion_tokens": 2000, - "temperature": 0.7, - "presence_penalty": 0.5, - "frequency_penalty": 0.5, "tools": TOOLS, "tool_choice": "auto", } diff --git a/reginaldCog/reginald.py b/reginaldCog/reginald.py index bbe676d..34bbd05 100644 --- a/reginaldCog/reginald.py +++ b/reginaldCog/reginald.py @@ -23,7 +23,6 @@ CALLABLE_FUNCTIONS = { DEFAULT_MODEL = "gpt-5-mini-2025-08-07" DEFAULT_MAX_COMPLETION_TOKENS = 2000 -DEFAULT_TEMPERATURE = 0.7 class ReginaldCog(PermissionsMixin, BlacklistMixin, MemoryMixin, commands.Cog): @@ -223,9 +222,10 @@ class ReginaldCog(PermissionsMixin, BlacklistMixin, MemoryMixin, commands.Cog): completion_args = { "model": model, "messages": messages, - # `max_completion_tokens` is the recommended limit field for modern/reasoning models. + # Keep modern token cap field and rely on model defaults for sampling controls. + # GPT-5 family compatibility notes: temperature/top_p/logprobs are not universally + # accepted across snapshots/reasoning settings. "max_completion_tokens": DEFAULT_MAX_COMPLETION_TOKENS, - "temperature": DEFAULT_TEMPERATURE, "tools": TOOLS, "tool_choice": "auto", }