From 8e576ae3322c7dcf3e3f579f733dfc236ff5188a Mon Sep 17 00:00:00 2001 From: k Date: Wed, 4 Mar 2026 00:55:29 -0500 Subject: [PATCH] Now handling temperature and max_tokens correctly --- bot.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/bot.py b/bot.py index 0c86086..4b0ad96 100755 --- a/bot.py +++ b/bot.py @@ -164,8 +164,10 @@ def completions(): messageArray = data.get("messages", []) stream = data.get("stream", False) model = data.get("model", "RatChat") + temp = data.get("temperature", 0.7) + maxGen = data.get("max_tokens", 50) chat = messageArray[-1]["content"] if messageArray else "" - msgobj = messageOBJ(chat) + msgobj = messageOBJ(chat,maxGen,temp) msg_q.put(msgobj) if stream: