From 5eca1a5d043e90e69dde6f8ffb19a67cb4c57eca Mon Sep 17 00:00:00 2001 From: Nawaz Dhandala Date: Tue, 16 Dec 2025 11:21:47 +0000 Subject: [PATCH] refactor: Remove maxTokens from LLMCompletionRequest and related usages --- Common/Server/Utils/LLM/LLMService.ts | 4 ---- 1 file changed, 4 deletions(-) diff --git a/Common/Server/Utils/LLM/LLMService.ts b/Common/Server/Utils/LLM/LLMService.ts index 23dbbb4a9e..e59bbfec60 100644 --- a/Common/Server/Utils/LLM/LLMService.ts +++ b/Common/Server/Utils/LLM/LLMService.ts @@ -15,7 +15,6 @@ export interface LLMMessage { export interface LLMCompletionRequest { messages: Array; - maxTokens?: number; temperature?: number; llmProviderConfig: LLMProviderConfig; } @@ -80,7 +79,6 @@ export default class LLMService { content: msg.content, }; }), - max_tokens: request.maxTokens || 4096, temperature: request.temperature ?? 0.7, }, headers: { @@ -154,7 +152,6 @@ export default class LLMService { const requestData: JSONObject = { model: modelName, messages: userMessages, - max_tokens: request.maxTokens || 4096, temperature: request.temperature ?? 0.7, }; @@ -244,7 +241,6 @@ export default class LLMService { stream: false, options: { temperature: request.temperature ?? 0.7, - num_predict: request.maxTokens || 4096, }, }, headers: {