Skip to content

Commit 10fac86

Browse files
committed
chore: increase llm_gateway default max tokens to 4096
1 parent b39e1cd commit 10fac86

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

src/uipath/_services/llm_gateway_service.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@ async def chat_completions(
207207
self,
208208
messages: List[Dict[str, str]],
209209
model: str = ChatModels.gpt_4o_mini_2024_07_18,
210-
max_tokens: int = 50,
210+
max_tokens: int = 4096,
211211
temperature: float = 0,
212212
response_format: Optional[Union[Dict[str, Any], type[BaseModel]]] = None,
213213
api_version: str = API_VERSION,
@@ -227,7 +227,7 @@ async def chat_completions(
227227
Defaults to ChatModels.gpt_4o_mini_2024_07_18.
228228
Available models are defined in the ChatModels class.
229229
max_tokens (int, optional): Maximum number of tokens to generate in the response.
230-
Defaults to 50. Higher values allow longer responses.
230+
Defaults to 4096. Higher values allow longer responses.
231231
temperature (float, optional): Temperature for sampling, between 0 and 1.
232232
Lower values (closer to 0) make output more deterministic and focused,
233233
higher values make it more creative and random. Defaults to 0.
@@ -350,7 +350,7 @@ async def chat_completions(
350350
self,
351351
messages: List[Dict[str, str]],
352352
model: str = ChatModels.gpt_4o_mini_2024_07_18,
353-
max_tokens: int = 250,
353+
max_tokens: int = 4096,
354354
temperature: float = 0,
355355
n: int = 1,
356356
frequency_penalty: float = 0,
@@ -377,7 +377,7 @@ async def chat_completions(
377377
Defaults to ChatModels.gpt_4o_mini_2024_07_18.
378378
Available models are defined in the ChatModels class.
379379
max_tokens (int, optional): Maximum number of tokens to generate in the response.
380-
Defaults to 250. Higher values allow longer responses.
380+
Defaults to 4096. Higher values allow longer responses.
381381
temperature (float, optional): Temperature for sampling, between 0 and 1.
382382
Lower values (closer to 0) make output more deterministic and focused,
383383
higher values make it more creative and random. Defaults to 0.

0 commit comments

Comments
 (0)