@@ -207,7 +207,7 @@ async def chat_completions(
207207 self ,
208208 messages : List [Dict [str , str ]],
209209 model : str = ChatModels .gpt_4o_mini_2024_07_18 ,
210- max_tokens : int = 50 ,
210+ max_tokens : int = 4096 ,
211211 temperature : float = 0 ,
212212 response_format : Optional [Union [Dict [str , Any ], type [BaseModel ]]] = None ,
213213 api_version : str = API_VERSION ,
@@ -227,7 +227,7 @@ async def chat_completions(
227227 Defaults to ChatModels.gpt_4o_mini_2024_07_18.
228228 Available models are defined in the ChatModels class.
229229 max_tokens (int, optional): Maximum number of tokens to generate in the response.
230- Defaults to 50 . Higher values allow longer responses.
230+ Defaults to 4096 . Higher values allow longer responses.
231231 temperature (float, optional): Temperature for sampling, between 0 and 1.
232232 Lower values (closer to 0) make output more deterministic and focused,
233233 higher values make it more creative and random. Defaults to 0.
@@ -350,7 +350,7 @@ async def chat_completions(
350350 self ,
351351 messages : List [Dict [str , str ]],
352352 model : str = ChatModels .gpt_4o_mini_2024_07_18 ,
353- max_tokens : int = 250 ,
353+ max_tokens : int = 4096 ,
354354 temperature : float = 0 ,
355355 n : int = 1 ,
356356 frequency_penalty : float = 0 ,
@@ -377,7 +377,7 @@ async def chat_completions(
377377 Defaults to ChatModels.gpt_4o_mini_2024_07_18.
378378 Available models are defined in the ChatModels class.
379379 max_tokens (int, optional): Maximum number of tokens to generate in the response.
380- Defaults to 250 . Higher values allow longer responses.
380+ Defaults to 4096 . Higher values allow longer responses.
381381 temperature (float, optional): Temperature for sampling, between 0 and 1.
382382 Lower values (closer to 0) make output more deterministic and focused,
383383 higher values make it more creative and random. Defaults to 0.
0 commit comments