@@ -78,6 +78,7 @@ def create(
7878 ],
7979 ],
8080 model : str ,
81+ background : bool | Omit = omit ,
8182 conversation : Optional [str ] | Omit = omit ,
8283 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
8384 include : Optional [
@@ -128,6 +129,9 @@ def create(
128129
129130 model: The underlying LLM used for completions.
130131
132+ background: Whether to run the model response in the background. When true, returns
133+ immediately with status 'queued'.
134+
131135 conversation: Optional ID of a conversation to add the response to.
132136
133137 guardrails: List of guardrails to apply during response generation.
@@ -197,6 +201,7 @@ def create(
197201 ],
198202 model : str ,
199203 stream : Literal [True ],
204+ background : bool | Omit = omit ,
200205 conversation : Optional [str ] | Omit = omit ,
201206 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
202207 include : Optional [
@@ -248,6 +253,9 @@ def create(
248253
249254 stream: Whether to stream the response.
250255
256+ background: Whether to run the model response in the background. When true, returns
257+ immediately with status 'queued'.
258+
251259 conversation: Optional ID of a conversation to add the response to.
252260
253261 guardrails: List of guardrails to apply during response generation.
@@ -315,6 +323,7 @@ def create(
315323 ],
316324 model : str ,
317325 stream : bool ,
326+ background : bool | Omit = omit ,
318327 conversation : Optional [str ] | Omit = omit ,
319328 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
320329 include : Optional [
@@ -366,6 +375,9 @@ def create(
366375
367376 stream: Whether to stream the response.
368377
378+ background: Whether to run the model response in the background. When true, returns
379+ immediately with status 'queued'.
380+
369381 conversation: Optional ID of a conversation to add the response to.
370382
371383 guardrails: List of guardrails to apply during response generation.
@@ -432,6 +444,7 @@ def create(
432444 ],
433445 ],
434446 model : str ,
447+ background : bool | Omit = omit ,
435448 conversation : Optional [str ] | Omit = omit ,
436449 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
437450 include : Optional [
@@ -480,6 +493,7 @@ def create(
480493 {
481494 "input" : input ,
482495 "model" : model ,
496+ "background" : background ,
483497 "conversation" : conversation ,
484498 "guardrails" : guardrails ,
485499 "include" : include ,
@@ -676,6 +690,7 @@ async def create(
676690 ],
677691 ],
678692 model : str ,
693+ background : bool | Omit = omit ,
679694 conversation : Optional [str ] | Omit = omit ,
680695 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
681696 include : Optional [
@@ -726,6 +741,9 @@ async def create(
726741
727742 model: The underlying LLM used for completions.
728743
744+ background: Whether to run the model response in the background. When true, returns
745+ immediately with status 'queued'.
746+
729747 conversation: Optional ID of a conversation to add the response to.
730748
731749 guardrails: List of guardrails to apply during response generation.
@@ -795,6 +813,7 @@ async def create(
795813 ],
796814 model : str ,
797815 stream : Literal [True ],
816+ background : bool | Omit = omit ,
798817 conversation : Optional [str ] | Omit = omit ,
799818 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
800819 include : Optional [
@@ -846,6 +865,9 @@ async def create(
846865
847866 stream: Whether to stream the response.
848867
868+ background: Whether to run the model response in the background. When true, returns
869+ immediately with status 'queued'.
870+
849871 conversation: Optional ID of a conversation to add the response to.
850872
851873 guardrails: List of guardrails to apply during response generation.
@@ -913,6 +935,7 @@ async def create(
913935 ],
914936 model : str ,
915937 stream : bool ,
938+ background : bool | Omit = omit ,
916939 conversation : Optional [str ] | Omit = omit ,
917940 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
918941 include : Optional [
@@ -964,6 +987,9 @@ async def create(
964987
965988 stream: Whether to stream the response.
966989
990+ background: Whether to run the model response in the background. When true, returns
991+ immediately with status 'queued'.
992+
967993 conversation: Optional ID of a conversation to add the response to.
968994
969995 guardrails: List of guardrails to apply during response generation.
@@ -1030,6 +1056,7 @@ async def create(
10301056 ],
10311057 ],
10321058 model : str ,
1059+ background : bool | Omit = omit ,
10331060 conversation : Optional [str ] | Omit = omit ,
10341061 guardrails : Optional [SequenceNotStr [response_create_params .Guardrail ]] | Omit = omit ,
10351062 include : Optional [
@@ -1078,6 +1105,7 @@ async def create(
10781105 {
10791106 "input" : input ,
10801107 "model" : model ,
1108+ "background" : background ,
10811109 "conversation" : conversation ,
10821110 "guardrails" : guardrails ,
10831111 "include" : include ,
0 commit comments