|
30 | 30 | from .._streaming import Stream, AsyncStream |
31 | 31 | from .._base_client import make_request_options |
32 | 32 | from ..types.embeddings_response import EmbeddingsResponse |
| 33 | +from ..types.response_format_param import ResponseFormatParam |
33 | 34 | from ..types.shared_params.message import Message |
34 | 35 | from ..types.inference_completion_response import InferenceCompletionResponse |
35 | 36 | from ..types.shared_params.sampling_params import SamplingParams |
@@ -66,7 +67,7 @@ def chat_completion( |
66 | 67 | messages: Iterable[Message], |
67 | 68 | model_id: str, |
68 | 69 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
69 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 70 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
70 | 71 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
71 | 72 | stream: Literal[False] | NotGiven = NOT_GIVEN, |
72 | 73 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
@@ -111,7 +112,7 @@ def chat_completion( |
111 | 112 | model_id: str, |
112 | 113 | stream: Literal[True], |
113 | 114 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
114 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 115 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
115 | 116 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
116 | 117 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
117 | 118 | tool_prompt_format: Literal["json", "function_tag", "python_list"] | NotGiven = NOT_GIVEN, |
@@ -155,7 +156,7 @@ def chat_completion( |
155 | 156 | model_id: str, |
156 | 157 | stream: bool, |
157 | 158 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
158 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 159 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
159 | 160 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
160 | 161 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
161 | 162 | tool_prompt_format: Literal["json", "function_tag", "python_list"] | NotGiven = NOT_GIVEN, |
@@ -198,7 +199,7 @@ def chat_completion( |
198 | 199 | messages: Iterable[Message], |
199 | 200 | model_id: str, |
200 | 201 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
201 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 202 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
202 | 203 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
203 | 204 | stream: Literal[False] | Literal[True] | NotGiven = NOT_GIVEN, |
204 | 205 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
@@ -259,7 +260,7 @@ def completion( |
259 | 260 | content: InterleavedContent, |
260 | 261 | model_id: str, |
261 | 262 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
262 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 263 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
263 | 264 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
264 | 265 | stream: Literal[False] | NotGiven = NOT_GIVEN, |
265 | 266 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
@@ -291,7 +292,7 @@ def completion( |
291 | 292 | model_id: str, |
292 | 293 | stream: Literal[True], |
293 | 294 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
294 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 295 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
295 | 296 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
296 | 297 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
297 | 298 | x_llama_stack_provider_data: str | NotGiven = NOT_GIVEN, |
@@ -322,7 +323,7 @@ def completion( |
322 | 323 | model_id: str, |
323 | 324 | stream: bool, |
324 | 325 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
325 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 326 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
326 | 327 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
327 | 328 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
328 | 329 | x_llama_stack_provider_data: str | NotGiven = NOT_GIVEN, |
@@ -352,7 +353,7 @@ def completion( |
352 | 353 | content: InterleavedContent, |
353 | 354 | model_id: str, |
354 | 355 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
355 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 356 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
356 | 357 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
357 | 358 | stream: Literal[False] | Literal[True] | NotGiven = NOT_GIVEN, |
358 | 359 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
@@ -476,7 +477,7 @@ async def chat_completion( |
476 | 477 | messages: Iterable[Message], |
477 | 478 | model_id: str, |
478 | 479 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
479 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 480 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
480 | 481 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
481 | 482 | stream: Literal[False] | NotGiven = NOT_GIVEN, |
482 | 483 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
@@ -521,7 +522,7 @@ async def chat_completion( |
521 | 522 | model_id: str, |
522 | 523 | stream: Literal[True], |
523 | 524 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
524 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 525 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
525 | 526 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
526 | 527 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
527 | 528 | tool_prompt_format: Literal["json", "function_tag", "python_list"] | NotGiven = NOT_GIVEN, |
@@ -565,7 +566,7 @@ async def chat_completion( |
565 | 566 | model_id: str, |
566 | 567 | stream: bool, |
567 | 568 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
568 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 569 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
569 | 570 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
570 | 571 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
571 | 572 | tool_prompt_format: Literal["json", "function_tag", "python_list"] | NotGiven = NOT_GIVEN, |
@@ -608,7 +609,7 @@ async def chat_completion( |
608 | 609 | messages: Iterable[Message], |
609 | 610 | model_id: str, |
610 | 611 | logprobs: inference_chat_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
611 | | - response_format: inference_chat_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 612 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
612 | 613 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
613 | 614 | stream: Literal[False] | Literal[True] | NotGiven = NOT_GIVEN, |
614 | 615 | tool_choice: Literal["auto", "required"] | NotGiven = NOT_GIVEN, |
@@ -669,7 +670,7 @@ async def completion( |
669 | 670 | content: InterleavedContent, |
670 | 671 | model_id: str, |
671 | 672 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
672 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 673 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
673 | 674 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
674 | 675 | stream: Literal[False] | NotGiven = NOT_GIVEN, |
675 | 676 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
@@ -701,7 +702,7 @@ async def completion( |
701 | 702 | model_id: str, |
702 | 703 | stream: Literal[True], |
703 | 704 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
704 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 705 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
705 | 706 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
706 | 707 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
707 | 708 | x_llama_stack_provider_data: str | NotGiven = NOT_GIVEN, |
@@ -732,7 +733,7 @@ async def completion( |
732 | 733 | model_id: str, |
733 | 734 | stream: bool, |
734 | 735 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
735 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 736 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
736 | 737 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
737 | 738 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
738 | 739 | x_llama_stack_provider_data: str | NotGiven = NOT_GIVEN, |
@@ -762,7 +763,7 @@ async def completion( |
762 | 763 | content: InterleavedContent, |
763 | 764 | model_id: str, |
764 | 765 | logprobs: inference_completion_params.Logprobs | NotGiven = NOT_GIVEN, |
765 | | - response_format: inference_completion_params.ResponseFormat | NotGiven = NOT_GIVEN, |
| 766 | + response_format: ResponseFormatParam | NotGiven = NOT_GIVEN, |
766 | 767 | sampling_params: SamplingParams | NotGiven = NOT_GIVEN, |
767 | 768 | stream: Literal[False] | Literal[True] | NotGiven = NOT_GIVEN, |
768 | 769 | x_llama_stack_client_version: str | NotGiven = NOT_GIVEN, |
|
0 commit comments