2020 ProxiesTypes ,
2121 RequestOptions ,
2222)
23- from ._utils import (
24- is_given ,
25- get_async_library ,
26- )
23+ from ._utils import is_given , get_async_library
2724from ._version import __version__
2825from .resources import (
2926 tools ,
4138 benchmarks ,
4239 toolgroups ,
4340 vector_dbs ,
41+ completions ,
4442 scoring_functions ,
4543 synthetic_data_generation ,
4644)
5149 SyncAPIClient ,
5250 AsyncAPIClient ,
5351)
52+ from .resources .chat import chat
5453from .resources .eval import eval
5554from .resources .agents import agents
5655from .resources .tool_runtime import tool_runtime
@@ -77,6 +76,8 @@ class LlamaStackClient(SyncAPIClient):
7776 eval : eval .EvalResource
7877 inspect : inspect .InspectResource
7978 inference : inference .InferenceResource
79+ chat : chat .ChatResource
80+ completions : completions .CompletionsResource
8081 vector_io : vector_io .VectorIoResource
8182 vector_dbs : vector_dbs .VectorDBsResource
8283 models : models .ModelsResource
@@ -157,6 +158,8 @@ def __init__(
157158 self .eval = eval .EvalResource (self )
158159 self .inspect = inspect .InspectResource (self )
159160 self .inference = inference .InferenceResource (self )
161+ self .chat = chat .ChatResource (self )
162+ self .completions = completions .CompletionsResource (self )
160163 self .vector_io = vector_io .VectorIoResource (self )
161164 self .vector_dbs = vector_dbs .VectorDBsResource (self )
162165 self .models = models .ModelsResource (self )
@@ -289,6 +292,8 @@ class AsyncLlamaStackClient(AsyncAPIClient):
289292 eval : eval .AsyncEvalResource
290293 inspect : inspect .AsyncInspectResource
291294 inference : inference .AsyncInferenceResource
295+ chat : chat .AsyncChatResource
296+ completions : completions .AsyncCompletionsResource
292297 vector_io : vector_io .AsyncVectorIoResource
293298 vector_dbs : vector_dbs .AsyncVectorDBsResource
294299 models : models .AsyncModelsResource
@@ -369,6 +374,8 @@ def __init__(
369374 self .eval = eval .AsyncEvalResource (self )
370375 self .inspect = inspect .AsyncInspectResource (self )
371376 self .inference = inference .AsyncInferenceResource (self )
377+ self .chat = chat .AsyncChatResource (self )
378+ self .completions = completions .AsyncCompletionsResource (self )
372379 self .vector_io = vector_io .AsyncVectorIoResource (self )
373380 self .vector_dbs = vector_dbs .AsyncVectorDBsResource (self )
374381 self .models = models .AsyncModelsResource (self )
@@ -502,6 +509,8 @@ def __init__(self, client: LlamaStackClient) -> None:
502509 self .eval = eval .EvalResourceWithRawResponse (client .eval )
503510 self .inspect = inspect .InspectResourceWithRawResponse (client .inspect )
504511 self .inference = inference .InferenceResourceWithRawResponse (client .inference )
512+ self .chat = chat .ChatResourceWithRawResponse (client .chat )
513+ self .completions = completions .CompletionsResourceWithRawResponse (client .completions )
505514 self .vector_io = vector_io .VectorIoResourceWithRawResponse (client .vector_io )
506515 self .vector_dbs = vector_dbs .VectorDBsResourceWithRawResponse (client .vector_dbs )
507516 self .models = models .ModelsResourceWithRawResponse (client .models )
@@ -529,6 +538,8 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
529538 self .eval = eval .AsyncEvalResourceWithRawResponse (client .eval )
530539 self .inspect = inspect .AsyncInspectResourceWithRawResponse (client .inspect )
531540 self .inference = inference .AsyncInferenceResourceWithRawResponse (client .inference )
541+ self .chat = chat .AsyncChatResourceWithRawResponse (client .chat )
542+ self .completions = completions .AsyncCompletionsResourceWithRawResponse (client .completions )
532543 self .vector_io = vector_io .AsyncVectorIoResourceWithRawResponse (client .vector_io )
533544 self .vector_dbs = vector_dbs .AsyncVectorDBsResourceWithRawResponse (client .vector_dbs )
534545 self .models = models .AsyncModelsResourceWithRawResponse (client .models )
@@ -558,6 +569,8 @@ def __init__(self, client: LlamaStackClient) -> None:
558569 self .eval = eval .EvalResourceWithStreamingResponse (client .eval )
559570 self .inspect = inspect .InspectResourceWithStreamingResponse (client .inspect )
560571 self .inference = inference .InferenceResourceWithStreamingResponse (client .inference )
572+ self .chat = chat .ChatResourceWithStreamingResponse (client .chat )
573+ self .completions = completions .CompletionsResourceWithStreamingResponse (client .completions )
561574 self .vector_io = vector_io .VectorIoResourceWithStreamingResponse (client .vector_io )
562575 self .vector_dbs = vector_dbs .VectorDBsResourceWithStreamingResponse (client .vector_dbs )
563576 self .models = models .ModelsResourceWithStreamingResponse (client .models )
@@ -587,6 +600,8 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
587600 self .eval = eval .AsyncEvalResourceWithStreamingResponse (client .eval )
588601 self .inspect = inspect .AsyncInspectResourceWithStreamingResponse (client .inspect )
589602 self .inference = inference .AsyncInferenceResourceWithStreamingResponse (client .inference )
603+ self .chat = chat .AsyncChatResourceWithStreamingResponse (client .chat )
604+ self .completions = completions .AsyncCompletionsResourceWithStreamingResponse (client .completions )
590605 self .vector_io = vector_io .AsyncVectorIoResourceWithStreamingResponse (client .vector_io )
591606 self .vector_dbs = vector_dbs .AsyncVectorDBsResourceWithStreamingResponse (client .vector_dbs )
592607 self .models = models .AsyncModelsResourceWithStreamingResponse (client .models )
0 commit comments