2323from ._utils import is_given , get_async_library
2424from ._version import __version__
2525from .resources import (
26+ files ,
2627 tools ,
2728 models ,
2829 routes ,
3334 datasets ,
3435 inference ,
3536 providers ,
36- responses ,
3737 telemetry ,
3838 vector_io ,
3939 benchmarks ,
40+ embeddings ,
4041 toolgroups ,
4142 vector_dbs ,
4243 completions ,
44+ vector_stores ,
4345 scoring_functions ,
4446 synthetic_data_generation ,
4547)
5355from .resources .chat import chat
5456from .resources .eval import eval
5557from .resources .agents import agents
58+ from .resources .responses import responses
5659from .resources .tool_runtime import tool_runtime
5760from .resources .post_training import post_training
5861
@@ -78,10 +81,12 @@ class LlamaStackClient(SyncAPIClient):
7881 eval : eval .EvalResource
7982 inspect : inspect .InspectResource
8083 inference : inference .InferenceResource
84+ embeddings : embeddings .EmbeddingsResource
8185 chat : chat .ChatResource
8286 completions : completions .CompletionsResource
8387 vector_io : vector_io .VectorIoResource
8488 vector_dbs : vector_dbs .VectorDBsResource
89+ vector_stores : vector_stores .VectorStoresResource
8590 models : models .ModelsResource
8691 post_training : post_training .PostTrainingResource
8792 providers : providers .ProvidersResource
@@ -93,6 +98,7 @@ class LlamaStackClient(SyncAPIClient):
9398 scoring : scoring .ScoringResource
9499 scoring_functions : scoring_functions .ScoringFunctionsResource
95100 benchmarks : benchmarks .BenchmarksResource
101+ files : files .FilesResource
96102 with_raw_response : LlamaStackClientWithRawResponse
97103 with_streaming_response : LlamaStackClientWithStreamedResponse
98104
@@ -161,10 +167,12 @@ def __init__(
161167 self .eval = eval .EvalResource (self )
162168 self .inspect = inspect .InspectResource (self )
163169 self .inference = inference .InferenceResource (self )
170+ self .embeddings = embeddings .EmbeddingsResource (self )
164171 self .chat = chat .ChatResource (self )
165172 self .completions = completions .CompletionsResource (self )
166173 self .vector_io = vector_io .VectorIoResource (self )
167174 self .vector_dbs = vector_dbs .VectorDBsResource (self )
175+ self .vector_stores = vector_stores .VectorStoresResource (self )
168176 self .models = models .ModelsResource (self )
169177 self .post_training = post_training .PostTrainingResource (self )
170178 self .providers = providers .ProvidersResource (self )
@@ -176,6 +184,7 @@ def __init__(
176184 self .scoring = scoring .ScoringResource (self )
177185 self .scoring_functions = scoring_functions .ScoringFunctionsResource (self )
178186 self .benchmarks = benchmarks .BenchmarksResource (self )
187+ self .files = files .FilesResource (self )
179188 self .with_raw_response = LlamaStackClientWithRawResponse (self )
180189 self .with_streaming_response = LlamaStackClientWithStreamedResponse (self )
181190
@@ -296,10 +305,12 @@ class AsyncLlamaStackClient(AsyncAPIClient):
296305 eval : eval .AsyncEvalResource
297306 inspect : inspect .AsyncInspectResource
298307 inference : inference .AsyncInferenceResource
308+ embeddings : embeddings .AsyncEmbeddingsResource
299309 chat : chat .AsyncChatResource
300310 completions : completions .AsyncCompletionsResource
301311 vector_io : vector_io .AsyncVectorIoResource
302312 vector_dbs : vector_dbs .AsyncVectorDBsResource
313+ vector_stores : vector_stores .AsyncVectorStoresResource
303314 models : models .AsyncModelsResource
304315 post_training : post_training .AsyncPostTrainingResource
305316 providers : providers .AsyncProvidersResource
@@ -311,6 +322,7 @@ class AsyncLlamaStackClient(AsyncAPIClient):
311322 scoring : scoring .AsyncScoringResource
312323 scoring_functions : scoring_functions .AsyncScoringFunctionsResource
313324 benchmarks : benchmarks .AsyncBenchmarksResource
325+ files : files .AsyncFilesResource
314326 with_raw_response : AsyncLlamaStackClientWithRawResponse
315327 with_streaming_response : AsyncLlamaStackClientWithStreamedResponse
316328
@@ -379,10 +391,12 @@ def __init__(
379391 self .eval = eval .AsyncEvalResource (self )
380392 self .inspect = inspect .AsyncInspectResource (self )
381393 self .inference = inference .AsyncInferenceResource (self )
394+ self .embeddings = embeddings .AsyncEmbeddingsResource (self )
382395 self .chat = chat .AsyncChatResource (self )
383396 self .completions = completions .AsyncCompletionsResource (self )
384397 self .vector_io = vector_io .AsyncVectorIoResource (self )
385398 self .vector_dbs = vector_dbs .AsyncVectorDBsResource (self )
399+ self .vector_stores = vector_stores .AsyncVectorStoresResource (self )
386400 self .models = models .AsyncModelsResource (self )
387401 self .post_training = post_training .AsyncPostTrainingResource (self )
388402 self .providers = providers .AsyncProvidersResource (self )
@@ -394,6 +408,7 @@ def __init__(
394408 self .scoring = scoring .AsyncScoringResource (self )
395409 self .scoring_functions = scoring_functions .AsyncScoringFunctionsResource (self )
396410 self .benchmarks = benchmarks .AsyncBenchmarksResource (self )
411+ self .files = files .AsyncFilesResource (self )
397412 self .with_raw_response = AsyncLlamaStackClientWithRawResponse (self )
398413 self .with_streaming_response = AsyncLlamaStackClientWithStreamedResponse (self )
399414
@@ -515,10 +530,12 @@ def __init__(self, client: LlamaStackClient) -> None:
515530 self .eval = eval .EvalResourceWithRawResponse (client .eval )
516531 self .inspect = inspect .InspectResourceWithRawResponse (client .inspect )
517532 self .inference = inference .InferenceResourceWithRawResponse (client .inference )
533+ self .embeddings = embeddings .EmbeddingsResourceWithRawResponse (client .embeddings )
518534 self .chat = chat .ChatResourceWithRawResponse (client .chat )
519535 self .completions = completions .CompletionsResourceWithRawResponse (client .completions )
520536 self .vector_io = vector_io .VectorIoResourceWithRawResponse (client .vector_io )
521537 self .vector_dbs = vector_dbs .VectorDBsResourceWithRawResponse (client .vector_dbs )
538+ self .vector_stores = vector_stores .VectorStoresResourceWithRawResponse (client .vector_stores )
522539 self .models = models .ModelsResourceWithRawResponse (client .models )
523540 self .post_training = post_training .PostTrainingResourceWithRawResponse (client .post_training )
524541 self .providers = providers .ProvidersResourceWithRawResponse (client .providers )
@@ -532,6 +549,7 @@ def __init__(self, client: LlamaStackClient) -> None:
532549 self .scoring = scoring .ScoringResourceWithRawResponse (client .scoring )
533550 self .scoring_functions = scoring_functions .ScoringFunctionsResourceWithRawResponse (client .scoring_functions )
534551 self .benchmarks = benchmarks .BenchmarksResourceWithRawResponse (client .benchmarks )
552+ self .files = files .FilesResourceWithRawResponse (client .files )
535553
536554
537555class AsyncLlamaStackClientWithRawResponse :
@@ -545,10 +563,12 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
545563 self .eval = eval .AsyncEvalResourceWithRawResponse (client .eval )
546564 self .inspect = inspect .AsyncInspectResourceWithRawResponse (client .inspect )
547565 self .inference = inference .AsyncInferenceResourceWithRawResponse (client .inference )
566+ self .embeddings = embeddings .AsyncEmbeddingsResourceWithRawResponse (client .embeddings )
548567 self .chat = chat .AsyncChatResourceWithRawResponse (client .chat )
549568 self .completions = completions .AsyncCompletionsResourceWithRawResponse (client .completions )
550569 self .vector_io = vector_io .AsyncVectorIoResourceWithRawResponse (client .vector_io )
551570 self .vector_dbs = vector_dbs .AsyncVectorDBsResourceWithRawResponse (client .vector_dbs )
571+ self .vector_stores = vector_stores .AsyncVectorStoresResourceWithRawResponse (client .vector_stores )
552572 self .models = models .AsyncModelsResourceWithRawResponse (client .models )
553573 self .post_training = post_training .AsyncPostTrainingResourceWithRawResponse (client .post_training )
554574 self .providers = providers .AsyncProvidersResourceWithRawResponse (client .providers )
@@ -564,6 +584,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
564584 client .scoring_functions
565585 )
566586 self .benchmarks = benchmarks .AsyncBenchmarksResourceWithRawResponse (client .benchmarks )
587+ self .files = files .AsyncFilesResourceWithRawResponse (client .files )
567588
568589
569590class LlamaStackClientWithStreamedResponse :
@@ -577,10 +598,12 @@ def __init__(self, client: LlamaStackClient) -> None:
577598 self .eval = eval .EvalResourceWithStreamingResponse (client .eval )
578599 self .inspect = inspect .InspectResourceWithStreamingResponse (client .inspect )
579600 self .inference = inference .InferenceResourceWithStreamingResponse (client .inference )
601+ self .embeddings = embeddings .EmbeddingsResourceWithStreamingResponse (client .embeddings )
580602 self .chat = chat .ChatResourceWithStreamingResponse (client .chat )
581603 self .completions = completions .CompletionsResourceWithStreamingResponse (client .completions )
582604 self .vector_io = vector_io .VectorIoResourceWithStreamingResponse (client .vector_io )
583605 self .vector_dbs = vector_dbs .VectorDBsResourceWithStreamingResponse (client .vector_dbs )
606+ self .vector_stores = vector_stores .VectorStoresResourceWithStreamingResponse (client .vector_stores )
584607 self .models = models .ModelsResourceWithStreamingResponse (client .models )
585608 self .post_training = post_training .PostTrainingResourceWithStreamingResponse (client .post_training )
586609 self .providers = providers .ProvidersResourceWithStreamingResponse (client .providers )
@@ -596,6 +619,7 @@ def __init__(self, client: LlamaStackClient) -> None:
596619 client .scoring_functions
597620 )
598621 self .benchmarks = benchmarks .BenchmarksResourceWithStreamingResponse (client .benchmarks )
622+ self .files = files .FilesResourceWithStreamingResponse (client .files )
599623
600624
601625class AsyncLlamaStackClientWithStreamedResponse :
@@ -609,10 +633,12 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
609633 self .eval = eval .AsyncEvalResourceWithStreamingResponse (client .eval )
610634 self .inspect = inspect .AsyncInspectResourceWithStreamingResponse (client .inspect )
611635 self .inference = inference .AsyncInferenceResourceWithStreamingResponse (client .inference )
636+ self .embeddings = embeddings .AsyncEmbeddingsResourceWithStreamingResponse (client .embeddings )
612637 self .chat = chat .AsyncChatResourceWithStreamingResponse (client .chat )
613638 self .completions = completions .AsyncCompletionsResourceWithStreamingResponse (client .completions )
614639 self .vector_io = vector_io .AsyncVectorIoResourceWithStreamingResponse (client .vector_io )
615640 self .vector_dbs = vector_dbs .AsyncVectorDBsResourceWithStreamingResponse (client .vector_dbs )
641+ self .vector_stores = vector_stores .AsyncVectorStoresResourceWithStreamingResponse (client .vector_stores )
616642 self .models = models .AsyncModelsResourceWithStreamingResponse (client .models )
617643 self .post_training = post_training .AsyncPostTrainingResourceWithStreamingResponse (client .post_training )
618644 self .providers = providers .AsyncProvidersResourceWithStreamingResponse (client .providers )
@@ -630,6 +656,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
630656 client .scoring_functions
631657 )
632658 self .benchmarks = benchmarks .AsyncBenchmarksResourceWithStreamingResponse (client .benchmarks )
659+ self .files = files .AsyncFilesResourceWithStreamingResponse (client .files )
633660
634661
635662Client = LlamaStackClient
0 commit comments