99
1010from tests .utils import assert_matches_type
1111from llama_stack_client import LlamaStackClient , AsyncLlamaStackClient
12- from llama_stack_client .types import EvaluateResponse
13- from llama_stack_client .types .eval import JobStatusResponse
12+ from llama_stack_client .types import Job , EvaluateResponse
1413
1514base_url = os .environ .get ("TEST_API_BASE_URL" , "http://127.0.0.1:4010" )
1615
@@ -120,7 +119,7 @@ def test_method_status(self, client: LlamaStackClient) -> None:
120119 job_id = "job_id" ,
121120 benchmark_id = "benchmark_id" ,
122121 )
123- assert_matches_type (JobStatusResponse , job , path = ["response" ])
122+ assert_matches_type (Job , job , path = ["response" ])
124123
125124 @parametrize
126125 def test_raw_response_status (self , client : LlamaStackClient ) -> None :
@@ -132,7 +131,7 @@ def test_raw_response_status(self, client: LlamaStackClient) -> None:
132131 assert response .is_closed is True
133132 assert response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
134133 job = response .parse ()
135- assert_matches_type (JobStatusResponse , job , path = ["response" ])
134+ assert_matches_type (Job , job , path = ["response" ])
136135
137136 @parametrize
138137 def test_streaming_response_status (self , client : LlamaStackClient ) -> None :
@@ -144,7 +143,7 @@ def test_streaming_response_status(self, client: LlamaStackClient) -> None:
144143 assert response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
145144
146145 job = response .parse ()
147- assert_matches_type (JobStatusResponse , job , path = ["response" ])
146+ assert_matches_type (Job , job , path = ["response" ])
148147
149148 assert cast (Any , response .is_closed ) is True
150149
@@ -268,7 +267,7 @@ async def test_method_status(self, async_client: AsyncLlamaStackClient) -> None:
268267 job_id = "job_id" ,
269268 benchmark_id = "benchmark_id" ,
270269 )
271- assert_matches_type (JobStatusResponse , job , path = ["response" ])
270+ assert_matches_type (Job , job , path = ["response" ])
272271
273272 @parametrize
274273 async def test_raw_response_status (self , async_client : AsyncLlamaStackClient ) -> None :
@@ -280,7 +279,7 @@ async def test_raw_response_status(self, async_client: AsyncLlamaStackClient) ->
280279 assert response .is_closed is True
281280 assert response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
282281 job = await response .parse ()
283- assert_matches_type (JobStatusResponse , job , path = ["response" ])
282+ assert_matches_type (Job , job , path = ["response" ])
284283
285284 @parametrize
286285 async def test_streaming_response_status (self , async_client : AsyncLlamaStackClient ) -> None :
@@ -292,7 +291,7 @@ async def test_streaming_response_status(self, async_client: AsyncLlamaStackClie
292291 assert response .http_request .headers .get ("X-Stainless-Lang" ) == "python"
293292
294293 job = await response .parse ()
295- assert_matches_type (JobStatusResponse , job , path = ["response" ])
294+ assert_matches_type (Job , job , path = ["response" ])
296295
297296 assert cast (Any , response .is_closed ) is True
298297
0 commit comments