Skip to content

Commit 05a0330

Browse files
feat: Implement include parameter specifically for adding logprobs in the output message
1 parent 05f867b commit 05a0330

18 files changed

+1109
-256
lines changed

.stats.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
configured_endpoints: 103
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-798a94e418defb2b9c8e7432a9f8a76635439b07e8fd3cc4a158a2b26dd60b90.yml
3-
openapi_spec_hash: 93b1520629590a9ecaa74d726c006a49
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-4cdfb92312df3e349a702847ffb5a464d108247790eec659fb4597712ca37653.yml
3+
openapi_spec_hash: 0888e1be36cd34c77dd131dc8cc21765
44
config_hash: 39578cfdeb4a10121f2cb3fa3e4d5e20

src/llama_stack_client/resources/responses/responses.py

Lines changed: 114 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -8,13 +8,13 @@
88

99
from __future__ import annotations
1010

11-
from typing import Dict, Union, Iterable, Optional
11+
from typing import Dict, List, Union, Iterable, Optional
1212
from typing_extensions import Literal, overload
1313

1414
import httpx
1515

1616
from ...types import response_list_params, response_create_params
17-
from ..._types import Body, Omit, Query, Headers, NotGiven, SequenceNotStr, omit, not_given
17+
from ..._types import Body, Omit, Query, Headers, NotGiven, omit, not_given
1818
from ..._utils import required_args, maybe_transform, async_maybe_transform
1919
from ..._compat import cached_property
2020
from ..._resource import SyncAPIResource, AsyncAPIResource
@@ -79,7 +79,20 @@ def create(
7979
],
8080
model: str,
8181
conversation: Optional[str] | Omit = omit,
82-
include: Optional[SequenceNotStr[str]] | Omit = omit,
82+
include: Optional[
83+
List[
84+
Literal[
85+
"web_search_call.action.sources",
86+
"code_interpreter_call.outputs",
87+
"computer_call_output.output.image_url",
88+
"file_search_call.results",
89+
"message.input_image.image_url",
90+
"message.output_text.logprobs",
91+
"reasoning.encrypted_content",
92+
]
93+
]
94+
]
95+
| Omit = omit,
8396
instructions: Optional[str] | Omit = omit,
8497
max_infer_iters: Optional[int] | Omit = omit,
8598
max_tool_calls: Optional[int] | Omit = omit,
@@ -130,7 +143,20 @@ def create(
130143
model: str,
131144
stream: Literal[True],
132145
conversation: Optional[str] | Omit = omit,
133-
include: Optional[SequenceNotStr[str]] | Omit = omit,
146+
include: Optional[
147+
List[
148+
Literal[
149+
"web_search_call.action.sources",
150+
"code_interpreter_call.outputs",
151+
"computer_call_output.output.image_url",
152+
"file_search_call.results",
153+
"message.input_image.image_url",
154+
"message.output_text.logprobs",
155+
"reasoning.encrypted_content",
156+
]
157+
]
158+
]
159+
| Omit = omit,
134160
instructions: Optional[str] | Omit = omit,
135161
max_infer_iters: Optional[int] | Omit = omit,
136162
max_tool_calls: Optional[int] | Omit = omit,
@@ -180,7 +206,20 @@ def create(
180206
model: str,
181207
stream: bool,
182208
conversation: Optional[str] | Omit = omit,
183-
include: Optional[SequenceNotStr[str]] | Omit = omit,
209+
include: Optional[
210+
List[
211+
Literal[
212+
"web_search_call.action.sources",
213+
"code_interpreter_call.outputs",
214+
"computer_call_output.output.image_url",
215+
"file_search_call.results",
216+
"message.input_image.image_url",
217+
"message.output_text.logprobs",
218+
"reasoning.encrypted_content",
219+
]
220+
]
221+
]
222+
| Omit = omit,
184223
instructions: Optional[str] | Omit = omit,
185224
max_infer_iters: Optional[int] | Omit = omit,
186225
max_tool_calls: Optional[int] | Omit = omit,
@@ -229,7 +268,20 @@ def create(
229268
],
230269
model: str,
231270
conversation: Optional[str] | Omit = omit,
232-
include: Optional[SequenceNotStr[str]] | Omit = omit,
271+
include: Optional[
272+
List[
273+
Literal[
274+
"web_search_call.action.sources",
275+
"code_interpreter_call.outputs",
276+
"computer_call_output.output.image_url",
277+
"file_search_call.results",
278+
"message.input_image.image_url",
279+
"message.output_text.logprobs",
280+
"reasoning.encrypted_content",
281+
]
282+
]
283+
]
284+
| Omit = omit,
233285
instructions: Optional[str] | Omit = omit,
234286
max_infer_iters: Optional[int] | Omit = omit,
235287
max_tool_calls: Optional[int] | Omit = omit,
@@ -434,7 +486,20 @@ async def create(
434486
],
435487
model: str,
436488
conversation: Optional[str] | Omit = omit,
437-
include: Optional[SequenceNotStr[str]] | Omit = omit,
489+
include: Optional[
490+
List[
491+
Literal[
492+
"web_search_call.action.sources",
493+
"code_interpreter_call.outputs",
494+
"computer_call_output.output.image_url",
495+
"file_search_call.results",
496+
"message.input_image.image_url",
497+
"message.output_text.logprobs",
498+
"reasoning.encrypted_content",
499+
]
500+
]
501+
]
502+
| Omit = omit,
438503
instructions: Optional[str] | Omit = omit,
439504
max_infer_iters: Optional[int] | Omit = omit,
440505
max_tool_calls: Optional[int] | Omit = omit,
@@ -485,7 +550,20 @@ async def create(
485550
model: str,
486551
stream: Literal[True],
487552
conversation: Optional[str] | Omit = omit,
488-
include: Optional[SequenceNotStr[str]] | Omit = omit,
553+
include: Optional[
554+
List[
555+
Literal[
556+
"web_search_call.action.sources",
557+
"code_interpreter_call.outputs",
558+
"computer_call_output.output.image_url",
559+
"file_search_call.results",
560+
"message.input_image.image_url",
561+
"message.output_text.logprobs",
562+
"reasoning.encrypted_content",
563+
]
564+
]
565+
]
566+
| Omit = omit,
489567
instructions: Optional[str] | Omit = omit,
490568
max_infer_iters: Optional[int] | Omit = omit,
491569
max_tool_calls: Optional[int] | Omit = omit,
@@ -535,7 +613,20 @@ async def create(
535613
model: str,
536614
stream: bool,
537615
conversation: Optional[str] | Omit = omit,
538-
include: Optional[SequenceNotStr[str]] | Omit = omit,
616+
include: Optional[
617+
List[
618+
Literal[
619+
"web_search_call.action.sources",
620+
"code_interpreter_call.outputs",
621+
"computer_call_output.output.image_url",
622+
"file_search_call.results",
623+
"message.input_image.image_url",
624+
"message.output_text.logprobs",
625+
"reasoning.encrypted_content",
626+
]
627+
]
628+
]
629+
| Omit = omit,
539630
instructions: Optional[str] | Omit = omit,
540631
max_infer_iters: Optional[int] | Omit = omit,
541632
max_tool_calls: Optional[int] | Omit = omit,
@@ -584,7 +675,20 @@ async def create(
584675
],
585676
model: str,
586677
conversation: Optional[str] | Omit = omit,
587-
include: Optional[SequenceNotStr[str]] | Omit = omit,
678+
include: Optional[
679+
List[
680+
Literal[
681+
"web_search_call.action.sources",
682+
"code_interpreter_call.outputs",
683+
"computer_call_output.output.image_url",
684+
"file_search_call.results",
685+
"message.input_image.image_url",
686+
"message.output_text.logprobs",
687+
"reasoning.encrypted_content",
688+
]
689+
]
690+
]
691+
| Omit = omit,
588692
instructions: Optional[str] | Omit = omit,
589693
max_infer_iters: Optional[int] | Omit = omit,
590694
max_tool_calls: Optional[int] | Omit = omit,

src/llama_stack_client/types/chat/completion_create_response.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -265,10 +265,10 @@ class ChoiceLogprobsContent(BaseModel):
265265

266266
logprob: float
267267

268-
top_logprobs: List[ChoiceLogprobsContentTopLogprob]
269-
270268
bytes: Optional[List[int]] = None
271269

270+
top_logprobs: Optional[List[ChoiceLogprobsContentTopLogprob]] = None
271+
272272

273273
class ChoiceLogprobsRefusalTopLogprob(BaseModel):
274274
"""
@@ -300,10 +300,10 @@ class ChoiceLogprobsRefusal(BaseModel):
300300

301301
logprob: float
302302

303-
top_logprobs: List[ChoiceLogprobsRefusalTopLogprob]
304-
305303
bytes: Optional[List[int]] = None
306304

305+
top_logprobs: Optional[List[ChoiceLogprobsRefusalTopLogprob]] = None
306+
307307

308308
class ChoiceLogprobs(BaseModel):
309309
"""

src/llama_stack_client/types/chat/completion_list_response.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -288,10 +288,10 @@ class DataChoiceLogprobsContent(BaseModel):
288288

289289
logprob: float
290290

291-
top_logprobs: List[DataChoiceLogprobsContentTopLogprob]
292-
293291
bytes: Optional[List[int]] = None
294292

293+
top_logprobs: Optional[List[DataChoiceLogprobsContentTopLogprob]] = None
294+
295295

296296
class DataChoiceLogprobsRefusalTopLogprob(BaseModel):
297297
"""
@@ -323,10 +323,10 @@ class DataChoiceLogprobsRefusal(BaseModel):
323323

324324
logprob: float
325325

326-
top_logprobs: List[DataChoiceLogprobsRefusalTopLogprob]
327-
328326
bytes: Optional[List[int]] = None
329327

328+
top_logprobs: Optional[List[DataChoiceLogprobsRefusalTopLogprob]] = None
329+
330330

331331
class DataChoiceLogprobs(BaseModel):
332332
"""

src/llama_stack_client/types/chat/completion_retrieve_response.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -283,10 +283,10 @@ class ChoiceLogprobsContent(BaseModel):
283283

284284
logprob: float
285285

286-
top_logprobs: List[ChoiceLogprobsContentTopLogprob]
287-
288286
bytes: Optional[List[int]] = None
289287

288+
top_logprobs: Optional[List[ChoiceLogprobsContentTopLogprob]] = None
289+
290290

291291
class ChoiceLogprobsRefusalTopLogprob(BaseModel):
292292
"""
@@ -318,10 +318,10 @@ class ChoiceLogprobsRefusal(BaseModel):
318318

319319
logprob: float
320320

321-
top_logprobs: List[ChoiceLogprobsRefusalTopLogprob]
322-
323321
bytes: Optional[List[int]] = None
324322

323+
top_logprobs: Optional[List[ChoiceLogprobsRefusalTopLogprob]] = None
324+
325325

326326
class ChoiceLogprobs(BaseModel):
327327
"""

src/llama_stack_client/types/chat_completion_chunk.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -93,10 +93,10 @@ class ChoiceLogprobsContent(BaseModel):
9393

9494
logprob: float
9595

96-
top_logprobs: List[ChoiceLogprobsContentTopLogprob]
97-
9896
bytes: Optional[List[int]] = None
9997

98+
top_logprobs: Optional[List[ChoiceLogprobsContentTopLogprob]] = None
99+
100100

101101
class ChoiceLogprobsRefusalTopLogprob(BaseModel):
102102
"""
@@ -128,10 +128,10 @@ class ChoiceLogprobsRefusal(BaseModel):
128128

129129
logprob: float
130130

131-
top_logprobs: List[ChoiceLogprobsRefusalTopLogprob]
132-
133131
bytes: Optional[List[int]] = None
134132

133+
top_logprobs: Optional[List[ChoiceLogprobsRefusalTopLogprob]] = None
134+
135135

136136
class ChoiceLogprobs(BaseModel):
137137
"""

src/llama_stack_client/types/completion_create_response.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -52,10 +52,10 @@ class ChoiceLogprobsContent(BaseModel):
5252

5353
logprob: float
5454

55-
top_logprobs: List[ChoiceLogprobsContentTopLogprob]
56-
5755
bytes: Optional[List[int]] = None
5856

57+
top_logprobs: Optional[List[ChoiceLogprobsContentTopLogprob]] = None
58+
5959

6060
class ChoiceLogprobsRefusalTopLogprob(BaseModel):
6161
"""
@@ -87,10 +87,10 @@ class ChoiceLogprobsRefusal(BaseModel):
8787

8888
logprob: float
8989

90-
top_logprobs: List[ChoiceLogprobsRefusalTopLogprob]
91-
9290
bytes: Optional[List[int]] = None
9391

92+
top_logprobs: Optional[List[ChoiceLogprobsRefusalTopLogprob]] = None
93+
9494

9595
class ChoiceLogprobs(BaseModel):
9696
"""

0 commit comments

Comments
 (0)