Skip to content

Commit 6c547e4

Browse files
feat(api): updates to vector_store, etc.
1 parent bb3cdd1 commit 6c547e4

24 files changed

+591
-1083
lines changed

.stats.yml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
configured_endpoints: 115
2-
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-8a12a05ba6892999ac506f69d5cbbc7218f28ee1a11bf8e0e548c603435bb643.yml
3-
openapi_spec_hash: 871ce212a98bdad4a44ec7fbf58d9fcb
4-
config_hash: 4c1ba9dc45c31189cd1b039d003a3544
1+
configured_endpoints: 111
2+
openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/llamastack%2Fllama-stack-client-15a929a0b71de779accc56bd09d1e5f580e216affdb408cf9890bc7a37847e9e.yml
3+
openapi_spec_hash: 5db9f7c7e80427cfa0298cbb01689559
4+
config_hash: 06758df5c4f261f9c97eafcef7e0028f

api.md

Lines changed: 0 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -234,26 +234,6 @@ Methods:
234234
- <code title="post /v1/vector-io/insert">client.vector_io.<a href="./src/llama_stack_client/resources/vector_io.py">insert</a>(\*\*<a href="src/llama_stack_client/types/vector_io_insert_params.py">params</a>) -> None</code>
235235
- <code title="post /v1/vector-io/query">client.vector_io.<a href="./src/llama_stack_client/resources/vector_io.py">query</a>(\*\*<a href="src/llama_stack_client/types/vector_io_query_params.py">params</a>) -> <a href="./src/llama_stack_client/types/query_chunks_response.py">QueryChunksResponse</a></code>
236236

237-
# VectorDBs
238-
239-
Types:
240-
241-
```python
242-
from llama_stack_client.types import (
243-
ListVectorDBsResponse,
244-
VectorDBRetrieveResponse,
245-
VectorDBListResponse,
246-
VectorDBRegisterResponse,
247-
)
248-
```
249-
250-
Methods:
251-
252-
- <code title="get /v1/vector-dbs/{vector_db_id}">client.vector_dbs.<a href="./src/llama_stack_client/resources/vector_dbs.py">retrieve</a>(vector_db_id) -> <a href="./src/llama_stack_client/types/vector_db_retrieve_response.py">VectorDBRetrieveResponse</a></code>
253-
- <code title="get /v1/vector-dbs">client.vector_dbs.<a href="./src/llama_stack_client/resources/vector_dbs.py">list</a>() -> <a href="./src/llama_stack_client/types/vector_db_list_response.py">VectorDBListResponse</a></code>
254-
- <code title="post /v1/vector-dbs">client.vector_dbs.<a href="./src/llama_stack_client/resources/vector_dbs.py">register</a>(\*\*<a href="src/llama_stack_client/types/vector_db_register_params.py">params</a>) -> <a href="./src/llama_stack_client/types/vector_db_register_response.py">VectorDBRegisterResponse</a></code>
255-
- <code title="delete /v1/vector-dbs/{vector_db_id}">client.vector_dbs.<a href="./src/llama_stack_client/resources/vector_dbs.py">unregister</a>(vector_db_id) -> None</code>
256-
257237
# VectorStores
258238

259239
Types:

src/llama_stack_client/_client.py

Lines changed: 0 additions & 38 deletions
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,6 @@
5151
benchmarks,
5252
embeddings,
5353
toolgroups,
54-
vector_dbs,
5554
completions,
5655
moderations,
5756
tool_runtime,
@@ -75,7 +74,6 @@
7574
from .resources.benchmarks import BenchmarksResource, AsyncBenchmarksResource
7675
from .resources.embeddings import EmbeddingsResource, AsyncEmbeddingsResource
7776
from .resources.toolgroups import ToolgroupsResource, AsyncToolgroupsResource
78-
from .resources.vector_dbs import VectorDBsResource, AsyncVectorDBsResource
7977
from .resources.alpha.alpha import AlphaResource, AsyncAlphaResource
8078
from .resources.completions import CompletionsResource, AsyncCompletionsResource
8179
from .resources.moderations import ModerationsResource, AsyncModerationsResource
@@ -225,12 +223,6 @@ def vector_io(self) -> VectorIoResource:
225223

226224
return VectorIoResource(self)
227225

228-
@cached_property
229-
def vector_dbs(self) -> VectorDBsResource:
230-
from .resources.vector_dbs import VectorDBsResource
231-
232-
return VectorDBsResource(self)
233-
234226
@cached_property
235227
def vector_stores(self) -> VectorStoresResource:
236228
from .resources.vector_stores import VectorStoresResource
@@ -553,12 +545,6 @@ def vector_io(self) -> AsyncVectorIoResource:
553545

554546
return AsyncVectorIoResource(self)
555547

556-
@cached_property
557-
def vector_dbs(self) -> AsyncVectorDBsResource:
558-
from .resources.vector_dbs import AsyncVectorDBsResource
559-
560-
return AsyncVectorDBsResource(self)
561-
562548
@cached_property
563549
def vector_stores(self) -> AsyncVectorStoresResource:
564550
from .resources.vector_stores import AsyncVectorStoresResource
@@ -830,12 +816,6 @@ def vector_io(self) -> vector_io.VectorIoResourceWithRawResponse:
830816

831817
return VectorIoResourceWithRawResponse(self._client.vector_io)
832818

833-
@cached_property
834-
def vector_dbs(self) -> vector_dbs.VectorDBsResourceWithRawResponse:
835-
from .resources.vector_dbs import VectorDBsResourceWithRawResponse
836-
837-
return VectorDBsResourceWithRawResponse(self._client.vector_dbs)
838-
839819
@cached_property
840820
def vector_stores(self) -> vector_stores.VectorStoresResourceWithRawResponse:
841821
from .resources.vector_stores import VectorStoresResourceWithRawResponse
@@ -993,12 +973,6 @@ def vector_io(self) -> vector_io.AsyncVectorIoResourceWithRawResponse:
993973

994974
return AsyncVectorIoResourceWithRawResponse(self._client.vector_io)
995975

996-
@cached_property
997-
def vector_dbs(self) -> vector_dbs.AsyncVectorDBsResourceWithRawResponse:
998-
from .resources.vector_dbs import AsyncVectorDBsResourceWithRawResponse
999-
1000-
return AsyncVectorDBsResourceWithRawResponse(self._client.vector_dbs)
1001-
1002976
@cached_property
1003977
def vector_stores(self) -> vector_stores.AsyncVectorStoresResourceWithRawResponse:
1004978
from .resources.vector_stores import AsyncVectorStoresResourceWithRawResponse
@@ -1158,12 +1132,6 @@ def vector_io(self) -> vector_io.VectorIoResourceWithStreamingResponse:
11581132

11591133
return VectorIoResourceWithStreamingResponse(self._client.vector_io)
11601134

1161-
@cached_property
1162-
def vector_dbs(self) -> vector_dbs.VectorDBsResourceWithStreamingResponse:
1163-
from .resources.vector_dbs import VectorDBsResourceWithStreamingResponse
1164-
1165-
return VectorDBsResourceWithStreamingResponse(self._client.vector_dbs)
1166-
11671135
@cached_property
11681136
def vector_stores(self) -> vector_stores.VectorStoresResourceWithStreamingResponse:
11691137
from .resources.vector_stores import VectorStoresResourceWithStreamingResponse
@@ -1323,12 +1291,6 @@ def vector_io(self) -> vector_io.AsyncVectorIoResourceWithStreamingResponse:
13231291

13241292
return AsyncVectorIoResourceWithStreamingResponse(self._client.vector_io)
13251293

1326-
@cached_property
1327-
def vector_dbs(self) -> vector_dbs.AsyncVectorDBsResourceWithStreamingResponse:
1328-
from .resources.vector_dbs import AsyncVectorDBsResourceWithStreamingResponse
1329-
1330-
return AsyncVectorDBsResourceWithStreamingResponse(self._client.vector_dbs)
1331-
13321294
@cached_property
13331295
def vector_stores(self) -> vector_stores.AsyncVectorStoresResourceWithStreamingResponse:
13341296
from .resources.vector_stores import AsyncVectorStoresResourceWithStreamingResponse

src/llama_stack_client/resources/__init__.py

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -144,14 +144,6 @@
144144
ToolgroupsResourceWithStreamingResponse,
145145
AsyncToolgroupsResourceWithStreamingResponse,
146146
)
147-
from .vector_dbs import (
148-
VectorDBsResource,
149-
AsyncVectorDBsResource,
150-
VectorDBsResourceWithRawResponse,
151-
AsyncVectorDBsResourceWithRawResponse,
152-
VectorDBsResourceWithStreamingResponse,
153-
AsyncVectorDBsResourceWithStreamingResponse,
154-
)
155147
from .completions import (
156148
CompletionsResource,
157149
AsyncCompletionsResource,
@@ -276,12 +268,6 @@
276268
"AsyncVectorIoResourceWithRawResponse",
277269
"VectorIoResourceWithStreamingResponse",
278270
"AsyncVectorIoResourceWithStreamingResponse",
279-
"VectorDBsResource",
280-
"AsyncVectorDBsResource",
281-
"VectorDBsResourceWithRawResponse",
282-
"AsyncVectorDBsResourceWithRawResponse",
283-
"VectorDBsResourceWithStreamingResponse",
284-
"AsyncVectorDBsResourceWithStreamingResponse",
285271
"VectorStoresResource",
286272
"AsyncVectorStoresResource",
287273
"VectorStoresResourceWithRawResponse",

src/llama_stack_client/resources/completions.py

Lines changed: 0 additions & 20 deletions
Original file line numberDiff line numberDiff line change
@@ -54,13 +54,11 @@ def create(
5454
best_of: int | Omit = omit,
5555
echo: bool | Omit = omit,
5656
frequency_penalty: float | Omit = omit,
57-
guided_choice: SequenceNotStr[str] | Omit = omit,
5857
logit_bias: Dict[str, float] | Omit = omit,
5958
logprobs: bool | Omit = omit,
6059
max_tokens: int | Omit = omit,
6160
n: int | Omit = omit,
6261
presence_penalty: float | Omit = omit,
63-
prompt_logprobs: int | Omit = omit,
6462
seed: int | Omit = omit,
6563
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
6664
stream: Literal[False] | Omit = omit,
@@ -139,13 +137,11 @@ def create(
139137
best_of: int | Omit = omit,
140138
echo: bool | Omit = omit,
141139
frequency_penalty: float | Omit = omit,
142-
guided_choice: SequenceNotStr[str] | Omit = omit,
143140
logit_bias: Dict[str, float] | Omit = omit,
144141
logprobs: bool | Omit = omit,
145142
max_tokens: int | Omit = omit,
146143
n: int | Omit = omit,
147144
presence_penalty: float | Omit = omit,
148-
prompt_logprobs: int | Omit = omit,
149145
seed: int | Omit = omit,
150146
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
151147
stream_options: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
@@ -223,13 +219,11 @@ def create(
223219
best_of: int | Omit = omit,
224220
echo: bool | Omit = omit,
225221
frequency_penalty: float | Omit = omit,
226-
guided_choice: SequenceNotStr[str] | Omit = omit,
227222
logit_bias: Dict[str, float] | Omit = omit,
228223
logprobs: bool | Omit = omit,
229224
max_tokens: int | Omit = omit,
230225
n: int | Omit = omit,
231226
presence_penalty: float | Omit = omit,
232-
prompt_logprobs: int | Omit = omit,
233227
seed: int | Omit = omit,
234228
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
235229
stream_options: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
@@ -306,13 +300,11 @@ def create(
306300
best_of: int | Omit = omit,
307301
echo: bool | Omit = omit,
308302
frequency_penalty: float | Omit = omit,
309-
guided_choice: SequenceNotStr[str] | Omit = omit,
310303
logit_bias: Dict[str, float] | Omit = omit,
311304
logprobs: bool | Omit = omit,
312305
max_tokens: int | Omit = omit,
313306
n: int | Omit = omit,
314307
presence_penalty: float | Omit = omit,
315-
prompt_logprobs: int | Omit = omit,
316308
seed: int | Omit = omit,
317309
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
318310
stream: Literal[False] | Literal[True] | Omit = omit,
@@ -337,13 +329,11 @@ def create(
337329
"best_of": best_of,
338330
"echo": echo,
339331
"frequency_penalty": frequency_penalty,
340-
"guided_choice": guided_choice,
341332
"logit_bias": logit_bias,
342333
"logprobs": logprobs,
343334
"max_tokens": max_tokens,
344335
"n": n,
345336
"presence_penalty": presence_penalty,
346-
"prompt_logprobs": prompt_logprobs,
347337
"seed": seed,
348338
"stop": stop,
349339
"stream": stream,
@@ -395,13 +385,11 @@ async def create(
395385
best_of: int | Omit = omit,
396386
echo: bool | Omit = omit,
397387
frequency_penalty: float | Omit = omit,
398-
guided_choice: SequenceNotStr[str] | Omit = omit,
399388
logit_bias: Dict[str, float] | Omit = omit,
400389
logprobs: bool | Omit = omit,
401390
max_tokens: int | Omit = omit,
402391
n: int | Omit = omit,
403392
presence_penalty: float | Omit = omit,
404-
prompt_logprobs: int | Omit = omit,
405393
seed: int | Omit = omit,
406394
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
407395
stream: Literal[False] | Omit = omit,
@@ -480,13 +468,11 @@ async def create(
480468
best_of: int | Omit = omit,
481469
echo: bool | Omit = omit,
482470
frequency_penalty: float | Omit = omit,
483-
guided_choice: SequenceNotStr[str] | Omit = omit,
484471
logit_bias: Dict[str, float] | Omit = omit,
485472
logprobs: bool | Omit = omit,
486473
max_tokens: int | Omit = omit,
487474
n: int | Omit = omit,
488475
presence_penalty: float | Omit = omit,
489-
prompt_logprobs: int | Omit = omit,
490476
seed: int | Omit = omit,
491477
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
492478
stream_options: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
@@ -564,13 +550,11 @@ async def create(
564550
best_of: int | Omit = omit,
565551
echo: bool | Omit = omit,
566552
frequency_penalty: float | Omit = omit,
567-
guided_choice: SequenceNotStr[str] | Omit = omit,
568553
logit_bias: Dict[str, float] | Omit = omit,
569554
logprobs: bool | Omit = omit,
570555
max_tokens: int | Omit = omit,
571556
n: int | Omit = omit,
572557
presence_penalty: float | Omit = omit,
573-
prompt_logprobs: int | Omit = omit,
574558
seed: int | Omit = omit,
575559
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
576560
stream_options: Dict[str, Union[bool, float, str, Iterable[object], object, None]] | Omit = omit,
@@ -647,13 +631,11 @@ async def create(
647631
best_of: int | Omit = omit,
648632
echo: bool | Omit = omit,
649633
frequency_penalty: float | Omit = omit,
650-
guided_choice: SequenceNotStr[str] | Omit = omit,
651634
logit_bias: Dict[str, float] | Omit = omit,
652635
logprobs: bool | Omit = omit,
653636
max_tokens: int | Omit = omit,
654637
n: int | Omit = omit,
655638
presence_penalty: float | Omit = omit,
656-
prompt_logprobs: int | Omit = omit,
657639
seed: int | Omit = omit,
658640
stop: Union[str, SequenceNotStr[str]] | Omit = omit,
659641
stream: Literal[False] | Literal[True] | Omit = omit,
@@ -678,13 +660,11 @@ async def create(
678660
"best_of": best_of,
679661
"echo": echo,
680662
"frequency_penalty": frequency_penalty,
681-
"guided_choice": guided_choice,
682663
"logit_bias": logit_bias,
683664
"logprobs": logprobs,
684665
"max_tokens": max_tokens,
685666
"n": n,
686667
"presence_penalty": presence_penalty,
687-
"prompt_logprobs": prompt_logprobs,
688668
"seed": seed,
689669
"stop": stop,
690670
"stream": stream,

src/llama_stack_client/resources/safety.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@
1919
)
2020
from .._base_client import make_request_options
2121
from ..types.run_shield_response import RunShieldResponse
22-
from ..types.shared_params.message import Message
2322

2423
__all__ = ["SafetyResource", "AsyncSafetyResource"]
2524

@@ -47,7 +46,7 @@ def with_streaming_response(self) -> SafetyResourceWithStreamingResponse:
4746
def run_shield(
4847
self,
4948
*,
50-
messages: Iterable[Message],
49+
messages: Iterable[safety_run_shield_params.Message],
5150
params: Dict[str, Union[bool, float, str, Iterable[object], object, None]],
5251
shield_id: str,
5352
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -116,7 +115,7 @@ def with_streaming_response(self) -> AsyncSafetyResourceWithStreamingResponse:
116115
async def run_shield(
117116
self,
118117
*,
119-
messages: Iterable[Message],
118+
messages: Iterable[safety_run_shield_params.Message],
120119
params: Dict[str, Union[bool, float, str, Iterable[object], object, None]],
121120
shield_id: str,
122121
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.

0 commit comments

Comments
 (0)