Skip to content

Commit 0e57233

Browse files
authored
chore(misc): update datasets, benchmarks to use alpha, beta prefixes (#3891)
This will be landed together with llamastack/llama-stack-client-python#282 (hence CI will be red on this one.) I have verified locally that tests pass with the updated version of the client-sdk.
1 parent 7918188 commit 0e57233

File tree

3 files changed

+15
-13
lines changed

3 files changed

+15
-13
lines changed

llama_stack/core/server/server.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -174,7 +174,9 @@ def __init__(self, config: StackRunConfig, *args, **kwargs):
174174

175175
@asynccontextmanager
176176
async def lifespan(app: StackApp):
177-
logger.info("Starting up")
177+
server_version = parse_version("llama-stack")
178+
179+
logger.info(f"Starting up Llama Stack server (version: {server_version})")
178180
assert app.stack is not None
179181
app.stack.create_registry_refresh_task()
180182
yield

tests/integration/datasets/test_datasets.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -78,18 +78,18 @@ def data_url_from_file(file_path: str) -> str:
7878
],
7979
)
8080
def test_register_and_iterrows(llama_stack_client, purpose, source, provider_id, limit):
81-
dataset = llama_stack_client.datasets.register(
81+
dataset = llama_stack_client.beta.datasets.register(
8282
purpose=purpose,
8383
source=source,
8484
)
8585
assert dataset.identifier is not None
8686
assert dataset.provider_id == provider_id
87-
iterrow_response = llama_stack_client.datasets.iterrows(dataset.identifier, limit=limit)
87+
iterrow_response = llama_stack_client.beta.datasets.iterrows(dataset.identifier, limit=limit)
8888
assert len(iterrow_response.data) == limit
8989

90-
dataset_list = llama_stack_client.datasets.list()
90+
dataset_list = llama_stack_client.beta.datasets.list()
9191
assert dataset.identifier in [d.identifier for d in dataset_list]
9292

93-
llama_stack_client.datasets.unregister(dataset.identifier)
94-
dataset_list = llama_stack_client.datasets.list()
93+
llama_stack_client.beta.datasets.unregister(dataset.identifier)
94+
dataset_list = llama_stack_client.beta.datasets.list()
9595
assert dataset.identifier not in [d.identifier for d in dataset_list]

tests/integration/eval/test_eval.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -17,17 +17,17 @@
1717

1818
@pytest.mark.parametrize("scoring_fn_id", ["basic::equality"])
1919
def test_evaluate_rows(llama_stack_client, text_model_id, scoring_fn_id):
20-
dataset = llama_stack_client.datasets.register(
20+
dataset = llama_stack_client.beta.datasets.register(
2121
purpose="eval/messages-answer",
2222
source={
2323
"type": "uri",
2424
"uri": data_url_from_file(Path(__file__).parent.parent / "datasets" / "test_dataset.csv"),
2525
},
2626
)
27-
response = llama_stack_client.datasets.list()
27+
response = llama_stack_client.beta.datasets.list()
2828
assert any(x.identifier == dataset.identifier for x in response)
2929

30-
rows = llama_stack_client.datasets.iterrows(
30+
rows = llama_stack_client.beta.datasets.iterrows(
3131
dataset_id=dataset.identifier,
3232
limit=3,
3333
)
@@ -37,12 +37,12 @@ def test_evaluate_rows(llama_stack_client, text_model_id, scoring_fn_id):
3737
scoring_fn_id,
3838
]
3939
benchmark_id = str(uuid.uuid4())
40-
llama_stack_client.benchmarks.register(
40+
llama_stack_client.alpha.benchmarks.register(
4141
benchmark_id=benchmark_id,
4242
dataset_id=dataset.identifier,
4343
scoring_functions=scoring_functions,
4444
)
45-
list_benchmarks = llama_stack_client.benchmarks.list()
45+
list_benchmarks = llama_stack_client.alpha.benchmarks.list()
4646
assert any(x.identifier == benchmark_id for x in list_benchmarks)
4747

4848
response = llama_stack_client.alpha.eval.evaluate_rows(
@@ -66,15 +66,15 @@ def test_evaluate_rows(llama_stack_client, text_model_id, scoring_fn_id):
6666

6767
@pytest.mark.parametrize("scoring_fn_id", ["basic::subset_of"])
6868
def test_evaluate_benchmark(llama_stack_client, text_model_id, scoring_fn_id):
69-
dataset = llama_stack_client.datasets.register(
69+
dataset = llama_stack_client.beta.datasets.register(
7070
purpose="eval/messages-answer",
7171
source={
7272
"type": "uri",
7373
"uri": data_url_from_file(Path(__file__).parent.parent / "datasets" / "test_dataset.csv"),
7474
},
7575
)
7676
benchmark_id = str(uuid.uuid4())
77-
llama_stack_client.benchmarks.register(
77+
llama_stack_client.alpha.benchmarks.register(
7878
benchmark_id=benchmark_id,
7979
dataset_id=dataset.identifier,
8080
scoring_functions=[scoring_fn_id],

0 commit comments

Comments
 (0)