3535class LlamaStackClient (SyncAPIClient ):
3636 agents : resources .AgentsResource
3737 batch_inferences : resources .BatchInferencesResource
38+ datasets : resources .DatasetsResource
39+ eval : resources .EvalResource
3840 inspect : resources .InspectResource
3941 inference : resources .InferenceResource
4042 memory : resources .MemoryResource
4143 memory_banks : resources .MemoryBanksResource
42- datasets : resources .DatasetsResource
4344 models : resources .ModelsResource
4445 post_training : resources .PostTrainingResource
4546 providers : resources .ProvidersResource
@@ -51,7 +52,7 @@ class LlamaStackClient(SyncAPIClient):
5152 datasetio : resources .DatasetioResource
5253 scoring : resources .ScoringResource
5354 scoring_functions : resources .ScoringFunctionsResource
54- eval : resources .EvalResource
55+ eval_tasks : resources .EvalTasksResource
5556 with_raw_response : LlamaStackClientWithRawResponse
5657 with_streaming_response : LlamaStackClientWithStreamedResponse
5758
@@ -85,7 +86,6 @@ def __init__(
8586 base_url = os .environ .get ("LLAMA_STACK_CLIENT_BASE_URL" )
8687 if base_url is None :
8788 base_url = f"http://any-hosted-llama-stack.com"
88-
8989 if provider_data :
9090 if default_headers is None :
9191 default_headers = {}
@@ -104,11 +104,12 @@ def __init__(
104104
105105 self .agents = resources .AgentsResource (self )
106106 self .batch_inferences = resources .BatchInferencesResource (self )
107+ self .datasets = resources .DatasetsResource (self )
108+ self .eval = resources .EvalResource (self )
107109 self .inspect = resources .InspectResource (self )
108110 self .inference = resources .InferenceResource (self )
109111 self .memory = resources .MemoryResource (self )
110112 self .memory_banks = resources .MemoryBanksResource (self )
111- self .datasets = resources .DatasetsResource (self )
112113 self .models = resources .ModelsResource (self )
113114 self .post_training = resources .PostTrainingResource (self )
114115 self .providers = resources .ProvidersResource (self )
@@ -120,7 +121,7 @@ def __init__(
120121 self .datasetio = resources .DatasetioResource (self )
121122 self .scoring = resources .ScoringResource (self )
122123 self .scoring_functions = resources .ScoringFunctionsResource (self )
123- self .eval = resources .EvalResource (self )
124+ self .eval_tasks = resources .EvalTasksResource (self )
124125 self .with_raw_response = LlamaStackClientWithRawResponse (self )
125126 self .with_streaming_response = LlamaStackClientWithStreamedResponse (self )
126127
@@ -224,11 +225,12 @@ def _make_status_error(
224225class AsyncLlamaStackClient (AsyncAPIClient ):
225226 agents : resources .AsyncAgentsResource
226227 batch_inferences : resources .AsyncBatchInferencesResource
228+ datasets : resources .AsyncDatasetsResource
229+ eval : resources .AsyncEvalResource
227230 inspect : resources .AsyncInspectResource
228231 inference : resources .AsyncInferenceResource
229232 memory : resources .AsyncMemoryResource
230233 memory_banks : resources .AsyncMemoryBanksResource
231- datasets : resources .AsyncDatasetsResource
232234 models : resources .AsyncModelsResource
233235 post_training : resources .AsyncPostTrainingResource
234236 providers : resources .AsyncProvidersResource
@@ -240,7 +242,7 @@ class AsyncLlamaStackClient(AsyncAPIClient):
240242 datasetio : resources .AsyncDatasetioResource
241243 scoring : resources .AsyncScoringResource
242244 scoring_functions : resources .AsyncScoringFunctionsResource
243- eval : resources .AsyncEvalResource
245+ eval_tasks : resources .AsyncEvalTasksResource
244246 with_raw_response : AsyncLlamaStackClientWithRawResponse
245247 with_streaming_response : AsyncLlamaStackClientWithStreamedResponse
246248
@@ -293,11 +295,12 @@ def __init__(
293295
294296 self .agents = resources .AsyncAgentsResource (self )
295297 self .batch_inferences = resources .AsyncBatchInferencesResource (self )
298+ self .datasets = resources .AsyncDatasetsResource (self )
299+ self .eval = resources .AsyncEvalResource (self )
296300 self .inspect = resources .AsyncInspectResource (self )
297301 self .inference = resources .AsyncInferenceResource (self )
298302 self .memory = resources .AsyncMemoryResource (self )
299303 self .memory_banks = resources .AsyncMemoryBanksResource (self )
300- self .datasets = resources .AsyncDatasetsResource (self )
301304 self .models = resources .AsyncModelsResource (self )
302305 self .post_training = resources .AsyncPostTrainingResource (self )
303306 self .providers = resources .AsyncProvidersResource (self )
@@ -309,7 +312,7 @@ def __init__(
309312 self .datasetio = resources .AsyncDatasetioResource (self )
310313 self .scoring = resources .AsyncScoringResource (self )
311314 self .scoring_functions = resources .AsyncScoringFunctionsResource (self )
312- self .eval = resources .AsyncEvalResource (self )
315+ self .eval_tasks = resources .AsyncEvalTasksResource (self )
313316 self .with_raw_response = AsyncLlamaStackClientWithRawResponse (self )
314317 self .with_streaming_response = AsyncLlamaStackClientWithStreamedResponse (self )
315318
@@ -414,11 +417,12 @@ class LlamaStackClientWithRawResponse:
414417 def __init__ (self , client : LlamaStackClient ) -> None :
415418 self .agents = resources .AgentsResourceWithRawResponse (client .agents )
416419 self .batch_inferences = resources .BatchInferencesResourceWithRawResponse (client .batch_inferences )
420+ self .datasets = resources .DatasetsResourceWithRawResponse (client .datasets )
421+ self .eval = resources .EvalResourceWithRawResponse (client .eval )
417422 self .inspect = resources .InspectResourceWithRawResponse (client .inspect )
418423 self .inference = resources .InferenceResourceWithRawResponse (client .inference )
419424 self .memory = resources .MemoryResourceWithRawResponse (client .memory )
420425 self .memory_banks = resources .MemoryBanksResourceWithRawResponse (client .memory_banks )
421- self .datasets = resources .DatasetsResourceWithRawResponse (client .datasets )
422426 self .models = resources .ModelsResourceWithRawResponse (client .models )
423427 self .post_training = resources .PostTrainingResourceWithRawResponse (client .post_training )
424428 self .providers = resources .ProvidersResourceWithRawResponse (client .providers )
@@ -432,18 +436,19 @@ def __init__(self, client: LlamaStackClient) -> None:
432436 self .datasetio = resources .DatasetioResourceWithRawResponse (client .datasetio )
433437 self .scoring = resources .ScoringResourceWithRawResponse (client .scoring )
434438 self .scoring_functions = resources .ScoringFunctionsResourceWithRawResponse (client .scoring_functions )
435- self .eval = resources .EvalResourceWithRawResponse (client .eval )
439+ self .eval_tasks = resources .EvalTasksResourceWithRawResponse (client .eval_tasks )
436440
437441
438442class AsyncLlamaStackClientWithRawResponse :
439443 def __init__ (self , client : AsyncLlamaStackClient ) -> None :
440444 self .agents = resources .AsyncAgentsResourceWithRawResponse (client .agents )
441445 self .batch_inferences = resources .AsyncBatchInferencesResourceWithRawResponse (client .batch_inferences )
446+ self .datasets = resources .AsyncDatasetsResourceWithRawResponse (client .datasets )
447+ self .eval = resources .AsyncEvalResourceWithRawResponse (client .eval )
442448 self .inspect = resources .AsyncInspectResourceWithRawResponse (client .inspect )
443449 self .inference = resources .AsyncInferenceResourceWithRawResponse (client .inference )
444450 self .memory = resources .AsyncMemoryResourceWithRawResponse (client .memory )
445451 self .memory_banks = resources .AsyncMemoryBanksResourceWithRawResponse (client .memory_banks )
446- self .datasets = resources .AsyncDatasetsResourceWithRawResponse (client .datasets )
447452 self .models = resources .AsyncModelsResourceWithRawResponse (client .models )
448453 self .post_training = resources .AsyncPostTrainingResourceWithRawResponse (client .post_training )
449454 self .providers = resources .AsyncProvidersResourceWithRawResponse (client .providers )
@@ -457,18 +462,19 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
457462 self .datasetio = resources .AsyncDatasetioResourceWithRawResponse (client .datasetio )
458463 self .scoring = resources .AsyncScoringResourceWithRawResponse (client .scoring )
459464 self .scoring_functions = resources .AsyncScoringFunctionsResourceWithRawResponse (client .scoring_functions )
460- self .eval = resources .AsyncEvalResourceWithRawResponse (client .eval )
465+ self .eval_tasks = resources .AsyncEvalTasksResourceWithRawResponse (client .eval_tasks )
461466
462467
463468class LlamaStackClientWithStreamedResponse :
464469 def __init__ (self , client : LlamaStackClient ) -> None :
465470 self .agents = resources .AgentsResourceWithStreamingResponse (client .agents )
466471 self .batch_inferences = resources .BatchInferencesResourceWithStreamingResponse (client .batch_inferences )
472+ self .datasets = resources .DatasetsResourceWithStreamingResponse (client .datasets )
473+ self .eval = resources .EvalResourceWithStreamingResponse (client .eval )
467474 self .inspect = resources .InspectResourceWithStreamingResponse (client .inspect )
468475 self .inference = resources .InferenceResourceWithStreamingResponse (client .inference )
469476 self .memory = resources .MemoryResourceWithStreamingResponse (client .memory )
470477 self .memory_banks = resources .MemoryBanksResourceWithStreamingResponse (client .memory_banks )
471- self .datasets = resources .DatasetsResourceWithStreamingResponse (client .datasets )
472478 self .models = resources .ModelsResourceWithStreamingResponse (client .models )
473479 self .post_training = resources .PostTrainingResourceWithStreamingResponse (client .post_training )
474480 self .providers = resources .ProvidersResourceWithStreamingResponse (client .providers )
@@ -482,18 +488,19 @@ def __init__(self, client: LlamaStackClient) -> None:
482488 self .datasetio = resources .DatasetioResourceWithStreamingResponse (client .datasetio )
483489 self .scoring = resources .ScoringResourceWithStreamingResponse (client .scoring )
484490 self .scoring_functions = resources .ScoringFunctionsResourceWithStreamingResponse (client .scoring_functions )
485- self .eval = resources .EvalResourceWithStreamingResponse (client .eval )
491+ self .eval_tasks = resources .EvalTasksResourceWithStreamingResponse (client .eval_tasks )
486492
487493
488494class AsyncLlamaStackClientWithStreamedResponse :
489495 def __init__ (self , client : AsyncLlamaStackClient ) -> None :
490496 self .agents = resources .AsyncAgentsResourceWithStreamingResponse (client .agents )
491497 self .batch_inferences = resources .AsyncBatchInferencesResourceWithStreamingResponse (client .batch_inferences )
498+ self .datasets = resources .AsyncDatasetsResourceWithStreamingResponse (client .datasets )
499+ self .eval = resources .AsyncEvalResourceWithStreamingResponse (client .eval )
492500 self .inspect = resources .AsyncInspectResourceWithStreamingResponse (client .inspect )
493501 self .inference = resources .AsyncInferenceResourceWithStreamingResponse (client .inference )
494502 self .memory = resources .AsyncMemoryResourceWithStreamingResponse (client .memory )
495503 self .memory_banks = resources .AsyncMemoryBanksResourceWithStreamingResponse (client .memory_banks )
496- self .datasets = resources .AsyncDatasetsResourceWithStreamingResponse (client .datasets )
497504 self .models = resources .AsyncModelsResourceWithStreamingResponse (client .models )
498505 self .post_training = resources .AsyncPostTrainingResourceWithStreamingResponse (client .post_training )
499506 self .providers = resources .AsyncProvidersResourceWithStreamingResponse (client .providers )
@@ -507,7 +514,7 @@ def __init__(self, client: AsyncLlamaStackClient) -> None:
507514 self .datasetio = resources .AsyncDatasetioResourceWithStreamingResponse (client .datasetio )
508515 self .scoring = resources .AsyncScoringResourceWithStreamingResponse (client .scoring )
509516 self .scoring_functions = resources .AsyncScoringFunctionsResourceWithStreamingResponse (client .scoring_functions )
510- self .eval = resources .AsyncEvalResourceWithStreamingResponse (client .eval )
517+ self .eval_tasks = resources .AsyncEvalTasksResourceWithStreamingResponse (client .eval_tasks )
511518
512519
513520Client = LlamaStackClient
0 commit comments