From 99a1a08edd200468851b970f0492a37c716c31fd Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Bar=C4=B1=C5=9F=20Can=20Durak?= <36421093+bcdurak@users.noreply.github.com> Date: Tue, 6 Feb 2024 14:35:51 +0100 Subject: [PATCH] Fixing the quickstart bug --- template/quickstart.ipynb | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/template/quickstart.ipynb b/template/quickstart.ipynb index 03ef6b2..ee14d95 100644 --- a/template/quickstart.ipynb +++ b/template/quickstart.ipynb @@ -628,8 +628,8 @@ " dataset_trn, dataset_tst = feature_engineering()\n", " else:\n", " # Load the datasets from an older pipeline\n", - " dataset_trn = client.get_artifact_version(id=train_dataset_id)\n", - " dataset_tst = client.get_artifact_version(id=test_dataset_id) \n", + " dataset_trn = client.get_artifact_version(name_id_or_prefix=train_dataset_id)\n", + " dataset_tst = client.get_artifact_version(name_id_or_prefix=test_dataset_id) \n", "\n", " trained_model = model_trainer(\n", " dataset_trn=dataset_trn,\n", @@ -970,8 +970,8 @@ "@pipeline\n", "def inference(preprocess_pipeline_id: UUID):\n", " \"\"\"Model batch inference pipeline\"\"\"\n", - " # random_state = client.get_artifact_version(id=preprocess_pipeline_id).metadata[\"random_state\"].value\n", - " # target = client.get_artifact_version(id=preprocess_pipeline_id).run_metadata['target'].value\n", + " # random_state = client.get_artifact_version(name_id_or_prefix=preprocess_pipeline_id).metadata[\"random_state\"].value\n", + " # target = client.get_artifact_version(name_id_or_prefix=preprocess_pipeline_id).run_metadata['target'].value\n", " random_state = 42\n", " target = \"target\"\n", "\n", @@ -981,7 +981,7 @@ " df_inference = inference_preprocessor(\n", " dataset_inf=df_inference,\n", " # We use the preprocess pipeline from the feature engineering pipeline\n", - " preprocess_pipeline=client.get_artifact_version(id=preprocess_pipeline_id),\n", + " preprocess_pipeline=client.get_artifact_version(name_id_or_prefix=preprocess_pipeline_id),\n", " target=target,\n", " )\n", " inference_predict(\n",