From 9c180013e5be41d90a617946121e15e071775cb6 Mon Sep 17 00:00:00 2001 From: Elastic Machine Date: Fri, 7 Jun 2024 06:07:21 +0000 Subject: [PATCH] Auto-generated API code --- elasticsearch/_async/client/inference.py | 10 ++++++++++ elasticsearch/_sync/client/inference.py | 10 ++++++++++ 2 files changed, 20 insertions(+) diff --git a/elasticsearch/_async/client/inference.py b/elasticsearch/_async/client/inference.py index c7669d1e6..696f7cca3 100644 --- a/elasticsearch/_async/client/inference.py +++ b/elasticsearch/_async/client/inference.py @@ -36,8 +36,10 @@ async def delete( str, ] ] = None, + dry_run: t.Optional[bool] = None, error_trace: t.Optional[bool] = None, filter_path: t.Optional[t.Union[str, t.Sequence[str]]] = None, + force: t.Optional[bool] = None, human: t.Optional[bool] = None, pretty: t.Optional[bool] = None, ) -> ObjectApiResponse[t.Any]: @@ -48,6 +50,10 @@ async def delete( :param inference_id: The inference Id :param task_type: The task type + :param dry_run: When true, the endpoint is not deleted, and a list of ingest + processors which reference this endpoint is returned + :param force: When true, the inference endpoint is forcefully deleted even if + it is still being used by ingest processors or semantic text fields """ if inference_id in SKIP_IN_PATH: raise ValueError("Empty value passed for parameter 'inference_id'") @@ -64,10 +70,14 @@ async def delete( else: raise ValueError("Couldn't find a path for the given parameters") __query: t.Dict[str, t.Any] = {} + if dry_run is not None: + __query["dry_run"] = dry_run if error_trace is not None: __query["error_trace"] = error_trace if filter_path is not None: __query["filter_path"] = filter_path + if force is not None: + __query["force"] = force if human is not None: __query["human"] = human if pretty is not None: diff --git a/elasticsearch/_sync/client/inference.py b/elasticsearch/_sync/client/inference.py index 7427833b0..45bade337 100644 --- a/elasticsearch/_sync/client/inference.py +++ b/elasticsearch/_sync/client/inference.py @@ -36,8 +36,10 @@ def delete( str, ] ] = None, + dry_run: t.Optional[bool] = None, error_trace: t.Optional[bool] = None, filter_path: t.Optional[t.Union[str, t.Sequence[str]]] = None, + force: t.Optional[bool] = None, human: t.Optional[bool] = None, pretty: t.Optional[bool] = None, ) -> ObjectApiResponse[t.Any]: @@ -48,6 +50,10 @@ def delete( :param inference_id: The inference Id :param task_type: The task type + :param dry_run: When true, the endpoint is not deleted, and a list of ingest + processors which reference this endpoint is returned + :param force: When true, the inference endpoint is forcefully deleted even if + it is still being used by ingest processors or semantic text fields """ if inference_id in SKIP_IN_PATH: raise ValueError("Empty value passed for parameter 'inference_id'") @@ -64,10 +70,14 @@ def delete( else: raise ValueError("Couldn't find a path for the given parameters") __query: t.Dict[str, t.Any] = {} + if dry_run is not None: + __query["dry_run"] = dry_run if error_trace is not None: __query["error_trace"] = error_trace if filter_path is not None: __query["filter_path"] = filter_path + if force is not None: + __query["force"] = force if human is not None: __query["human"] = human if pretty is not None: