From 804ef48218f875432437922e32706d1321019539 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Date: Tue, 12 Dec 2023 16:41:52 +0000 Subject: [PATCH] feat(bigquery): update the api #### bigquery:v2 The following keys were added: - schemas.RemoteModelInfo.properties.speechRecognizer (Total Keys: 2) - schemas.VectorSearchStatistics (Total Keys: 5) --- docs/dyn/bigquery_v2.models.html | 4 ++ .../documents/bigquery.v2.json | 43 +++++++++++++++++-- 2 files changed, 44 insertions(+), 3 deletions(-) diff --git a/docs/dyn/bigquery_v2.models.html b/docs/dyn/bigquery_v2.models.html index 608a360819..e9dbd9bf7e 100644 --- a/docs/dyn/bigquery_v2.models.html +++ b/docs/dyn/bigquery_v2.models.html @@ -801,6 +801,7 @@

Method Details

"maxBatchingRows": "A String", # Output only. Max number of rows in each batch sent to the remote service. If unset, the number of rows in each batch is set dynamically. "remoteModelVersion": "A String", # Output only. The model version for LLM. "remoteServiceType": "A String", # Output only. The remote service type for remote model. + "speechRecognizer": "A String", # Output only. The name of the speech recognizer to use for speech recognition. The expected format is `projects/{project}/locations/{location}/recognizers/{recognizer}`. Customers can specify this field at model creation. If not specified, a default recognizer `projects/{model project}/locations/global/recognizers/_` will be used. See more details at [recognizers](https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers) }, "trainingRuns": [ # Information for all training runs in increasing order of start_time. { # Information about a single training query run for the model. @@ -1821,6 +1822,7 @@

Method Details

"maxBatchingRows": "A String", # Output only. Max number of rows in each batch sent to the remote service. If unset, the number of rows in each batch is set dynamically. "remoteModelVersion": "A String", # Output only. The model version for LLM. "remoteServiceType": "A String", # Output only. The remote service type for remote model. + "speechRecognizer": "A String", # Output only. The name of the speech recognizer to use for speech recognition. The expected format is `projects/{project}/locations/{location}/recognizers/{recognizer}`. Customers can specify this field at model creation. If not specified, a default recognizer `projects/{model project}/locations/global/recognizers/_` will be used. See more details at [recognizers](https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers) }, "trainingRuns": [ # Information for all training runs in increasing order of start_time. { # Information about a single training query run for the model. @@ -2854,6 +2856,7 @@

Method Details

"maxBatchingRows": "A String", # Output only. Max number of rows in each batch sent to the remote service. If unset, the number of rows in each batch is set dynamically. "remoteModelVersion": "A String", # Output only. The model version for LLM. "remoteServiceType": "A String", # Output only. The remote service type for remote model. + "speechRecognizer": "A String", # Output only. The name of the speech recognizer to use for speech recognition. The expected format is `projects/{project}/locations/{location}/recognizers/{recognizer}`. Customers can specify this field at model creation. If not specified, a default recognizer `projects/{model project}/locations/global/recognizers/_` will be used. See more details at [recognizers](https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers) }, "trainingRuns": [ # Information for all training runs in increasing order of start_time. { # Information about a single training query run for the model. @@ -3862,6 +3865,7 @@

Method Details

"maxBatchingRows": "A String", # Output only. Max number of rows in each batch sent to the remote service. If unset, the number of rows in each batch is set dynamically. "remoteModelVersion": "A String", # Output only. The model version for LLM. "remoteServiceType": "A String", # Output only. The remote service type for remote model. + "speechRecognizer": "A String", # Output only. The name of the speech recognizer to use for speech recognition. The expected format is `projects/{project}/locations/{location}/recognizers/{recognizer}`. Customers can specify this field at model creation. If not specified, a default recognizer `projects/{model project}/locations/global/recognizers/_` will be used. See more details at [recognizers](https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers) }, "trainingRuns": [ # Information for all training runs in increasing order of start_time. { # Information about a single training query run for the model. diff --git a/googleapiclient/discovery_cache/documents/bigquery.v2.json b/googleapiclient/discovery_cache/documents/bigquery.v2.json index 13a5347219..7682dde74f 100644 --- a/googleapiclient/discovery_cache/documents/bigquery.v2.json +++ b/googleapiclient/discovery_cache/documents/bigquery.v2.json @@ -1686,7 +1686,7 @@ } } }, - "revision": "20231021", + "revision": "20231202", "rootUrl": "https://bigquery.googleapis.com/", "schemas": { "AggregateClassificationMetrics": { @@ -5966,16 +5966,23 @@ "REMOTE_SERVICE_TYPE_UNSPECIFIED", "CLOUD_AI_TRANSLATE_V3", "CLOUD_AI_VISION_V1", - "CLOUD_AI_NATURAL_LANGUAGE_V1" + "CLOUD_AI_NATURAL_LANGUAGE_V1", + "CLOUD_AI_SPEECH_TO_TEXT_V2" ], "enumDescriptions": [ "Unspecified remote service type.", "V3 Cloud AI Translation API. See more details at [Cloud Translation API] (https://cloud.google.com/translate/docs/reference/rest).", "V1 Cloud AI Vision API See more details at [Cloud Vision API] (https://cloud.google.com/vision/docs/reference/rest).", - "V1 Cloud AI Natural Language API. See more details at [REST Resource: documents](https://cloud.google.com/natural-language/docs/reference/rest/v1/documents)." + "V1 Cloud AI Natural Language API. See more details at [REST Resource: documents](https://cloud.google.com/natural-language/docs/reference/rest/v1/documents).", + "V2 Speech-to-Text API. See more details at [Google Cloud Speech-to-Text V2 API](https://cloud.google.com/speech-to-text/v2/docs)" ], "readOnly": true, "type": "string" + }, + "speechRecognizer": { + "description": "Output only. The name of the speech recognizer to use for speech recognition. The expected format is `projects/{project}/locations/{location}/recognizers/{recognizer}`. Customers can specify this field at model creation. If not specified, a default recognizer `projects/{model project}/locations/global/recognizers/_` will be used. See more details at [recognizers](https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers)", + "readOnly": true, + "type": "string" } }, "type": "object" @@ -8273,6 +8280,36 @@ }, "type": "object" }, + "VectorSearchStatistics": { + "description": "Statistics for a vector search query. Populated as part of JobStatistics2.", + "id": "VectorSearchStatistics", + "properties": { + "indexUnusedReasons": { + "description": "When `indexUsageMode` is `UNUSED` or `PARTIALLY_USED`, this field explains why indexes were not used in all or part of the vector search query. If `indexUsageMode` is `FULLY_USED`, this field is not populated.", + "items": { + "$ref": "IndexUnusedReason" + }, + "type": "array" + }, + "indexUsageMode": { + "description": "Specifies the index usage mode for the query.", + "enum": [ + "INDEX_USAGE_MODE_UNSPECIFIED", + "UNUSED", + "PARTIALLY_USED", + "FULLY_USED" + ], + "enumDescriptions": [ + "Index usage mode not specified.", + "No vector indexes were used in the vector search query. See [`indexUnusedReasons`] (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason) for detailed reasons.", + "Part of the vector search query used vector indexes. See [`indexUnusedReasons`] (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason) for why other parts of the query did not use vector indexes.", + "The entire vector search query used vector indexes." + ], + "type": "string" + } + }, + "type": "object" + }, "ViewDefinition": { "id": "ViewDefinition", "properties": {