From 74fd10779b27f6c5f93aa7bb01d9ad4bcd15c951 Mon Sep 17 00:00:00 2001 From: Serge Smertin <259697+nfx@users.noreply.github.com> Date: Mon, 25 Mar 2024 21:11:25 +0100 Subject: [PATCH] Make `R CMD` check pass (#34) --- .codegen.json | 3 + .codegen/service.R.tmpl | 13 +- Makefile | 7 +- R/account_access_control_proxy.R | 11 +- R/alerts.R | 22 ++- R/apps.R | 19 ++- R/artifact_allowlists.R | 6 + R/automatic_cluster_update.R | 7 +- R/catalogs.R | 14 ++ R/clean_rooms.R | 14 ++ R/cluster_policies.R | 30 ++++- R/clusters.R | 102 +++++++++++--- R/command_execution.R | 51 ++++--- R/connections.R | 16 ++- R/credentials_manager.R | 3 + R/csp_enablement.R | 7 +- R/current_user.R | 4 +- R/dashboard_widgets.R | 13 +- R/dashboards.R | 27 +++- R/data_sources.R | 4 +- R/dbfs.R | 30 +++++ R/dbsql_permissions.R | 11 +- R/default_namespace.R | 7 + R/esm_enablement.R | 7 +- R/experiments.R | 99 ++++++++++++-- R/external_locations.R | 14 ++ R/files.R | 26 +++- R/functions.R | 15 +++ R/git_credentials.R | 16 ++- R/global_init_scripts.R | 16 ++- R/grants.R | 9 ++ R/groups.R | 30 ++++- R/instance_pools.R | 32 ++++- R/instance_profiles.R | 13 +- R/ip_access_lists.R | 25 +++- R/jobs.R | 85 ++++++++++-- R/lakehouse_monitors.R | 24 ++++ R/lakeview.R | 18 +++ R/libraries.R | 13 +- R/metastores.R | 33 ++++- R/model_registry.R | 125 ++++++++++++++++-- R/model_versions.R | 15 +++ R/online_tables.R | 8 ++ R/permission_migration.R | 3 + R/permissions.R | 16 ++- R/pipelines.R | 64 +++++++-- R/policy_families.R | 7 +- R/providers.R | 17 +++ R/queries.R | 24 +++- R/query_history.R | 2 + R/query_visualizations.R | 13 +- R/recipient_activation.R | 6 + R/recipients.R | 20 +++ R/registered_models.R | 20 +++ R/repos.R | 30 ++++- R/restrict_workspace_admins.R | 9 +- R/schemas.R | 15 +++ R/secrets.R | 34 ++++- R/service_principals.R | 26 +++- R/serving_endpoints.R | 58 +++++++- R/shares.R | 22 ++- R/statement_execution.R | 14 +- R/storage_credentials.R | 16 +++ R/system_schemas.R | 9 ++ R/table_constraints.R | 6 + R/tables.R | 20 ++- R/token_management.R | 27 +++- R/tokens.R | 9 +- R/users.R | 42 ++++-- R/vector_search_endpoints.R | 15 ++- R/vector_search_indexes.R | 24 ++++ R/volumes.R | 17 ++- R/warehouses.R | 57 +++++++- R/workspace.R | 34 ++++- R/workspace_bindings.R | 12 ++ R/workspace_conf.R | 7 +- README.md | 1 - man/accountAccessControlProxyUpdateRuleSet.Rd | 2 +- man/alertsDelete.Rd | 2 +- man/alertsGet.Rd | 2 +- man/alertsList.Rd | 5 +- man/alertsUpdate.Rd | 2 +- man/appsGetApps.Rd | 5 +- man/automaticClusterUpdateUpdate.Rd | 2 +- man/clusterPoliciesSetPermissions.Rd | 2 +- man/clusterPoliciesUpdatePermissions.Rd | 2 +- man/clustersCreate.Rd | 10 +- man/clustersDelete.Rd | 4 + man/clustersEdit.Rd | 10 +- man/clustersListNodeTypes.Rd | 5 +- man/clustersListZones.Rd | 5 +- man/clustersResize.Rd | 4 + man/clustersRestart.Rd | 4 + man/clustersSetPermissions.Rd | 2 +- man/clustersSparkVersions.Rd | 5 +- man/clustersStart.Rd | 4 + man/clustersUpdatePermissions.Rd | 2 +- man/commandExecutionCancel.Rd | 10 +- man/commandExecutionCommandStatus.Rd | 6 +- man/commandExecutionContextStatus.Rd | 4 +- man/commandExecutionCreate.Rd | 6 +- man/commandExecutionDestroy.Rd | 4 +- man/commandExecutionExecute.Rd | 6 +- man/connectionsList.Rd | 5 +- man/cspEnablementUpdate.Rd | 2 +- man/currentUserMe.Rd | 5 +- man/dashboardWidgetsCreate.Rd | 2 +- man/dashboardWidgetsUpdate.Rd | 2 +- man/dashboardsCreate.Rd | 2 +- man/dashboardsDelete.Rd | 2 +- man/dashboardsGet.Rd | 2 +- man/dashboardsRestore.Rd | 2 +- man/dashboardsUpdate.Rd | 2 +- man/dataSourcesList.Rd | 5 +- man/dbsqlPermissionsSet.Rd | 2 +- man/esmEnablementUpdate.Rd | 2 +- man/experimentsGetHistory.Rd | 2 +- man/experimentsGetRun.Rd | 2 +- man/experimentsListArtifacts.Rd | 2 +- man/experimentsLogMetric.Rd | 2 +- man/experimentsLogParam.Rd | 2 +- man/experimentsSetPermissions.Rd | 2 +- man/experimentsSetTag.Rd | 2 +- man/experimentsUpdatePermissions.Rd | 2 +- man/experimentsUpdateRun.Rd | 2 +- man/filesUpload.Rd | 2 +- man/gitCredentialsList.Rd | 5 +- man/globalInitScriptsList.Rd | 5 +- man/groupsCreate.Rd | 6 +- man/groupsPatch.Rd | 2 +- man/groupsUpdate.Rd | 6 +- man/instancePoolsList.Rd | 5 +- man/instancePoolsSetPermissions.Rd | 2 +- man/instancePoolsUpdatePermissions.Rd | 2 +- man/instanceProfilesList.Rd | 5 +- man/ipAccessListsCreate.Rd | 2 +- man/ipAccessListsList.Rd | 5 +- man/ipAccessListsReplace.Rd | 2 +- man/ipAccessListsUpdate.Rd | 2 +- man/jobsCancelRun.Rd | 4 + man/jobsGetRun.Rd | 4 + man/jobsRepairRun.Rd | 6 +- man/jobsRunNow.Rd | 6 +- man/jobsSetPermissions.Rd | 2 +- man/jobsSubmit.Rd | 6 +- man/jobsUpdatePermissions.Rd | 2 +- man/librariesAllClusterStatuses.Rd | 5 +- man/metastoresCurrent.Rd | 5 +- man/metastoresList.Rd | 5 +- man/metastoresSummary.Rd | 5 +- man/modelRegistryCreateWebhook.Rd | 4 +- man/modelRegistryDeleteComment.Rd | 2 +- man/modelRegistryGetModel.Rd | 4 +- man/modelRegistrySetPermissions.Rd | 2 +- man/modelRegistryTransitionStage.Rd | 4 +- man/modelRegistryUpdatePermissions.Rd | 2 +- man/modelRegistryUpdateWebhook.Rd | 4 +- man/permissionsSet.Rd | 2 +- man/permissionsUpdate.Rd | 2 +- man/pipelinesCreate.Rd | 2 +- man/pipelinesDelete.Rd | 2 +- man/pipelinesGet.Rd | 2 +- man/pipelinesListPipelineEvents.Rd | 4 +- man/pipelinesSetPermissions.Rd | 2 +- man/pipelinesStartUpdate.Rd | 4 +- man/pipelinesStop.Rd | 6 +- man/pipelinesUpdatePermissions.Rd | 2 +- man/policyFamiliesGet.Rd | 2 +- man/queriesDelete.Rd | 2 +- man/queriesGet.Rd | 2 +- man/queriesRestore.Rd | 2 +- man/queriesUpdate.Rd | 2 +- man/queryVisualizationsUpdate.Rd | 4 +- man/reposSetPermissions.Rd | 2 +- man/reposUpdatePermissions.Rd | 2 +- man/restrictWorkspaceAdminsUpdate.Rd | 2 +- man/secretsListScopes.Rd | 5 +- man/servicePrincipalsCreate.Rd | 4 +- man/servicePrincipalsPatch.Rd | 2 +- man/servicePrincipalsUpdate.Rd | 4 +- man/servingEndpointsCreate.Rd | 4 + man/servingEndpointsList.Rd | 7 +- man/servingEndpointsSetPermissions.Rd | 2 +- man/servingEndpointsUpdateConfig.Rd | 4 + man/servingEndpointsUpdatePermissions.Rd | 2 +- man/sharesList.Rd | 5 +- ...tementExecutionGetStatementResultChunkN.Rd | 2 +- man/tablesUpdate.Rd | 2 +- man/tokenManagementGetPermissionLevels.Rd | 5 +- man/tokenManagementGetPermissions.Rd | 5 +- man/tokenManagementSetPermissions.Rd | 2 +- man/tokenManagementUpdatePermissions.Rd | 2 +- man/tokensList.Rd | 5 +- man/usersCreate.Rd | 4 +- man/usersGetPermissionLevels.Rd | 5 +- man/usersGetPermissions.Rd | 5 +- man/usersPatch.Rd | 2 +- man/usersSetPermissions.Rd | 2 +- man/usersUpdate.Rd | 4 +- man/usersUpdatePermissions.Rd | 2 +- man/vectorSearchEndpointsCreateEndpoint.Rd | 4 + man/volumesCreate.Rd | 2 +- man/warehousesCreate.Rd | 4 + man/warehousesEdit.Rd | 4 + man/warehousesGetWorkspaceWarehouseConfig.Rd | 5 +- man/warehousesSetPermissions.Rd | 2 +- man/warehousesStart.Rd | 4 + man/warehousesStop.Rd | 4 + man/warehousesUpdatePermissions.Rd | 2 +- man/workspaceConfGetStatus.Rd | 2 +- man/workspaceSetPermissions.Rd | 2 +- man/workspaceUpdatePermissions.Rd | 2 +- 212 files changed, 1874 insertions(+), 328 deletions(-) diff --git a/.codegen.json b/.codegen.json index f0a16adf..21890003 100644 --- a/.codegen.json +++ b/.codegen.json @@ -14,6 +14,9 @@ "Rscript -e \"devtools::install_dev_deps('.')\"" ], "post_generate": [ + "gsed -i -- 's/\\[Deprecated, use run_id instead\\]/Deprecated, use run_id instead./g' **/*.R", + "gsed -i -- 's/\\[MLflow endpoint\\]/MLflow endpoint/g' **/*.R", + "gsed -i -- \"s/\\['timestamp asc'\\]/timestamp asc/g\" **/*.R", "Rscript -e \"devtools::document()\"" ] } diff --git a/.codegen/service.R.tmpl b/.codegen/service.R.tmpl index a8ae5ca1..2f34d39c 100644 --- a/.codegen/service.R.tmpl +++ b/.codegen/service.R.tmpl @@ -3,16 +3,21 @@ #' @importFrom stats runif NULL -{{range .Methods}}{{.Comment "#' " 80}}{{if .Request}} +{{range .Methods}}{{.Comment "#' " 80}} +#' @param client Required. Instance of DatabricksClient() +{{if .Request}} {{if and .Wait (not .IsCrudRead)}}#' #' @description #' This is a long-running operation, which blocks until {{.Service.TitleName}} on Databricks reach #' {{range $i, $s := .Wait.Success}}{{if $i}} or {{end}}{{.Content}}{{end}} state with the timeout of {{.Wait.Timeout}} minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks {{.Service.TitleName}} is reported to console. You can change this behavior #' by changing the `callback` parameter. -{{end}}#' @param client Required. Instance of DatabricksClient() -#'{{range .Request.Fields}} -#' @param {{.SnakeName}} {{if .Required}}Required. {{end}}{{.Summary}}{{end}} +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. +{{end}} +{{if .Request.RequiredFields}}#' +{{end}}#'{{range .Request.Fields}} +#' @param {{.SnakeName}} {{if .Required}}Required. {{end}}{{with .Summary}}{{.}}{{else}}This field has no description yet.{{end}}{{end}} {{end}}{{if .Pagination}}#' #' @return `data.frame` with all of the response pages. {{end}}#' diff --git a/Makefile b/Makefile index 4a6b0ca2..d5ff3cdf 100644 --- a/Makefile +++ b/Makefile @@ -4,11 +4,6 @@ deps: @echo "✅ Installing dependencies" @Rscript -e "if (!require(devtools)) install.packages('devtools', repos = 'https://cran.rstudio.com')" @Rscript -e "devtools::install_dev_deps('.')" - -gen: # invoked by Databricks employees - @echo "✅ Regenerating files from OpenAPI" - @oac - @Rscript -e "devtools::document()" tarball: @echo "✅ Building tarball" @@ -20,7 +15,7 @@ test: fmt: @Rscript .codegen/format.R -check: gen +check: @echo "✅ run R CMD check on packages" @set -e ; \ cleanup () { rm *.tar.gz; rm -fr *.Rcheck; } ; \ diff --git a/R/account_access_control_proxy.R b/R/account_access_control_proxy.R index a74f196d..686ef62d 100755 --- a/R/account_access_control_proxy.R +++ b/R/account_access_control_proxy.R @@ -9,6 +9,9 @@ NULL #' is grantable if the rule set on the resource can contain an access rule of #' the role. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param resource Required. The resource name for which assignable roles will be listed. #' @@ -26,6 +29,9 @@ accountAccessControlProxyGetAssignableRolesForResource <- function(client, resou #' contains a list of access rules on the said resource. Currently only a #' default rule set for each resource is supported. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param etag Required. Etag used for versioning. #' @param name Required. The ruleset name associated with the request. @@ -43,9 +49,12 @@ accountAccessControlProxyGetRuleSet <- function(client, name, etag) { #' the current version of the rule set before modifying it. This pattern helps #' prevent conflicts between concurrent updates. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the rule set. -#' @param rule_set Required. +#' @param rule_set Required. This field has no description yet. #' #' @rdname accountAccessControlProxyUpdateRuleSet #' @export diff --git a/R/alerts.R b/R/alerts.R index 38705aad..c669829f 100755 --- a/R/alerts.R +++ b/R/alerts.R @@ -9,6 +9,9 @@ NULL #' a query, evaluates a condition of its result, and notifies users or #' notification destinations if the condition was met. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the alert. #' @param options Required. Alert configuration options. @@ -30,8 +33,11 @@ alertsCreate <- function(client, name, options, query_id, parent = NULL, rearm = #' restored. **Note:** Unlike queries and dashboards, alerts cannot be moved to #' the trash. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param alert_id Required. +#' @param alert_id Required. This field has no description yet. #' #' @rdname alertsDelete #' @export @@ -44,8 +50,11 @@ alertsDelete <- function(client, alert_id) { #' #' Gets an alert. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param alert_id Required. +#' @param alert_id Required. This field has no description yet. #' #' @rdname alertsGet #' @export @@ -56,7 +65,9 @@ alertsGet <- function(client, alert_id) { #' Get alerts. #' -#' Gets a list of alerts.#' +#' Gets a list of alerts. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname alertsList #' @export alertsList <- function(client) { @@ -67,8 +78,11 @@ alertsList <- function(client) { #' #' Updates an alert. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param alert_id Required. +#' @param alert_id Required. This field has no description yet. #' @param name Required. Name of the alert. #' @param options Required. Alert configuration options. #' @param query_id Required. Query ID. diff --git a/R/apps.R b/R/apps.R index 648e5668..56fc90fc 100755 --- a/R/apps.R +++ b/R/apps.R @@ -7,6 +7,9 @@ NULL #' #' Creates and deploys an application. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param manifest Required. Manifest that specifies the application requirements. #' @param resources Information passed at app deployment time to fulfill app dependencies. @@ -22,6 +25,9 @@ appsCreate <- function(client, manifest, resources = NULL) { #' #' Delete an application definition #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of an application. #' @@ -36,6 +42,9 @@ appsDeleteApp <- function(client, name) { #' #' Get an application definition #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of an application. #' @@ -50,6 +59,9 @@ appsGetApp <- function(client, name) { #' #' Get deployment status for an application #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param deployment_id Required. The deployment id for an application. #' @param include_app_log Boolean flag to include application logs. @@ -64,7 +76,9 @@ appsGetAppDeploymentStatus <- function(client, deployment_id, include_app_log = #' List all applications. #' -#' List all available applications#' +#' List all available applications +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname appsGetApps #' @export appsGetApps <- function(client) { @@ -75,6 +89,9 @@ appsGetApps <- function(client) { #' #' Get deployment events for an application #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of an application. #' diff --git a/R/artifact_allowlists.R b/R/artifact_allowlists.R index 2c8be9b7..75de9822 100755 --- a/R/artifact_allowlists.R +++ b/R/artifact_allowlists.R @@ -8,6 +8,9 @@ NULL #' Get the artifact allowlist of a certain artifact type. The caller must be a #' metastore admin or have the **MANAGE ALLOWLIST** privilege on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param artifact_type Required. The artifact type of the allowlist. #' @@ -25,6 +28,9 @@ artifactAllowlistsGet <- function(client, artifact_type) { #' allowlist is replaced with the new allowlist. The caller must be a metastore #' admin or have the **MANAGE ALLOWLIST** privilege on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param artifact_matchers Required. A list of allowed artifact match patterns. #' @param artifact_type Required. The artifact type of the allowlist. diff --git a/R/automatic_cluster_update.R b/R/automatic_cluster_update.R index be842bd9..8f65b93b 100755 --- a/R/automatic_cluster_update.R +++ b/R/automatic_cluster_update.R @@ -7,6 +7,8 @@ NULL #' #' Gets the automatic cluster update setting. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -26,10 +28,13 @@ automaticClusterUpdateGet <- function(client, etag = NULL) { #' If the setting is updated concurrently, `PATCH` fails with 409 and the #' request must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_missing Required. This should always be set to true for Settings API. #' @param field_mask Required. Field mask is required to be passed into the PATCH request. -#' @param setting Required. +#' @param setting Required. This field has no description yet. #' #' @rdname automaticClusterUpdateUpdate #' @export diff --git a/R/catalogs.R b/R/catalogs.R index 6c18537d..c330ed95 100755 --- a/R/catalogs.R +++ b/R/catalogs.R @@ -8,6 +8,9 @@ NULL #' Creates a new catalog instance in the parent metastore if the caller is a #' metastore admin or has the **CREATE_CATALOG** privilege. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param connection_name The name of the connection to an external data source. @@ -33,6 +36,9 @@ catalogsCreate <- function(client, name, comment = NULL, connection_name = NULL, #' Deletes the catalog that matches the supplied name. The caller must be a #' metastore admin or the owner of the catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param force Force deletion even if the catalog is not empty. #' @param name Required. The name of the catalog. @@ -51,6 +57,9 @@ catalogsDelete <- function(client, name, force = NULL) { #' admin, the owner of the catalog, or a user that has the **USE_CATALOG** #' privilege set for their account. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_browse Whether to include catalogs in the response for which the principal can only access selective metadata for. #' @param name Required. The name of the catalog. @@ -70,6 +79,8 @@ catalogsGet <- function(client, name, include_browse = NULL) { #' retrieved. There is no guarantee of a specific ordering of the elements in #' the array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param include_browse Whether to include catalogs in the response for which the principal can only access selective metadata for. #' @@ -91,6 +102,9 @@ catalogsList <- function(client, include_browse = NULL) { #' the owner of the catalog, or a metastore admin (when changing the owner field #' of the catalog). #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param enable_predictive_optimization Whether predictive optimization should be enabled for this object and objects under it. diff --git a/R/clean_rooms.R b/R/clean_rooms.R index fbfe3ffa..ba491a13 100755 --- a/R/clean_rooms.R +++ b/R/clean_rooms.R @@ -8,6 +8,9 @@ NULL #' Creates a new clean room with specified colaborators. The caller must be a #' metastore admin or have the **CREATE_CLEAN_ROOM** privilege on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param name Required. Name of the clean room. @@ -25,6 +28,9 @@ cleanRoomsCreate <- function(client, name, remote_detailed_info, comment = NULL) #' Deletes a data object clean room from the metastore. The caller must be an #' owner of the clean room. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the clean room. #' @@ -40,6 +46,9 @@ cleanRoomsDelete <- function(client, name) { #' Gets a data object clean room from the metastore. The caller must be a #' metastore admin or the owner of the clean room. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_remote_details Whether to include remote details (central) on the clean room. #' @param name Required. The name of the clean room. @@ -58,6 +67,8 @@ cleanRoomsGet <- function(client, name, include_remote_details = NULL) { #' be a metastore admin or the owner of the clean room. There is no guarantee of #' a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param max_results Maximum number of clean rooms to return. #' @param page_token Opaque pagination token to go to next page based on previous query. @@ -104,6 +115,9 @@ cleanRoomsList <- function(client, max_results = NULL, page_token = NULL) { #' #' Table removals through **update** do not require additional privileges. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_updates Array of shared data object updates. #' @param comment User-provided free-form text description. diff --git a/R/cluster_policies.R b/R/cluster_policies.R index 5686b494..ea5d82fb 100755 --- a/R/cluster_policies.R +++ b/R/cluster_policies.R @@ -7,6 +7,9 @@ NULL #' #' Creates a new policy with prescribed settings. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param definition Policy definition document expressed in [Databricks Cluster Policy Definition Language](https://docs.databricks.com/administration-guide/clusters/policy-definition.html). #' @param description Additional human-readable description of the cluster policy. @@ -32,6 +35,9 @@ clusterPoliciesCreate <- function(client, name, definition = NULL, description = #' Delete a policy for a cluster. Clusters governed by this policy can still #' run, but cannot be edited. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param policy_id Required. The ID of the policy to delete. #' @@ -47,6 +53,9 @@ clusterPoliciesDelete <- function(client, policy_id) { #' Update an existing policy for cluster. This operation may make some clusters #' governed by the previous policy invalid. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param definition Policy definition document expressed in [Databricks Cluster Policy Definition Language](https://docs.databricks.com/administration-guide/clusters/policy-definition.html). #' @param description Additional human-readable description of the cluster policy. @@ -73,6 +82,9 @@ clusterPoliciesEdit <- function(client, policy_id, name, definition = NULL, desc #' Get a cluster policy entity. Creation and editing is available to admins #' only. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param policy_id Required. Canonical unique identifier for the cluster policy. #' @@ -87,6 +99,9 @@ clusterPoliciesGet <- function(client, policy_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_policy_id Required. The cluster policy for which to get or manage permissions. #' @@ -103,6 +118,9 @@ clusterPoliciesGetPermissionLevels <- function(client, cluster_policy_id) { #' Gets the permissions of a cluster policy. Cluster policies can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_policy_id Required. The cluster policy for which to get or manage permissions. #' @@ -118,6 +136,8 @@ clusterPoliciesGetPermissions <- function(client, cluster_policy_id) { #' #' Returns a list of policies accessible by the requesting user. #' @param client Required. Instance of DatabricksClient() + + #' #' @param sort_column The cluster policy attribute to sort by. #' @param sort_order The order in which the policies get listed. @@ -139,8 +159,11 @@ clusterPoliciesList <- function(client, sort_column = NULL, sort_order = NULL) { #' Sets permissions on a cluster policy. Cluster policies can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param cluster_policy_id Required. The cluster policy for which to get or manage permissions. #' #' @rdname clusterPoliciesSetPermissions @@ -156,8 +179,11 @@ clusterPoliciesSetPermissions <- function(client, cluster_policy_id, access_cont #' Updates the permissions on a cluster policy. Cluster policies can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param cluster_policy_id Required. The cluster policy for which to get or manage permissions. #' #' @rdname clusterPoliciesUpdatePermissions diff --git a/R/clusters.R b/R/clusters.R index aa08a2e0..1a6c7ec1 100755 --- a/R/clusters.R +++ b/R/clusters.R @@ -9,6 +9,9 @@ NULL #' terminated to perform this operation. The service principal application ID #' can be supplied as an argument to `owner_username`. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. . #' @param owner_username Required. New owner of the cluster_id after this RPC. @@ -30,15 +33,20 @@ clustersChangeOwner <- function(client, cluster_id, owner_username) { #' If Databricks acquires at least 85% of the requested on-demand nodes, cluster #' creation will succeed. Otherwise the cluster will terminate with an #' informative error message. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' -#' @param apply_policy_default_values +#' +#' @param apply_policy_default_values This field has no description yet. #' @param autoscale Parameters needed in order to automatically scale clusters up and down based on load. #' @param autotermination_minutes Automatically terminates the cluster after it is inactive for this time in minutes. #' @param aws_attributes Attributes related to clusters running on Amazon Web Services. @@ -48,7 +56,7 @@ clustersChangeOwner <- function(client, cluster_id, owner_username) { #' @param cluster_source Determines whether the cluster was created by a user through the UI, created by the Databricks Jobs Scheduler, or through an API request. #' @param custom_tags Additional tags for cluster resources. #' @param data_security_mode Data security mode decides what data governance model to use when accessing data from a cluster. -#' @param docker_image +#' @param docker_image This field has no description yet. #' @param driver_instance_pool_id The optional ID of the instance pool for the driver of the cluster belongs. #' @param driver_node_type_id The node type of the Spark driver. #' @param enable_elastic_disk Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space when its Spark workers are running low on disk space. @@ -65,7 +73,7 @@ clustersChangeOwner <- function(client, cluster_id, owner_username) { #' @param spark_env_vars An object containing a set of optional, user-specified environment variable key-value pairs. #' @param spark_version Required. The Spark version of the cluster, e.g. #' @param ssh_public_keys SSH public key contents that will be added to each Spark node in this cluster. -#' @param workload_type +#' @param workload_type This field has no description yet. #' #' @rdname clustersCreate #' @export @@ -133,13 +141,18 @@ clustersCreate <- function(client, spark_version, apply_policy_default_values = #' asynchronously. Once the termination has completed, the cluster will be in a #' `TERMINATED` state. If the cluster is already in a `TERMINATING` or #' `TERMINATED` state, nothing will happen. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' TERMINATED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param cluster_id Required. The cluster to be terminated. #' @@ -198,15 +211,20 @@ clustersDelete <- function(client, cluster_id, timeout = 20, callback = cli_repo #' state will be rejected with an `INVALID_STATE` error code. #' #' Clusters created by the Databricks Jobs service cannot be edited. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' -#' @param apply_policy_default_values +#' +#' @param apply_policy_default_values This field has no description yet. #' @param autoscale Parameters needed in order to automatically scale clusters up and down based on load. #' @param autotermination_minutes Automatically terminates the cluster after it is inactive for this time in minutes. #' @param aws_attributes Attributes related to clusters running on Amazon Web Services. @@ -217,7 +235,7 @@ clustersDelete <- function(client, cluster_id, timeout = 20, callback = cli_repo #' @param cluster_source Determines whether the cluster was created by a user through the UI, created by the Databricks Jobs Scheduler, or through an API request. #' @param custom_tags Additional tags for cluster resources. #' @param data_security_mode Data security mode decides what data governance model to use when accessing data from a cluster. -#' @param docker_image +#' @param docker_image This field has no description yet. #' @param driver_instance_pool_id The optional ID of the instance pool for the driver of the cluster belongs. #' @param driver_node_type_id The node type of the Spark driver. #' @param enable_elastic_disk Autoscaling Local Storage: when enabled, this cluster will dynamically acquire additional disk space when its Spark workers are running low on disk space. @@ -234,7 +252,7 @@ clustersDelete <- function(client, cluster_id, timeout = 20, callback = cli_repo #' @param spark_env_vars An object containing a set of optional, user-specified environment variable key-value pairs. #' @param spark_version Required. The Spark version of the cluster, e.g. #' @param ssh_public_keys SSH public key contents that will be added to each Spark node in this cluster. -#' @param workload_type +#' @param workload_type This field has no description yet. #' #' @rdname clustersEdit #' @export @@ -301,6 +319,9 @@ clustersEdit <- function(client, cluster_id, spark_version, apply_policy_default #' paginated. If there are more events to read, the response includes all the #' nparameters necessary to request the next page of events. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. The ID of the cluster to retrieve events about. #' @param end_time The end time in epoch milliseconds. @@ -341,6 +362,9 @@ clustersEvents <- function(client, cluster_id, end_time = NULL, event_types = NU #' Retrieves the information for a cluster given its identifier. Clusters can be #' described while they are running, or up to 60 days after they are terminated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. The cluster about which to retrieve information. #' @@ -355,6 +379,9 @@ clustersGet <- function(client, cluster_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. The cluster for which to get or manage permissions. #' @@ -371,6 +398,9 @@ clustersGetPermissionLevels <- function(client, cluster_id) { #' Gets the permissions of a cluster. Clusters can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. The cluster for which to get or manage permissions. #' @@ -393,6 +423,8 @@ clustersGetPermissions <- function(client, cluster_id) { #' clusters, all 45 terminated all-purpose clusters, and the 30 most recently #' terminated job clusters. #' @param client Required. Instance of DatabricksClient() + + #' #' @param can_use_client Filter clusters based on what type of client it can be used for. #' @@ -411,7 +443,9 @@ clustersList <- function(client, can_use_client = NULL) { #' List node types. #' #' Returns a list of supported Spark node types. These node types can be used to -#' launch a cluster.#' +#' launch a cluster. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname clustersListNodeTypes #' @export clustersListNodeTypes <- function(client) { @@ -421,7 +455,9 @@ clustersListNodeTypes <- function(client) { #' List availability zones. #' #' Returns a list of availability zones where clusters can be created in (For -#' example, us-west-2a). These zones can be used to launch a cluster.#' +#' example, us-west-2a). These zones can be used to launch a cluster. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname clustersListZones #' @export clustersListZones <- function(client) { @@ -437,6 +473,9 @@ clustersListZones <- function(client) { #' cluster list, and API users can no longer perform any action on permanently #' deleted clusters. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. The cluster to be deleted. #' @@ -453,6 +492,9 @@ clustersPermanentDelete <- function(client, cluster_id) { #' ListClusters API. Pinning a cluster that is already pinned will have no #' effect. This API can only be called by workspace admins. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. . #' @@ -467,13 +509,18 @@ clustersPin <- function(client, cluster_id) { #' #' Resizes a cluster to have a desired number of workers. This will fail unless #' the cluster is in a `RUNNING` state. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param autoscale Parameters needed in order to automatically scale clusters up and down based on load. #' @param cluster_id Required. The cluster to be resized. @@ -525,13 +572,18 @@ clustersResize <- function(client, cluster_id, autoscale = NULL, num_workers = N #' #' Restarts a Spark cluster with the supplied ID. If the cluster is not #' currently in a `RUNNING` state, nothing will happen. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param cluster_id Required. The cluster to be started. #' @param restart_user . @@ -583,8 +635,11 @@ clustersRestart <- function(client, cluster_id, restart_user = NULL, timeout = 2 #' Sets permissions on a cluster. Clusters can inherit permissions from their #' root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param cluster_id Required. The cluster for which to get or manage permissions. #' #' @rdname clustersSetPermissions @@ -598,7 +653,9 @@ clustersSetPermissions <- function(client, cluster_id, access_control_list = NUL #' List available Spark versions. #' #' Returns the list of available Spark versions. These versions can be used to -#' launch a cluster.#' +#' launch a cluster. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname clustersSparkVersions #' @export clustersSparkVersions <- function(client) { @@ -615,13 +672,18 @@ clustersSparkVersions <- function(client) { #' autoscaling cluster, the current cluster starts with the minimum number of #' nodes. * If the cluster is not currently in a `TERMINATED` state, nothing #' will happen. * Clusters launched to run a job cannot be started. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Clusters on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Clusters is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param cluster_id Required. The cluster to be started. #' @@ -672,6 +734,9 @@ clustersStart <- function(client, cluster_id, timeout = 20, callback = cli_repor #' ListClusters API. Unpinning a cluster that is not pinned will have no effect. #' This API can only be called by workspace admins. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. . #' @@ -687,8 +752,11 @@ clustersUnpin <- function(client, cluster_id) { #' Updates the permissions on a cluster. Clusters can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param cluster_id Required. The cluster for which to get or manage permissions. #' #' @rdname clustersUpdatePermissions diff --git a/R/command_execution.R b/R/command_execution.R index 7b9cd8e2..dea790ed 100755 --- a/R/command_execution.R +++ b/R/command_execution.R @@ -8,17 +8,21 @@ NULL #' Cancels a currently running command within an execution context. #' #' The command ID is obtained from a prior successful call to __execute__. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Command Execution on Databricks reach #' Cancelled state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Command Execution is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' -#' @param cluster_id -#' @param command_id -#' @param context_id +#' @param cluster_id This field has no description yet. +#' @param command_id This field has no description yet. +#' @param context_id This field has no description yet. #' #' @rdname commandExecutionCancel #' @export @@ -74,10 +78,13 @@ commandExecutionCancel <- function(client, cluster_id = NULL, command_id = NULL, #' #' The command ID is obtained from a prior successful call to __execute__. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param cluster_id Required. -#' @param command_id Required. -#' @param context_id Required. +#' +#' @param cluster_id Required. This field has no description yet. +#' @param command_id Required. This field has no description yet. +#' @param context_id Required. This field has no description yet. #' #' @rdname commandExecutionCommandStatus #' @export @@ -90,9 +97,12 @@ commandExecutionCommandStatus <- function(client, cluster_id, context_id, comman #' #' Gets the status for an execution context. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param cluster_id Required. -#' @param context_id Required. +#' +#' @param cluster_id Required. This field has no description yet. +#' @param context_id Required. This field has no description yet. #' #' @rdname commandExecutionContextStatus #' @export @@ -106,16 +116,20 @@ commandExecutionContextStatus <- function(client, cluster_id, context_id) { #' Creates an execution context for running cluster commands. #' #' If successful, this method returns the ID of the new execution context. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Command Execution on Databricks reach #' Running state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Command Execution is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' #' @param cluster_id Running cluster id. -#' @param language +#' @param language This field has no description yet. #' #' @rdname commandExecutionCreate #' @export @@ -164,9 +178,12 @@ commandExecutionCreate <- function(client, cluster_id = NULL, language = NULL, t #' #' Deletes an execution context. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param cluster_id Required. -#' @param context_id Required. +#' +#' @param cluster_id Required. This field has no description yet. +#' @param context_id Required. This field has no description yet. #' #' @rdname commandExecutionDestroy #' @export @@ -182,18 +199,22 @@ commandExecutionDestroy <- function(client, cluster_id, context_id) { #' #' If successful, it returns an ID for tracking the status of the command's #' execution. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Command Execution on Databricks reach #' Finished or Error state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Command Execution is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' #' @param cluster_id Running cluster id. #' @param command Executable code. #' @param context_id Running context id. -#' @param language +#' @param language This field has no description yet. #' #' @rdname commandExecutionExecute #' @export diff --git a/R/connections.R b/R/connections.R index 42936713..3f5d987b 100755 --- a/R/connections.R +++ b/R/connections.R @@ -11,6 +11,9 @@ NULL #' specify connection details and configurations for interaction with the #' external server. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param connection_type Required. The type of connection. @@ -32,6 +35,9 @@ connectionsCreate <- function(client, name, connection_type, options, comment = #' #' Deletes the connection that matches the supplied name. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the connection to be deleted. #' @@ -46,6 +52,9 @@ connectionsDelete <- function(client, name) { #' #' Gets a connection from it's name. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the connection. #' @@ -58,7 +67,9 @@ connectionsGet <- function(client, name) { #' List connections. #' -#' List all connections.#' +#' List all connections. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname connectionsList @@ -74,6 +85,9 @@ connectionsList <- function(client) { #' #' Updates the connection that matches the supplied name. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the connection. #' @param new_name New name for the connection. diff --git a/R/credentials_manager.R b/R/credentials_manager.R index ca2c8cac..15bfa7c0 100755 --- a/R/credentials_manager.R +++ b/R/credentials_manager.R @@ -8,6 +8,9 @@ NULL #' Exchange tokens with an Identity Provider to get a new access token. It #' allows specifying scopes to determine token permissions. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param partition_id Required. The partition of Credentials store. #' @param scopes Required. Array of scopes for the token request. diff --git a/R/csp_enablement.R b/R/csp_enablement.R index 03ed60c2..68fc14bd 100755 --- a/R/csp_enablement.R +++ b/R/csp_enablement.R @@ -7,6 +7,8 @@ NULL #' #' Gets the compliance security profile setting. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -26,10 +28,13 @@ cspEnablementGet <- function(client, etag = NULL) { #' request. If the setting is updated concurrently, `PATCH` fails with 409 and #' the request must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_missing Required. This should always be set to true for Settings API. #' @param field_mask Required. Field mask is required to be passed into the PATCH request. -#' @param setting Required. +#' @param setting Required. This field has no description yet. #' #' @rdname cspEnablementUpdate #' @export diff --git a/R/current_user.R b/R/current_user.R index 70ad3b8e..1773b00f 100755 --- a/R/current_user.R +++ b/R/current_user.R @@ -5,7 +5,9 @@ NULL #' Get current user info. #' -#' Get details about the current method caller's identity.#' +#' Get details about the current method caller's identity. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname currentUserMe #' @export currentUserMe <- function(client) { diff --git a/R/dashboard_widgets.R b/R/dashboard_widgets.R index ded29790..27fc3f44 100755 --- a/R/dashboard_widgets.R +++ b/R/dashboard_widgets.R @@ -5,9 +5,12 @@ NULL #' Add widget to a dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. Dashboard ID returned by :method:dashboards/create. -#' @param options Required. +#' @param options Required. This field has no description yet. #' @param text If this is a textbox widget, the application displays this text. #' @param visualization_id Query Vizualization ID returned by :method:queryvisualizations/create. #' @param width Required. Width of a widget. @@ -23,6 +26,9 @@ dashboardWidgetsCreate <- function(client, dashboard_id, options, width, text = #' Remove widget. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Widget ID returned by :method:dashboardwidgets/create. #' @@ -35,10 +41,13 @@ dashboardWidgetsDelete <- function(client, id) { #' Update existing widget. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. Dashboard ID returned by :method:dashboards/create. #' @param id Required. Widget ID returned by :method:dashboardwidgets/create. -#' @param options Required. +#' @param options Required. This field has no description yet. #' @param text If this is a textbox widget, the application displays this text. #' @param visualization_id Query Vizualization ID returned by :method:queryvisualizations/create. #' @param width Required. Width of a widget. diff --git a/R/dashboards.R b/R/dashboards.R index d82683fc..5f48958b 100755 --- a/R/dashboards.R +++ b/R/dashboards.R @@ -5,13 +5,16 @@ NULL #' Create a dashboard object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_filters_enabled Indicates whether the dashboard filters are enabled. #' @param is_favorite Indicates whether this dashboard object should appear in the current user's favorites list. #' @param name Required. The title of this dashboard that appears in list views and at the top of the dashboard page. #' @param parent The identifier of the workspace folder containing the object. #' @param run_as_role Sets the **Run as** role for the object. -#' @param tags +#' @param tags This field has no description yet. #' #' @rdname dashboardsCreate #' @export @@ -27,8 +30,11 @@ dashboardsCreate <- function(client, name, dashboard_filters_enabled = NULL, is_ #' Moves a dashboard to the trash. Trashed dashboards do not appear in list #' views or searches, and cannot be shared. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param dashboard_id Required. +#' +#' @param dashboard_id Required. This field has no description yet. #' #' @rdname dashboardsDelete #' @export @@ -42,8 +48,11 @@ dashboardsDelete <- function(client, dashboard_id) { #' Returns a JSON representation of a dashboard object, including its #' visualization and query objects. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param dashboard_id Required. +#' @param dashboard_id Required. This field has no description yet. #' #' @rdname dashboardsGet #' @export @@ -59,6 +68,8 @@ dashboardsGet <- function(client, dashboard_id) { #' ### **Warning: Calling this API concurrently 10 or more times could result in #' throttling, service degradation, or a temporary ban.** #' @param client Required. Instance of DatabricksClient() + + #' #' @param order Name of dashboard attribute to order by. #' @param page Page number to retrieve. @@ -93,8 +104,11 @@ dashboardsList <- function(client, order = NULL, page = NULL, page_size = NULL, #' #' A restored dashboard appears in list views and searches and can be shared. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param dashboard_id Required. +#' +#' @param dashboard_id Required. This field has no description yet. #' #' @rdname dashboardsRestore #' @export @@ -111,8 +125,11 @@ dashboardsRestore <- function(client, dashboard_id) { #' #' **Note**: You cannot undo this operation. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param dashboard_id Required. +#' @param dashboard_id Required. This field has no description yet. #' @param name The title of this dashboard that appears in list views and at the top of the dashboard page. #' @param run_as_role Sets the **Run as** role for the object. #' diff --git a/R/data_sources.R b/R/data_sources.R index 1108dd2e..4bc66a5b 100755 --- a/R/data_sources.R +++ b/R/data_sources.R @@ -7,7 +7,9 @@ NULL #' #' Retrieves a full list of SQL warehouses available in this workspace. All #' fields that appear in this API response are enumerated for clarity. However, -#' you need only a SQL warehouse's `id` to create new queries against it.#' +#' you need only a SQL warehouse's `id` to create new queries against it. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname dataSourcesList #' @export dataSourcesList <- function(client) { diff --git a/R/dbfs.R b/R/dbfs.R index 38e66893..b33b1363 100755 --- a/R/dbfs.R +++ b/R/dbfs.R @@ -12,6 +12,9 @@ NULL #' If the block of data exceeds 1 MB, this call will throw an exception with #' ``MAX_BLOCK_SIZE_EXCEEDED``. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param data Required. The base64-encoded data to append to the stream. #' @param handle Required. The handle on an open stream. @@ -28,6 +31,9 @@ dbfsAddBlock <- function(client, handle, data) { #' Closes the stream specified by the input handle. If the handle does not #' exist, this call throws an exception with ``RESOURCE_DOES_NOT_EXIST``. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param handle Required. The handle on an open stream. #' @@ -51,6 +57,9 @@ dbfsClose <- function(client, handle) { #' ``add-block`` calls with the handle you have. 3. Issue a ``close`` call with #' the handle you have. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param overwrite The flag that specifies whether to overwrite existing file/files. #' @param path Required. The path of the new file. @@ -83,6 +92,9 @@ dbfsCreate <- function(client, path, overwrite = NULL) { #' such as selective deletes, and the possibility to automate periodic delete #' jobs. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The path of the file or directory to delete. #' @param recursive Whether or not to recursively delete the directory's contents. @@ -99,6 +111,9 @@ dbfsDelete <- function(client, path, recursive = NULL) { #' Gets the file information for a file or directory. If the file or directory #' does not exist, this call throws an exception with `RESOURCE_DOES_NOT_EXIST`. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The path of the file or directory. #' @@ -123,6 +138,9 @@ dbfsGetStatus <- function(client, path) { #' system utility (dbutils.fs)](/dev-tools/databricks-utils.html#dbutils-fs), #' which provides the same functionality without timing out. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The path of the file or directory. #' @@ -146,6 +164,9 @@ dbfsList <- function(client, path) { #' this operation fails, it might have succeeded in creating some of the #' necessary parent directories. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The path of the new directory. #' @@ -164,6 +185,9 @@ dbfsMkdirs <- function(client, path) { #' this call throws an exception with `RESOURCE_ALREADY_EXISTS`. If the given #' source path is a directory, this call always recursively moves all files. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param destination_path Required. The destination path of the file or directory. #' @param source_path Required. The source path of the file or directory. @@ -190,6 +214,9 @@ dbfsMove <- function(client, source_path, destination_path) { #' If you want to upload large files, use the streaming upload. For details, see #' :method:dbfs/create, :method:dbfs/addBlock, :method:dbfs/close. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param contents This parameter might be absent, and instead a posted file will be used. #' @param overwrite The flag that specifies whether to overwrite existing file/files. @@ -213,6 +240,9 @@ dbfsPut <- function(client, path, contents = NULL, overwrite = NULL) { #' If `offset + length` exceeds the number of bytes in a file, it reads the #' contents until the end of file. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param length The number of bytes to read starting from the offset. #' @param offset The offset to read from in bytes. diff --git a/R/dbsql_permissions.R b/R/dbsql_permissions.R index c1545303..f53b73d1 100755 --- a/R/dbsql_permissions.R +++ b/R/dbsql_permissions.R @@ -8,6 +8,9 @@ NULL #' Gets a JSON representation of the access control list (ACL) for a specified #' object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param object_id Required. Object ID. #' @param object_type Required. The type of object permissions to check. @@ -25,8 +28,11 @@ dbsqlPermissionsGet <- function(client, object_type, object_id) { #' Sets the access control list (ACL) for a specified object. This operation #' will complete rewrite the ACL. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param object_id Required. Object ID. #' @param object_type Required. The type of object permission to set. #' @@ -43,6 +49,9 @@ dbsqlPermissionsSet <- function(client, object_type, object_id, access_control_l #' Transfers ownership of a dashboard, query, or alert to an active user. #' Requires an admin API key. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param new_owner Email address for the new owner, who must exist in the workspace. #' @param object_id Required. The ID of the object on which to change ownership. diff --git a/R/default_namespace.R b/R/default_namespace.R index b0bef3f3..284611da 100755 --- a/R/default_namespace.R +++ b/R/default_namespace.R @@ -11,6 +11,8 @@ NULL #' setting is updated/deleted concurrently, `DELETE` fails with 409 and the #' request must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -26,6 +28,8 @@ defaultNamespaceDelete <- function(client, etag = NULL) { #' #' Gets the default namespace setting. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -47,6 +51,9 @@ defaultNamespaceGet <- function(client, etag = NULL) { #' request. If the setting is updated concurrently, `PATCH` fails with 409 and #' the request must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_missing Required. This should always be set to true for Settings API. #' @param field_mask Required. Field mask is required to be passed into the PATCH request. diff --git a/R/esm_enablement.R b/R/esm_enablement.R index 1c83a13b..7839e49b 100755 --- a/R/esm_enablement.R +++ b/R/esm_enablement.R @@ -7,6 +7,8 @@ NULL #' #' Gets the enhanced security monitoring setting. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -26,10 +28,13 @@ esmEnablementGet <- function(client, etag = NULL) { #' request. If the setting is updated concurrently, `PATCH` fails with 409 and #' the request must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_missing Required. This should always be set to true for Settings API. #' @param field_mask Required. Field mask is required to be passed into the PATCH request. -#' @param setting Required. +#' @param setting Required. This field has no description yet. #' #' @rdname esmEnablementUpdate #' @export diff --git a/R/experiments.R b/R/experiments.R index fa10a097..5af5fece 100755 --- a/R/experiments.R +++ b/R/experiments.R @@ -12,6 +12,9 @@ NULL #' #' Throws `RESOURCE_ALREADY_EXISTS` if a experiment with the given name exists. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param artifact_location Location where all artifacts for the experiment are stored. #' @param name Required. Experiment name. @@ -31,6 +34,8 @@ experimentsCreateExperiment <- function(client, name, artifact_location = NULL, #' `mlflowParam`, `mlflowMetric` and `mlflowRunTag` associated with a single #' execution. #' @param client Required. Instance of DatabricksClient() + + #' #' @param experiment_id ID of the associated experiment. #' @param start_time Unix timestamp in milliseconds of when the run started. @@ -52,6 +57,9 @@ experimentsCreateRun <- function(client, experiment_id = NULL, start_time = NULL #' for deletion. If the experiment uses FileStore, artifacts associated with #' experiment are also deleted. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. ID of the associated experiment. #' @@ -66,6 +74,9 @@ experimentsDeleteExperiment <- function(client, experiment_id) { #' #' Marks a run for deletion. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. ID of the run to delete. #' @@ -83,6 +94,9 @@ experimentsDeleteRun <- function(client, run_id) { #' from a Databricks Notebook in Python, you can use the client code snippet on #' https://learn.microsoft.com/en-us/azure/databricks/mlflow/runs#bulk-delete. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. The ID of the experiment containing the runs to delete. #' @param max_runs An optional positive integer indicating the maximum number of runs to delete. @@ -100,6 +114,9 @@ experimentsDeleteRuns <- function(client, experiment_id, max_timestamp_millis, m #' Deletes a tag on a run. Tags are run metadata that can be updated during a #' run and after a run completes. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param run_id Required. ID of the run that the tag was logged under. @@ -123,6 +140,9 @@ experimentsDeleteTag <- function(client, run_id, key) { #' Throws `RESOURCE_DOES_NOT_EXIST` if no experiment with the specified name #' exists. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_name Required. Name of the associated experiment. #' @@ -137,6 +157,9 @@ experimentsGetByName <- function(client, experiment_name) { #' #' Gets metadata for an experiment. This method works on deleted experiments. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. ID of the associated experiment. #' @@ -151,12 +174,15 @@ experimentsGetExperiment <- function(client, experiment_id) { #' #' Gets a list of all values for the specified metric for a given run. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param max_results Maximum number of Metric records to return per paginated request. #' @param metric_key Required. Name of the metric. #' @param page_token Token indicating the page of metric histories to fetch. #' @param run_id ID of the run from which to fetch metric values. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run from which to fetch metric values. +#' @param run_uuid Deprecated, use run_id instead. ID of the run from which to fetch metric values. #' #' @return `data.frame` with all of the response pages. #' @@ -188,6 +214,9 @@ experimentsGetHistory <- function(client, metric_key, max_results = NULL, page_t #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. The experiment for which to get or manage permissions. #' @@ -204,6 +233,9 @@ experimentsGetPermissionLevels <- function(client, experiment_id) { #' Gets the permissions of an experiment. Experiments can inherit permissions #' from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. The experiment for which to get or manage permissions. #' @@ -223,9 +255,12 @@ experimentsGetPermissions <- function(client, experiment_id) { #' If there are multiple values with the latest timestamp, return the maximum of #' these values. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. ID of the run to fetch. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run to fetch. +#' @param run_uuid Deprecated, use run_id instead. ID of the run to fetch. #' #' @rdname experimentsGetRun #' @export @@ -239,11 +274,13 @@ experimentsGetRun <- function(client, run_id, run_uuid = NULL) { #' List artifacts for a run. Takes an optional `artifact_path` prefix. If it is #' specified, the response contains only artifacts with the specified prefix.', #' @param client Required. Instance of DatabricksClient() + + #' #' @param page_token Token indicating the page of artifact results to fetch. #' @param path Filter artifacts matching this path (a relative path from the root artifact directory). #' @param run_id ID of the run whose artifacts to list. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run whose artifacts to list. +#' @param run_uuid Deprecated, use run_id instead. ID of the run whose artifacts to list. #' #' @return `data.frame` with all of the response pages. #' @@ -274,6 +311,8 @@ experimentsListArtifacts <- function(client, page_token = NULL, path = NULL, run #' #' Gets a list of all experiments. #' @param client Required. Instance of DatabricksClient() + + #' #' @param max_results Maximum number of experiments desired. #' @param page_token Token indicating the page of experiments to fetch. @@ -344,6 +383,8 @@ experimentsListExperiments <- function(client, max_results = NULL, page_token = #' * Metric keys, param keys, and tag keys can be up to 250 characters in length #' * Parameter and tag values can be up to 250 characters in length #' @param client Required. Instance of DatabricksClient() + + #' #' @param metrics Metrics to log. #' @param params Params to log. @@ -363,6 +404,8 @@ experimentsLogBatch <- function(client, metrics = NULL, params = NULL, run_id = #' **NOTE:** Experimental: This API may change or be removed in a future release #' without warning. #' @param client Required. Instance of DatabricksClient() + + #' #' @param datasets Dataset inputs. #' @param run_id ID of the run to log under. @@ -380,10 +423,13 @@ experimentsLogInputs <- function(client, datasets = NULL, run_id = NULL) { #' value) with an associated timestamp. Examples include the various metrics #' that represent ML model accuracy. A metric can be logged multiple times. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the metric. #' @param run_id ID of the run under which to log the metric. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run under which to log the metric. +#' @param run_uuid Deprecated, use run_id instead. ID of the run under which to log the metric. #' @param step Step at which to log the metric. #' @param timestamp Required. Unix timestamp in milliseconds at the time metric was logged. #' @param value Required. Double value of the metric being logged. @@ -402,6 +448,8 @@ experimentsLogMetric <- function(client, key, value, timestamp, run_id = NULL, r #' **NOTE:** Experimental: This API may change or be removed in a future release #' without warning. #' @param client Required. Instance of DatabricksClient() + + #' #' @param model_json MLmodel file in json format. #' @param run_id ID of the run to log under. @@ -420,10 +468,13 @@ experimentsLogModel <- function(client, model_json = NULL, run_id = NULL) { #' constant dates and values used in an ETL pipeline. A param can be logged only #' once for a run. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the param. #' @param run_id ID of the run under which to log the param. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run under which to log the param. +#' @param run_uuid Deprecated, use run_id instead. ID of the run under which to log the param. #' @param value Required. String value of the param being logged. #' #' @rdname experimentsLogParam @@ -442,6 +493,9 @@ experimentsLogParam <- function(client, key, value, run_id = NULL, run_uuid = NU #' Throws `RESOURCE_DOES_NOT_EXIST` if experiment was never created or was #' permanently deleted. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. ID of the associated experiment. #' @@ -456,6 +510,9 @@ experimentsRestoreExperiment <- function(client, experiment_id) { #' #' Restores a deleted run. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. ID of the run to restore. #' @@ -473,6 +530,9 @@ experimentsRestoreRun <- function(client, run_id) { #' from a Databricks Notebook in Python, you can use the client code snippet on #' https://learn.microsoft.com/en-us/azure/databricks/mlflow/runs#bulk-restore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. The ID of the experiment containing the runs to restore. #' @param max_runs An optional positive integer indicating the maximum number of runs to restore. @@ -489,6 +549,8 @@ experimentsRestoreRuns <- function(client, experiment_id, min_timestamp_millis, #' #' Searches for experiments that satisfy specified search criteria. #' @param client Required. Instance of DatabricksClient() + + #' #' @param filter String representing a SQL filter condition (e.g. #' @param max_results Maximum number of experiments desired. @@ -528,6 +590,8 @@ experimentsSearchExperiments <- function(client, filter = NULL, max_results = NU #' #' Search expressions can use `mlflowMetric` and `mlflowParam` keys.', #' @param client Required. Instance of DatabricksClient() + + #' #' @param experiment_ids List of experiment IDs to search over. #' @param filter A filter expression over params, metrics, and tags, that allows returning a subset of runs. @@ -567,6 +631,9 @@ experimentsSearchRuns <- function(client, experiment_ids = NULL, filter = NULL, #' Sets a tag on an experiment. Experiment tags are metadata that can be #' updated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. ID of the experiment under which to log the tag. #' @param key Required. Name of the tag. @@ -584,8 +651,11 @@ experimentsSetExperimentTag <- function(client, experiment_id, key, value) { #' Sets permissions on an experiment. Experiments can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param experiment_id Required. The experiment for which to get or manage permissions. #' #' @rdname experimentsSetPermissions @@ -601,10 +671,13 @@ experimentsSetPermissions <- function(client, experiment_id, access_control_list #' Sets a tag on a run. Tags are run metadata that can be updated during a run #' and after a run completes. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param run_id ID of the run under which to log the tag. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run under which to log the tag. +#' @param run_uuid Deprecated, use run_id instead. ID of the run under which to log the tag. #' @param value Required. String value of the tag being logged. #' #' @rdname experimentsSetTag @@ -618,6 +691,9 @@ experimentsSetTag <- function(client, key, value, run_id = NULL, run_uuid = NULL #' #' Updates experiment metadata. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param experiment_id Required. ID of the associated experiment. #' @param new_name If provided, the experiment's name is changed to the new name. @@ -634,8 +710,11 @@ experimentsUpdateExperiment <- function(client, experiment_id, new_name = NULL) #' Updates the permissions on an experiment. Experiments can inherit permissions #' from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param experiment_id Required. The experiment for which to get or manage permissions. #' #' @rdname experimentsUpdatePermissions @@ -650,10 +729,12 @@ experimentsUpdatePermissions <- function(client, experiment_id, access_control_l #' #' Updates run metadata. #' @param client Required. Instance of DatabricksClient() + + #' #' @param end_time Unix timestamp in milliseconds of when the run ended. #' @param run_id ID of the run to update. -#' @param run_uuid [Deprecated, use run_id instead] ID of the run to update. +#' @param run_uuid Deprecated, use run_id instead. ID of the run to update. #' @param status Updated status of the run. #' #' @rdname experimentsUpdateRun diff --git a/R/external_locations.R b/R/external_locations.R index a4a3c4c6..de502617 100755 --- a/R/external_locations.R +++ b/R/external_locations.R @@ -9,6 +9,9 @@ NULL #' metastore admin or have the **CREATE_EXTERNAL_LOCATION** privilege on both #' the metastore and the associated storage credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param access_point The AWS access point to use when accesing s3 for this external location. #' @param comment User-provided free-form text description. @@ -34,6 +37,9 @@ externalLocationsCreate <- function(client, name, url, credential_name, access_p #' Deletes the specified external location from the metastore. The caller must #' be the owner of the external location. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param force Force deletion even if there are dependent external tables or mounts. #' @param name Required. Name of the external location. @@ -52,6 +58,9 @@ externalLocationsDelete <- function(client, name, force = NULL) { #' metastore admin, the owner of the external location, or a user that has some #' privilege on the external location. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_browse Whether to include external locations in the response for which the principal can only access selective metadata for. #' @param name Required. Name of the external location. @@ -72,6 +81,8 @@ externalLocationsGet <- function(client, name, include_browse = NULL) { #' location. There is no guarantee of a specific ordering of the elements in the #' array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param include_browse Whether to include external locations in the response for which the principal can only access selective metadata for. #' @param max_results Maximum number of external locations to return. @@ -108,6 +119,9 @@ externalLocationsList <- function(client, include_browse = NULL, max_results = N #' of the external location, or be a metastore admin. In the second case, the #' admin can only update the name of the external location. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param access_point The AWS access point to use when accesing s3 for this external location. #' @param comment User-provided free-form text description. diff --git a/R/files.R b/R/files.R index 9465fa59..9d851e31 100755 --- a/R/files.R +++ b/R/files.R @@ -10,6 +10,9 @@ NULL #' an existing directory, returns a success response; this method is idempotent #' (it will succeed if the directory already exists). #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param directory_path Required. The absolute path of a directory. #' @@ -24,6 +27,9 @@ filesCreateDirectory <- function(client, directory_path) { #' #' Deletes a file. If the request is successful, there is no response body. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param file_path Required. The absolute path of the file. #' @@ -42,6 +48,9 @@ filesDelete <- function(client, file_path) { #' be done by listing the directory contents and deleting each file and #' subdirectory recursively. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param directory_path Required. The absolute path of a directory. #' @@ -57,6 +66,9 @@ filesDeleteDirectory <- function(client, directory_path) { #' Downloads a file of up to 5 GiB. The file contents are the response body. #' This is a standard HTTP file download, not a JSON RPC. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param file_path Required. The absolute path of the file. #' @@ -79,6 +91,9 @@ filesDownload <- function(client, file_path) { #' will create the directory if it does not exist, and is idempotent (it will #' succeed if the directory already exists). #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param directory_path Required. The absolute path of a directory. #' @@ -94,6 +109,9 @@ filesGetDirectoryMetadata <- function(client, directory_path) { #' Get the metadata of a file. The response HTTP headers contain the metadata. #' There is no response body. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param file_path Required. The absolute path of the file. #' @@ -109,6 +127,9 @@ filesGetMetadata <- function(client, file_path) { #' Returns the contents of a directory. If there is no directory at the #' specified path, the API returns a HTTP 404 error. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param directory_path Required. The absolute path of a directory. #' @param page_size The maximum number of directory entries to return. @@ -148,8 +169,11 @@ filesListDirectoryContents <- function(client, directory_path, page_size = NULL, #' exactly the bytes sent in the request body. If the request is successful, #' there is no response body. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param contents +#' @param contents This field has no description yet. #' @param file_path Required. The absolute path of the file. #' @param overwrite If true, an existing file will be overwritten. #' diff --git a/R/functions.R b/R/functions.R index 72f4d145..2fad039e 100755 --- a/R/functions.R +++ b/R/functions.R @@ -11,6 +11,9 @@ NULL #' created: - **USE_CATALOG** on the function's parent catalog - **USE_SCHEMA** #' and **CREATE_FUNCTION** on the function's parent schema #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param function_info Required. Partial __FunctionInfo__ specifying the function to be created. #' @@ -31,6 +34,9 @@ functionsCreate <- function(client, function_info) { #' privilege on its parent catalog and the **USE_SCHEMA** privilege on its #' parent schema #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param force Force deletion even if the function is notempty. #' @param name Required. The fully-qualified name of the function (of the form __catalog_name__.__schema_name__.__function__name__). @@ -53,6 +59,9 @@ functionsDelete <- function(client, name, force = NULL) { #' catalog, the **USE_SCHEMA** privilege on the function's parent schema, and #' the **EXECUTE** privilege on the function itself #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_browse Whether to include functions in the response for which the principal can only access selective metadata for. #' @param name Required. The fully-qualified name of the function (of the form __catalog_name__.__schema_name__.__function__name__). @@ -75,6 +84,9 @@ functionsGet <- function(client, name, include_browse = NULL) { #' is the owner. There is no guarantee of a specific ordering of the elements in #' the array. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. Name of parent catalog for functions of interest. #' @param include_browse Whether to include functions in the response for which the principal can only access selective metadata for. @@ -119,6 +131,9 @@ functionsList <- function(client, catalog_name, schema_name, include_browse = NU #' privilege on its parent catalog as well as the **USE_SCHEMA** privilege on #' the function's parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The fully-qualified name of the function (of the form __catalog_name__.__schema_name__.__function__name__). #' @param owner Username of current owner of function. diff --git a/R/git_credentials.R b/R/git_credentials.R index 155fbb53..913435ed 100755 --- a/R/git_credentials.R +++ b/R/git_credentials.R @@ -10,6 +10,9 @@ NULL #' exists will fail. Use the PATCH endpoint to update existing credentials, or #' the DELETE endpoint to delete existing credentials. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param git_provider Required. Git provider. #' @param git_username Git username. @@ -26,6 +29,9 @@ gitCredentialsCreate <- function(client, git_provider, git_username = NULL, pers #' #' Deletes the specified Git credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param credential_id Required. The ID for the corresponding credential to access. #' @@ -40,6 +46,9 @@ gitCredentialsDelete <- function(client, credential_id) { #' #' Gets the Git credential with the specified credential ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param credential_id Required. The ID for the corresponding credential to access. #' @@ -53,7 +62,9 @@ gitCredentialsGet <- function(client, credential_id) { #' Get Git credentials. #' #' Lists the calling user's Git credentials. One credential per user is -#' supported.#' +#' supported. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname gitCredentialsList @@ -69,6 +80,9 @@ gitCredentialsList <- function(client) { #' #' Updates the specified Git credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param credential_id Required. The ID for the corresponding credential to access. #' @param git_provider Git provider. diff --git a/R/global_init_scripts.R b/R/global_init_scripts.R index 8c3bb83f..087a6cc6 100755 --- a/R/global_init_scripts.R +++ b/R/global_init_scripts.R @@ -7,6 +7,9 @@ NULL #' #' Creates a new global init script in this workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param enabled Specifies whether the script is enabled. #' @param name Required. The name of the script. @@ -24,6 +27,9 @@ globalInitScriptsCreate <- function(client, name, script, enabled = NULL, positi #' #' Deletes a global init script. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param script_id Required. The ID of the global init script. #' @@ -38,6 +44,9 @@ globalInitScriptsDelete <- function(client, script_id) { #' #' Gets all the details of a script, including its Base64-encoded contents. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param script_id Required. The ID of the global init script. #' @@ -53,7 +62,9 @@ globalInitScriptsGet <- function(client, script_id) { #' Get a list of all global init scripts for this workspace. This returns all #' properties for each script but **not** the script contents. To retrieve the #' contents of a script, use the [get a global init -#' script](:method:globalinitscripts/get) operation.#' +#' script](:method:globalinitscripts/get) operation. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname globalInitScriptsList @@ -70,6 +81,9 @@ globalInitScriptsList <- function(client) { #' Updates a global init script, specifying only the fields to change. All #' fields are optional. Unspecified fields retain their current value. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param enabled Specifies whether the script is enabled. #' @param name Required. The name of the script. diff --git a/R/grants.R b/R/grants.R index 07cd7ffa..df2a53b2 100755 --- a/R/grants.R +++ b/R/grants.R @@ -7,6 +7,9 @@ NULL #' #' Gets the permissions for a securable. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of securable. #' @param principal If provided, only the permissions for the specified principal (user or group) are returned. @@ -24,6 +27,9 @@ grantsGet <- function(client, securable_type, full_name, principal = NULL) { #' #' Gets the effective permissions for a securable. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of securable. #' @param principal If provided, only the effective permissions for the specified principal (user or group) are returned. @@ -41,6 +47,9 @@ grantsGetEffective <- function(client, securable_type, full_name, principal = NU #' #' Updates the permissions for a securable. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param changes Array of permissions change objects. #' @param full_name Required. Full name of securable. diff --git a/R/groups.R b/R/groups.R index 019bc6f7..f82e680f 100755 --- a/R/groups.R +++ b/R/groups.R @@ -8,13 +8,15 @@ NULL #' Creates a group in the Databricks workspace with a unique name, using the #' supplied group details. #' @param client Required. Instance of DatabricksClient() + + #' #' @param display_name String that represents a human-readable group name. #' @param entitlements Entitlements assigned to the group. -#' @param external_id -#' @param groups +#' @param external_id This field has no description yet. +#' @param groups This field has no description yet. #' @param id Databricks group ID. -#' @param members +#' @param members This field has no description yet. #' @param meta Container for the group identifier. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the group. @@ -33,6 +35,9 @@ groupsCreate <- function(client, display_name = NULL, entitlements = NULL, exter #' #' Deletes a group from the Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a group in the Databricks workspace. #' @@ -47,6 +52,9 @@ groupsDelete <- function(client, id) { #' #' Gets the information for a specific group in the Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a group in the Databricks workspace. #' @@ -61,6 +69,8 @@ groupsGet <- function(client, id) { #' #' Gets all details of the groups associated with the Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + #' #' @param attributes Comma-separated list of attributes to return in response. #' @param count Desired number of results per page. @@ -100,9 +110,12 @@ groupsList <- function(client, attributes = NULL, count = NULL, excluded_attribu #' #' Partially updates the details of a group. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a group in the Databricks workspace. -#' @param operations +#' @param operations This field has no description yet. #' @param schemas The schema of the patch request. #' #' @rdname groupsPatch @@ -116,13 +129,16 @@ groupsPatch <- function(client, id, operations = NULL, schemas = NULL) { #' #' Updates the details of a group by replacing the entire group entity. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param display_name String that represents a human-readable group name. #' @param entitlements Entitlements assigned to the group. -#' @param external_id -#' @param groups +#' @param external_id This field has no description yet. +#' @param groups This field has no description yet. #' @param id Databricks group ID. -#' @param members +#' @param members This field has no description yet. #' @param meta Container for the group identifier. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the group. diff --git a/R/instance_pools.R b/R/instance_pools.R index 594990b5..2401c050 100755 --- a/R/instance_pools.R +++ b/R/instance_pools.R @@ -7,6 +7,9 @@ NULL #' #' Creates a new instance pool using idle and ready-to-use cloud instances. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param aws_attributes Attributes related to instance pools running on Amazon Web Services. #' @param azure_attributes Attributes related to instance pools running on Azure. @@ -42,6 +45,9 @@ instancePoolsCreate <- function(client, instance_pool_name, node_type_id, aws_at #' Deletes the instance pool permanently. The idle instances in the pool are #' terminated asynchronously. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param instance_pool_id Required. The instance pool to be terminated. #' @@ -56,6 +62,9 @@ instancePoolsDelete <- function(client, instance_pool_id) { #' #' Modifies the configuration of an existing instance pool. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param custom_tags Additional tags for pool resources. #' @param idle_instance_autotermination_minutes Automatically terminates the extra instances in the pool cache after they are inactive for this time in minutes if min_idle_instances requirement is already met. @@ -80,6 +89,9 @@ instancePoolsEdit <- function(client, instance_pool_id, instance_pool_name, node #' #' Retrieve the information for an instance pool based on its identifier. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param instance_pool_id Required. The canonical unique identifier for the instance pool. #' @@ -94,6 +106,9 @@ instancePoolsGet <- function(client, instance_pool_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param instance_pool_id Required. The instance pool for which to get or manage permissions. #' @@ -110,6 +125,9 @@ instancePoolsGetPermissionLevels <- function(client, instance_pool_id) { #' Gets the permissions of an instance pool. Instance pools can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param instance_pool_id Required. The instance pool for which to get or manage permissions. #' @@ -123,7 +141,9 @@ instancePoolsGetPermissions <- function(client, instance_pool_id) { #' List instance pool info. #' -#' Gets a list of instance pools with their statistics.#' +#' Gets a list of instance pools with their statistics. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname instancePoolsList @@ -140,8 +160,11 @@ instancePoolsList <- function(client) { #' Sets permissions on an instance pool. Instance pools can inherit permissions #' from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param instance_pool_id Required. The instance pool for which to get or manage permissions. #' #' @rdname instancePoolsSetPermissions @@ -157,8 +180,11 @@ instancePoolsSetPermissions <- function(client, instance_pool_id, access_control #' Updates the permissions on an instance pool. Instance pools can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param instance_pool_id Required. The instance pool for which to get or manage permissions. #' #' @rdname instancePoolsUpdatePermissions diff --git a/R/instance_profiles.R b/R/instance_profiles.R index 823e9d74..ef905b30 100755 --- a/R/instance_profiles.R +++ b/R/instance_profiles.R @@ -8,6 +8,9 @@ NULL #' In the UI, you can select the instance profile when launching clusters. This #' API is only available to admin users. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param iam_role_arn The AWS IAM role ARN of the role associated with the instance profile. #' @param instance_profile_arn Required. The AWS ARN of the instance profile to register with Databricks. @@ -41,6 +44,9 @@ instanceProfilesAdd <- function(client, instance_profile_arn, iam_role_arn = NUL #' [Databricks SQL Serverless]: https://docs.databricks.com/sql/admin/serverless.html #' [Enable serverless SQL warehouses]: https://docs.databricks.com/sql/admin/serverless.html #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param iam_role_arn The AWS IAM role ARN of the role associated with the instance profile. #' @param instance_profile_arn Required. The AWS ARN of the instance profile to register with Databricks. @@ -59,7 +65,9 @@ instanceProfilesEdit <- function(client, instance_profile_arn, iam_role_arn = NU #' #' List the instance profiles that the calling user can use to launch a cluster. #' -#' This API is available to all users.#' +#' This API is available to all users. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname instanceProfilesList @@ -78,6 +86,9 @@ instanceProfilesList <- function(client) { #' #' This API is only accessible to admin users. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param instance_profile_arn Required. The ARN of the instance profile to remove. #' diff --git a/R/ip_access_lists.R b/R/ip_access_lists.R index 400ef046..f710ccfd 100755 --- a/R/ip_access_lists.R +++ b/R/ip_access_lists.R @@ -22,8 +22,11 @@ NULL #' IP access list has no effect until you enable the feature. See #' :method:workspaceconf/setStatus #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param ip_addresses +#' @param ip_addresses This field has no description yet. #' @param label Required. Label for the IP access list. #' @param list_type Required. Type of IP access list. #' @@ -38,6 +41,9 @@ ipAccessListsCreate <- function(client, label, list_type, ip_addresses = NULL) { #' #' Deletes an IP access list, specified by its list ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param ip_access_list_id Required. The ID for the corresponding IP access list. #' @@ -52,6 +58,9 @@ ipAccessListsDelete <- function(client, ip_access_list_id) { #' #' Gets an IP access list, specified by its list ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param ip_access_list_id Required. The ID for the corresponding IP access list. #' @@ -64,7 +73,9 @@ ipAccessListsGet <- function(client, ip_access_list_id) { #' Get access lists. #' -#' Gets all IP access lists for the specified workspace.#' +#' Gets all IP access lists for the specified workspace. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname ipAccessListsList @@ -91,10 +102,13 @@ ipAccessListsList <- function(client) { #' Note that your resulting IP access list has no effect until you enable the #' feature. See :method:workspaceconf/setStatus. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param enabled Required. Specifies whether this IP access list is enabled. #' @param ip_access_list_id Required. The ID for the corresponding IP access list. -#' @param ip_addresses +#' @param ip_addresses This field has no description yet. #' @param label Required. Label for the IP access list. #' @param list_type Required. Type of IP access list. #' @@ -127,10 +141,13 @@ ipAccessListsReplace <- function(client, ip_access_list_id, label, list_type, en #' resulting IP access list has no effect until you enable the feature. See #' :method:workspaceconf/setStatus. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param enabled Specifies whether this IP access list is enabled. #' @param ip_access_list_id Required. The ID for the corresponding IP access list. -#' @param ip_addresses +#' @param ip_addresses This field has no description yet. #' @param label Label for the IP access list. #' @param list_type Type of IP access list. #' diff --git a/R/jobs.R b/R/jobs.R index 7b9f39ae..3303f872 100755 --- a/R/jobs.R +++ b/R/jobs.R @@ -8,6 +8,8 @@ NULL #' Cancels all active runs of a job. The runs are canceled asynchronously, so it #' doesn't prevent new runs from being started. #' @param client Required. Instance of DatabricksClient() + + #' #' @param all_queued_runs Optional boolean parameter to cancel all queued runs. #' @param job_id The canonical identifier of the job to cancel all runs of. @@ -23,13 +25,18 @@ jobsCancelAllRuns <- function(client, all_queued_runs = NULL, job_id = NULL) { #' #' Cancels a job run or a task run. The run is canceled asynchronously, so it #' may still be running when this request completes. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Jobs on Databricks reach #' TERMINATED or SKIPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Jobs is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param run_id Required. This field is required. #' @@ -82,6 +89,8 @@ jobsCancelRun <- function(client, run_id, timeout = 20, callback = cli_reporter) #' #' Create a new job. #' @param client Required. Instance of DatabricksClient() + + #' #' @param access_control_list List of permissions to set on the job. #' @param compute A list of compute requirements that can be referenced by tasks of this job. @@ -129,6 +138,9 @@ jobsCreate <- function(client, access_control_list = NULL, compute = NULL, conti #' #' Deletes a job. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param job_id Required. The canonical identifier of the job to delete. #' @@ -143,6 +155,9 @@ jobsDelete <- function(client, job_id) { #' #' Deletes a non-active run. Returns an error if the run is active. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. The canonical identifier of the run for which to retrieve the metadata. #' @@ -157,6 +172,9 @@ jobsDeleteRun <- function(client, run_id) { #' #' Export and retrieve the job run task. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. The canonical identifier for the run. #' @param views_to_export Which views to export (CODE, DASHBOARDS, or ALL). @@ -172,6 +190,9 @@ jobsExportRun <- function(client, run_id, views_to_export = NULL) { #' #' Retrieves the details for a single job. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param job_id Required. The canonical identifier of the job to retrieve information about. #' @@ -186,6 +207,9 @@ jobsGet <- function(client, job_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param job_id Required. The job for which to get or manage permissions. #' @@ -202,6 +226,9 @@ jobsGetPermissionLevels <- function(client, job_id) { #' Gets the permissions of a job. Jobs can inherit permissions from their root #' object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param job_id Required. The job for which to get or manage permissions. #' @@ -215,13 +242,18 @@ jobsGetPermissions <- function(client, job_id) { #' Get a single job run. #' #' Retrieve the metadata of a run. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Jobs on Databricks reach #' TERMINATED or SKIPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Jobs is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param include_history Whether to include the repair history in the response. #' @param include_resolved_values Whether to include resolved parameter values in the response. @@ -287,6 +319,9 @@ jobsGetRun <- function(client, run_id, include_history = NULL, include_resolved_ #' automatically removed after 60 days. If you to want to reference them beyond #' 60 days, you must save old run results before they expire. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param run_id Required. The canonical identifier for the run. #' @@ -301,6 +336,8 @@ jobsGetRunOutput <- function(client, run_id) { #' #' Retrieves a list of jobs. #' @param client Required. Instance of DatabricksClient() + + #' #' @param expand_tasks Whether to include task and cluster details in the response. #' @param limit The number of jobs to return. @@ -338,6 +375,8 @@ jobsList <- function(client, expand_tasks = NULL, limit = NULL, name = NULL, off #' #' List runs in descending order by start time. #' @param client Required. Instance of DatabricksClient() + + #' #' @param active_only If active_only is `true`, only active runs are included in the results; otherwise, lists both active and completed runs. #' @param completed_only If completed_only is `true`, only completed runs are included in the results; otherwise, lists both active and completed runs. @@ -383,20 +422,25 @@ jobsListRuns <- function(client, active_only = NULL, completed_only = NULL, expa #' Re-run one or more tasks. Tasks are re-run as part of the original job run. #' They use the current job and task settings, and can be viewed in the history #' for the original job run. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Jobs on Databricks reach #' TERMINATED or SKIPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Jobs is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param dbt_commands An array of commands to execute for jobs with the dbt task, for example `'dbt_commands': ['dbt deps', 'dbt seed', 'dbt run']`. #' @param jar_params A list of parameters for jobs with Spark JAR tasks, for example `'jar_params': ['john doe', '35']`. #' @param job_parameters Job-level parameters used in the run. #' @param latest_repair_id The ID of the latest repair. #' @param notebook_params A map from keys to values for jobs with notebook task, for example `'notebook_params': {'name': 'john doe', 'age': '35'}`. -#' @param pipeline_params +#' @param pipeline_params This field has no description yet. #' @param python_named_params A map from keys to values for jobs with Python wheel task, for example `'python_named_params': {'name': 'task', 'data': 'dbfs:/path/to/data.json'}`. #' @param python_params A list of parameters for jobs with Python tasks, for example `'python_params': ['john doe', '35']`. #' @param rerun_all_failed_tasks If true, repair all failed tasks. @@ -465,6 +509,9 @@ jobsRepairRun <- function(client, run_id, dbt_commands = NULL, jar_params = NULL #' Overwrite all settings for the given job. Use the [_Update_ #' endpoint](:method:jobs/update) to update job settings partially. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param job_id Required. The canonical identifier of the job to reset. #' @param new_settings Required. The new settings of the job. @@ -479,13 +526,18 @@ jobsReset <- function(client, job_id, new_settings) { #' Trigger a new job run. #' #' Run a job and return the `run_id` of the triggered run. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Jobs on Databricks reach #' TERMINATED or SKIPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Jobs is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param dbt_commands An array of commands to execute for jobs with the dbt task, for example `'dbt_commands': ['dbt deps', 'dbt seed', 'dbt run']`. #' @param idempotency_token An optional token to guarantee the idempotency of job run requests. @@ -493,7 +545,7 @@ jobsReset <- function(client, job_id, new_settings) { #' @param job_id Required. The ID of the job to be executed. #' @param job_parameters Job-level parameters used in the run. #' @param notebook_params A map from keys to values for jobs with notebook task, for example `'notebook_params': {'name': 'john doe', 'age': '35'}`. -#' @param pipeline_params +#' @param pipeline_params This field has no description yet. #' @param python_named_params A map from keys to values for jobs with Python wheel task, for example `'python_named_params': {'name': 'task', 'data': 'dbfs:/path/to/data.json'}`. #' @param python_params A list of parameters for jobs with Python tasks, for example `'python_params': ['john doe', '35']`. #' @param queue The queue settings of the run. @@ -557,8 +609,11 @@ jobsRunNow <- function(client, job_id, dbt_commands = NULL, idempotency_token = #' Sets permissions on a job. Jobs can inherit permissions from their root #' object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param job_id Required. The job for which to get or manage permissions. #' #' @rdname jobsSetPermissions @@ -574,13 +629,17 @@ jobsSetPermissions <- function(client, job_id, access_control_list = NULL) { #' without creating a job. Runs submitted using this endpoint don’t display in #' the UI. Use the `jobs/runs/get` API to check the run state after the job is #' submitted. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Jobs on Databricks reach #' TERMINATED or SKIPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Jobs is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' #' @param access_control_list List of permissions to set on the job. #' @param email_notifications An optional set of email addresses notified when the run begins or completes. @@ -590,7 +649,7 @@ jobsSetPermissions <- function(client, job_id, access_control_list = NULL) { #' @param notification_settings Optional notification settings that are used when sending notifications to each of the `email_notifications` and `webhook_notifications` for this run. #' @param queue The queue settings of the one-time run. #' @param run_name An optional name for the run. -#' @param tasks +#' @param tasks This field has no description yet. #' @param timeout_seconds An optional timeout applied to each run of this job. #' @param webhook_notifications A collection of system notification IDs to notify when the run begins or completes. #' @@ -650,6 +709,9 @@ jobsSubmit <- function(client, access_control_list = NULL, email_notifications = #' Add, update, or remove specific settings of an existing job. Use the [_Reset_ #' endpoint](:method:jobs/reset) to overwrite all job settings. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param fields_to_remove Remove top-level fields in the job settings. #' @param job_id Required. The canonical identifier of the job to update. @@ -667,8 +729,11 @@ jobsUpdate <- function(client, job_id, fields_to_remove = NULL, new_settings = N #' Updates the permissions on a job. Jobs can inherit permissions from their #' root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param job_id Required. The job for which to get or manage permissions. #' #' @rdname jobsUpdatePermissions diff --git a/R/lakehouse_monitors.R b/R/lakehouse_monitors.R index 93580cc7..be7edc9e 100755 --- a/R/lakehouse_monitors.R +++ b/R/lakehouse_monitors.R @@ -16,6 +16,9 @@ NULL #' Additionally, the call must be made from the workspace where the monitor was #' created. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @param refresh_id Required. ID of the refresh. @@ -43,6 +46,9 @@ lakehouseMonitorsCancelRefresh <- function(client, full_name, refresh_id) { #' Workspace assets, such as the dashboard, will be created in the workspace #' where this call was made. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param assets_dir Required. The directory to store monitoring assets (e.g. #' @param baseline_table_name Name of the baseline table from which drift metrics are computed from. @@ -90,6 +96,9 @@ lakehouseMonitorsCreate <- function(client, full_name, assets_dir, output_schema #' Note that the metric tables and dashboard will not be deleted as part of this #' call; those assets must be manually cleaned up (if desired). #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -116,6 +125,9 @@ lakehouseMonitorsDelete <- function(client, full_name) { #' dashboard) may be filtered out if the caller is in a different workspace than #' where the monitor was created. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -140,6 +152,9 @@ lakehouseMonitorsGet <- function(client, full_name) { #' Additionally, the call must be made from the workspace where the monitor was #' created. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @param refresh_id Required. ID of the refresh. @@ -166,6 +181,9 @@ lakehouseMonitorsGetRefresh <- function(client, full_name, refresh_id) { #' Additionally, the call must be made from the workspace where the monitor was #' created. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -191,6 +209,9 @@ lakehouseMonitorsListRefreshes <- function(client, full_name) { #' Additionally, the call must be made from the workspace where the monitor was #' created. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -218,6 +239,9 @@ lakehouseMonitorsRunRefresh <- function(client, full_name) { #' Certain configuration fields, such as output asset identifiers, cannot be #' updated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param baseline_table_name Name of the baseline table from which drift metrics are computed from. #' @param custom_metrics Custom metrics to compute on the monitored table. diff --git a/R/lakeview.R b/R/lakeview.R index ca3e46df..a2e8a439 100755 --- a/R/lakeview.R +++ b/R/lakeview.R @@ -7,6 +7,9 @@ NULL #' #' Create a draft dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param display_name Required. The display name of the dashboard. #' @param parent_path The workspace path of the folder containing the dashboard. @@ -26,6 +29,9 @@ lakeviewCreate <- function(client, display_name, parent_path = NULL, serialized_ #' #' Get a draft dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. UUID identifying the dashboard. #' @@ -40,6 +46,9 @@ lakeviewGet <- function(client, dashboard_id) { #' #' Get the current published dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. UUID identifying the dashboard to be published. #' @@ -55,6 +64,9 @@ lakeviewGetPublished <- function(client, dashboard_id) { #' #' Publish the current draft dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. UUID identifying the dashboard to be published. #' @param embed_credentials Flag to indicate if the publisher's credentials should be embedded in the published dashboard. @@ -72,6 +84,9 @@ lakeviewPublish <- function(client, dashboard_id, embed_credentials = NULL, ware #' #' Trash a dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. UUID identifying the dashboard. #' @@ -86,6 +101,9 @@ lakeviewTrash <- function(client, dashboard_id) { #' #' Update a draft dashboard. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dashboard_id Required. UUID identifying the dashboard. #' @param display_name The display name of the dashboard. diff --git a/R/libraries.R b/R/libraries.R index 090a585c..b4efd7d7 100755 --- a/R/libraries.R +++ b/R/libraries.R @@ -8,7 +8,9 @@ NULL #' Get the status of all libraries on all clusters. A status will be available #' for all libraries installed on this cluster via the API or the libraries UI #' as well as libraries set to be installed on all clusters via the libraries -#' UI.#' +#' UI. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname librariesAllClusterStatuses #' @export librariesAllClusterStatuses <- function(client) { @@ -33,6 +35,9 @@ librariesAllClusterStatuses <- function(client) { #' clusters, but now marked for removal. Within this group there is no order #' guarantee. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. Unique identifier of the cluster whose status should be retrieved. #' @@ -57,6 +62,9 @@ librariesClusterStatus <- function(client, cluster_id) { #' union of the libraries specified via this method and the libraries set to be #' installed on all clusters via the libraries UI. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. Unique identifier for the cluster on which to install these libraries. #' @param libraries Required. The libraries to install. @@ -74,6 +82,9 @@ librariesInstall <- function(client, cluster_id, libraries) { #' uninstalled until the cluster is restarted. Uninstalling libraries that are #' not installed on the cluster will have no impact but is not an error. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cluster_id Required. Unique identifier for the cluster on which to uninstall these libraries. #' @param libraries Required. The libraries to uninstall. diff --git a/R/metastores.R b/R/metastores.R index 99c608ed..f1a0e8cd 100755 --- a/R/metastores.R +++ b/R/metastores.R @@ -9,6 +9,9 @@ NULL #' __workspace_id__ exists, it will be overwritten by the new __metastore_id__ #' and __default_catalog_name__. The caller must be an account admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param default_catalog_name Required. The name of the default catalog in the metastore. #' @param metastore_id Required. The unique ID of the metastore. @@ -30,6 +33,9 @@ metastoresAssign <- function(client, workspace_id, metastore_id, default_catalog #' field is set to the empty string (**''**), the ownership is assigned to the #' System User instead. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The user-specified name of the metastore. #' @param region Cloud region which the metastore serves (e.g., `us-west-2`, `westus`). @@ -44,7 +50,9 @@ metastoresCreate <- function(client, name, region = NULL, storage_root = NULL) { #' Get metastore assignment for workspace. #' -#' Gets the metastore assignment for the workspace being accessed.#' +#' Gets the metastore assignment for the workspace being accessed. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname metastoresCurrent #' @export metastoresCurrent <- function(client) { @@ -55,6 +63,9 @@ metastoresCurrent <- function(client) { #' #' Deletes a metastore. The caller must be a metastore admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param force Force deletion even if the metastore is not empty. #' @param id Required. Unique ID of the metastore. @@ -72,6 +83,9 @@ metastoresDelete <- function(client, id, force = NULL) { #' Gets a metastore that matches the supplied ID. The caller must be a metastore #' admin to retrieve this info. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID of the metastore. #' @@ -86,7 +100,9 @@ metastoresGet <- function(client, id) { #' #' Gets an array of the available metastores (as __MetastoreInfo__ objects). The #' caller must be an admin to retrieve this info. There is no guarantee of a -#' specific ordering of the elements in the array.#' +#' specific ordering of the elements in the array. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname metastoresList @@ -101,7 +117,9 @@ metastoresList <- function(client) { #' Get a metastore summary. #' #' Gets information about a metastore. This summary includes the storage -#' credential, the cloud vendor, the cloud region, and the global metastore ID.#' +#' credential, the cloud vendor, the cloud region, and the global metastore ID. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname metastoresSummary #' @export metastoresSummary <- function(client) { @@ -112,6 +130,9 @@ metastoresSummary <- function(client) { #' #' Deletes a metastore assignment. The caller must be an account administrator. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param metastore_id Required. Query for the ID of the metastore to delete. #' @param workspace_id Required. A workspace ID. @@ -130,6 +151,9 @@ metastoresUnassign <- function(client, workspace_id, metastore_id) { #' admin. If the __owner__ field is set to the empty string (**''**), the #' ownership is updated to the System User. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param delta_sharing_organization_name The organization name of a Delta Sharing entity, to be used in Databricks-to-Databricks Delta Sharing as the official name. #' @param delta_sharing_recipient_token_lifetime_in_seconds The lifetime of delta sharing recipient token in seconds. @@ -161,6 +185,9 @@ metastoresUpdate <- function(client, id, delta_sharing_organization_name = NULL, #' admin to update __metastore_id__; otherwise, the caller can be a Workspace #' admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param default_catalog_name The name of the default catalog for the metastore. #' @param metastore_id The unique ID of the metastore. diff --git a/R/model_registry.R b/R/model_registry.R index 6172d410..17165f40 100755 --- a/R/model_registry.R +++ b/R/model_registry.R @@ -7,6 +7,9 @@ NULL #' #' Approves a model version stage transition request. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param archive_existing_versions Required. Specifies whether to archive all current model versions in the target stage. #' @param comment User-provided comment on the action. @@ -29,6 +32,9 @@ modelRegistryApproveTransitionRequest <- function(client, name, version, stage, #' user or programmatically to display relevant information about the model. For #' example, test results or deployment errors. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment Required. User-provided comment on the action. #' @param name Required. Name of the model. @@ -48,6 +54,9 @@ modelRegistryCreateComment <- function(client, name, version, comment) { #' Throws `RESOURCE_ALREADY_EXISTS` if a registered model with the given name #' exists. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description Optional description for registered model. #' @param name Required. Register models under this name. @@ -64,6 +73,9 @@ modelRegistryCreateModel <- function(client, name, description = NULL, tags = NU #' #' Creates a model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description Optional description for model version. #' @param name Required. Register model under this name. @@ -85,6 +97,9 @@ modelRegistryCreateModelVersion <- function(client, name, source, description = #' #' Creates a model version stage transition request. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided comment on the action. #' @param name Required. Name of the model. @@ -104,11 +119,14 @@ modelRegistryCreateTransitionRequest <- function(client, name, version, stage, c #' #' Creates a registry webhook. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description User-specified description for the webhook. #' @param events Required. Events that can trigger a registry webhook: * `MODEL_VERSION_CREATED`: A new model version was created for the associated model. -#' @param http_url_spec -#' @param job_spec +#' @param http_url_spec This field has no description yet. +#' @param job_spec This field has no description yet. #' @param model_name Name of the model whose events would trigger this webhook. #' @param status Enable or disable triggering the webhook, or put the webhook into test mode. #' @@ -125,8 +143,11 @@ modelRegistryCreateWebhook <- function(client, events, description = NULL, http_ #' #' Deletes a comment on a model version. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param id Required. +#' +#' @param id Required. This field has no description yet. #' #' @rdname modelRegistryDeleteComment #' @export @@ -139,6 +160,9 @@ modelRegistryDeleteComment <- function(client, id) { #' #' Deletes a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Registered model unique name identifier. #' @@ -153,6 +177,9 @@ modelRegistryDeleteModel <- function(client, name) { #' #' Deletes the tag for a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param name Required. Name of the registered model that the tag was logged under. @@ -168,6 +195,9 @@ modelRegistryDeleteModelTag <- function(client, name, key) { #' #' Deletes a model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the registered model. #' @param version Required. Model version number. @@ -183,6 +213,9 @@ modelRegistryDeleteModelVersion <- function(client, name, version) { #' #' Deletes a model version tag. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param name Required. Name of the registered model that the tag was logged under. @@ -199,6 +232,9 @@ modelRegistryDeleteModelVersionTag <- function(client, name, version, key) { #' #' Cancels a model version stage transition request. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided comment on the action. #' @param creator Required. Username of the user who created this request. @@ -221,6 +257,8 @@ modelRegistryDeleteTransitionRequest <- function(client, name, version, stage, c #' #' Deletes a registry webhook. #' @param client Required. Instance of DatabricksClient() + + #' #' @param id Webhook ID required to delete a registry webhook. #' @@ -235,6 +273,9 @@ modelRegistryDeleteWebhook <- function(client, id = NULL) { #' #' Gets the latest version of a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Registered model unique name identifier. #' @param stages List of stages. @@ -255,11 +296,14 @@ modelRegistryGetLatestVersions <- function(client, name, stages = NULL) { #' Get model. #' #' Get the details of a model. This is a Databricks workspace version of the -#' [MLflow endpoint] that also returns the model's Databricks workspace ID and +#' MLflow endpoint that also returns the model's Databricks workspace ID and #' the permission level of the requesting user on the model. #' -#' [MLflow endpoint]: https://www.mlflow.org/docs/latest/rest-api.html#get-registeredmodel +#' MLflow endpoint: https://www.mlflow.org/docs/latest/rest-api.html#get-registeredmodel #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Registered model unique name identifier. #' @@ -274,6 +318,9 @@ modelRegistryGetModel <- function(client, name) { #' #' Get a model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the registered model. #' @param version Required. Model version number. @@ -289,6 +336,9 @@ modelRegistryGetModelVersion <- function(client, name, version) { #' #' Gets a URI to download the model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the registered model. #' @param version Required. Model version number. @@ -304,6 +354,9 @@ modelRegistryGetModelVersionDownloadUri <- function(client, name, version) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param registered_model_id Required. The registered model for which to get or manage permissions. #' @@ -320,6 +373,9 @@ modelRegistryGetPermissionLevels <- function(client, registered_model_id) { #' Gets the permissions of a registered model. Registered models can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param registered_model_id Required. The registered model for which to get or manage permissions. #' @@ -336,6 +392,8 @@ modelRegistryGetPermissions <- function(client, registered_model_id) { #' Lists all available registered models, up to the limit specified in #' __max_results__. #' @param client Required. Instance of DatabricksClient() + + #' #' @param max_results Maximum number of registered models desired. #' @param page_token Pagination token to go to the next page based on a previous query. @@ -368,6 +426,9 @@ modelRegistryListModels <- function(client, max_results = NULL, page_token = NUL #' #' Gets a list of all open stage transition requests for the model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the model. #' @param version Required. Version of the model. @@ -390,6 +451,8 @@ modelRegistryListTransitionRequests <- function(client, name, version) { #' #' Lists all registry webhooks. #' @param client Required. Instance of DatabricksClient() + + #' #' @param events If `events` is specified, any webhook with one or more of the specified trigger events is included in the output. #' @param model_name If not specified, all webhooks associated with the specified events are listed, regardless of their associated model. @@ -423,6 +486,9 @@ modelRegistryListWebhooks <- function(client, events = NULL, model_name = NULL, #' #' Rejects a model version stage transition request. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided comment on the action. #' @param name Required. Name of the model. @@ -440,6 +506,9 @@ modelRegistryRejectTransitionRequest <- function(client, name, version, stage, c #' #' Renames a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Registered model unique name identifier. #' @param new_name If provided, updates the name for this `registered_model`. @@ -455,6 +524,8 @@ modelRegistryRenameModel <- function(client, name, new_name = NULL) { #' #' Searches for specific model versions based on the supplied __filter__. #' @param client Required. Instance of DatabricksClient() + + #' #' @param filter String filter condition, like 'name='my-model-name''. #' @param max_results Maximum number of models desired. @@ -491,6 +562,8 @@ modelRegistrySearchModelVersions <- function(client, filter = NULL, max_results #' #' Search for registered models based on the specified __filter__. #' @param client Required. Instance of DatabricksClient() + + #' #' @param filter String filter condition, like 'name LIKE 'my-model-name''. #' @param max_results Maximum number of models desired. @@ -527,6 +600,9 @@ modelRegistrySearchModels <- function(client, filter = NULL, max_results = NULL, #' #' Sets a tag on a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param name Required. Unique name of the model. @@ -543,6 +619,9 @@ modelRegistrySetModelTag <- function(client, name, key, value) { #' #' Sets a model version tag. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the tag. #' @param name Required. Unique name of the model. @@ -561,8 +640,11 @@ modelRegistrySetModelVersionTag <- function(client, name, version, key, value) { #' Sets permissions on a registered model. Registered models can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param registered_model_id Required. The registered model for which to get or manage permissions. #' #' @rdname modelRegistrySetPermissions @@ -579,6 +661,9 @@ modelRegistrySetPermissions <- function(client, registered_model_id, access_cont #' #' Tests a registry webhook. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param event If `event` is specified, the test trigger uses the specified event. #' @param id Required. Webhook ID. @@ -593,11 +678,14 @@ modelRegistryTestRegistryWebhook <- function(client, id, event = NULL) { #' Transition a stage. #' #' Transition a model version's stage. This is a Databricks workspace version of -#' the [MLflow endpoint] that also accepts a comment associated with the +#' the MLflow endpoint that also accepts a comment associated with the #' transition to be recorded.', #' -#' [MLflow endpoint]: https://www.mlflow.org/docs/latest/rest-api.html#transition-modelversion-stage +#' MLflow endpoint: https://www.mlflow.org/docs/latest/rest-api.html#transition-modelversion-stage #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param archive_existing_versions Required. Specifies whether to archive all current model versions in the target stage. #' @param comment User-provided comment on the action. @@ -619,6 +707,9 @@ modelRegistryTransitionStage <- function(client, name, version, stage, archive_e #' #' Post an edit to a comment on a model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment Required. User-provided comment on the action. #' @param id Required. Unique identifier of an activity. @@ -634,6 +725,9 @@ modelRegistryUpdateComment <- function(client, id, comment) { #' #' Updates a registered model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description If provided, updates the description for this `registered_model`. #' @param name Required. Registered model unique name identifier. @@ -649,6 +743,9 @@ modelRegistryUpdateModel <- function(client, name, description = NULL) { #' #' Updates the model version. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description If provided, updates the description for this `registered_model`. #' @param name Required. Name of the registered model. @@ -666,8 +763,11 @@ modelRegistryUpdateModelVersion <- function(client, name, version, description = #' Updates the permissions on a registered model. Registered models can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param registered_model_id Required. The registered model for which to get or manage permissions. #' #' @rdname modelRegistryUpdatePermissions @@ -684,12 +784,15 @@ modelRegistryUpdatePermissions <- function(client, registered_model_id, access_c #' #' Updates a registry webhook. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description User-specified description for the webhook. #' @param events Events that can trigger a registry webhook: * `MODEL_VERSION_CREATED`: A new model version was created for the associated model. -#' @param http_url_spec +#' @param http_url_spec This field has no description yet. #' @param id Required. Webhook ID. -#' @param job_spec +#' @param job_spec This field has no description yet. #' @param status Enable or disable triggering the webhook, or put the webhook into test mode. #' #' @rdname modelRegistryUpdateWebhook diff --git a/R/model_versions.R b/R/model_versions.R index ce882484..95d55b2c 100755 --- a/R/model_versions.R +++ b/R/model_versions.R @@ -13,6 +13,9 @@ NULL #' **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** #' privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. The three-level (fully qualified) name of the model version. #' @param version Required. The integer version number of the model version. @@ -34,6 +37,9 @@ modelVersionsDelete <- function(client, full_name, version) { #' must also be the owner or have the **USE_CATALOG** privilege on the parent #' catalog and the **USE_SCHEMA** privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. The three-level (fully qualified) name of the model version. #' @param include_browse Whether to include model versions in the response for which the principal can only access selective metadata for. @@ -56,6 +62,9 @@ modelVersionsGet <- function(client, full_name, version, include_browse = NULL) #' be the owner or have the **USE_CATALOG** privilege on the parent catalog and #' the **USE_SCHEMA** privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param alias Required. The name of the alias. #' @param full_name Required. The three-level (fully qualified) name of the registered model. @@ -83,6 +92,9 @@ modelVersionsGetByAlias <- function(client, full_name, alias) { #' #' There is no guarantee of a specific ordering of the elements in the response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. The full three-level name of the registered model under which to list model versions. #' @param include_browse Whether to include model versions in the response for which the principal can only access selective metadata for. @@ -126,6 +138,9 @@ modelVersionsList <- function(client, full_name, include_browse = NULL, max_resu #' #' Currently only the comment of the model version can be updated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment The comment attached to the model version. #' @param full_name Required. The three-level (fully qualified) name of the model version. diff --git a/R/online_tables.R b/R/online_tables.R index e1b69f6f..c185de84 100755 --- a/R/online_tables.R +++ b/R/online_tables.R @@ -7,6 +7,8 @@ NULL #' #' Create a new Online Table. #' @param client Required. Instance of DatabricksClient() + + #' #' @param name Full three-part (catalog, schema, table) name of the table. #' @param spec Specification of the online table. @@ -24,6 +26,9 @@ onlineTablesCreate <- function(client, name = NULL, spec = NULL) { #' table. If the source Delta table was deleted or modified since this Online #' Table was created, this will lose the data forever! #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Full three-part (catalog, schema, table) name of the table. #' @@ -38,6 +43,9 @@ onlineTablesDelete <- function(client, name) { #' #' Get information about an existing online table and its status. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Full three-part (catalog, schema, table) name of the table. #' diff --git a/R/permission_migration.R b/R/permission_migration.R index 423c4d6a..0c865b2d 100755 --- a/R/permission_migration.R +++ b/R/permission_migration.R @@ -8,6 +8,9 @@ NULL #' Migrate a batch of permissions from a workspace local group to an account #' group. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param from_workspace_group_name Required. The name of the workspace group that permissions will be migrated from. #' @param size The maximum number of permissions that will be migrated. diff --git a/R/permissions.R b/R/permissions.R index 93336a3c..820d0a2a 100755 --- a/R/permissions.R +++ b/R/permissions.R @@ -8,6 +8,9 @@ NULL #' Gets the permissions of an object. Objects can inherit permissions from their #' parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param request_object_id Required. The id of the request object. #' @param request_object_type Required. The type of the request object. @@ -24,6 +27,9 @@ permissionsGet <- function(client, request_object_type, request_object_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param request_object_id Required. . #' @param request_object_type Required. . @@ -41,8 +47,11 @@ permissionsGetPermissionLevels <- function(client, request_object_type, request_ #' Sets permissions on an object. Objects can inherit permissions from their #' parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param request_object_id Required. The id of the request object. #' @param request_object_type Required. The type of the request object. #' @@ -59,8 +68,11 @@ permissionsSet <- function(client, request_object_type, request_object_id, acces #' Updates the permissions on an object. Objects can inherit permissions from #' their parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param request_object_id Required. The id of the request object. #' @param request_object_type Required. The type of the request object. #' diff --git a/R/pipelines.R b/R/pipelines.R index 098df59f..890224dd 100755 --- a/R/pipelines.R +++ b/R/pipelines.R @@ -8,6 +8,8 @@ NULL #' Creates a new data processing pipeline based on the requested configuration. #' If successful, this method returns the ID of the new pipeline. #' @param client Required. Instance of DatabricksClient() + + #' #' @param allow_duplicate_names If false, deployment will fail if name conflicts with that of another pipeline. #' @param catalog A catalog in Unity Catalog to publish data from this pipeline to. @@ -16,7 +18,7 @@ NULL #' @param configuration String-String configuration for this pipeline execution. #' @param continuous Whether the pipeline is continuous or triggered. #' @param development Whether the pipeline is in Development mode. -#' @param dry_run +#' @param dry_run This field has no description yet. #' @param edition Pipeline product edition. #' @param filters Filters on which Pipeline packages to include in the deployed graph. #' @param id Unique identifier for this pipeline. @@ -49,8 +51,11 @@ pipelinesCreate <- function(client, allow_duplicate_names = NULL, catalog = NULL #' #' Deletes a pipeline. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param pipeline_id Required. +#' @param pipeline_id Required. This field has no description yet. #' #' @rdname pipelinesDelete #' @export @@ -61,8 +66,11 @@ pipelinesDelete <- function(client, pipeline_id) { #' Get a pipeline. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param pipeline_id Required. +#' @param pipeline_id Required. This field has no description yet. #' #' @rdname pipelinesGet #' @export @@ -75,6 +83,9 @@ pipelinesGet <- function(client, pipeline_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param pipeline_id Required. The pipeline for which to get or manage permissions. #' @@ -91,6 +102,9 @@ pipelinesGetPermissionLevels <- function(client, pipeline_id) { #' Gets the permissions of a pipeline. Pipelines can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param pipeline_id Required. The pipeline for which to get or manage permissions. #' @@ -105,6 +119,9 @@ pipelinesGetPermissions <- function(client, pipeline_id) { #' #' Gets an update from an active pipeline. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param pipeline_id Required. The ID of the pipeline. #' @param update_id Required. The ID of the update. @@ -121,12 +138,15 @@ pipelinesGetUpdate <- function(client, pipeline_id, update_id) { #' #' Retrieves events for a pipeline. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param filter Criteria to select a subset of results, expressed using a SQL-like syntax. #' @param max_results Max number of entries to return in a single page. -#' @param order_by A string indicating a sort order by timestamp for the results, for example, ['timestamp asc']. +#' @param order_by A string indicating a sort order by timestamp for the results, for example, timestamp asc. #' @param page_token Page token returned by previous call. -#' @param pipeline_id Required. +#' @param pipeline_id Required. This field has no description yet. #' #' @return `data.frame` with all of the response pages. #' @@ -159,6 +179,8 @@ pipelinesListPipelineEvents <- function(client, pipeline_id, filter = NULL, max_ #' #' Lists pipelines defined in the Delta Live Tables system. #' @param client Required. Instance of DatabricksClient() + + #' #' @param filter Select a subset of results based on the specified criteria. #' @param max_results The maximum number of entries to return in a single page. @@ -195,6 +217,9 @@ pipelinesListPipelines <- function(client, filter = NULL, max_results = NULL, or #' #' List updates for an active pipeline. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param max_results Max number of entries to return in a single page. #' @param page_token Page token returned by previous call. @@ -215,8 +240,11 @@ pipelinesListUpdates <- function(client, pipeline_id, max_results = NULL, page_t #' Sets permissions on a pipeline. Pipelines can inherit permissions from their #' root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param pipeline_id Required. The pipeline for which to get or manage permissions. #' #' @rdname pipelinesSetPermissions @@ -233,11 +261,14 @@ pipelinesSetPermissions <- function(client, pipeline_id, access_control_list = N #' for the pipeline, the request will fail and the active update will remain #' running. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param cause +#' +#' @param cause This field has no description yet. #' @param full_refresh If true, this update will reset all tables before running. #' @param full_refresh_selection A list of tables to update with fullRefresh. -#' @param pipeline_id Required. +#' @param pipeline_id Required. This field has no description yet. #' @param refresh_selection A list of tables to update without fullRefresh. #' @param validate_only If true, this update only validates the correctness of pipeline source code but does not materialize or publish any datasets. #' @@ -255,15 +286,20 @@ pipelinesStartUpdate <- function(client, pipeline_id, cause = NULL, full_refresh #' #' Stops the pipeline by canceling the active update. If there is no active #' update for the pipeline, this request is a no-op. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Pipelines on Databricks reach #' IDLE state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Pipelines is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' -#' @param pipeline_id Required. +#' +#' @param pipeline_id Required. This field has no description yet. #' #' @rdname pipelinesStop #' @export @@ -311,6 +347,9 @@ pipelinesStop <- function(client, pipeline_id, timeout = 20, callback = cli_repo #' #' Updates a pipeline with the supplied configuration. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_duplicate_names If false, deployment will fail if name has changed and conflicts the name of another pipeline. #' @param catalog A catalog in Unity Catalog to publish data from this pipeline to. @@ -354,8 +393,11 @@ pipelinesUpdate <- function(client, pipeline_id, allow_duplicate_names = NULL, c #' Updates the permissions on a pipeline. Pipelines can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param pipeline_id Required. The pipeline for which to get or manage permissions. #' #' @rdname pipelinesUpdatePermissions diff --git a/R/policy_families.R b/R/policy_families.R index f5ab1d8e..8d3983c3 100755 --- a/R/policy_families.R +++ b/R/policy_families.R @@ -7,8 +7,11 @@ NULL #' #' Retrieve the information for an policy family based on its identifier. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param policy_family_id Required. +#' @param policy_family_id Required. This field has no description yet. #' #' @rdname policyFamiliesGet #' @export @@ -21,6 +24,8 @@ policyFamiliesGet <- function(client, policy_family_id) { #' #' Retrieve a list of policy families. This API is paginated. #' @param client Required. Instance of DatabricksClient() + + #' #' @param max_results The max number of policy families to return. #' @param page_token A token that can be used to get the next page of results. diff --git a/R/providers.R b/R/providers.R index aeb768ca..6da2e8f6 100755 --- a/R/providers.R +++ b/R/providers.R @@ -8,6 +8,9 @@ NULL #' Creates a new authentication provider minimally based on a name and #' authentication type. The caller must be an admin on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param authentication_type Required. The delta sharing authentication type. #' @param comment Description about the provider. @@ -27,6 +30,9 @@ providersCreate <- function(client, name, authentication_type, comment = NULL, r #' Deletes an authentication provider, if the caller is a metastore admin or is #' the owner of the provider. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the provider. #' @@ -43,6 +49,9 @@ providersDelete <- function(client, name) { #' the provider, and must either be a metastore admin or the owner of the #' provider. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the provider. #' @@ -60,6 +69,8 @@ providersGet <- function(client, name) { #' the caller are not included in the response. There is no guarantee of a #' specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param data_provider_global_metastore_id If not provided, all providers will be returned. #' @@ -81,6 +92,9 @@ providersList <- function(client, data_provider_global_metastore_id = NULL) { #' #' * the caller is a metastore admin, or * the caller is the owner. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the provider in which to list shares. #' @@ -104,6 +118,9 @@ providersListShares <- function(client, name) { #' provider name, the caller must be both a metastore admin and the owner of the #' provider. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment Description about the provider. #' @param name Required. Name of the provider. diff --git a/R/queries.R b/R/queries.R index d3acc412..a028589b 100755 --- a/R/queries.R +++ b/R/queries.R @@ -15,6 +15,8 @@ NULL #' #' **Note**: You cannot add a visualization until you create the query. #' @param client Required. Instance of DatabricksClient() + + #' #' @param data_source_id Data source ID maps to the ID of the data source used by the resource and is distinct from the warehouse ID. #' @param description General description that conveys additional information about this query such as usage notes. @@ -39,8 +41,11 @@ queriesCreate <- function(client, data_source_id = NULL, description = NULL, nam #' searches and list views, and they cannot be used for alerts. The trash is #' deleted after 30 days. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param query_id Required. +#' +#' @param query_id Required. This field has no description yet. #' #' @rdname queriesDelete #' @export @@ -54,8 +59,11 @@ queriesDelete <- function(client, query_id) { #' Retrieve a query object definition along with contextual permissions #' information about the currently authenticated user. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param query_id Required. +#' @param query_id Required. This field has no description yet. #' #' @rdname queriesGet #' @export @@ -72,6 +80,8 @@ queriesGet <- function(client, query_id) { #' ### **Warning: Calling this API concurrently 10 or more times could result in #' throttling, service degradation, or a temporary ban.** #' @param client Required. Instance of DatabricksClient() + + #' #' @param order Name of query attribute to order by. #' @param page Page number to retrieve. @@ -107,8 +117,11 @@ queriesList <- function(client, order = NULL, page = NULL, page_size = NULL, q = #' Restore a query that has been moved to the trash. A restored query appears in #' list views and searches. You can use restored queries for alerts. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param query_id Required. +#' +#' @param query_id Required. This field has no description yet. #' #' @rdname queriesRestore #' @export @@ -123,13 +136,16 @@ queriesRestore <- function(client, query_id) { #' #' **Note**: You cannot undo this operation. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param data_source_id Data source ID maps to the ID of the data source used by the resource and is distinct from the warehouse ID. #' @param description General description that conveys additional information about this query such as usage notes. #' @param name The title of this query that appears in list views, widget headings, and on the query page. #' @param options Exclusively used for storing a list parameter definitions. #' @param query The text of the query to be run. -#' @param query_id Required. +#' @param query_id Required. This field has no description yet. #' @param run_as_role Sets the **Run as** role for the object. #' #' @rdname queriesUpdate diff --git a/R/query_history.R b/R/query_history.R index 6fc1a687..eb3445a7 100755 --- a/R/query_history.R +++ b/R/query_history.R @@ -9,6 +9,8 @@ NULL #' #' You can filter by user ID, warehouse ID, status, and time range. #' @param client Required. Instance of DatabricksClient() + + #' #' @param filter_by A filter to limit query history results. #' @param include_metrics Whether to include metrics about query. diff --git a/R/query_visualizations.R b/R/query_visualizations.R index 2f5a817d..241b9cc8 100755 --- a/R/query_visualizations.R +++ b/R/query_visualizations.R @@ -5,6 +5,9 @@ NULL #' Add visualization to a query. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param description A short description of this visualization. #' @param name The name of the visualization that appears on dashboards and the query screen. @@ -23,6 +26,9 @@ queryVisualizationsCreate <- function(client, query_id, type, options, descripti #' Remove visualization. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Widget ID returned by :method:queryvizualisations/create. #' @@ -35,14 +41,17 @@ queryVisualizationsDelete <- function(client, id) { #' Edit existing visualization. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param created_at +#' @param created_at This field has no description yet. #' @param description A short description of this visualization. #' @param id The UUID for this visualization. #' @param name The name of the visualization that appears on dashboards and the query screen. #' @param options The options object varies widely from one visualization type to the next and is unsupported. #' @param type The type of visualization: chart, table, pivot table, and so on. -#' @param updated_at +#' @param updated_at This field has no description yet. #' #' @rdname queryVisualizationsUpdate #' @export diff --git a/R/recipient_activation.R b/R/recipient_activation.R index a6103ea6..82856107 100755 --- a/R/recipient_activation.R +++ b/R/recipient_activation.R @@ -7,6 +7,9 @@ NULL #' #' Gets an activation URL for a share. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param activation_url Required. The one time activation url. #' @@ -23,6 +26,9 @@ recipientActivationGetActivationUrlInfo <- function(client, activation_url) { #' Retrieve access token with an activation url. This is a public API without #' any authentication. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param activation_url Required. The one time activation url. #' diff --git a/R/recipients.R b/R/recipients.R index 8b82d9c2..3aac12af 100755 --- a/R/recipients.R +++ b/R/recipients.R @@ -9,6 +9,9 @@ NULL #' metastore. The caller must be a metastore admin or has the #' **CREATE_RECIPIENT** privilege on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param authentication_type Required. The delta sharing authentication type. #' @param comment Description about the recipient. @@ -34,6 +37,9 @@ recipientsCreate <- function(client, name, authentication_type, comment = NULL, #' Deletes the specified recipient from the metastore. The caller must be the #' owner of the recipient. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the recipient. #' @@ -50,6 +56,9 @@ recipientsDelete <- function(client, name) { #' #' * the caller is the owner of the share recipient, or: * is a metastore admin #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the recipient. #' @@ -67,6 +76,8 @@ recipientsGet <- function(client, name) { #' * the caller is a metastore admin, or * the caller is the owner. There is no #' guarantee of a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param data_recipient_global_metastore_id If not provided, all recipients will be returned. #' @@ -87,6 +98,9 @@ recipientsList <- function(client, data_recipient_global_metastore_id = NULL) { #' Refreshes the specified recipient's delta sharing authentication token with #' the provided token info. The caller must be the owner of the recipient. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param existing_token_expire_in_seconds Required. The expiration time of the bearer token in ISO 8601 format. #' @param name Required. The name of the recipient. @@ -104,6 +118,9 @@ recipientsRotateToken <- function(client, name, existing_token_expire_in_seconds #' Gets the share permissions for the specified Recipient. The caller must be a #' metastore admin or the owner of the Recipient. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the Recipient. #' @@ -122,6 +139,9 @@ recipientsSharePermissions <- function(client, name) { #' updated, the user must be both a metastore admin and the owner of the #' recipient. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment Description about the recipient. #' @param ip_access_list IP Access List. diff --git a/R/registered_models.R b/R/registered_models.R index 759044c0..f1d4e74f 100755 --- a/R/registered_models.R +++ b/R/registered_models.R @@ -18,6 +18,9 @@ NULL #' caller must have the **CREATE MODEL** or **CREATE FUNCTION** privilege on the #' parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. The name of the catalog where the schema and the registered model reside. #' @param comment The comment attached to the registered model. @@ -44,6 +47,9 @@ registeredModelsCreate <- function(client, catalog_name, schema_name, name, comm #' **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** #' privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. The three-level (fully qualified) name of the registered model. #' @@ -63,6 +69,9 @@ registeredModelsDelete <- function(client, full_name) { #' **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** #' privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param alias Required. The name of the alias. #' @param full_name Required. The three-level (fully qualified) name of the registered model. @@ -84,6 +93,9 @@ registeredModelsDeleteAlias <- function(client, full_name, alias) { #' be the owner or have the **USE_CATALOG** privilege on the parent catalog and #' the **USE_SCHEMA** privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. The three-level (fully qualified) name of the registered model. #' @param include_browse Whether to include registered models in the response for which the principal can only access selective metadata for. @@ -111,6 +123,8 @@ registeredModelsGet <- function(client, full_name, include_browse = NULL) { #' #' There is no guarantee of a specific ordering of the elements in the response. #' @param client Required. Instance of DatabricksClient() + + #' #' @param catalog_name The identifier of the catalog under which to list registered models. #' @param include_browse Whether to include registered models in the response for which the principal can only access selective metadata for. @@ -153,6 +167,9 @@ registeredModelsList <- function(client, catalog_name = NULL, include_browse = N #' **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** #' privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param alias Required. The name of the alias. #' @param full_name Required. Full name of the registered model. @@ -178,6 +195,9 @@ registeredModelsSetAlias <- function(client, full_name, alias, version_num) { #' Currently only the name, the owner or the comment of the registered model can #' be updated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment The comment attached to the registered model. #' @param full_name Required. The three-level (fully qualified) name of the registered model. diff --git a/R/repos.R b/R/repos.R index 2e1e00ff..419f7c09 100755 --- a/R/repos.R +++ b/R/repos.R @@ -9,6 +9,9 @@ NULL #' specified. Note that repos created programmatically must be linked to a #' remote Git repo, unlike repos created in the browser. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Desired path for the repo in the workspace. #' @param provider Required. Git provider. @@ -27,6 +30,9 @@ reposCreate <- function(client, url, provider, path = NULL, sparse_checkout = NU #' #' Deletes the specified repo. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param repo_id Required. The ID for the corresponding repo to access. #' @@ -41,6 +47,9 @@ reposDelete <- function(client, repo_id) { #' #' Returns the repo with the given repo ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param repo_id Required. The ID for the corresponding repo to access. #' @@ -55,6 +64,9 @@ reposGet <- function(client, repo_id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param repo_id Required. The repo for which to get or manage permissions. #' @@ -71,6 +83,9 @@ reposGetPermissionLevels <- function(client, repo_id) { #' Gets the permissions of a repo. Repos can inherit permissions from their root #' object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param repo_id Required. The repo for which to get or manage permissions. #' @@ -86,6 +101,8 @@ reposGetPermissions <- function(client, repo_id) { #' Returns repos that the calling user has Manage permissions on. Results are #' paginated with each page containing twenty repos. #' @param client Required. Instance of DatabricksClient() + + #' #' @param next_page_token Token used to get the next page of results. #' @param path_prefix Filters repos that have paths starting with the given path prefix. @@ -119,8 +136,11 @@ reposList <- function(client, next_page_token = NULL, path_prefix = NULL) { #' Sets permissions on a repo. Repos can inherit permissions from their root #' object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param repo_id Required. The repo for which to get or manage permissions. #' #' @rdname reposSetPermissions @@ -135,6 +155,9 @@ reposSetPermissions <- function(client, repo_id, access_control_list = NULL) { #' Updates the repo to a different branch or tag, or updates the repo to the #' latest commit on the same branch. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param branch Branch that the local version of the repo is checked out to. #' @param repo_id Required. The ID for the corresponding repo to access. @@ -153,8 +176,11 @@ reposUpdate <- function(client, repo_id, branch = NULL, sparse_checkout = NULL, #' Updates the permissions on a repo. Repos can inherit permissions from their #' root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param repo_id Required. The repo for which to get or manage permissions. #' #' @rdname reposUpdatePermissions diff --git a/R/restrict_workspace_admins.R b/R/restrict_workspace_admins.R index 52419bd3..dd7ab990 100755 --- a/R/restrict_workspace_admins.R +++ b/R/restrict_workspace_admins.R @@ -12,6 +12,8 @@ NULL #' 409 and the request must be retried by using the fresh etag in the 409 #' response. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -27,6 +29,8 @@ restrictWorkspaceAdminsDelete <- function(client, etag = NULL) { #' #' Gets the restrict workspace admins setting. #' @param client Required. Instance of DatabricksClient() + + #' #' @param etag etag used for versioning. #' @@ -46,10 +50,13 @@ restrictWorkspaceAdminsGet <- function(client, etag = NULL) { #' the setting is updated concurrently, `PATCH` fails with 409 and the request #' must be retried by using the fresh etag in the 409 response. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param allow_missing Required. This should always be set to true for Settings API. #' @param field_mask Required. Field mask is required to be passed into the PATCH request. -#' @param setting Required. +#' @param setting Required. This field has no description yet. #' #' @rdname restrictWorkspaceAdminsUpdate #' @export diff --git a/R/schemas.R b/R/schemas.R index 85e50680..bf0b997b 100755 --- a/R/schemas.R +++ b/R/schemas.R @@ -9,6 +9,9 @@ NULL #' metastore admin, or have the **CREATE_SCHEMA** privilege in the parent #' catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. Name of parent catalog. #' @param comment User-provided free-form text description. @@ -30,6 +33,9 @@ schemasCreate <- function(client, name, catalog_name, comment = NULL, properties #' Deletes the specified schema from the parent catalog. The caller must be the #' owner of the schema or an owner of the parent catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the schema. #' @@ -46,6 +52,9 @@ schemasDelete <- function(client, full_name) { #' metastore admin, the owner of the schema, or a user that has the #' **USE_SCHEMA** privilege on the schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the schema. #' @param include_browse Whether to include schemas in the response for which the principal can only access selective metadata for. @@ -66,6 +75,9 @@ schemasGet <- function(client, full_name, include_browse = NULL) { #' for which the caller has the **USE_SCHEMA** privilege) will be retrieved. #' There is no guarantee of a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. Parent catalog for schemas of interest. #' @param include_browse Whether to include schemas in the response for which the principal can only access selective metadata for. @@ -106,6 +118,9 @@ schemasList <- function(client, catalog_name, include_browse = NULL, max_results #' the caller must be a metastore admin or have the **CREATE_SCHEMA** privilege #' on the parent catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param enable_predictive_optimization Whether predictive optimization should be enabled for this object and objects under it. diff --git a/R/secrets.R b/R/secrets.R index 87f627d0..0c143e97 100755 --- a/R/secrets.R +++ b/R/secrets.R @@ -8,6 +8,9 @@ NULL #' The scope name must consist of alphanumeric characters, dashes, underscores, #' and periods, and may not exceed 128 characters. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param backend_azure_keyvault The metadata for the secret scope if the type is `AZURE_KEYVAULT`. #' @param initial_manage_principal The principal that is initially granted `MANAGE` permission to the created scope. @@ -32,6 +35,9 @@ secretsCreateScope <- function(client, scope, backend_azure_keyvault = NULL, ini #' Throws `PERMISSION_DENIED` if the user does not have permission to make this #' API call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param principal Required. The principal to remove an existing ACL from. #' @param scope Required. The name of the scope to remove permissions from. @@ -51,6 +57,9 @@ secretsDeleteAcl <- function(client, scope, principal) { #' `PERMISSION_DENIED` if the user does not have permission to make this API #' call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param scope Required. Name of the scope to delete. #' @@ -70,6 +79,9 @@ secretsDeleteScope <- function(client, scope) { #' Throws `PERMISSION_DENIED` if the user does not have permission to make this #' API call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. Name of the secret to delete. #' @param scope Required. The name of the scope that contains the secret to delete. @@ -90,6 +102,9 @@ secretsDeleteSecret <- function(client, scope, key) { #' `PERMISSION_DENIED` if the user does not have permission to make this API #' call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param principal Required. The principal to fetch ACL information for. #' @param scope Required. The name of the scope to fetch ACL information from. @@ -116,6 +131,9 @@ secretsGetAcl <- function(client, scope, principal) { #' this API call. Throws ``RESOURCE_DOES_NOT_EXIST`` if no such secret or secret #' scope exists. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param key Required. The key to fetch secret for. #' @param scope Required. The name of the scope to fetch secret information from. @@ -136,6 +154,9 @@ secretsGetSecret <- function(client, scope, key) { #' `PERMISSION_DENIED` if the user does not have permission to make this API #' call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param scope Required. The name of the scope to fetch ACL information from. #' @@ -156,7 +177,9 @@ secretsListAcls <- function(client, scope) { #' Lists all secret scopes available in the workspace. #' #' Throws `PERMISSION_DENIED` if the user does not have permission to make this -#' API call.#' +#' API call. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname secretsListScopes @@ -179,6 +202,9 @@ secretsListScopes <- function(client) { #' `PERMISSION_DENIED` if the user does not have permission to make this API #' call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param scope Required. The name of the scope to list secrets within. #' @@ -223,6 +249,9 @@ secretsListSecrets <- function(client, scope) { #' Throws `PERMISSION_DENIED` if the user does not have permission to make this #' API call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param permission Required. The permission level applied to the principal. #' @param principal Required. The principal in which the permission is applied. @@ -257,6 +286,9 @@ secretsPutAcl <- function(client, scope, principal, permission) { #' Throws `PERMISSION_DENIED` if the user does not have permission to make this #' API call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param bytes_value If specified, value will be stored as bytes. #' @param key Required. A unique name to identify the secret. diff --git a/R/service_principals.R b/R/service_principals.R index 68413370..d78fa8e7 100755 --- a/R/service_principals.R +++ b/R/service_principals.R @@ -7,13 +7,15 @@ NULL #' #' Creates a new service principal in the Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + #' #' @param active If this user is active. #' @param application_id UUID relating to the service principal. #' @param display_name String that represents a concatenation of given and family names. #' @param entitlements Entitlements assigned to the service principal. -#' @param external_id -#' @param groups +#' @param external_id This field has no description yet. +#' @param groups This field has no description yet. #' @param id Databricks service principal ID. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the List response. @@ -33,6 +35,9 @@ servicePrincipalsCreate <- function(client, active = NULL, application_id = NULL #' #' Delete a single service principal in the Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a service principal in the Databricks workspace. #' @@ -49,6 +54,9 @@ servicePrincipalsDelete <- function(client, id) { #' Gets the details for a single service principal define in the Databricks #' workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a service principal in the Databricks workspace. #' @@ -63,6 +71,8 @@ servicePrincipalsGet <- function(client, id) { #' #' Gets the set of service principals associated with a Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + #' #' @param attributes Comma-separated list of attributes to return in response. #' @param count Desired number of results per page. @@ -103,9 +113,12 @@ servicePrincipalsList <- function(client, attributes = NULL, count = NULL, exclu #' Partially updates the details of a single service principal in the Databricks #' workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a service principal in the Databricks workspace. -#' @param operations +#' @param operations This field has no description yet. #' @param schemas The schema of the patch request. #' #' @rdname servicePrincipalsPatch @@ -122,13 +135,16 @@ servicePrincipalsPatch <- function(client, id, operations = NULL, schemas = NULL #' #' This action replaces the existing service principal with the same name. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param active If this user is active. #' @param application_id UUID relating to the service principal. #' @param display_name String that represents a concatenation of given and family names. #' @param entitlements Entitlements assigned to the service principal. -#' @param external_id -#' @param groups +#' @param external_id This field has no description yet. +#' @param groups This field has no description yet. #' @param id Databricks service principal ID. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the List response. diff --git a/R/serving_endpoints.R b/R/serving_endpoints.R index 84b14c43..562c0e57 100755 --- a/R/serving_endpoints.R +++ b/R/serving_endpoints.R @@ -7,6 +7,9 @@ NULL #' #' Retrieves the build logs associated with the provided served model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint that the served model belongs to. #' @param served_model_name Required. The name of the served model that build logs will be retrieved for. @@ -20,13 +23,18 @@ servingEndpointsBuildLogs <- function(client, name, served_model_name) { } #' Create a new serving endpoint. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Serving Endpoints on Databricks reach #' NOT_UPDATING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Serving Endpoints is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param config Required. The core config of the serving endpoint. #' @param name Required. The name of the serving endpoint. @@ -78,6 +86,9 @@ servingEndpointsCreate <- function(client, name, config, rate_limits = NULL, tag #' Delete a serving endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint. #' @@ -93,6 +104,9 @@ servingEndpointsDelete <- function(client, name) { #' Retrieves the metrics associated with the provided serving endpoint in either #' Prometheus or OpenMetrics exposition format. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint to retrieve metrics for. #' @@ -107,6 +121,9 @@ servingEndpointsExportMetrics <- function(client, name) { #' #' Retrieves the details for a single serving endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint. #' @@ -121,6 +138,9 @@ servingEndpointsGet <- function(client, name) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param serving_endpoint_id Required. The serving endpoint for which to get or manage permissions. #' @@ -137,6 +157,9 @@ servingEndpointsGetPermissionLevels <- function(client, serving_endpoint_id) { #' Gets the permissions of a serving endpoint. Serving endpoints can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param serving_endpoint_id Required. The serving endpoint for which to get or manage permissions. #' @@ -148,7 +171,9 @@ servingEndpointsGetPermissions <- function(client, serving_endpoint_id) { sep = "")) } -#' Get all serving endpoints.#' +#' Get all serving endpoints. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname servingEndpointsList @@ -164,6 +189,9 @@ servingEndpointsList <- function(client) { #' #' Retrieves the service logs associated with the provided served model. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint that the served model belongs to. #' @param served_model_name Required. The name of the served model that logs will be retrieved for. @@ -181,6 +209,9 @@ servingEndpointsLogs <- function(client, name, served_model_name) { #' Used to batch add and delete tags from a serving endpoint with a single API #' call. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param add_tags List of endpoint tags to add. #' @param delete_tags List of tag keys to delete. @@ -199,6 +230,9 @@ servingEndpointsPatch <- function(client, name, add_tags = NULL, delete_tags = N #' Used to update the rate limits of a serving endpoint. NOTE: only external and #' foundation model endpoints are supported as of now. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the serving endpoint whose rate limits are being updated. #' @param rate_limits The list of endpoint rate limits. @@ -213,6 +247,9 @@ servingEndpointsPut <- function(client, name, rate_limits = NULL) { #' Query a serving endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param dataframe_records Pandas Dataframe input in the records orientation. #' @param dataframe_split Pandas Dataframe input in the split orientation. @@ -247,8 +284,11 @@ servingEndpointsQuery <- function(client, name, dataframe_records = NULL, datafr #' Sets permissions on a serving endpoint. Serving endpoints can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param serving_endpoint_id Required. The serving endpoint for which to get or manage permissions. #' #' @rdname servingEndpointsSetPermissions @@ -265,13 +305,18 @@ servingEndpointsSetPermissions <- function(client, serving_endpoint_id, access_c #' compute configuration of those served entities, and the endpoint's traffic #' config. An endpoint that already has an update in progress can not be updated #' until the current update completes or fails. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Serving Endpoints on Databricks reach #' NOT_UPDATING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Serving Endpoints is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param auto_capture_config Configuration for Inference Tables which automatically logs requests and responses to Unity Catalog. #' @param name Required. The name of the serving endpoint to update. @@ -330,8 +375,11 @@ servingEndpointsUpdateConfig <- function(client, name, auto_capture_config = NUL #' Updates the permissions on a serving endpoint. Serving endpoints can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param serving_endpoint_id Required. The serving endpoint for which to get or manage permissions. #' #' @rdname servingEndpointsUpdatePermissions diff --git a/R/shares.R b/R/shares.R index 9c01af23..7e478fb9 100755 --- a/R/shares.R +++ b/R/shares.R @@ -9,6 +9,9 @@ NULL #' creation with **update**. The caller must be a metastore admin or have the #' **CREATE_SHARE** privilege on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param name Required. Name of the share. @@ -25,6 +28,9 @@ sharesCreate <- function(client, name, comment = NULL) { #' Deletes a data object share from the metastore. The caller must be an owner #' of the share. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the share. #' @@ -40,6 +46,9 @@ sharesDelete <- function(client, name) { #' Gets a data object share from the metastore. The caller must be a metastore #' admin or the owner of the share. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_shared_data Query for data to include in the share. #' @param name Required. The name of the share. @@ -55,7 +64,9 @@ sharesGet <- function(client, name, include_shared_data = NULL) { #' #' Gets an array of data object shares from the metastore. The caller must be a #' metastore admin or the owner of the share. There is no guarantee of a -#' specific ordering of the elements in the array.#' +#' specific ordering of the elements in the array. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname sharesList @@ -72,6 +83,9 @@ sharesList <- function(client) { #' Gets the permissions for a data share from the metastore. The caller must be #' a metastore admin or the owner of the share. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the share. #' @@ -101,6 +115,9 @@ sharesSharePermissions <- function(client, name) { #' #' Table removals through **update** do not require additional privileges. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment User-provided free-form text description. #' @param name Required. The name of the share. @@ -124,6 +141,9 @@ sharesUpdate <- function(client, name, comment = NULL, new_name = NULL, owner = #' For new recipient grants, the user must also be the owner of the recipients. #' recipient revocations do not require additional privileges. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param changes Array of permission changes. #' @param name Required. The name of the share. diff --git a/R/statement_execution.R b/R/statement_execution.R index 4a232390..1bf8b5ec 100755 --- a/R/statement_execution.R +++ b/R/statement_execution.R @@ -8,6 +8,9 @@ NULL #' Requests that an executing statement be canceled. Callers must poll for #' status to see the terminal state. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param statement_id Required. The statement ID is returned upon successfully submitting a SQL statement, and is a required reference for all subsequent calls. #' @@ -21,6 +24,9 @@ statementExecutionCancelExecution <- function(client, statement_id) { #' Execute a SQL statement. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param byte_limit Applies the given byte limit to the statement's result size. #' @param catalog Sets default catalog for statement execution, similar to [`USE CATALOG`](https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-use-catalog.html) in SQL. @@ -58,6 +64,9 @@ statementExecutionExecuteStatement <- function(client, statement, warehouse_id, #' **NOTE** This call currently might take up to 5 seconds to get the latest #' status and result. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param statement_id Required. The statement ID is returned upon successfully submitting a SQL statement, and is a required reference for all subsequent calls. #' @@ -79,8 +88,11 @@ statementExecutionGetStatement <- function(client, statement_id) { #' similarly includes the `next_chunk_index` and `next_chunk_internal_link` #' fields for simple iteration through the result set. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param chunk_index Required. +#' @param chunk_index Required. This field has no description yet. #' @param statement_id Required. The statement ID is returned upon successfully submitting a SQL statement, and is a required reference for all subsequent calls. #' #' @rdname statementExecutionGetStatementResultChunkN diff --git a/R/storage_credentials.R b/R/storage_credentials.R index 41f6a1d2..da37aa43 100755 --- a/R/storage_credentials.R +++ b/R/storage_credentials.R @@ -7,6 +7,9 @@ NULL #' #' Creates a new storage credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param aws_iam_role The AWS IAM role configuration. #' @param azure_managed_identity The Azure managed identity configuration. @@ -35,6 +38,9 @@ storageCredentialsCreate <- function(client, name, aws_iam_role = NULL, azure_ma #' Deletes a storage credential from the metastore. The caller must be an owner #' of the storage credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param force Force deletion even if there are dependent external locations or external tables. #' @param name Required. Name of the storage credential. @@ -53,6 +59,9 @@ storageCredentialsDelete <- function(client, name, force = NULL) { #' admin, the owner of the storage credential, or have some permission on the #' storage credential. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. Name of the storage credential. #' @@ -71,6 +80,8 @@ storageCredentialsGet <- function(client, name) { #' credentials is unrestricted. There is no guarantee of a specific ordering of #' the elements in the array. #' @param client Required. Instance of DatabricksClient() + + #' #' @param max_results Maximum number of storage credentials to return. #' @param page_token Opaque pagination token to go to next page based on previous query. @@ -103,6 +114,9 @@ storageCredentialsList <- function(client, max_results = NULL, page_token = NULL #' #' Updates a storage credential on the metastore. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param aws_iam_role The AWS IAM role configuration. #' @param azure_managed_identity The Azure managed identity configuration. @@ -147,6 +161,8 @@ storageCredentialsUpdate <- function(client, name, aws_iam_role = NULL, azure_ma #' the **CREATE_EXTERNAL_LOCATION** privilege on the metastore and the storage #' credential. #' @param client Required. Instance of DatabricksClient() + + #' #' @param aws_iam_role The AWS IAM role configuration. #' @param azure_managed_identity The Azure managed identity configuration. diff --git a/R/system_schemas.R b/R/system_schemas.R index f8fbda1f..ed6a0561 100755 --- a/R/system_schemas.R +++ b/R/system_schemas.R @@ -8,6 +8,9 @@ NULL #' Disables the system schema and removes it from the system catalog. The caller #' must be an account admin or a metastore admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param metastore_id Required. The metastore ID under which the system schema lives. #' @param schema_name Required. Full name of the system schema. @@ -25,6 +28,9 @@ systemSchemasDisable <- function(client, metastore_id, schema_name) { #' Enables the system schema and adds it to the system catalog. The caller must #' be an account admin or a metastore admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param metastore_id Required. The metastore ID under which the system schema lives. #' @param schema_name Required. Full name of the system schema. @@ -42,6 +48,9 @@ systemSchemasEnable <- function(client, metastore_id, schema_name) { #' Gets an array of system schemas for a metastore. The caller must be an #' account admin or a metastore admin. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param metastore_id Required. The ID for the metastore in which the system schema resides. #' diff --git a/R/table_constraints.R b/R/table_constraints.R index d96140cf..c0284f1d 100755 --- a/R/table_constraints.R +++ b/R/table_constraints.R @@ -16,6 +16,9 @@ NULL #' referenced parent table's schema, and be the owner of the referenced parent #' table. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param constraint Required. A table constraint, as defined by *one* of the following fields being set: __primary_key_constraint__, __foreign_key_constraint__, __named_table_constraint__. #' @param full_name_arg Required. The full name of the table referenced by the constraint. @@ -39,6 +42,9 @@ tableConstraintsCreate <- function(client, full_name_arg, constraint) { #' **USE_CATALOG** privilege on the table's catalog, the **USE_SCHEMA** #' privilege on the table's schema, and be the owner of the table. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param cascade Required. If true, try deleting all child constraints of the current constraint. #' @param constraint_name Required. The name of the constraint to delete. diff --git a/R/tables.R b/R/tables.R index 528da3ee..cfbd5d2e 100755 --- a/R/tables.R +++ b/R/tables.R @@ -11,6 +11,9 @@ NULL #' table and have the **USE_CATALOG** privilege on the parent catalog and the #' **USE_SCHEMA** privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -32,6 +35,9 @@ tablesDelete <- function(client, full_name) { #' SELECT privilege on the table. * Have BROWSE privilege on the parent catalog #' * Have BROWSE privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @@ -53,6 +59,9 @@ tablesExists <- function(client, full_name) { #' parent schema, and either be the table owner or have the SELECT privilege on #' the table. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. #' @param include_browse Whether to include tables in the response for which the principal can only access selective metadata for. @@ -75,6 +84,9 @@ tablesGet <- function(client, full_name, include_browse = NULL, include_delta_me #' catalog and the **USE_SCHEMA** privilege on the parent schema. There is no #' guarantee of a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. Name of parent catalog for tables of interest. #' @param include_browse Whether to include tables in the response for which the principal can only access selective metadata for. @@ -127,6 +139,9 @@ tablesList <- function(client, catalog_name, schema_name, include_browse = NULL, #' #' There is no guarantee of a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. Name of parent catalog for tables of interest. #' @param max_results Maximum number of summaries for tables to return. @@ -168,9 +183,12 @@ tablesListSummaries <- function(client, catalog_name, max_results = NULL, page_t #' **USE_CATALOG** privilege on the parent catalog and the **USE_SCHEMA** #' privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param full_name Required. Full name of the table. -#' @param owner +#' @param owner This field has no description yet. #' #' @rdname tablesUpdate #' @export diff --git a/R/token_management.R b/R/token_management.R index 4893c981..d7f923e0 100755 --- a/R/token_management.R +++ b/R/token_management.R @@ -7,6 +7,9 @@ NULL #' #' Creates a token on behalf of a service principal. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param application_id Required. Application ID of the service principal. #' @param comment Comment that describes the purpose of the token. @@ -24,6 +27,9 @@ tokenManagementCreateOboToken <- function(client, application_id, comment = NULL #' #' Deletes a token, specified by its ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param token_id Required. The ID of the token to get. #' @@ -38,6 +44,9 @@ tokenManagementDelete <- function(client, token_id) { #' #' Gets information about a token, specified by its ID. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param token_id Required. The ID of the token to get. #' @@ -50,7 +59,9 @@ tokenManagementGet <- function(client, token_id) { #' Get token permission levels. #' -#' Gets the permission levels that a user can have on an object.#' +#' Gets the permission levels that a user can have on an object. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname tokenManagementGetPermissionLevels #' @export tokenManagementGetPermissionLevels <- function(client) { @@ -60,7 +71,9 @@ tokenManagementGetPermissionLevels <- function(client) { #' Get token permissions. #' #' Gets the permissions of all tokens. Tokens can inherit permissions from their -#' root object.#' +#' root object. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname tokenManagementGetPermissions #' @export tokenManagementGetPermissions <- function(client) { @@ -71,6 +84,8 @@ tokenManagementGetPermissions <- function(client) { #' #' Lists all tokens associated with the specified workspace or user. #' @param client Required. Instance of DatabricksClient() + + #' #' @param created_by_id User ID of the user that created the token. #' @param created_by_username Username of the user that created the token. @@ -92,8 +107,10 @@ tokenManagementList <- function(client, created_by_id = NULL, created_by_usernam #' Sets permissions on all tokens. Tokens can inherit permissions from their #' root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' #' @rdname tokenManagementSetPermissions #' @export @@ -107,8 +124,10 @@ tokenManagementSetPermissions <- function(client, access_control_list = NULL) { #' Updates the permissions on all tokens. Tokens can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' #' @rdname tokenManagementUpdatePermissions #' @export diff --git a/R/tokens.R b/R/tokens.R index 0acb4c38..d1bd4413 100755 --- a/R/tokens.R +++ b/R/tokens.R @@ -10,6 +10,8 @@ NULL #' authenticated token. If the user's token quota is exceeded, this call returns #' an error **QUOTA_EXCEEDED**. #' @param client Required. Instance of DatabricksClient() + + #' #' @param comment Optional description to attach to the token. #' @param lifetime_seconds The lifetime of the token, in seconds. @@ -28,6 +30,9 @@ tokensCreate <- function(client, comment = NULL, lifetime_seconds = NULL) { #' If a token with the specified ID is not valid, this call returns an error #' **RESOURCE_DOES_NOT_EXIST**. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param token_id Required. The ID of the token to be revoked. #' @@ -40,7 +45,9 @@ tokensDelete <- function(client, token_id) { #' List tokens. #' -#' Lists all the valid tokens for a user-workspace pair.#' +#' Lists all the valid tokens for a user-workspace pair. +#' @param client Required. Instance of DatabricksClient() +#' #' @return `data.frame` with all of the response pages. #' #' @rdname tokensList diff --git a/R/users.R b/R/users.R index 1b1d5917..f2ef8028 100755 --- a/R/users.R +++ b/R/users.R @@ -8,15 +8,17 @@ NULL #' Creates a new user in the Databricks workspace. This new user will also be #' added to the Databricks account. #' @param client Required. Instance of DatabricksClient() + + #' #' @param active If this user is active. #' @param display_name String that represents a concatenation of given and family names. #' @param emails All the emails associated with the Databricks user. #' @param entitlements Entitlements assigned to the user. #' @param external_id External ID is not currently supported. -#' @param groups +#' @param groups This field has no description yet. #' @param id Databricks user ID. -#' @param name +#' @param name This field has no description yet. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the user. #' @param user_name Email address of the Databricks user. @@ -37,6 +39,9 @@ usersCreate <- function(client, active = NULL, display_name = NULL, emails = NUL #' Deletes a user. Deleting a user from a Databricks workspace also removes #' objects associated with the user. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a user in the Databricks workspace. #' @@ -51,6 +56,9 @@ usersDelete <- function(client, id) { #' #' Gets information for a specific user in Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param attributes Comma-separated list of attributes to return in response. #' @param count Desired number of results per page. @@ -72,7 +80,9 @@ usersGet <- function(client, id, attributes = NULL, count = NULL, excluded_attri #' Get password permission levels. #' -#' Gets the permission levels that a user can have on an object.#' +#' Gets the permission levels that a user can have on an object. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname usersGetPermissionLevels #' @export usersGetPermissionLevels <- function(client) { @@ -82,7 +92,9 @@ usersGetPermissionLevels <- function(client) { #' Get password permissions. #' #' Gets the permissions of all passwords. Passwords can inherit permissions from -#' their root object.#' +#' their root object. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname usersGetPermissions #' @export usersGetPermissions <- function(client) { @@ -93,6 +105,8 @@ usersGetPermissions <- function(client) { #' #' Gets details for all the users associated with a Databricks workspace. #' @param client Required. Instance of DatabricksClient() + + #' #' @param attributes Comma-separated list of attributes to return in response. #' @param count Desired number of results per page. @@ -133,9 +147,12 @@ usersList <- function(client, attributes = NULL, count = NULL, excluded_attribut #' Partially updates a user resource by applying the supplied operations on #' specific user attributes. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Unique ID for a user in the Databricks workspace. -#' @param operations +#' @param operations This field has no description yet. #' @param schemas The schema of the patch request. #' #' @rdname usersPatch @@ -150,8 +167,10 @@ usersPatch <- function(client, id, operations = NULL, schemas = NULL) { #' Sets permissions on all passwords. Passwords can inherit permissions from #' their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' #' @rdname usersSetPermissions #' @export @@ -164,15 +183,18 @@ usersSetPermissions <- function(client, access_control_list = NULL) { #' #' Replaces a user's information with the data supplied in request. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param active If this user is active. #' @param display_name String that represents a concatenation of given and family names. #' @param emails All the emails associated with the Databricks user. #' @param entitlements Entitlements assigned to the user. #' @param external_id External ID is not currently supported. -#' @param groups +#' @param groups This field has no description yet. #' @param id Databricks user ID. -#' @param name +#' @param name This field has no description yet. #' @param roles Corresponds to AWS instance profile/arn role. #' @param schemas The schema of the user. #' @param user_name Email address of the Databricks user. @@ -193,8 +215,10 @@ usersUpdate <- function(client, id, active = NULL, display_name = NULL, emails = #' Updates the permissions on all passwords. Passwords can inherit permissions #' from their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' #' @rdname usersUpdatePermissions #' @export diff --git a/R/vector_search_endpoints.R b/R/vector_search_endpoints.R index e36cd617..32e5b9e5 100755 --- a/R/vector_search_endpoints.R +++ b/R/vector_search_endpoints.R @@ -6,13 +6,18 @@ NULL #' Create an endpoint. #' #' Create a new endpoint. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Vector Search Endpoints on Databricks reach #' ONLINE state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Vector Search Endpoints is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param endpoint_type Required. Type of endpoint. #' @param name Required. Name of endpoint. @@ -65,6 +70,9 @@ vectorSearchEndpointsCreateEndpoint <- function(client, name, endpoint_type, tim #' Delete an endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param endpoint_name Required. Name of the endpoint. #' @@ -78,6 +86,9 @@ vectorSearchEndpointsDeleteEndpoint <- function(client, endpoint_name) { #' Get an endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param endpoint_name Required. Name of the endpoint. #' @@ -90,6 +101,8 @@ vectorSearchEndpointsGetEndpoint <- function(client, endpoint_name) { #' List all endpoints. #' @param client Required. Instance of DatabricksClient() + + #' #' @param page_token Token for pagination. #' diff --git a/R/vector_search_indexes.R b/R/vector_search_indexes.R index b8b68d07..3bc2fc43 100755 --- a/R/vector_search_indexes.R +++ b/R/vector_search_indexes.R @@ -7,6 +7,9 @@ NULL #' #' Create a new index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param delta_sync_index_spec Specification for Delta Sync Index. #' @param direct_access_index_spec Specification for Direct Vector Access Index. @@ -28,6 +31,9 @@ vectorSearchIndexesCreateIndex <- function(client, name, endpoint_name, primary_ #' #' Handles the deletion of data from a specified vector index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param index_name Required. Name of the vector index where data is to be deleted. #' @param primary_keys Required. List of primary keys for the data to be deleted. @@ -44,6 +50,9 @@ vectorSearchIndexesDeleteDataVectorIndex <- function(client, index_name, primary #' #' Delete an index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param index_name Required. Name of the index. #' @@ -58,6 +67,9 @@ vectorSearchIndexesDeleteIndex <- function(client, index_name) { #' #' Get an index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param index_name Required. Name of the index. #' @@ -72,6 +84,9 @@ vectorSearchIndexesGetIndex <- function(client, index_name) { #' #' List all indexes in the given endpoint. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param endpoint_name Required. Name of the endpoint. #' @param page_token Token for pagination. @@ -104,6 +119,9 @@ vectorSearchIndexesListIndexes <- function(client, endpoint_name, page_token = N #' #' Query the specified vector index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param columns Required. List of column names to include in the response. #' @param filters_json JSON string representing query filters. @@ -127,6 +145,9 @@ vectorSearchIndexesQueryIndex <- function(client, index_name, columns, filters_j #' #' Triggers a synchronization process for a specified vector index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param index_name Required. Name of the vector index to synchronize. #' @@ -142,6 +163,9 @@ vectorSearchIndexesSyncIndex <- function(client, index_name) { #' #' Handles the upserting of data into a specified vector index. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param index_name Required. Name of the vector index where data is to be upserted. #' @param inputs_json Required. JSON string representing the data to be upserted. diff --git a/R/volumes.R b/R/volumes.R index 93613738..7878a332 100755 --- a/R/volumes.R +++ b/R/volumes.R @@ -24,13 +24,16 @@ NULL #' storage location. - The specified storage location is not under the location #' of other tables, nor volumes, or catalogs or schemas. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. The name of the catalog where the schema and the volume are. #' @param comment The comment attached to the volume. #' @param name Required. The name of the volume. #' @param schema_name Required. The name of the schema where the volume is. #' @param storage_location The storage location on the cloud. -#' @param volume_type Required. +#' @param volume_type Required. This field has no description yet. #' #' @rdname volumesCreate #' @export @@ -50,6 +53,9 @@ volumesCreate <- function(client, catalog_name, schema_name, name, volume_type, #' privilege on the parent catalog and the **USE_SCHEMA** privilege on the #' parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The three-level (fully qualified) name of the volume. #' @@ -74,6 +80,9 @@ volumesDelete <- function(client, name) { #' #' There is no guarantee of a specific ordering of the elements in the array. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param catalog_name Required. The identifier of the catalog. #' @param include_browse Whether to include volumes in the response for which the principal can only access selective metadata for. @@ -116,6 +125,9 @@ volumesList <- function(client, catalog_name, schema_name, include_browse = NULL #' be the owner or have the **USE_CATALOG** privilege on the parent catalog and #' the **USE_SCHEMA** privilege on the parent schema. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param include_browse Whether to include volumes in the response for which the principal can only access selective metadata for. #' @param name Required. The three-level (fully qualified) name of the volume. @@ -139,6 +151,9 @@ volumesRead <- function(client, name, include_browse = NULL) { #' Currently only the name, the owner or the comment of the volume could be #' updated. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param comment The comment attached to the volume. #' @param name Required. The three-level (fully qualified) name of the volume. diff --git a/R/warehouses.R b/R/warehouses.R index a85ede67..c1fd81d0 100755 --- a/R/warehouses.R +++ b/R/warehouses.R @@ -6,13 +6,17 @@ NULL #' Create a warehouse. #' #' Creates a new SQL warehouse. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Warehouses on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Warehouses is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + #' #' @param auto_stop_mins The amount of time in minutes that a SQL warehouse must be idle (i.e., no RUNNING queries) before it is automatically stopped. #' @param channel Channel Details. @@ -84,6 +88,9 @@ warehousesCreate <- function(client, auto_stop_mins = NULL, channel = NULL, clus #' #' Deletes a SQL warehouse. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Required. #' @@ -97,13 +104,18 @@ warehousesDelete <- function(client, id) { #' Update a warehouse. #' #' Updates the configuration for a SQL warehouse. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Warehouses on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Warehouses is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param auto_stop_mins The amount of time in minutes that a SQL warehouse must be idle (i.e., no RUNNING queries) before it is automatically stopped. #' @param channel Channel Details. @@ -177,6 +189,9 @@ warehousesEdit <- function(client, id, auto_stop_mins = NULL, channel = NULL, cl #' #' Gets the information for a single SQL warehouse. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param id Required. Required. #' @@ -191,6 +206,9 @@ warehousesGet <- function(client, id) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param warehouse_id Required. The SQL warehouse for which to get or manage permissions. #' @@ -207,6 +225,9 @@ warehousesGetPermissionLevels <- function(client, warehouse_id) { #' Gets the permissions of a SQL warehouse. SQL warehouses can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param warehouse_id Required. The SQL warehouse for which to get or manage permissions. #' @@ -220,7 +241,9 @@ warehousesGetPermissions <- function(client, warehouse_id) { #' Get the workspace configuration. #' #' Gets the workspace level configuration that is shared by all SQL warehouses -#' in a workspace.#' +#' in a workspace. +#' @param client Required. Instance of DatabricksClient() +#' #' @rdname warehousesGetWorkspaceWarehouseConfig #' @export warehousesGetWorkspaceWarehouseConfig <- function(client) { @@ -231,6 +254,8 @@ warehousesGetWorkspaceWarehouseConfig <- function(client) { #' #' Lists all SQL warehouses that a user has manager permissions on. #' @param client Required. Instance of DatabricksClient() + + #' #' @param run_as_user_id Service Principal which will be used to fetch the list of warehouses. #' @@ -251,8 +276,11 @@ warehousesList <- function(client, run_as_user_id = NULL) { #' Sets permissions on a SQL warehouse. SQL warehouses can inherit permissions #' from their root object. #' @param client Required. Instance of DatabricksClient() + + #' -#' @param access_control_list +#' +#' @param access_control_list This field has no description yet. #' @param warehouse_id Required. The SQL warehouse for which to get or manage permissions. #' #' @rdname warehousesSetPermissions @@ -268,6 +296,8 @@ warehousesSetPermissions <- function(client, warehouse_id, access_control_list = #' Sets the workspace level configuration that is shared by all SQL warehouses #' in a workspace. #' @param client Required. Instance of DatabricksClient() + + #' #' @param channel Optional: Channel selection details. #' @param config_param Deprecated: Use sql_configuration_parameters. @@ -295,13 +325,18 @@ warehousesSetWorkspaceWarehouseConfig <- function(client, channel = NULL, config #' Start a warehouse. #' #' Starts a SQL warehouse. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Warehouses on Databricks reach #' RUNNING state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Warehouses is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param id Required. Required. #' @@ -353,13 +388,18 @@ warehousesStart <- function(client, id, timeout = 20, callback = cli_reporter) { #' Stop a warehouse. #' #' Stops a SQL warehouse. +#' @param client Required. Instance of DatabricksClient() + #' #' @description #' This is a long-running operation, which blocks until Warehouses on Databricks reach #' STOPPED state with the timeout of 20 minutes, that you can change via `timeout` parameter. #' By default, the state of Databricks Warehouses is reported to console. You can change this behavior #' by changing the `callback` parameter. -#' @param client Required. Instance of DatabricksClient() +#' @param timeout Time to wait for the operation to complete in minutes. +#' @param callback Function to report the status of the operation. By default, it reports to console. + +#' #' #' @param id Required. Required. #' @@ -408,8 +448,11 @@ warehousesStop <- function(client, id, timeout = 20, callback = cli_reporter) { #' Updates the permissions on a SQL warehouse. SQL warehouses can inherit #' permissions from their root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param warehouse_id Required. The SQL warehouse for which to get or manage permissions. #' #' @rdname warehousesUpdatePermissions diff --git a/R/workspace.R b/R/workspace.R index 576d5d25..a7f7e96e 100755 --- a/R/workspace.R +++ b/R/workspace.R @@ -14,6 +14,9 @@ NULL #' Object deletion cannot be undone and deleting a directory recursively is not #' atomic. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The absolute path of the notebook or directory. #' @param recursive The flag that specifies whether to delete the object recursively. @@ -36,6 +39,9 @@ workspaceDelete <- function(client, path, recursive = NULL) { #' `MAX_NOTEBOOK_SIZE_EXCEEDED`. Currently, this API does not support exporting #' a library. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param format This specifies the format of the exported file. #' @param path Required. The absolute path of the object or directory. @@ -51,6 +57,9 @@ workspaceExport <- function(client, path, format = NULL) { #' #' Gets the permission levels that a user can have on an object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param workspace_object_id Required. The workspace object for which to get or manage permissions. #' @param workspace_object_type Required. The workspace object type for which to get or manage permissions. @@ -68,6 +77,9 @@ workspaceGetPermissionLevels <- function(client, workspace_object_type, workspac #' Gets the permissions of a workspace object. Workspace objects can inherit #' permissions from their parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param workspace_object_id Required. The workspace object for which to get or manage permissions. #' @param workspace_object_type Required. The workspace object type for which to get or manage permissions. @@ -85,6 +97,9 @@ workspaceGetPermissions <- function(client, workspace_object_type, workspace_obj #' Gets the status of an object or a directory. If `path` does not exist, this #' call returns an error `RESOURCE_DOES_NOT_EXIST`. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The absolute path of the notebook or directory. #' @@ -104,6 +119,9 @@ workspaceGetStatus <- function(client, path) { #' the `language` field unset. To import a single file as `SOURCE`, you must set #' the `language` field. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param content The base64-encoded content. #' @param format This specifies the format of the file to be imported. @@ -126,6 +144,9 @@ workspaceImport <- function(client, path, content = NULL, format = NULL, languag #' the input path does not exist, this call returns an error #' `RESOURCE_DOES_NOT_EXIST`. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param notebooks_modified_after UTC timestamp in milliseconds. #' @param path Required. The absolute path of the notebook or directory. @@ -151,6 +172,9 @@ workspaceList <- function(client, path, notebooks_modified_after = NULL) { #' Note that if this operation fails it may have succeeded in creating some of #' the necessary parent directories. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param path Required. The absolute path of the directory. #' @@ -166,8 +190,11 @@ workspaceMkdirs <- function(client, path) { #' Sets permissions on a workspace object. Workspace objects can inherit #' permissions from their parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param workspace_object_id Required. The workspace object for which to get or manage permissions. #' @param workspace_object_type Required. The workspace object type for which to get or manage permissions. #' @@ -185,8 +212,11 @@ workspaceSetPermissions <- function(client, workspace_object_type, workspace_obj #' Updates the permissions on a workspace object. Workspace objects can inherit #' permissions from their parent objects or root object. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param access_control_list +#' @param access_control_list This field has no description yet. #' @param workspace_object_id Required. The workspace object for which to get or manage permissions. #' @param workspace_object_type Required. The workspace object type for which to get or manage permissions. #' diff --git a/R/workspace_bindings.R b/R/workspace_bindings.R index 890bfb94..6638ccbb 100755 --- a/R/workspace_bindings.R +++ b/R/workspace_bindings.R @@ -8,6 +8,9 @@ NULL #' Gets workspace bindings of the catalog. The caller must be a metastore admin #' or an owner of the catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param name Required. The name of the catalog. #' @@ -24,6 +27,9 @@ workspaceBindingsGet <- function(client, name) { #' Gets workspace bindings of the securable. The caller must be a metastore #' admin or an owner of the securable. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param securable_name Required. The name of the securable. #' @param securable_type Required. The type of the securable. @@ -41,6 +47,9 @@ workspaceBindingsGetBindings <- function(client, securable_type, securable_name) #' Updates workspace bindings of the catalog. The caller must be a metastore #' admin or an owner of the catalog. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param assign_workspaces A list of workspace IDs. #' @param name Required. The name of the catalog. @@ -59,6 +68,9 @@ workspaceBindingsUpdate <- function(client, name, assign_workspaces = NULL, unas #' Updates workspace bindings of the securable. The caller must be a metastore #' admin or an owner of the securable. #' @param client Required. Instance of DatabricksClient() + + +#' #' #' @param add List of workspace bindings. #' @param remove List of workspace bindings. diff --git a/R/workspace_conf.R b/R/workspace_conf.R index 6c4f76a1..8420211a 100755 --- a/R/workspace_conf.R +++ b/R/workspace_conf.R @@ -7,8 +7,11 @@ NULL #' #' Gets the configuration status for a workspace. #' @param client Required. Instance of DatabricksClient() + + +#' #' -#' @param keys Required. +#' @param keys Required. This field has no description yet. #' #' @rdname workspaceConfGetStatus #' @export @@ -22,6 +25,8 @@ workspaceConfGetStatus <- function(client, keys) { #' Sets the configuration status for a workspace, including enabling or #' disabling it. #' @param client Required. Instance of DatabricksClient() + + #' #' #' @rdname workspaceConfSetStatus diff --git a/README.md b/README.md index fe50691a..6ea20a15 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,6 @@ [![R-CMD-check](https://github.com/databrickslabs/databricks-sdk-r/actions/workflows/R-CMD-check.yaml/badge.svg)](https://github.com/databrickslabs/databricks-sdk-r/actions/workflows/R-CMD-check.yaml) -[![lines of code](https://tokei.rs/b1/github/databrickslabs/databricks-sdk-r)]([https://codecov.io/github/databrickslabs/databricks-sdk-r](https://github.com/databrickslabs/databricks-sdk-r)) ## Authentication diff --git a/man/accountAccessControlProxyUpdateRuleSet.Rd b/man/accountAccessControlProxyUpdateRuleSet.Rd index 4038618d..5e8d95e6 100644 --- a/man/accountAccessControlProxyUpdateRuleSet.Rd +++ b/man/accountAccessControlProxyUpdateRuleSet.Rd @@ -11,7 +11,7 @@ accountAccessControlProxyUpdateRuleSet(client, name, rule_set) \item{name}{Required. Name of the rule set.} -\item{rule_set}{Required.} +\item{rule_set}{Required. This field has no description yet.} } \description{ Replace the rules of a rule set. First, use a GET rule set request to read diff --git a/man/alertsDelete.Rd b/man/alertsDelete.Rd index ba6f268d..5188f58a 100644 --- a/man/alertsDelete.Rd +++ b/man/alertsDelete.Rd @@ -9,7 +9,7 @@ alertsDelete(client, alert_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{alert_id}{Required.} +\item{alert_id}{Required. This field has no description yet.} } \description{ Deletes an alert. Deleted alerts are no longer accessible and cannot be diff --git a/man/alertsGet.Rd b/man/alertsGet.Rd index 2ab5adab..76639d17 100644 --- a/man/alertsGet.Rd +++ b/man/alertsGet.Rd @@ -9,7 +9,7 @@ alertsGet(client, alert_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{alert_id}{Required.} +\item{alert_id}{Required. This field has no description yet.} } \description{ Gets an alert. diff --git a/man/alertsList.Rd b/man/alertsList.Rd index 33db0e6f..6b5f3827 100644 --- a/man/alertsList.Rd +++ b/man/alertsList.Rd @@ -6,6 +6,9 @@ \usage{ alertsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -Gets a list of alerts.#' +Gets a list of alerts. } diff --git a/man/alertsUpdate.Rd b/man/alertsUpdate.Rd index 8a97190f..1906b415 100644 --- a/man/alertsUpdate.Rd +++ b/man/alertsUpdate.Rd @@ -9,7 +9,7 @@ alertsUpdate(client, alert_id, name, options, query_id, rearm = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{alert_id}{Required.} +\item{alert_id}{Required. This field has no description yet.} \item{name}{Required. Name of the alert.} diff --git a/man/appsGetApps.Rd b/man/appsGetApps.Rd index b99cfe11..083b3c04 100644 --- a/man/appsGetApps.Rd +++ b/man/appsGetApps.Rd @@ -6,6 +6,9 @@ \usage{ appsGetApps(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -List all available applications#' +List all available applications } diff --git a/man/automaticClusterUpdateUpdate.Rd b/man/automaticClusterUpdateUpdate.Rd index 73628cd0..8b67120e 100644 --- a/man/automaticClusterUpdateUpdate.Rd +++ b/man/automaticClusterUpdateUpdate.Rd @@ -11,7 +11,7 @@ automaticClusterUpdateUpdate(client, allow_missing, setting, field_mask) \item{allow_missing}{Required. This should always be set to true for Settings API.} -\item{setting}{Required.} +\item{setting}{Required. This field has no description yet.} \item{field_mask}{Required. Field mask is required to be passed into the PATCH request.} } diff --git a/man/clusterPoliciesSetPermissions.Rd b/man/clusterPoliciesSetPermissions.Rd index 192400ca..e73aff44 100644 --- a/man/clusterPoliciesSetPermissions.Rd +++ b/man/clusterPoliciesSetPermissions.Rd @@ -15,7 +15,7 @@ clusterPoliciesSetPermissions( \item{cluster_policy_id}{Required. The cluster policy for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a cluster policy. Cluster policies can inherit diff --git a/man/clusterPoliciesUpdatePermissions.Rd b/man/clusterPoliciesUpdatePermissions.Rd index fd30afb9..bdcf4cca 100644 --- a/man/clusterPoliciesUpdatePermissions.Rd +++ b/man/clusterPoliciesUpdatePermissions.Rd @@ -15,7 +15,7 @@ clusterPoliciesUpdatePermissions( \item{cluster_policy_id}{Required. The cluster policy for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a cluster policy. Cluster policies can inherit diff --git a/man/clustersCreate.Rd b/man/clustersCreate.Rd index 43dbba51..d9859c9b 100644 --- a/man/clustersCreate.Rd +++ b/man/clustersCreate.Rd @@ -43,7 +43,7 @@ clustersCreate( \item{spark_version}{Required. The Spark version of the cluster, e.g.} -\item{apply_policy_default_values}{} +\item{apply_policy_default_values}{This field has no description yet.} \item{autoscale}{Parameters needed in order to automatically scale clusters up and down based on load.} @@ -63,7 +63,7 @@ clustersCreate( \item{data_security_mode}{Data security mode decides what data governance model to use when accessing data from a cluster.} -\item{docker_image}{} +\item{docker_image}{This field has no description yet.} \item{driver_instance_pool_id}{The optional ID of the instance pool for the driver of the cluster belongs.} @@ -95,7 +95,11 @@ clustersCreate( \item{ssh_public_keys}{SSH public key contents that will be added to each Spark node in this cluster.} -\item{workload_type}{} +\item{workload_type}{This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersDelete.Rd b/man/clustersDelete.Rd index f5c501f9..8caf005c 100644 --- a/man/clustersDelete.Rd +++ b/man/clustersDelete.Rd @@ -10,6 +10,10 @@ clustersDelete(client, cluster_id, timeout = 20, callback = cli_reporter) \item{client}{Required. Instance of DatabricksClient()} \item{cluster_id}{Required. The cluster to be terminated.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersEdit.Rd b/man/clustersEdit.Rd index d47d3dc1..aacbf435 100644 --- a/man/clustersEdit.Rd +++ b/man/clustersEdit.Rd @@ -46,7 +46,7 @@ clustersEdit( \item{spark_version}{Required. The Spark version of the cluster, e.g.} -\item{apply_policy_default_values}{} +\item{apply_policy_default_values}{This field has no description yet.} \item{autoscale}{Parameters needed in order to automatically scale clusters up and down based on load.} @@ -66,7 +66,7 @@ clustersEdit( \item{data_security_mode}{Data security mode decides what data governance model to use when accessing data from a cluster.} -\item{docker_image}{} +\item{docker_image}{This field has no description yet.} \item{driver_instance_pool_id}{The optional ID of the instance pool for the driver of the cluster belongs.} @@ -98,7 +98,11 @@ clustersEdit( \item{ssh_public_keys}{SSH public key contents that will be added to each Spark node in this cluster.} -\item{workload_type}{} +\item{workload_type}{This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersListNodeTypes.Rd b/man/clustersListNodeTypes.Rd index 9e8f38a2..d6a0a883 100644 --- a/man/clustersListNodeTypes.Rd +++ b/man/clustersListNodeTypes.Rd @@ -6,7 +6,10 @@ \usage{ clustersListNodeTypes(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Returns a list of supported Spark node types. These node types can be used to -launch a cluster.#' +launch a cluster. } diff --git a/man/clustersListZones.Rd b/man/clustersListZones.Rd index 278b6e61..ef1fde04 100644 --- a/man/clustersListZones.Rd +++ b/man/clustersListZones.Rd @@ -6,7 +6,10 @@ \usage{ clustersListZones(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Returns a list of availability zones where clusters can be created in (For -example, us-west-2a). These zones can be used to launch a cluster.#' +example, us-west-2a). These zones can be used to launch a cluster. } diff --git a/man/clustersResize.Rd b/man/clustersResize.Rd index 6f826774..9bedada6 100644 --- a/man/clustersResize.Rd +++ b/man/clustersResize.Rd @@ -21,6 +21,10 @@ clustersResize( \item{autoscale}{Parameters needed in order to automatically scale clusters up and down based on load.} \item{num_workers}{Number of worker nodes that this cluster should have.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersRestart.Rd b/man/clustersRestart.Rd index c19ce23b..40d44c66 100644 --- a/man/clustersRestart.Rd +++ b/man/clustersRestart.Rd @@ -18,6 +18,10 @@ clustersRestart( \item{cluster_id}{Required. The cluster to be started.} \item{restart_user}{\if{html}{\out{}}.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersSetPermissions.Rd b/man/clustersSetPermissions.Rd index 305559c9..4f6f75f5 100644 --- a/man/clustersSetPermissions.Rd +++ b/man/clustersSetPermissions.Rd @@ -11,7 +11,7 @@ clustersSetPermissions(client, cluster_id, access_control_list = NULL) \item{cluster_id}{Required. The cluster for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a cluster. Clusters can inherit permissions from their diff --git a/man/clustersSparkVersions.Rd b/man/clustersSparkVersions.Rd index 5e786218..495da3ca 100644 --- a/man/clustersSparkVersions.Rd +++ b/man/clustersSparkVersions.Rd @@ -6,7 +6,10 @@ \usage{ clustersSparkVersions(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Returns the list of available Spark versions. These versions can be used to -launch a cluster.#' +launch a cluster. } diff --git a/man/clustersStart.Rd b/man/clustersStart.Rd index 10be514f..908a27e0 100644 --- a/man/clustersStart.Rd +++ b/man/clustersStart.Rd @@ -10,6 +10,10 @@ clustersStart(client, cluster_id, timeout = 20, callback = cli_reporter) \item{client}{Required. Instance of DatabricksClient()} \item{cluster_id}{Required. The cluster to be started.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Clusters on Databricks reach diff --git a/man/clustersUpdatePermissions.Rd b/man/clustersUpdatePermissions.Rd index 9b2aa76a..f2dcc403 100644 --- a/man/clustersUpdatePermissions.Rd +++ b/man/clustersUpdatePermissions.Rd @@ -11,7 +11,7 @@ clustersUpdatePermissions(client, cluster_id, access_control_list = NULL) \item{cluster_id}{Required. The cluster for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a cluster. Clusters can inherit permissions from diff --git a/man/commandExecutionCancel.Rd b/man/commandExecutionCancel.Rd index 54a6f874..38316eb9 100644 --- a/man/commandExecutionCancel.Rd +++ b/man/commandExecutionCancel.Rd @@ -16,11 +16,15 @@ commandExecutionCancel( \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{cluster_id}{} +\item{cluster_id}{This field has no description yet.} -\item{command_id}{} +\item{command_id}{This field has no description yet.} -\item{context_id}{} +\item{context_id}{This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Command Execution on Databricks reach diff --git a/man/commandExecutionCommandStatus.Rd b/man/commandExecutionCommandStatus.Rd index 170357c8..dce67ac8 100644 --- a/man/commandExecutionCommandStatus.Rd +++ b/man/commandExecutionCommandStatus.Rd @@ -9,11 +9,11 @@ commandExecutionCommandStatus(client, cluster_id, context_id, command_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{cluster_id}{Required.} +\item{cluster_id}{Required. This field has no description yet.} -\item{context_id}{Required.} +\item{context_id}{Required. This field has no description yet.} -\item{command_id}{Required.} +\item{command_id}{Required. This field has no description yet.} } \description{ Gets the status of and, if available, the results from a currently executing diff --git a/man/commandExecutionContextStatus.Rd b/man/commandExecutionContextStatus.Rd index dea4e21c..f1fa685c 100644 --- a/man/commandExecutionContextStatus.Rd +++ b/man/commandExecutionContextStatus.Rd @@ -9,9 +9,9 @@ commandExecutionContextStatus(client, cluster_id, context_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{cluster_id}{Required.} +\item{cluster_id}{Required. This field has no description yet.} -\item{context_id}{Required.} +\item{context_id}{Required. This field has no description yet.} } \description{ Gets the status for an execution context. diff --git a/man/commandExecutionCreate.Rd b/man/commandExecutionCreate.Rd index 032d1dca..fae681e0 100644 --- a/man/commandExecutionCreate.Rd +++ b/man/commandExecutionCreate.Rd @@ -17,7 +17,11 @@ commandExecutionCreate( \item{cluster_id}{Running cluster id.} -\item{language}{} +\item{language}{This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Command Execution on Databricks reach diff --git a/man/commandExecutionDestroy.Rd b/man/commandExecutionDestroy.Rd index eed9a7b6..2eb3e069 100644 --- a/man/commandExecutionDestroy.Rd +++ b/man/commandExecutionDestroy.Rd @@ -9,9 +9,9 @@ commandExecutionDestroy(client, cluster_id, context_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{cluster_id}{Required.} +\item{cluster_id}{Required. This field has no description yet.} -\item{context_id}{Required.} +\item{context_id}{Required. This field has no description yet.} } \description{ Deletes an execution context. diff --git a/man/commandExecutionExecute.Rd b/man/commandExecutionExecute.Rd index 995c8ce2..2d3a9e1f 100644 --- a/man/commandExecutionExecute.Rd +++ b/man/commandExecutionExecute.Rd @@ -23,7 +23,11 @@ commandExecutionExecute( \item{context_id}{Running context id.} -\item{language}{} +\item{language}{This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Command Execution on Databricks reach diff --git a/man/connectionsList.Rd b/man/connectionsList.Rd index 7f87b723..fc0c2597 100644 --- a/man/connectionsList.Rd +++ b/man/connectionsList.Rd @@ -6,9 +6,12 @@ \usage{ connectionsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ -List all connections.#' +List all connections. } diff --git a/man/cspEnablementUpdate.Rd b/man/cspEnablementUpdate.Rd index 4cae98c6..711b3887 100644 --- a/man/cspEnablementUpdate.Rd +++ b/man/cspEnablementUpdate.Rd @@ -11,7 +11,7 @@ cspEnablementUpdate(client, allow_missing, setting, field_mask) \item{allow_missing}{Required. This should always be set to true for Settings API.} -\item{setting}{Required.} +\item{setting}{Required. This field has no description yet.} \item{field_mask}{Required. Field mask is required to be passed into the PATCH request.} } diff --git a/man/currentUserMe.Rd b/man/currentUserMe.Rd index b8dd1330..92e7723c 100644 --- a/man/currentUserMe.Rd +++ b/man/currentUserMe.Rd @@ -6,6 +6,9 @@ \usage{ currentUserMe(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -Get details about the current method caller's identity.#' +Get details about the current method caller's identity. } diff --git a/man/dashboardWidgetsCreate.Rd b/man/dashboardWidgetsCreate.Rd index c3cea57b..b00dd167 100644 --- a/man/dashboardWidgetsCreate.Rd +++ b/man/dashboardWidgetsCreate.Rd @@ -18,7 +18,7 @@ dashboardWidgetsCreate( \item{dashboard_id}{Required. Dashboard ID returned by :method:dashboards/create.} -\item{options}{Required.} +\item{options}{Required. This field has no description yet.} \item{width}{Required. Width of a widget.} diff --git a/man/dashboardWidgetsUpdate.Rd b/man/dashboardWidgetsUpdate.Rd index c4d258d4..2bccc282 100644 --- a/man/dashboardWidgetsUpdate.Rd +++ b/man/dashboardWidgetsUpdate.Rd @@ -21,7 +21,7 @@ dashboardWidgetsUpdate( \item{dashboard_id}{Required. Dashboard ID returned by :method:dashboards/create.} -\item{options}{Required.} +\item{options}{Required. This field has no description yet.} \item{width}{Required. Width of a widget.} diff --git a/man/dashboardsCreate.Rd b/man/dashboardsCreate.Rd index b55c8d5c..51cc81dc 100644 --- a/man/dashboardsCreate.Rd +++ b/man/dashboardsCreate.Rd @@ -27,7 +27,7 @@ dashboardsCreate( \item{run_as_role}{Sets the \strong{Run as} role for the object.} -\item{tags}{} +\item{tags}{This field has no description yet.} } \description{ Create a dashboard object. diff --git a/man/dashboardsDelete.Rd b/man/dashboardsDelete.Rd index 685062b1..98fc4f06 100644 --- a/man/dashboardsDelete.Rd +++ b/man/dashboardsDelete.Rd @@ -9,7 +9,7 @@ dashboardsDelete(client, dashboard_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{dashboard_id}{Required.} +\item{dashboard_id}{Required. This field has no description yet.} } \description{ Moves a dashboard to the trash. Trashed dashboards do not appear in list diff --git a/man/dashboardsGet.Rd b/man/dashboardsGet.Rd index ca57a9db..bc6a7faf 100644 --- a/man/dashboardsGet.Rd +++ b/man/dashboardsGet.Rd @@ -9,7 +9,7 @@ dashboardsGet(client, dashboard_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{dashboard_id}{Required.} +\item{dashboard_id}{Required. This field has no description yet.} } \description{ Returns a JSON representation of a dashboard object, including its diff --git a/man/dashboardsRestore.Rd b/man/dashboardsRestore.Rd index f5925e7e..6f524f8c 100644 --- a/man/dashboardsRestore.Rd +++ b/man/dashboardsRestore.Rd @@ -9,7 +9,7 @@ dashboardsRestore(client, dashboard_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{dashboard_id}{Required.} +\item{dashboard_id}{Required. This field has no description yet.} } \description{ A restored dashboard appears in list views and searches and can be shared. diff --git a/man/dashboardsUpdate.Rd b/man/dashboardsUpdate.Rd index a842d681..49d47532 100644 --- a/man/dashboardsUpdate.Rd +++ b/man/dashboardsUpdate.Rd @@ -9,7 +9,7 @@ dashboardsUpdate(client, dashboard_id, name = NULL, run_as_role = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{dashboard_id}{Required.} +\item{dashboard_id}{Required. This field has no description yet.} \item{name}{The title of this dashboard that appears in list views and at the top of the dashboard page.} diff --git a/man/dataSourcesList.Rd b/man/dataSourcesList.Rd index ac5c2ffc..e4951710 100644 --- a/man/dataSourcesList.Rd +++ b/man/dataSourcesList.Rd @@ -6,8 +6,11 @@ \usage{ dataSourcesList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Retrieves a full list of SQL warehouses available in this workspace. All fields that appear in this API response are enumerated for clarity. However, -you need only a SQL warehouse's \code{id} to create new queries against it.#' +you need only a SQL warehouse's \code{id} to create new queries against it. } diff --git a/man/dbsqlPermissionsSet.Rd b/man/dbsqlPermissionsSet.Rd index ffb74a31..ab832651 100644 --- a/man/dbsqlPermissionsSet.Rd +++ b/man/dbsqlPermissionsSet.Rd @@ -13,7 +13,7 @@ dbsqlPermissionsSet(client, object_type, object_id, access_control_list = NULL) \item{object_id}{Required. Object ID.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets the access control list (ACL) for a specified object. This operation diff --git a/man/esmEnablementUpdate.Rd b/man/esmEnablementUpdate.Rd index 86361467..6ab4529d 100644 --- a/man/esmEnablementUpdate.Rd +++ b/man/esmEnablementUpdate.Rd @@ -11,7 +11,7 @@ esmEnablementUpdate(client, allow_missing, setting, field_mask) \item{allow_missing}{Required. This should always be set to true for Settings API.} -\item{setting}{Required.} +\item{setting}{Required. This field has no description yet.} \item{field_mask}{Required. Field mask is required to be passed into the PATCH request.} } diff --git a/man/experimentsGetHistory.Rd b/man/experimentsGetHistory.Rd index 5944dfc5..46719f96 100644 --- a/man/experimentsGetHistory.Rd +++ b/man/experimentsGetHistory.Rd @@ -24,7 +24,7 @@ experimentsGetHistory( \item{run_id}{ID of the run from which to fetch metric values.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run from which to fetch metric values.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run from which to fetch metric values.} } \value{ \code{data.frame} with all of the response pages. diff --git a/man/experimentsGetRun.Rd b/man/experimentsGetRun.Rd index d6a8bf1c..706f6371 100644 --- a/man/experimentsGetRun.Rd +++ b/man/experimentsGetRun.Rd @@ -11,7 +11,7 @@ experimentsGetRun(client, run_id, run_uuid = NULL) \item{run_id}{Required. ID of the run to fetch.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run to fetch.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run to fetch.} } \description{ Gets the metadata, metrics, params, and tags for a run. In the case where diff --git a/man/experimentsListArtifacts.Rd b/man/experimentsListArtifacts.Rd index a4e917f1..145a38fd 100644 --- a/man/experimentsListArtifacts.Rd +++ b/man/experimentsListArtifacts.Rd @@ -21,7 +21,7 @@ experimentsListArtifacts( \item{run_id}{ID of the run whose artifacts to list.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run whose artifacts to list.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run whose artifacts to list.} } \value{ \code{data.frame} with all of the response pages. diff --git a/man/experimentsLogMetric.Rd b/man/experimentsLogMetric.Rd index 2af1323b..08811edd 100644 --- a/man/experimentsLogMetric.Rd +++ b/man/experimentsLogMetric.Rd @@ -25,7 +25,7 @@ experimentsLogMetric( \item{run_id}{ID of the run under which to log the metric.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run under which to log the metric.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run under which to log the metric.} \item{step}{Step at which to log the metric.} } diff --git a/man/experimentsLogParam.Rd b/man/experimentsLogParam.Rd index 50af1ec0..0fbe4883 100644 --- a/man/experimentsLogParam.Rd +++ b/man/experimentsLogParam.Rd @@ -15,7 +15,7 @@ experimentsLogParam(client, key, value, run_id = NULL, run_uuid = NULL) \item{run_id}{ID of the run under which to log the param.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run under which to log the param.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run under which to log the param.} } \description{ Logs a param used for a run. A param is a key-value pair (string key, string diff --git a/man/experimentsSetPermissions.Rd b/man/experimentsSetPermissions.Rd index 02ab81b0..9b3df7de 100644 --- a/man/experimentsSetPermissions.Rd +++ b/man/experimentsSetPermissions.Rd @@ -11,7 +11,7 @@ experimentsSetPermissions(client, experiment_id, access_control_list = NULL) \item{experiment_id}{Required. The experiment for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on an experiment. Experiments can inherit permissions from diff --git a/man/experimentsSetTag.Rd b/man/experimentsSetTag.Rd index 5469bef7..dd0843e2 100644 --- a/man/experimentsSetTag.Rd +++ b/man/experimentsSetTag.Rd @@ -15,7 +15,7 @@ experimentsSetTag(client, key, value, run_id = NULL, run_uuid = NULL) \item{run_id}{ID of the run under which to log the tag.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run under which to log the tag.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run under which to log the tag.} } \description{ Sets a tag on a run. Tags are run metadata that can be updated during a run diff --git a/man/experimentsUpdatePermissions.Rd b/man/experimentsUpdatePermissions.Rd index 1dcb12f8..5fb8acfe 100644 --- a/man/experimentsUpdatePermissions.Rd +++ b/man/experimentsUpdatePermissions.Rd @@ -11,7 +11,7 @@ experimentsUpdatePermissions(client, experiment_id, access_control_list = NULL) \item{experiment_id}{Required. The experiment for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on an experiment. Experiments can inherit permissions diff --git a/man/experimentsUpdateRun.Rd b/man/experimentsUpdateRun.Rd index 4d8b9570..0cbb95d3 100644 --- a/man/experimentsUpdateRun.Rd +++ b/man/experimentsUpdateRun.Rd @@ -19,7 +19,7 @@ experimentsUpdateRun( \item{run_id}{ID of the run to update.} -\item{run_uuid}{\link{Deprecated, use run_id instead} ID of the run to update.} +\item{run_uuid}{Deprecated, use run_id instead. ID of the run to update.} \item{status}{Updated status of the run.} } diff --git a/man/filesUpload.Rd b/man/filesUpload.Rd index daf9d831..c581457f 100644 --- a/man/filesUpload.Rd +++ b/man/filesUpload.Rd @@ -11,7 +11,7 @@ filesUpload(client, file_path, contents, overwrite = NULL) \item{file_path}{Required. The absolute path of the file.} -\item{contents}{} +\item{contents}{This field has no description yet.} \item{overwrite}{If true, an existing file will be overwritten.} } diff --git a/man/gitCredentialsList.Rd b/man/gitCredentialsList.Rd index ec911b06..70e54119 100644 --- a/man/gitCredentialsList.Rd +++ b/man/gitCredentialsList.Rd @@ -6,10 +6,13 @@ \usage{ gitCredentialsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ Lists the calling user's Git credentials. One credential per user is -supported.#' +supported. } diff --git a/man/globalInitScriptsList.Rd b/man/globalInitScriptsList.Rd index b9f4845f..a828e74b 100644 --- a/man/globalInitScriptsList.Rd +++ b/man/globalInitScriptsList.Rd @@ -6,11 +6,14 @@ \usage{ globalInitScriptsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ Get a list of all global init scripts for this workspace. This returns all properties for each script but \strong{not} the script contents. To retrieve the -contents of a script, use the \href{:method:globalinitscripts/get}{get a global init script} operation.#' +contents of a script, use the \href{:method:globalinitscripts/get}{get a global init script} operation. } diff --git a/man/groupsCreate.Rd b/man/groupsCreate.Rd index 3f0321d2..fd2b1df9 100644 --- a/man/groupsCreate.Rd +++ b/man/groupsCreate.Rd @@ -24,13 +24,13 @@ groupsCreate( \item{entitlements}{Entitlements assigned to the group.} -\item{external_id}{} +\item{external_id}{This field has no description yet.} -\item{groups}{} +\item{groups}{This field has no description yet.} \item{id}{Databricks group ID.} -\item{members}{} +\item{members}{This field has no description yet.} \item{meta}{Container for the group identifier.} diff --git a/man/groupsPatch.Rd b/man/groupsPatch.Rd index a5fca33e..88c184fd 100644 --- a/man/groupsPatch.Rd +++ b/man/groupsPatch.Rd @@ -11,7 +11,7 @@ groupsPatch(client, id, operations = NULL, schemas = NULL) \item{id}{Required. Unique ID for a group in the Databricks workspace.} -\item{operations}{} +\item{operations}{This field has no description yet.} \item{schemas}{The schema of the patch request.} } diff --git a/man/groupsUpdate.Rd b/man/groupsUpdate.Rd index 2681da27..b8dd4035 100644 --- a/man/groupsUpdate.Rd +++ b/man/groupsUpdate.Rd @@ -26,11 +26,11 @@ groupsUpdate( \item{entitlements}{Entitlements assigned to the group.} -\item{external_id}{} +\item{external_id}{This field has no description yet.} -\item{groups}{} +\item{groups}{This field has no description yet.} -\item{members}{} +\item{members}{This field has no description yet.} \item{meta}{Container for the group identifier.} diff --git a/man/instancePoolsList.Rd b/man/instancePoolsList.Rd index fe87ecbc..672a19a3 100644 --- a/man/instancePoolsList.Rd +++ b/man/instancePoolsList.Rd @@ -6,9 +6,12 @@ \usage{ instancePoolsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ -Gets a list of instance pools with their statistics.#' +Gets a list of instance pools with their statistics. } diff --git a/man/instancePoolsSetPermissions.Rd b/man/instancePoolsSetPermissions.Rd index c596c30f..0485180f 100644 --- a/man/instancePoolsSetPermissions.Rd +++ b/man/instancePoolsSetPermissions.Rd @@ -15,7 +15,7 @@ instancePoolsSetPermissions( \item{instance_pool_id}{Required. The instance pool for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on an instance pool. Instance pools can inherit permissions diff --git a/man/instancePoolsUpdatePermissions.Rd b/man/instancePoolsUpdatePermissions.Rd index af8a1acf..99354aa1 100644 --- a/man/instancePoolsUpdatePermissions.Rd +++ b/man/instancePoolsUpdatePermissions.Rd @@ -15,7 +15,7 @@ instancePoolsUpdatePermissions( \item{instance_pool_id}{Required. The instance pool for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on an instance pool. Instance pools can inherit diff --git a/man/instanceProfilesList.Rd b/man/instanceProfilesList.Rd index 67870fe7..11afd5b4 100644 --- a/man/instanceProfilesList.Rd +++ b/man/instanceProfilesList.Rd @@ -6,6 +6,9 @@ \usage{ instanceProfilesList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } @@ -13,5 +16,5 @@ instanceProfilesList(client) List the instance profiles that the calling user can use to launch a cluster. } \details{ -This API is available to all users.#' +This API is available to all users. } diff --git a/man/ipAccessListsCreate.Rd b/man/ipAccessListsCreate.Rd index 13feb8f6..4ea9e40e 100644 --- a/man/ipAccessListsCreate.Rd +++ b/man/ipAccessListsCreate.Rd @@ -13,7 +13,7 @@ ipAccessListsCreate(client, label, list_type, ip_addresses = NULL) \item{list_type}{Required. Type of IP access list.} -\item{ip_addresses}{} +\item{ip_addresses}{This field has no description yet.} } \description{ Creates an IP access list for this workspace. diff --git a/man/ipAccessListsList.Rd b/man/ipAccessListsList.Rd index e6df46b5..0e27cf2d 100644 --- a/man/ipAccessListsList.Rd +++ b/man/ipAccessListsList.Rd @@ -6,9 +6,12 @@ \usage{ ipAccessListsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ -Gets all IP access lists for the specified workspace.#' +Gets all IP access lists for the specified workspace. } diff --git a/man/ipAccessListsReplace.Rd b/man/ipAccessListsReplace.Rd index 830671cd..c1a6bad0 100644 --- a/man/ipAccessListsReplace.Rd +++ b/man/ipAccessListsReplace.Rd @@ -24,7 +24,7 @@ ipAccessListsReplace( \item{enabled}{Required. Specifies whether this IP access list is enabled.} -\item{ip_addresses}{} +\item{ip_addresses}{This field has no description yet.} } \description{ Replaces an IP access list, specified by its ID. diff --git a/man/ipAccessListsUpdate.Rd b/man/ipAccessListsUpdate.Rd index 35c93ae6..aef947f7 100644 --- a/man/ipAccessListsUpdate.Rd +++ b/man/ipAccessListsUpdate.Rd @@ -20,7 +20,7 @@ ipAccessListsUpdate( \item{enabled}{Specifies whether this IP access list is enabled.} -\item{ip_addresses}{} +\item{ip_addresses}{This field has no description yet.} \item{label}{Label for the IP access list.} diff --git a/man/jobsCancelRun.Rd b/man/jobsCancelRun.Rd index 01d199b5..c5f96a55 100644 --- a/man/jobsCancelRun.Rd +++ b/man/jobsCancelRun.Rd @@ -10,6 +10,10 @@ jobsCancelRun(client, run_id, timeout = 20, callback = cli_reporter) \item{client}{Required. Instance of DatabricksClient()} \item{run_id}{Required. This field is required.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Jobs on Databricks reach diff --git a/man/jobsGetRun.Rd b/man/jobsGetRun.Rd index 235dc9f1..9722416b 100644 --- a/man/jobsGetRun.Rd +++ b/man/jobsGetRun.Rd @@ -21,6 +21,10 @@ jobsGetRun( \item{include_history}{Whether to include the repair history in the response.} \item{include_resolved_values}{Whether to include resolved parameter values in the response.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Jobs on Databricks reach diff --git a/man/jobsRepairRun.Rd b/man/jobsRepairRun.Rd index 3c5103a2..e8f7f491 100644 --- a/man/jobsRepairRun.Rd +++ b/man/jobsRepairRun.Rd @@ -39,7 +39,7 @@ jobsRepairRun( \item{notebook_params}{A map from keys to values for jobs with notebook task, for example \verb{'notebook_params': \{'name': 'john doe', 'age': '35'\}}.} -\item{pipeline_params}{} +\item{pipeline_params}{This field has no description yet.} \item{python_named_params}{A map from keys to values for jobs with Python wheel task, for example \verb{'python_named_params': \{'name': 'task', 'data': 'dbfs:/path/to/data.json'\}}.} @@ -54,6 +54,10 @@ jobsRepairRun( \item{spark_submit_params}{A list of parameters for jobs with spark submit task, for example \verb{'spark_submit_params': ['--class', 'org.apache.spark.examples.SparkPi']}.} \item{sql_params}{A map from keys to values for jobs with SQL task, for example \verb{'sql_params': \{'name': 'john doe', 'age': '35'\}}.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Jobs on Databricks reach diff --git a/man/jobsRunNow.Rd b/man/jobsRunNow.Rd index f58fce18..ce4277b0 100644 --- a/man/jobsRunNow.Rd +++ b/man/jobsRunNow.Rd @@ -37,7 +37,7 @@ jobsRunNow( \item{notebook_params}{A map from keys to values for jobs with notebook task, for example \verb{'notebook_params': \{'name': 'john doe', 'age': '35'\}}.} -\item{pipeline_params}{} +\item{pipeline_params}{This field has no description yet.} \item{python_named_params}{A map from keys to values for jobs with Python wheel task, for example \verb{'python_named_params': \{'name': 'task', 'data': 'dbfs:/path/to/data.json'\}}.} @@ -48,6 +48,10 @@ jobsRunNow( \item{spark_submit_params}{A list of parameters for jobs with spark submit task, for example \verb{'spark_submit_params': ['--class', 'org.apache.spark.examples.SparkPi']}.} \item{sql_params}{A map from keys to values for jobs with SQL task, for example \verb{'sql_params': \{'name': 'john doe', 'age': '35'\}}.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Jobs on Databricks reach diff --git a/man/jobsSetPermissions.Rd b/man/jobsSetPermissions.Rd index 405e7298..01062527 100644 --- a/man/jobsSetPermissions.Rd +++ b/man/jobsSetPermissions.Rd @@ -11,7 +11,7 @@ jobsSetPermissions(client, job_id, access_control_list = NULL) \item{job_id}{Required. The job for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a job. Jobs can inherit permissions from their root diff --git a/man/jobsSubmit.Rd b/man/jobsSubmit.Rd index 2426eaf4..01d2839c 100644 --- a/man/jobsSubmit.Rd +++ b/man/jobsSubmit.Rd @@ -40,11 +40,15 @@ jobsSubmit( \item{run_name}{An optional name for the run.} -\item{tasks}{} +\item{tasks}{This field has no description yet.} \item{timeout_seconds}{An optional timeout applied to each run of this job.} \item{webhook_notifications}{A collection of system notification IDs to notify when the run begins or completes.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Jobs on Databricks reach diff --git a/man/jobsUpdatePermissions.Rd b/man/jobsUpdatePermissions.Rd index 27ec60d8..194c4ece 100644 --- a/man/jobsUpdatePermissions.Rd +++ b/man/jobsUpdatePermissions.Rd @@ -11,7 +11,7 @@ jobsUpdatePermissions(client, job_id, access_control_list = NULL) \item{job_id}{Required. The job for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a job. Jobs can inherit permissions from their diff --git a/man/librariesAllClusterStatuses.Rd b/man/librariesAllClusterStatuses.Rd index c40e5cd4..79ac5240 100644 --- a/man/librariesAllClusterStatuses.Rd +++ b/man/librariesAllClusterStatuses.Rd @@ -6,9 +6,12 @@ \usage{ librariesAllClusterStatuses(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Get the status of all libraries on all clusters. A status will be available for all libraries installed on this cluster via the API or the libraries UI as well as libraries set to be installed on all clusters via the libraries -UI.#' +UI. } diff --git a/man/metastoresCurrent.Rd b/man/metastoresCurrent.Rd index d3e61fe1..0d0e2087 100644 --- a/man/metastoresCurrent.Rd +++ b/man/metastoresCurrent.Rd @@ -6,6 +6,9 @@ \usage{ metastoresCurrent(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -Gets the metastore assignment for the workspace being accessed.#' +Gets the metastore assignment for the workspace being accessed. } diff --git a/man/metastoresList.Rd b/man/metastoresList.Rd index 789aab63..7e9a471a 100644 --- a/man/metastoresList.Rd +++ b/man/metastoresList.Rd @@ -6,11 +6,14 @@ \usage{ metastoresList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ Gets an array of the available metastores (as \strong{MetastoreInfo} objects). The caller must be an admin to retrieve this info. There is no guarantee of a -specific ordering of the elements in the array.#' +specific ordering of the elements in the array. } diff --git a/man/metastoresSummary.Rd b/man/metastoresSummary.Rd index 0d6b2e99..4f354d35 100644 --- a/man/metastoresSummary.Rd +++ b/man/metastoresSummary.Rd @@ -6,7 +6,10 @@ \usage{ metastoresSummary(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Gets information about a metastore. This summary includes the storage -credential, the cloud vendor, the cloud region, and the global metastore ID.#' +credential, the cloud vendor, the cloud region, and the global metastore ID. } diff --git a/man/modelRegistryCreateWebhook.Rd b/man/modelRegistryCreateWebhook.Rd index ddfee0e4..7e84a7c3 100644 --- a/man/modelRegistryCreateWebhook.Rd +++ b/man/modelRegistryCreateWebhook.Rd @@ -21,9 +21,9 @@ modelRegistryCreateWebhook( \item{description}{User-specified description for the webhook.} -\item{http_url_spec}{} +\item{http_url_spec}{This field has no description yet.} -\item{job_spec}{} +\item{job_spec}{This field has no description yet.} \item{model_name}{Name of the model whose events would trigger this webhook.} diff --git a/man/modelRegistryDeleteComment.Rd b/man/modelRegistryDeleteComment.Rd index 3341a363..363779c3 100644 --- a/man/modelRegistryDeleteComment.Rd +++ b/man/modelRegistryDeleteComment.Rd @@ -9,7 +9,7 @@ modelRegistryDeleteComment(client, id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{id}{Required.} +\item{id}{Required. This field has no description yet.} } \description{ Deletes a comment on a model version. diff --git a/man/modelRegistryGetModel.Rd b/man/modelRegistryGetModel.Rd index 82e6a516..dd6cc5bb 100644 --- a/man/modelRegistryGetModel.Rd +++ b/man/modelRegistryGetModel.Rd @@ -13,9 +13,9 @@ modelRegistryGetModel(client, name) } \description{ Get the details of a model. This is a Databricks workspace version of the -\link{MLflow endpoint} that also returns the model's Databricks workspace ID and +MLflow endpoint that also returns the model's Databricks workspace ID and the permission level of the requesting user on the model. } \details{ - +MLflow endpoint: https://www.mlflow.org/docs/latest/rest-api.html#get-registeredmodel } diff --git a/man/modelRegistrySetPermissions.Rd b/man/modelRegistrySetPermissions.Rd index b3fd171b..b2d8b329 100644 --- a/man/modelRegistrySetPermissions.Rd +++ b/man/modelRegistrySetPermissions.Rd @@ -15,7 +15,7 @@ modelRegistrySetPermissions( \item{registered_model_id}{Required. The registered model for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a registered model. Registered models can inherit diff --git a/man/modelRegistryTransitionStage.Rd b/man/modelRegistryTransitionStage.Rd index 3414d70a..962b3fc1 100644 --- a/man/modelRegistryTransitionStage.Rd +++ b/man/modelRegistryTransitionStage.Rd @@ -28,9 +28,9 @@ modelRegistryTransitionStage( } \description{ Transition a model version's stage. This is a Databricks workspace version of -the \link{MLflow endpoint} that also accepts a comment associated with the +the MLflow endpoint that also accepts a comment associated with the transition to be recorded.', } \details{ - +MLflow endpoint: https://www.mlflow.org/docs/latest/rest-api.html#transition-modelversion-stage } diff --git a/man/modelRegistryUpdatePermissions.Rd b/man/modelRegistryUpdatePermissions.Rd index 6fab5029..e732ab36 100644 --- a/man/modelRegistryUpdatePermissions.Rd +++ b/man/modelRegistryUpdatePermissions.Rd @@ -15,7 +15,7 @@ modelRegistryUpdatePermissions( \item{registered_model_id}{Required. The registered model for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a registered model. Registered models can inherit diff --git a/man/modelRegistryUpdateWebhook.Rd b/man/modelRegistryUpdateWebhook.Rd index 37371697..36c5bca0 100644 --- a/man/modelRegistryUpdateWebhook.Rd +++ b/man/modelRegistryUpdateWebhook.Rd @@ -23,9 +23,9 @@ modelRegistryUpdateWebhook( \item{events}{Events that can trigger a registry webhook: * \code{MODEL_VERSION_CREATED}: A new model version was created for the associated model.} -\item{http_url_spec}{} +\item{http_url_spec}{This field has no description yet.} -\item{job_spec}{} +\item{job_spec}{This field has no description yet.} \item{status}{Enable or disable triggering the webhook, or put the webhook into test mode.} } diff --git a/man/permissionsSet.Rd b/man/permissionsSet.Rd index c3296771..ca9023d5 100644 --- a/man/permissionsSet.Rd +++ b/man/permissionsSet.Rd @@ -18,7 +18,7 @@ permissionsSet( \item{request_object_id}{Required. The id of the request object.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on an object. Objects can inherit permissions from their diff --git a/man/permissionsUpdate.Rd b/man/permissionsUpdate.Rd index 5c8a5a79..99fc0062 100644 --- a/man/permissionsUpdate.Rd +++ b/man/permissionsUpdate.Rd @@ -18,7 +18,7 @@ permissionsUpdate( \item{request_object_id}{Required. The id of the request object.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on an object. Objects can inherit permissions from diff --git a/man/pipelinesCreate.Rd b/man/pipelinesCreate.Rd index 0a3d6cbd..5923aadb 100644 --- a/man/pipelinesCreate.Rd +++ b/man/pipelinesCreate.Rd @@ -44,7 +44,7 @@ pipelinesCreate( \item{development}{Whether the pipeline is in Development mode.} -\item{dry_run}{} +\item{dry_run}{This field has no description yet.} \item{edition}{Pipeline product edition.} diff --git a/man/pipelinesDelete.Rd b/man/pipelinesDelete.Rd index 21478d27..d65d11b3 100644 --- a/man/pipelinesDelete.Rd +++ b/man/pipelinesDelete.Rd @@ -9,7 +9,7 @@ pipelinesDelete(client, pipeline_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{pipeline_id}{Required.} +\item{pipeline_id}{Required. This field has no description yet.} } \description{ Deletes a pipeline. diff --git a/man/pipelinesGet.Rd b/man/pipelinesGet.Rd index ca4f8a7d..1919e34d 100644 --- a/man/pipelinesGet.Rd +++ b/man/pipelinesGet.Rd @@ -9,7 +9,7 @@ pipelinesGet(client, pipeline_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{pipeline_id}{Required.} +\item{pipeline_id}{Required. This field has no description yet.} } \description{ Get a pipeline. diff --git a/man/pipelinesListPipelineEvents.Rd b/man/pipelinesListPipelineEvents.Rd index 6415748e..057d027f 100644 --- a/man/pipelinesListPipelineEvents.Rd +++ b/man/pipelinesListPipelineEvents.Rd @@ -16,13 +16,13 @@ pipelinesListPipelineEvents( \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{pipeline_id}{Required.} +\item{pipeline_id}{Required. This field has no description yet.} \item{filter}{Criteria to select a subset of results, expressed using a SQL-like syntax.} \item{max_results}{Max number of entries to return in a single page.} -\item{order_by}{A string indicating a sort order by timestamp for the results, for example, \link{'timestamp asc'}.} +\item{order_by}{A string indicating a sort order by timestamp for the results, for example, timestamp asc.} \item{page_token}{Page token returned by previous call.} } diff --git a/man/pipelinesSetPermissions.Rd b/man/pipelinesSetPermissions.Rd index 46a18fec..4fd9ee3b 100644 --- a/man/pipelinesSetPermissions.Rd +++ b/man/pipelinesSetPermissions.Rd @@ -11,7 +11,7 @@ pipelinesSetPermissions(client, pipeline_id, access_control_list = NULL) \item{pipeline_id}{Required. The pipeline for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a pipeline. Pipelines can inherit permissions from their diff --git a/man/pipelinesStartUpdate.Rd b/man/pipelinesStartUpdate.Rd index ceaa4305..13090e27 100644 --- a/man/pipelinesStartUpdate.Rd +++ b/man/pipelinesStartUpdate.Rd @@ -17,9 +17,9 @@ pipelinesStartUpdate( \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{pipeline_id}{Required.} +\item{pipeline_id}{Required. This field has no description yet.} -\item{cause}{} +\item{cause}{This field has no description yet.} \item{full_refresh}{If true, this update will reset all tables before running.} diff --git a/man/pipelinesStop.Rd b/man/pipelinesStop.Rd index 2f4438af..25e02641 100644 --- a/man/pipelinesStop.Rd +++ b/man/pipelinesStop.Rd @@ -9,7 +9,11 @@ pipelinesStop(client, pipeline_id, timeout = 20, callback = cli_reporter) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{pipeline_id}{Required.} +\item{pipeline_id}{Required. This field has no description yet.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Pipelines on Databricks reach diff --git a/man/pipelinesUpdatePermissions.Rd b/man/pipelinesUpdatePermissions.Rd index 27b2b4f5..26ab8a0c 100644 --- a/man/pipelinesUpdatePermissions.Rd +++ b/man/pipelinesUpdatePermissions.Rd @@ -11,7 +11,7 @@ pipelinesUpdatePermissions(client, pipeline_id, access_control_list = NULL) \item{pipeline_id}{Required. The pipeline for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a pipeline. Pipelines can inherit permissions from diff --git a/man/policyFamiliesGet.Rd b/man/policyFamiliesGet.Rd index f28b3ad9..30a09b43 100644 --- a/man/policyFamiliesGet.Rd +++ b/man/policyFamiliesGet.Rd @@ -9,7 +9,7 @@ policyFamiliesGet(client, policy_family_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{policy_family_id}{Required.} +\item{policy_family_id}{Required. This field has no description yet.} } \description{ Retrieve the information for an policy family based on its identifier. diff --git a/man/queriesDelete.Rd b/man/queriesDelete.Rd index 86a180dc..af12d62f 100644 --- a/man/queriesDelete.Rd +++ b/man/queriesDelete.Rd @@ -9,7 +9,7 @@ queriesDelete(client, query_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{query_id}{Required.} +\item{query_id}{Required. This field has no description yet.} } \description{ Moves a query to the trash. Trashed queries immediately disappear from diff --git a/man/queriesGet.Rd b/man/queriesGet.Rd index 39f67f20..511cd3c6 100644 --- a/man/queriesGet.Rd +++ b/man/queriesGet.Rd @@ -9,7 +9,7 @@ queriesGet(client, query_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{query_id}{Required.} +\item{query_id}{Required. This field has no description yet.} } \description{ Retrieve a query object definition along with contextual permissions diff --git a/man/queriesRestore.Rd b/man/queriesRestore.Rd index d9fa2f30..748b9e4d 100644 --- a/man/queriesRestore.Rd +++ b/man/queriesRestore.Rd @@ -9,7 +9,7 @@ queriesRestore(client, query_id) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{query_id}{Required.} +\item{query_id}{Required. This field has no description yet.} } \description{ Restore a query that has been moved to the trash. A restored query appears in diff --git a/man/queriesUpdate.Rd b/man/queriesUpdate.Rd index 8248cbfd..4fad2113 100644 --- a/man/queriesUpdate.Rd +++ b/man/queriesUpdate.Rd @@ -18,7 +18,7 @@ queriesUpdate( \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{query_id}{Required.} +\item{query_id}{Required. This field has no description yet.} \item{data_source_id}{Data source ID maps to the ID of the data source used by the resource and is distinct from the warehouse ID.} diff --git a/man/queryVisualizationsUpdate.Rd b/man/queryVisualizationsUpdate.Rd index 2dbc599f..695df94e 100644 --- a/man/queryVisualizationsUpdate.Rd +++ b/man/queryVisualizationsUpdate.Rd @@ -20,7 +20,7 @@ queryVisualizationsUpdate( \item{id}{The UUID for this visualization.} -\item{created_at}{} +\item{created_at}{This field has no description yet.} \item{description}{A short description of this visualization.} @@ -30,7 +30,7 @@ queryVisualizationsUpdate( \item{type}{The type of visualization: chart, table, pivot table, and so on.} -\item{updated_at}{} +\item{updated_at}{This field has no description yet.} } \description{ Edit existing visualization. diff --git a/man/reposSetPermissions.Rd b/man/reposSetPermissions.Rd index 50394a84..8b6bbf58 100644 --- a/man/reposSetPermissions.Rd +++ b/man/reposSetPermissions.Rd @@ -11,7 +11,7 @@ reposSetPermissions(client, repo_id, access_control_list = NULL) \item{repo_id}{Required. The repo for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a repo. Repos can inherit permissions from their root diff --git a/man/reposUpdatePermissions.Rd b/man/reposUpdatePermissions.Rd index d74aa3a7..a02961f9 100644 --- a/man/reposUpdatePermissions.Rd +++ b/man/reposUpdatePermissions.Rd @@ -11,7 +11,7 @@ reposUpdatePermissions(client, repo_id, access_control_list = NULL) \item{repo_id}{Required. The repo for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a repo. Repos can inherit permissions from their diff --git a/man/restrictWorkspaceAdminsUpdate.Rd b/man/restrictWorkspaceAdminsUpdate.Rd index 73d73a1e..5d2afd21 100644 --- a/man/restrictWorkspaceAdminsUpdate.Rd +++ b/man/restrictWorkspaceAdminsUpdate.Rd @@ -11,7 +11,7 @@ restrictWorkspaceAdminsUpdate(client, allow_missing, setting, field_mask) \item{allow_missing}{Required. This should always be set to true for Settings API.} -\item{setting}{Required.} +\item{setting}{Required. This field has no description yet.} \item{field_mask}{Required. Field mask is required to be passed into the PATCH request.} } diff --git a/man/secretsListScopes.Rd b/man/secretsListScopes.Rd index f7d3b294..6c4ddde2 100644 --- a/man/secretsListScopes.Rd +++ b/man/secretsListScopes.Rd @@ -6,6 +6,9 @@ \usage{ secretsListScopes(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } @@ -14,5 +17,5 @@ Lists all secret scopes available in the workspace. } \details{ Throws \code{PERMISSION_DENIED} if the user does not have permission to make this -API call.#' +API call. } diff --git a/man/servicePrincipalsCreate.Rd b/man/servicePrincipalsCreate.Rd index 353e9216..79664d77 100644 --- a/man/servicePrincipalsCreate.Rd +++ b/man/servicePrincipalsCreate.Rd @@ -28,9 +28,9 @@ servicePrincipalsCreate( \item{entitlements}{Entitlements assigned to the service principal.} -\item{external_id}{} +\item{external_id}{This field has no description yet.} -\item{groups}{} +\item{groups}{This field has no description yet.} \item{id}{Databricks service principal ID.} diff --git a/man/servicePrincipalsPatch.Rd b/man/servicePrincipalsPatch.Rd index 3f2165cc..b035a979 100644 --- a/man/servicePrincipalsPatch.Rd +++ b/man/servicePrincipalsPatch.Rd @@ -11,7 +11,7 @@ servicePrincipalsPatch(client, id, operations = NULL, schemas = NULL) \item{id}{Required. Unique ID for a service principal in the Databricks workspace.} -\item{operations}{} +\item{operations}{This field has no description yet.} \item{schemas}{The schema of the patch request.} } diff --git a/man/servicePrincipalsUpdate.Rd b/man/servicePrincipalsUpdate.Rd index 362cac06..64d73133 100644 --- a/man/servicePrincipalsUpdate.Rd +++ b/man/servicePrincipalsUpdate.Rd @@ -30,9 +30,9 @@ servicePrincipalsUpdate( \item{entitlements}{Entitlements assigned to the service principal.} -\item{external_id}{} +\item{external_id}{This field has no description yet.} -\item{groups}{} +\item{groups}{This field has no description yet.} \item{roles}{Corresponds to AWS instance profile/arn role.} diff --git a/man/servingEndpointsCreate.Rd b/man/servingEndpointsCreate.Rd index 447b9517..784b0109 100644 --- a/man/servingEndpointsCreate.Rd +++ b/man/servingEndpointsCreate.Rd @@ -24,6 +24,10 @@ servingEndpointsCreate( \item{rate_limits}{Rate limits to be applied to the serving endpoint.} \item{tags}{Tags to be attached to the serving endpoint and automatically propagated to billing logs.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Serving Endpoints on Databricks reach diff --git a/man/servingEndpointsList.Rd b/man/servingEndpointsList.Rd index 3e64955d..7c3d4d73 100644 --- a/man/servingEndpointsList.Rd +++ b/man/servingEndpointsList.Rd @@ -2,13 +2,16 @@ % Please edit documentation in R/serving_endpoints.R \name{servingEndpointsList} \alias{servingEndpointsList} -\title{Get all serving endpoints.#'} +\title{Get all serving endpoints.} \usage{ servingEndpointsList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ -Get all serving endpoints.#' +Get all serving endpoints. } diff --git a/man/servingEndpointsSetPermissions.Rd b/man/servingEndpointsSetPermissions.Rd index 98b37df4..34440150 100644 --- a/man/servingEndpointsSetPermissions.Rd +++ b/man/servingEndpointsSetPermissions.Rd @@ -15,7 +15,7 @@ servingEndpointsSetPermissions( \item{serving_endpoint_id}{Required. The serving endpoint for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a serving endpoint. Serving endpoints can inherit diff --git a/man/servingEndpointsUpdateConfig.Rd b/man/servingEndpointsUpdateConfig.Rd index d2eed47e..569512b5 100644 --- a/man/servingEndpointsUpdateConfig.Rd +++ b/man/servingEndpointsUpdateConfig.Rd @@ -27,6 +27,10 @@ servingEndpointsUpdateConfig( \item{served_models}{(Deprecated, use served_entities instead) A list of served models for the endpoint to serve.} \item{traffic_config}{The traffic config defining how invocations to the serving endpoint should be routed.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Serving Endpoints on Databricks reach diff --git a/man/servingEndpointsUpdatePermissions.Rd b/man/servingEndpointsUpdatePermissions.Rd index 0215ec5e..54f631af 100644 --- a/man/servingEndpointsUpdatePermissions.Rd +++ b/man/servingEndpointsUpdatePermissions.Rd @@ -15,7 +15,7 @@ servingEndpointsUpdatePermissions( \item{serving_endpoint_id}{Required. The serving endpoint for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a serving endpoint. Serving endpoints can inherit diff --git a/man/sharesList.Rd b/man/sharesList.Rd index e91bf6cf..bfa1dca4 100644 --- a/man/sharesList.Rd +++ b/man/sharesList.Rd @@ -6,11 +6,14 @@ \usage{ sharesList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ Gets an array of data object shares from the metastore. The caller must be a metastore admin or the owner of the share. There is no guarantee of a -specific ordering of the elements in the array.#' +specific ordering of the elements in the array. } diff --git a/man/statementExecutionGetStatementResultChunkN.Rd b/man/statementExecutionGetStatementResultChunkN.Rd index a290d461..1b105f6c 100644 --- a/man/statementExecutionGetStatementResultChunkN.Rd +++ b/man/statementExecutionGetStatementResultChunkN.Rd @@ -11,7 +11,7 @@ statementExecutionGetStatementResultChunkN(client, statement_id, chunk_index) \item{statement_id}{Required. The statement ID is returned upon successfully submitting a SQL statement, and is a required reference for all subsequent calls.} -\item{chunk_index}{Required.} +\item{chunk_index}{Required. This field has no description yet.} } \description{ After the statement execution has \code{SUCCEEDED}, this request can be used to diff --git a/man/tablesUpdate.Rd b/man/tablesUpdate.Rd index f0d99b80..04a3960b 100644 --- a/man/tablesUpdate.Rd +++ b/man/tablesUpdate.Rd @@ -11,7 +11,7 @@ tablesUpdate(client, full_name, owner = NULL) \item{full_name}{Required. Full name of the table.} -\item{owner}{} +\item{owner}{This field has no description yet.} } \description{ Change the owner of the table. The caller must be the owner of the parent diff --git a/man/tokenManagementGetPermissionLevels.Rd b/man/tokenManagementGetPermissionLevels.Rd index 003ae8b9..4ed42b6e 100644 --- a/man/tokenManagementGetPermissionLevels.Rd +++ b/man/tokenManagementGetPermissionLevels.Rd @@ -6,6 +6,9 @@ \usage{ tokenManagementGetPermissionLevels(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -Gets the permission levels that a user can have on an object.#' +Gets the permission levels that a user can have on an object. } diff --git a/man/tokenManagementGetPermissions.Rd b/man/tokenManagementGetPermissions.Rd index 3ef65e3d..c2fee2d0 100644 --- a/man/tokenManagementGetPermissions.Rd +++ b/man/tokenManagementGetPermissions.Rd @@ -6,7 +6,10 @@ \usage{ tokenManagementGetPermissions(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Gets the permissions of all tokens. Tokens can inherit permissions from their -root object.#' +root object. } diff --git a/man/tokenManagementSetPermissions.Rd b/man/tokenManagementSetPermissions.Rd index a7c764c5..fffdecd6 100644 --- a/man/tokenManagementSetPermissions.Rd +++ b/man/tokenManagementSetPermissions.Rd @@ -9,7 +9,7 @@ tokenManagementSetPermissions(client, access_control_list = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on all tokens. Tokens can inherit permissions from their diff --git a/man/tokenManagementUpdatePermissions.Rd b/man/tokenManagementUpdatePermissions.Rd index 73b81272..b6fd0bb0 100644 --- a/man/tokenManagementUpdatePermissions.Rd +++ b/man/tokenManagementUpdatePermissions.Rd @@ -9,7 +9,7 @@ tokenManagementUpdatePermissions(client, access_control_list = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on all tokens. Tokens can inherit permissions from diff --git a/man/tokensList.Rd b/man/tokensList.Rd index 4e08aa69..a989b9eb 100644 --- a/man/tokensList.Rd +++ b/man/tokensList.Rd @@ -6,9 +6,12 @@ \usage{ tokensList(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \value{ \code{data.frame} with all of the response pages. } \description{ -Lists all the valid tokens for a user-workspace pair.#' +Lists all the valid tokens for a user-workspace pair. } diff --git a/man/usersCreate.Rd b/man/usersCreate.Rd index 683f0799..4669a1d2 100644 --- a/man/usersCreate.Rd +++ b/man/usersCreate.Rd @@ -32,11 +32,11 @@ usersCreate( \item{external_id}{External ID is not currently supported.} -\item{groups}{} +\item{groups}{This field has no description yet.} \item{id}{Databricks user ID.} -\item{name}{} +\item{name}{This field has no description yet.} \item{roles}{Corresponds to AWS instance profile/arn role.} diff --git a/man/usersGetPermissionLevels.Rd b/man/usersGetPermissionLevels.Rd index 799a89e2..f2cb6276 100644 --- a/man/usersGetPermissionLevels.Rd +++ b/man/usersGetPermissionLevels.Rd @@ -6,6 +6,9 @@ \usage{ usersGetPermissionLevels(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ -Gets the permission levels that a user can have on an object.#' +Gets the permission levels that a user can have on an object. } diff --git a/man/usersGetPermissions.Rd b/man/usersGetPermissions.Rd index 7edcc02b..8ded373f 100644 --- a/man/usersGetPermissions.Rd +++ b/man/usersGetPermissions.Rd @@ -6,7 +6,10 @@ \usage{ usersGetPermissions(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Gets the permissions of all passwords. Passwords can inherit permissions from -their root object.#' +their root object. } diff --git a/man/usersPatch.Rd b/man/usersPatch.Rd index f604fdef..5b79ce0a 100644 --- a/man/usersPatch.Rd +++ b/man/usersPatch.Rd @@ -11,7 +11,7 @@ usersPatch(client, id, operations = NULL, schemas = NULL) \item{id}{Required. Unique ID for a user in the Databricks workspace.} -\item{operations}{} +\item{operations}{This field has no description yet.} \item{schemas}{The schema of the patch request.} } diff --git a/man/usersSetPermissions.Rd b/man/usersSetPermissions.Rd index 3f7a8ca0..b70448f5 100644 --- a/man/usersSetPermissions.Rd +++ b/man/usersSetPermissions.Rd @@ -9,7 +9,7 @@ usersSetPermissions(client, access_control_list = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on all passwords. Passwords can inherit permissions from diff --git a/man/usersUpdate.Rd b/man/usersUpdate.Rd index daf38534..91b27b0d 100644 --- a/man/usersUpdate.Rd +++ b/man/usersUpdate.Rd @@ -34,9 +34,9 @@ usersUpdate( \item{external_id}{External ID is not currently supported.} -\item{groups}{} +\item{groups}{This field has no description yet.} -\item{name}{} +\item{name}{This field has no description yet.} \item{roles}{Corresponds to AWS instance profile/arn role.} diff --git a/man/usersUpdatePermissions.Rd b/man/usersUpdatePermissions.Rd index c2aa3e22..2e0720df 100644 --- a/man/usersUpdatePermissions.Rd +++ b/man/usersUpdatePermissions.Rd @@ -9,7 +9,7 @@ usersUpdatePermissions(client, access_control_list = NULL) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on all passwords. Passwords can inherit permissions diff --git a/man/vectorSearchEndpointsCreateEndpoint.Rd b/man/vectorSearchEndpointsCreateEndpoint.Rd index 145e33b0..7e15f739 100644 --- a/man/vectorSearchEndpointsCreateEndpoint.Rd +++ b/man/vectorSearchEndpointsCreateEndpoint.Rd @@ -18,6 +18,10 @@ vectorSearchEndpointsCreateEndpoint( \item{name}{Required. Name of endpoint.} \item{endpoint_type}{Required. Type of endpoint.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Vector Search Endpoints on Databricks reach diff --git a/man/volumesCreate.Rd b/man/volumesCreate.Rd index f0f11b27..73bc25ea 100644 --- a/man/volumesCreate.Rd +++ b/man/volumesCreate.Rd @@ -23,7 +23,7 @@ volumesCreate( \item{name}{Required. The name of the volume.} -\item{volume_type}{Required.} +\item{volume_type}{Required. This field has no description yet.} \item{comment}{The comment attached to the volume.} diff --git a/man/warehousesCreate.Rd b/man/warehousesCreate.Rd index a7c4f60b..f2a1b78f 100644 --- a/man/warehousesCreate.Rd +++ b/man/warehousesCreate.Rd @@ -51,6 +51,10 @@ warehousesCreate( \item{tags}{A set of key-value pairs that will be tagged on all resources (e.g., AWS instances and EBS volumes) associated with this SQL warehouse.} \item{warehouse_type}{Warehouse type: \code{PRO} or \code{CLASSIC}.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Warehouses on Databricks reach diff --git a/man/warehousesEdit.Rd b/man/warehousesEdit.Rd index 3a25bd47..7d2f7abd 100644 --- a/man/warehousesEdit.Rd +++ b/man/warehousesEdit.Rd @@ -54,6 +54,10 @@ warehousesEdit( \item{tags}{A set of key-value pairs that will be tagged on all resources (e.g., AWS instances and EBS volumes) associated with this SQL warehouse.} \item{warehouse_type}{Warehouse type: \code{PRO} or \code{CLASSIC}.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Warehouses on Databricks reach diff --git a/man/warehousesGetWorkspaceWarehouseConfig.Rd b/man/warehousesGetWorkspaceWarehouseConfig.Rd index 608fc83c..8b077293 100644 --- a/man/warehousesGetWorkspaceWarehouseConfig.Rd +++ b/man/warehousesGetWorkspaceWarehouseConfig.Rd @@ -6,7 +6,10 @@ \usage{ warehousesGetWorkspaceWarehouseConfig(client) } +\arguments{ +\item{client}{Required. Instance of DatabricksClient()} +} \description{ Gets the workspace level configuration that is shared by all SQL warehouses -in a workspace.#' +in a workspace. } diff --git a/man/warehousesSetPermissions.Rd b/man/warehousesSetPermissions.Rd index 5b46b87d..95cdded2 100644 --- a/man/warehousesSetPermissions.Rd +++ b/man/warehousesSetPermissions.Rd @@ -11,7 +11,7 @@ warehousesSetPermissions(client, warehouse_id, access_control_list = NULL) \item{warehouse_id}{Required. The SQL warehouse for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a SQL warehouse. SQL warehouses can inherit permissions diff --git a/man/warehousesStart.Rd b/man/warehousesStart.Rd index 71cb2ef1..50af94d9 100644 --- a/man/warehousesStart.Rd +++ b/man/warehousesStart.Rd @@ -10,6 +10,10 @@ warehousesStart(client, id, timeout = 20, callback = cli_reporter) \item{client}{Required. Instance of DatabricksClient()} \item{id}{Required. Required.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Warehouses on Databricks reach diff --git a/man/warehousesStop.Rd b/man/warehousesStop.Rd index 33d9a2c3..c29a212f 100644 --- a/man/warehousesStop.Rd +++ b/man/warehousesStop.Rd @@ -10,6 +10,10 @@ warehousesStop(client, id, timeout = 20, callback = cli_reporter) \item{client}{Required. Instance of DatabricksClient()} \item{id}{Required. Required.} + +\item{timeout}{Time to wait for the operation to complete in minutes.} + +\item{callback}{Function to report the status of the operation. By default, it reports to console.} } \description{ This is a long-running operation, which blocks until Warehouses on Databricks reach diff --git a/man/warehousesUpdatePermissions.Rd b/man/warehousesUpdatePermissions.Rd index bd756883..e4bd9383 100644 --- a/man/warehousesUpdatePermissions.Rd +++ b/man/warehousesUpdatePermissions.Rd @@ -11,7 +11,7 @@ warehousesUpdatePermissions(client, warehouse_id, access_control_list = NULL) \item{warehouse_id}{Required. The SQL warehouse for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a SQL warehouse. SQL warehouses can inherit diff --git a/man/workspaceConfGetStatus.Rd b/man/workspaceConfGetStatus.Rd index 57337ca5..c02b2102 100644 --- a/man/workspaceConfGetStatus.Rd +++ b/man/workspaceConfGetStatus.Rd @@ -9,7 +9,7 @@ workspaceConfGetStatus(client, keys) \arguments{ \item{client}{Required. Instance of DatabricksClient()} -\item{keys}{Required.} +\item{keys}{Required. This field has no description yet.} } \description{ Gets the configuration status for a workspace. diff --git a/man/workspaceSetPermissions.Rd b/man/workspaceSetPermissions.Rd index 29052a9c..828ddfb4 100644 --- a/man/workspaceSetPermissions.Rd +++ b/man/workspaceSetPermissions.Rd @@ -18,7 +18,7 @@ workspaceSetPermissions( \item{workspace_object_id}{Required. The workspace object for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Sets permissions on a workspace object. Workspace objects can inherit diff --git a/man/workspaceUpdatePermissions.Rd b/man/workspaceUpdatePermissions.Rd index 54f4d5e5..5bbe9810 100644 --- a/man/workspaceUpdatePermissions.Rd +++ b/man/workspaceUpdatePermissions.Rd @@ -18,7 +18,7 @@ workspaceUpdatePermissions( \item{workspace_object_id}{Required. The workspace object for which to get or manage permissions.} -\item{access_control_list}{} +\item{access_control_list}{This field has no description yet.} } \description{ Updates the permissions on a workspace object. Workspace objects can inherit