From 37aceecc38bf6ba7591c073db8cdcabea2137310 Mon Sep 17 00:00:00 2001 From: Yoshi Automation Bot Date: Fri, 21 Feb 2020 09:02:35 -0800 Subject: [PATCH] feat(dataproc): Add SparkRJob, PrestoJob, LifecycleConfig and ReservationAffinity * Add ClusterConfig#lifecycle_config (LifecycleConfig) * Add GceClusterConfig#reservation_affinity (ReservationAffinity) * Add SparkRJob * Add PrestoJob --- .../v1/autoscaling_policies_services_pb.rb | 3 +- .../dataproc/v1/cluster_controller_client.rb | 3 +- .../google/cloud/dataproc/v1/clusters_pb.rb | 25 +++++- .../cloud/dataproc/v1/clusters_services_pb.rb | 13 ++- .../doc/google/cloud/dataproc/v1/clusters.rb | 72 ++++++++++++++++- .../v1/doc/google/cloud/dataproc/v1/jobs.rb | 79 +++++++++++++++++-- .../cloud/dataproc/v1/workflow_templates.rb | 6 -- .../lib/google/cloud/dataproc/v1/jobs_pb.rb | 24 ++++++ .../cloud/dataproc/v1/jobs_services_pb.rb | 6 +- .../v1/workflow_template_service_client.rb | 4 +- .../v1/workflow_templates_services_pb.rb | 11 ++- google-cloud-dataproc/synth.metadata | 28 +++++-- 12 files changed, 229 insertions(+), 45 deletions(-) diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb index 5bb784195dc1..3c99ffe8867c 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/autoscaling_policies_services_pb.rb @@ -1,7 +1,7 @@ # Generated by the protocol buffer compiler. DO NOT EDIT! # Source: google/cloud/dataproc/v1/autoscaling_policies.proto for package 'google.cloud.dataproc.v1' # Original file comments: -# Copyright 2019 Google LLC. +# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -15,7 +15,6 @@ # See the License for the specific language governing permissions and # limitations under the License. # -# require 'grpc' diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/cluster_controller_client.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/cluster_controller_client.rb index 4774169ed03b..75146167ce9e 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/cluster_controller_client.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/cluster_controller_client.rb @@ -401,7 +401,8 @@ def create_cluster \ # interrupting jobs in progress. Timeout specifies how long to wait for jobs # in progress to finish before forcefully removing nodes (and potentially # interrupting jobs). Default timeout is 0 (for forceful decommission), and - # the maximum allowed timeout is 1 day. + # the maximum allowed timeout is 1 day. (see JSON representation of + # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)). # # Only supported on Dataproc image versions 1.2 and higher. # A hash of the same form as `Google::Protobuf::Duration` diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_pb.rb index ad487ef08afa..00e0b13bed47 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_pb.rb @@ -7,7 +7,6 @@ require 'google/api/annotations_pb' require 'google/api/client_pb' require 'google/api/field_behavior_pb' -require 'google/cloud/dataproc/v1/operations_pb' require 'google/cloud/dataproc/v1/shared_pb' require 'google/longrunning/operations_pb' require 'google/protobuf/duration_pb' @@ -35,6 +34,7 @@ optional :encryption_config, :message, 15, "google.cloud.dataproc.v1.EncryptionConfig" optional :autoscaling_config, :message, 18, "google.cloud.dataproc.v1.AutoscalingConfig" optional :security_config, :message, 16, "google.cloud.dataproc.v1.SecurityConfig" + optional :lifecycle_config, :message, 17, "google.cloud.dataproc.v1.LifecycleConfig" end add_message "google.cloud.dataproc.v1.AutoscalingConfig" do optional :policy_uri, :string, 1 @@ -51,6 +51,7 @@ repeated :service_account_scopes, :string, 3 repeated :tags, :string, 4 map :metadata, :string, :string, 5 + optional :reservation_affinity, :message, 11, "google.cloud.dataproc.v1.ReservationAffinity" end add_message "google.cloud.dataproc.v1.InstanceGroupConfig" do optional :num_instances, :int32, 1 @@ -124,6 +125,14 @@ map :properties, :string, :string, 2 repeated :optional_components, :enum, 3, "google.cloud.dataproc.v1.Component" end + add_message "google.cloud.dataproc.v1.LifecycleConfig" do + optional :idle_delete_ttl, :message, 1, "google.protobuf.Duration" + optional :idle_start_time, :message, 4, "google.protobuf.Timestamp" + oneof :ttl do + optional :auto_delete_time, :message, 2, "google.protobuf.Timestamp" + optional :auto_delete_ttl, :message, 3, "google.protobuf.Duration" + end + end add_message "google.cloud.dataproc.v1.ClusterMetrics" do map :hdfs_metrics, :string, :int64, 1 map :yarn_metrics, :string, :int64, 2 @@ -174,6 +183,17 @@ add_message "google.cloud.dataproc.v1.DiagnoseClusterResults" do optional :output_uri, :string, 1 end + add_message "google.cloud.dataproc.v1.ReservationAffinity" do + optional :consume_reservation_type, :enum, 1, "google.cloud.dataproc.v1.ReservationAffinity.Type" + optional :key, :string, 2 + repeated :values, :string, 3 + end + add_enum "google.cloud.dataproc.v1.ReservationAffinity.Type" do + value :TYPE_UNSPECIFIED, 0 + value :NO_RESERVATION, 1 + value :ANY_RESERVATION, 2 + value :SPECIFIC_RESERVATION, 3 + end end module Google @@ -196,6 +216,7 @@ module V1 SecurityConfig = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SecurityConfig").msgclass KerberosConfig = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.KerberosConfig").msgclass SoftwareConfig = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SoftwareConfig").msgclass + LifecycleConfig = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.LifecycleConfig").msgclass ClusterMetrics = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ClusterMetrics").msgclass CreateClusterRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.CreateClusterRequest").msgclass UpdateClusterRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.UpdateClusterRequest").msgclass @@ -205,6 +226,8 @@ module V1 ListClustersResponse = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ListClustersResponse").msgclass DiagnoseClusterRequest = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.DiagnoseClusterRequest").msgclass DiagnoseClusterResults = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.DiagnoseClusterResults").msgclass + ReservationAffinity = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ReservationAffinity").msgclass + ReservationAffinity::Type = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.ReservationAffinity.Type").enummodule end end end diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_services_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_services_pb.rb index 41f9dc8c0080..7af0dba51a24 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_services_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/clusters_services_pb.rb @@ -1,7 +1,7 @@ # Generated by the protocol buffer compiler. DO NOT EDIT! # Source: google/cloud/dataproc/v1/clusters.proto for package 'google.cloud.dataproc.v1' # Original file comments: -# Copyright 2019 Google LLC. +# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -15,7 +15,6 @@ # See the License for the specific language governing permissions and # limitations under the License. # -# require 'grpc' @@ -38,15 +37,15 @@ class Service # Creates a cluster in a project. The returned # [Operation.metadata][google.longrunning.Operation.metadata] will be - # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). + # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). rpc :CreateCluster, CreateClusterRequest, Google::Longrunning::Operation # Updates a cluster in a project. The returned # [Operation.metadata][google.longrunning.Operation.metadata] will be - # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). + # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). rpc :UpdateCluster, UpdateClusterRequest, Google::Longrunning::Operation # Deletes a cluster in a project. The returned # [Operation.metadata][google.longrunning.Operation.metadata] will be - # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). + # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). rpc :DeleteCluster, DeleteClusterRequest, Google::Longrunning::Operation # Gets the resource representation for a cluster in a project. rpc :GetCluster, GetClusterRequest, Cluster @@ -54,11 +53,11 @@ class Service rpc :ListClusters, ListClustersRequest, ListClustersResponse # Gets cluster diagnostic information. The returned # [Operation.metadata][google.longrunning.Operation.metadata] will be - # [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). + # [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#clusteroperationmetadata). # After the operation completes, # [Operation.response][google.longrunning.Operation.response] # contains - # [DiagnoseClusterResults](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#diagnoseclusterresults). + # [DiagnoseClusterResults](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#diagnoseclusterresults). rpc :DiagnoseCluster, DiagnoseClusterRequest, Google::Longrunning::Operation end diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/clusters.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/clusters.rb index a9d4ab6ab4b0..b9a45e66cbc7 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/clusters.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/clusters.rb @@ -113,6 +113,9 @@ class Cluster; end # @!attribute [rw] security_config # @return [Google::Cloud::Dataproc::V1::SecurityConfig] # Optional. Security settings for the cluster. + # @!attribute [rw] lifecycle_config + # @return [Google::Cloud::Dataproc::V1::LifecycleConfig] + # Optional. Lifecycle setting for the cluster. class ClusterConfig; end # Autoscaling Policy config associated with the cluster. @@ -220,9 +223,12 @@ class EncryptionConfig; end # The Compute Engine metadata entries to add to all instances (see # [Project and instance # metadata](https://cloud.google.com/compute/docs/storing-retrieving-metadata#project_and_instance_metadata)). + # @!attribute [rw] reservation_affinity + # @return [Google::Cloud::Dataproc::V1::ReservationAffinity] + # Optional. Reservation Affinity for consuming Zonal reservation. class GceClusterConfig; end - # Optional. The config settings for Compute Engine resources in + # The config settings for Compute Engine resources in # an instance group, such as a master or worker group. # @!attribute [rw] num_instances # @return [Integer] @@ -337,7 +343,10 @@ class DiskConfig; end # @!attribute [rw] execution_timeout # @return [Google::Protobuf::Duration] # Optional. Amount of time executable has to complete. Default is - # 10 minutes. Cluster creation fails with an explanatory error message (the + # 10 minutes (see JSON representation of + # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)). + # + # Cluster creation fails with an explanatory error message (the # name of the executable that caused the error and the exceeded timeout # period) if the executable is not completed at end of the timeout period. class NodeInitializationAction; end @@ -351,7 +360,8 @@ class NodeInitializationAction; end # Optional. Output only. Details of cluster's state. # @!attribute [rw] state_start_time # @return [Google::Protobuf::Timestamp] - # Output only. Time when this state was entered. + # Output only. Time when this state was entered (see JSON representation of + # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)). # @!attribute [rw] substate # @return [Google::Cloud::Dataproc::V1::ClusterStatus::Substate] # Output only. Additional state information that includes @@ -508,6 +518,32 @@ class KerberosConfig; end # Optional. The set of components to activate on the cluster. class SoftwareConfig; end + # Specifies the cluster auto-delete schedule configuration. + # @!attribute [rw] idle_delete_ttl + # @return [Google::Protobuf::Duration] + # Optional. The duration to keep the cluster alive while idling (when no jobs + # are running). Passing this threshold will cause the cluster to be + # deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON + # representation of + # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json). + # @!attribute [rw] auto_delete_time + # @return [Google::Protobuf::Timestamp] + # Optional. The time when cluster will be auto-deleted (see JSON representation of + # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)). + # @!attribute [rw] auto_delete_ttl + # @return [Google::Protobuf::Duration] + # Optional. The lifetime duration of cluster. The cluster will be + # auto-deleted at the end of this period. Minimum value is 10 minutes; + # maximum value is 14 days (see JSON representation of + # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)). + # @!attribute [rw] idle_start_time + # @return [Google::Protobuf::Timestamp] + # Output only. The time when cluster became idle (most recent job finished) + # and became eligible for deletion due to idleness (see JSON representation + # of + # [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)). + class LifecycleConfig; end + # Contains cluster daemon metrics, such as HDFS and YARN stats. # # **Beta Feature**: This report is available for testing purposes only. It may @@ -567,7 +603,8 @@ class CreateClusterRequest; end # interrupting jobs in progress. Timeout specifies how long to wait for jobs # in progress to finish before forcefully removing nodes (and potentially # interrupting jobs). Default timeout is 0 (for forceful decommission), and - # the maximum allowed timeout is 1 day. + # the maximum allowed timeout is 1 day. (see JSON representation of + # [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)). # # Only supported on Dataproc image versions 1.2 and higher. # @!attribute [rw] update_mask @@ -749,6 +786,33 @@ class DiagnoseClusterRequest; end # The output report is a plain text file with a summary of collected # diagnostics. class DiagnoseClusterResults; end + + # Reservation Affinity for consuming Zonal reservation. + # @!attribute [rw] consume_reservation_type + # @return [Google::Cloud::Dataproc::V1::ReservationAffinity::Type] + # Optional. Type of reservation to consume + # @!attribute [rw] key + # @return [String] + # Optional. Corresponds to the label key of reservation resource. + # @!attribute [rw] values + # @return [Array] + # Optional. Corresponds to the label values of reservation resource. + class ReservationAffinity + # Indicates whether to consume capacity from an reservation or not. + module Type + TYPE_UNSPECIFIED = 0 + + # Do not consume from any allocated capacity. + NO_RESERVATION = 1 + + # Consume any reservation available. + ANY_RESERVATION = 2 + + # Must consume from a specific reservation. Must specify key value fields + # for specifying the reservations. + SPECIFIC_RESERVATION = 3 + end + end end end end diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/jobs.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/jobs.rb index 3c751694ac8c..3875567e1952 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/jobs.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/jobs.rb @@ -289,6 +289,67 @@ class SparkSqlJob; end # Optional. The runtime log config for job execution. class PigJob; end + # A Dataproc job for running + # [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html) + # applications on YARN. + # @!attribute [rw] main_r_file_uri + # @return [String] + # Required. The HCFS URI of the main R file to use as the driver. + # Must be a .R file. + # @!attribute [rw] args + # @return [Array] + # Optional. The arguments to pass to the driver. Do not include arguments, + # such as `--conf`, that can be set as job properties, since a collision may + # occur that causes an incorrect job submission. + # @!attribute [rw] file_uris + # @return [Array] + # Optional. HCFS URIs of files to be copied to the working directory of + # R drivers and distributed tasks. Useful for naively parallel tasks. + # @!attribute [rw] archive_uris + # @return [Array] + # Optional. HCFS URIs of archives to be extracted in the working directory of + # Spark drivers and tasks. Supported file types: + # .jar, .tar, .tar.gz, .tgz, and .zip. + # @!attribute [rw] properties + # @return [Hash{String => String}] + # Optional. A mapping of property names to values, used to configure SparkR. + # Properties that conflict with values set by the Dataproc API may be + # overwritten. Can include properties set in + # /etc/spark/conf/spark-defaults.conf and classes in user code. + # @!attribute [rw] logging_config + # @return [Google::Cloud::Dataproc::V1::LoggingConfig] + # Optional. The runtime log config for job execution. + class SparkRJob; end + + # A Dataproc job for running [Presto](https://prestosql.io/) queries + # @!attribute [rw] query_file_uri + # @return [String] + # The HCFS URI of the script that contains SQL queries. + # @!attribute [rw] query_list + # @return [Google::Cloud::Dataproc::V1::QueryList] + # A list of queries. + # @!attribute [rw] continue_on_failure + # @return [true, false] + # Optional. Whether to continue executing queries if a query fails. + # The default value is `false`. Setting to `true` can be useful when + # executing independent parallel queries. + # @!attribute [rw] output_format + # @return [String] + # Optional. The format in which query output will be displayed. See the + # Presto documentation for supported output formats + # @!attribute [rw] client_tags + # @return [Array] + # Optional. Presto client tags to attach to this query + # @!attribute [rw] properties + # @return [Hash{String => String}] + # Optional. A mapping of property names to values. Used to set Presto + # [session properties](https://prestodb.io/docs/current/sql/set-session.html) + # Equivalent to using the --session flag in the Presto CLI + # @!attribute [rw] logging_config + # @return [Google::Cloud::Dataproc::V1::LoggingConfig] + # Optional. The runtime log config for job execution. + class PrestoJob; end + # Dataproc job config. # @!attribute [rw] cluster_name # @return [String] @@ -460,22 +521,28 @@ module State # run the job. # @!attribute [rw] hadoop_job # @return [Google::Cloud::Dataproc::V1::HadoopJob] - # Job is a Hadoop job. + # Optional. Job is a Hadoop job. # @!attribute [rw] spark_job # @return [Google::Cloud::Dataproc::V1::SparkJob] - # Job is a Spark job. + # Optional. Job is a Spark job. # @!attribute [rw] pyspark_job # @return [Google::Cloud::Dataproc::V1::PySparkJob] - # Job is a Pyspark job. + # Optional. Job is a PySpark job. # @!attribute [rw] hive_job # @return [Google::Cloud::Dataproc::V1::HiveJob] - # Job is a Hive job. + # Optional. Job is a Hive job. # @!attribute [rw] pig_job # @return [Google::Cloud::Dataproc::V1::PigJob] - # Job is a Pig job. + # Optional. Job is a Pig job. + # @!attribute [rw] spark_r_job + # @return [Google::Cloud::Dataproc::V1::SparkRJob] + # Optional. Job is a SparkR job. # @!attribute [rw] spark_sql_job # @return [Google::Cloud::Dataproc::V1::SparkSqlJob] - # Job is a SparkSql job. + # Optional. Job is a SparkSql job. + # @!attribute [rw] presto_job + # @return [Google::Cloud::Dataproc::V1::PrestoJob] + # Optional. Job is a Presto job. # @!attribute [rw] status # @return [Google::Cloud::Dataproc::V1::JobStatus] # Output only. The job status. Additional application-specific diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/workflow_templates.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/workflow_templates.rb index a3d6cc34e878..a6a2e792aadc 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/workflow_templates.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/doc/google/cloud/dataproc/v1/workflow_templates.rb @@ -146,22 +146,16 @@ class ClusterSelector; end # or hyphen. Must consist of between 3 and 50 characters. # @!attribute [rw] hadoop_job # @return [Google::Cloud::Dataproc::V1::HadoopJob] - # Job is a Hadoop job. # @!attribute [rw] spark_job # @return [Google::Cloud::Dataproc::V1::SparkJob] - # Job is a Spark job. # @!attribute [rw] pyspark_job # @return [Google::Cloud::Dataproc::V1::PySparkJob] - # Job is a Pyspark job. # @!attribute [rw] hive_job # @return [Google::Cloud::Dataproc::V1::HiveJob] - # Job is a Hive job. # @!attribute [rw] pig_job # @return [Google::Cloud::Dataproc::V1::PigJob] - # Job is a Pig job. # @!attribute [rw] spark_sql_job # @return [Google::Cloud::Dataproc::V1::SparkSqlJob] - # Job is a SparkSql job. # @!attribute [rw] labels # @return [Hash{String => String}] # Optional. The labels to associate with this job. diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_pb.rb index 23dc6f9beebe..cdc202239048 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_pb.rb @@ -7,6 +7,7 @@ require 'google/api/annotations_pb' require 'google/api/client_pb' require 'google/api/field_behavior_pb' +require 'google/longrunning/operations_pb' require 'google/protobuf/empty_pb' require 'google/protobuf/field_mask_pb' require 'google/protobuf/timestamp_pb' @@ -93,6 +94,25 @@ optional :query_list, :message, 2, "google.cloud.dataproc.v1.QueryList" end end + add_message "google.cloud.dataproc.v1.SparkRJob" do + optional :main_r_file_uri, :string, 1 + repeated :args, :string, 2 + repeated :file_uris, :string, 3 + repeated :archive_uris, :string, 4 + map :properties, :string, :string, 5 + optional :logging_config, :message, 6, "google.cloud.dataproc.v1.LoggingConfig" + end + add_message "google.cloud.dataproc.v1.PrestoJob" do + optional :continue_on_failure, :bool, 3 + optional :output_format, :string, 4 + repeated :client_tags, :string, 5 + map :properties, :string, :string, 6 + optional :logging_config, :message, 7, "google.cloud.dataproc.v1.LoggingConfig" + oneof :queries do + optional :query_file_uri, :string, 1 + optional :query_list, :message, 2, "google.cloud.dataproc.v1.QueryList" + end + end add_message "google.cloud.dataproc.v1.JobPlacement" do optional :cluster_name, :string, 1 optional :cluster_uuid, :string, 2 @@ -159,7 +179,9 @@ optional :pyspark_job, :message, 5, "google.cloud.dataproc.v1.PySparkJob" optional :hive_job, :message, 6, "google.cloud.dataproc.v1.HiveJob" optional :pig_job, :message, 7, "google.cloud.dataproc.v1.PigJob" + optional :spark_r_job, :message, 21, "google.cloud.dataproc.v1.SparkRJob" optional :spark_sql_job, :message, 12, "google.cloud.dataproc.v1.SparkSqlJob" + optional :presto_job, :message, 23, "google.cloud.dataproc.v1.PrestoJob" end end add_message "google.cloud.dataproc.v1.JobScheduling" do @@ -226,6 +248,8 @@ module V1 HiveJob = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.HiveJob").msgclass SparkSqlJob = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkSqlJob").msgclass PigJob = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PigJob").msgclass + SparkRJob = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.SparkRJob").msgclass + PrestoJob = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.PrestoJob").msgclass JobPlacement = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.JobPlacement").msgclass JobStatus = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.JobStatus").msgclass JobStatus::State = Google::Protobuf::DescriptorPool.generated_pool.lookup("google.cloud.dataproc.v1.JobStatus.State").enummodule diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_services_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_services_pb.rb index ad8199b02b06..22aef84618d3 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_services_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/jobs_services_pb.rb @@ -1,7 +1,7 @@ # Generated by the protocol buffer compiler. DO NOT EDIT! # Source: google/cloud/dataproc/v1/jobs.proto for package 'google.cloud.dataproc.v1' # Original file comments: -# Copyright 2019 Google LLC. +# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -44,9 +44,9 @@ class Service rpc :UpdateJob, UpdateJobRequest, Job # Starts a job cancellation request. To access the job resource # after cancellation, call - # [regions/\\{region}/jobs.list](/dataproc/docs/reference/rest/v1/projects.regions.jobs/list) + # [regions/\\{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/list) # or - # [regions/\\{region}/jobs.get](/dataproc/docs/reference/rest/v1/projects.regions.jobs/get). + # [regions/\\{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.jobs/get). rpc :CancelJob, CancelJobRequest, Job # Deletes the job from the project. If the job is active, the delete fails, # and the response returns `FAILED_PRECONDITION`. diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_template_service_client.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_template_service_client.rb index 0bf9ca25050c..cd5aec130eac 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_template_service_client.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_template_service_client.rb @@ -404,7 +404,7 @@ def get_workflow_template \ # The {Google::Longrunning::Operation#metadata Operation#metadata} will be # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). # Also see [Using - # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). + # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). # # On successful completion, # {Google::Longrunning::Operation#response Operation#response} will be @@ -522,7 +522,7 @@ def instantiate_workflow_template \ # The {Google::Longrunning::Operation#metadata Operation#metadata} will be # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). # Also see [Using - # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). + # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). # # On successful completion, # {Google::Longrunning::Operation#response Operation#response} will be diff --git a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb index 603847f6c5ef..fc60edbc5b11 100644 --- a/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb +++ b/google-cloud-dataproc/lib/google/cloud/dataproc/v1/workflow_templates_services_pb.rb @@ -1,7 +1,7 @@ # Generated by the protocol buffer compiler. DO NOT EDIT! # Source: google/cloud/dataproc/v1/workflow_templates.proto for package 'google.cloud.dataproc.v1' # Original file comments: -# Copyright 2019 Google LLC. +# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. @@ -15,7 +15,6 @@ # See the License for the specific language governing permissions and # limitations under the License. # -# require 'grpc' @@ -56,9 +55,9 @@ class Service # clusters to be deleted. # # The [Operation.metadata][google.longrunning.Operation.metadata] will be - # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). + # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). # Also see [Using - # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). + # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). # # On successful completion, # [Operation.response][google.longrunning.Operation.response] will be @@ -81,9 +80,9 @@ class Service # clusters to be deleted. # # The [Operation.metadata][google.longrunning.Operation.metadata] will be - # [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). + # [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata). # Also see [Using - # WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). + # WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata). # # On successful completion, # [Operation.response][google.longrunning.Operation.response] will be diff --git a/google-cloud-dataproc/synth.metadata b/google-cloud-dataproc/synth.metadata index 9c0d3f2c457a..765197101c95 100644 --- a/google-cloud-dataproc/synth.metadata +++ b/google-cloud-dataproc/synth.metadata @@ -1,27 +1,41 @@ { - "updateTime": "2020-02-01T11:38:54.395210Z", + "updateTime": "2020-02-21T11:41:46.919086Z", "sources": [ { "generator": { "name": "artman", - "version": "0.44.4", - "dockerImage": "googleapis/artman@sha256:19e945954fc960a4bdfee6cb34695898ab21a8cf0bac063ee39b91f00a1faec8" + "version": "0.45.0", + "dockerImage": "googleapis/artman@sha256:6aec9c34db0e4be221cdaf6faba27bdc07cfea846808b3d3b964dfce3a9a0f9b" + } + }, + { + "git": { + "name": ".", + "remote": "https://github.com/googleapis/google-cloud-ruby.git", + "sha": "8d7d444e7189735b15961087c598a79882444fd2" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "b5cbe4a4ba64ab19e6627573ff52057a1657773d", - "internalRef": "292647187", - "log": "b5cbe4a4ba64ab19e6627573ff52057a1657773d\nSecurityCenter v1p1beta1: move file-level option on top to workaround protobuf.js bug.\n\nPiperOrigin-RevId: 292647187\n\nb224b317bf20c6a4fbc5030b4a969c3147f27ad3\nAdds API definitions for bigqueryreservation v1beta1.\n\nPiperOrigin-RevId: 292634722\n\nc1468702f9b17e20dd59007c0804a089b83197d2\nSynchronize new proto/yaml changes.\n\nPiperOrigin-RevId: 292626173\n\nffdfa4f55ab2f0afc11d0eb68f125ccbd5e404bd\nvision: v1p3beta1 publish annotations and retry config\n\nPiperOrigin-RevId: 292605599\n\n78f61482cd028fc1d9892aa5d89d768666a954cd\nvision: v1p1beta1 publish annotations and retry config\n\nPiperOrigin-RevId: 292605125\n\n60bb5a294a604fd1778c7ec87b265d13a7106171\nvision: v1p2beta1 publish annotations and retry config\n\nPiperOrigin-RevId: 292604980\n\n3bcf7aa79d45eb9ec29ab9036e9359ea325a7fc3\nvision: v1p4beta1 publish annotations and retry config\n\nPiperOrigin-RevId: 292604656\n\n2717b8a1c762b26911b45ecc2e4ee01d98401b28\nFix dataproc artman client library generation.\n\nPiperOrigin-RevId: 292555664\n\n7ac66d9be8a7d7de4f13566d8663978c9ee9dcd7\nAdd Dataproc Autoscaling API to V1.\n\nPiperOrigin-RevId: 292450564\n\n5d932b2c1be3a6ef487d094e3cf5c0673d0241dd\n- Improve documentation\n- Add a client_id field to StreamingPullRequest\n\nPiperOrigin-RevId: 292434036\n\neaff9fa8edec3e914995ce832b087039c5417ea7\nmonitoring: v3 publish annotations and client retry config\n\nPiperOrigin-RevId: 292425288\n\n70958bab8c5353870d31a23fb2c40305b050d3fe\nBigQuery Storage Read API v1 clients.\n\nPiperOrigin-RevId: 292407644\n\n7a15e7fe78ff4b6d5c9606a3264559e5bde341d1\nUpdate backend proto for Google Cloud Endpoints\n\nPiperOrigin-RevId: 292391607\n\n3ca2c014e24eb5111c8e7248b1e1eb833977c83d\nbazel: Add --flaky_test_attempts=3 argument to prevent CI failures caused by flaky tests\n\nPiperOrigin-RevId: 292382559\n\n9933347c1f677e81e19a844c2ef95bfceaf694fe\nbazel:Integrate latest protoc-java-resource-names-plugin changes (fix for PyYAML dependency in bazel rules)\n\nPiperOrigin-RevId: 292376626\n\nb835ab9d2f62c88561392aa26074c0b849fb0bd3\nasset: v1p2beta1 add client config annotations\n\n* remove unintentionally exposed RPCs\n* remove messages relevant to removed RPCs\n\nPiperOrigin-RevId: 292369593\n\nc1246a29e22b0f98e800a536b5b0da2d933a55f2\nUpdating v1 protos with the latest inline documentation (in comments) and config options. Also adding a per-service .yaml file.\n\nPiperOrigin-RevId: 292310790\n\nb491d07cadaae7cde5608321f913e5ca1459b32d\nRevert accidental local_repository change\n\nPiperOrigin-RevId: 292245373\n\naf3400a8cb6110025198b59a0f7d018ae3cda700\nUpdate gapic-generator dependency (prebuilt PHP binary support).\n\nPiperOrigin-RevId: 292243997\n\n341fd5690fae36f36cf626ef048fbcf4bbe7cee6\ngrafeas: v1 add resource_definition for the grafeas.io/Project and change references for Project.\n\nPiperOrigin-RevId: 292221998\n\n42e915ec2ece1cd37a590fbcd10aa2c0fb0e5b06\nUpdate the gapic-generator, protoc-java-resource-name-plugin and protoc-docs-plugin to the latest commit.\n\nPiperOrigin-RevId: 292182368\n\nf035f47250675d31492a09f4a7586cfa395520a7\nFix grafeas build and update build.sh script to include gerafeas.\n\nPiperOrigin-RevId: 292168753\n\n26ccb214b7bc4a716032a6266bcb0a9ca55d6dbb\nasset: v1p1beta1 add client config annotations and retry config\n\nPiperOrigin-RevId: 292154210\n\n974ee5c0b5d03e81a50dafcedf41e0efebb5b749\nasset: v1beta1 add client config annotations\n\nPiperOrigin-RevId: 292152573\n\ncf3b61102ed5f36b827bc82ec39be09525f018c8\n Fix to protos for v1p1beta1 release of Cloud Security Command Center\n\nPiperOrigin-RevId: 292034635\n\n4e1cfaa7c0fede9e65d64213ca3da1b1255816c0\nUpdate the public proto to support UTF-8 encoded id for CatalogService API, increase the ListCatalogItems deadline to 300s and some minor documentation change\n\nPiperOrigin-RevId: 292030970\n\n9c483584f8fd5a1b862ae07973f4cc7bb3e46648\nasset: add annotations to v1p1beta1\n\nPiperOrigin-RevId: 292009868\n\ne19209fac29731d0baf6d9ac23da1164f7bdca24\nAdd the google.rpc.context.AttributeContext message to the open source\ndirectories.\n\nPiperOrigin-RevId: 291999930\n\nae5662960573f279502bf98a108a35ba1175e782\noslogin API: move file level option on top of the file to avoid protobuf.js bug.\n\nPiperOrigin-RevId: 291990506\n\neba3897fff7c49ed85d3c47fc96fe96e47f6f684\nAdd cc_proto_library and cc_grpc_library targets for Spanner and IAM protos.\n\nPiperOrigin-RevId: 291988651\n\n8e981acfd9b97ea2f312f11bbaa7b6c16e412dea\nBeta launch for PersonDetection and FaceDetection features.\n\nPiperOrigin-RevId: 291821782\n\n994e067fae3b21e195f7da932b08fff806d70b5d\nasset: add annotations to v1p2beta1\n\nPiperOrigin-RevId: 291815259\n\n244e1d2c89346ca2e0701b39e65552330d68545a\nAdd Playable Locations service\n\nPiperOrigin-RevId: 291806349\n\n909f8f67963daf45dd88d020877fb9029b76788d\nasset: add annotations to v1beta2\n\nPiperOrigin-RevId: 291805301\n\n3c39a1d6e23c1ef63c7fba4019c25e76c40dfe19\nKMS: add file-level message for CryptoKeyPath, it is defined in gapic yaml but not\nin proto files.\n\nPiperOrigin-RevId: 291420695\n\nc6f3f350b8387f8d1b85ed4506f30187ebaaddc3\ncontaineranalysis: update v1beta1 and bazel build with annotations\n\nPiperOrigin-RevId: 291401900\n\n92887d74b44e4e636252b7b8477d0d2570cd82db\nfix: fix the location of grpc config file.\n\nPiperOrigin-RevId: 291396015\n\ne26cab8afd19d396b929039dac5d874cf0b5336c\nexpr: add default_host and method_signature annotations to CelService\n\nPiperOrigin-RevId: 291240093\n\n06093ae3952441c34ec176d1f7431b8765cec0be\nirm: fix v1alpha2 bazel build by adding missing proto imports\n\nPiperOrigin-RevId: 291227940\n\na8a2514af326e4673063f9a3c9d0ef1091c87e6c\nAdd proto annotation for cloud/irm API\n\nPiperOrigin-RevId: 291217859\n\n8d16f76de065f530d395a4c7eabbf766d6a120fd\nGenerate Memcache v1beta2 API protos and gRPC ServiceConfig files\n\nPiperOrigin-RevId: 291008516\n\n3af1dabd93df9a9f17bf3624d3b875c11235360b\ngrafeas: Add containeranalysis default_host to Grafeas service\n\nPiperOrigin-RevId: 290965849\n\nbe2663fa95e31cba67d0cd62611a6674db9f74b7\nfix(google/maps/roads): add missing opening bracket\n\nPiperOrigin-RevId: 290964086\n\nfacc26550a0af0696e0534bc9cae9df14275aa7c\nUpdating v2 protos with the latest inline documentation (in comments) and adding a per-service .yaml file.\n\nPiperOrigin-RevId: 290952261\n\n" + "sha": "e5bc9566ae057fb4c92f8b7e047f1c8958235b53", + "internalRef": "296357191", + "log": "e5bc9566ae057fb4c92f8b7e047f1c8958235b53\nDeprecate the endpoint_uris field, as it is unused.\n\nPiperOrigin-RevId: 296357191\n\n8c12e2b4dca94e12bff9f538bdac29524ff7ef7a\nUpdate Dataproc v1 client.\n\nPiperOrigin-RevId: 296336662\n\n17567c4a1ef0a9b50faa87024d66f8acbb561089\nRemoving erroneous comment, a la https://github.com/googleapis/java-speech/pull/103\n\nPiperOrigin-RevId: 296332968\n\n3eaaaf8626ce5b0c0bc7eee05e143beffa373b01\nAdd BUILD.bazel for v1 secretmanager.googleapis.com\n\nPiperOrigin-RevId: 296274723\n\ne76149c3d992337f85eeb45643106aacae7ede82\nMove securitycenter v1 to use generate from annotations.\n\nPiperOrigin-RevId: 296266862\n\n203740c78ac69ee07c3bf6be7408048751f618f8\nAdd StackdriverLoggingConfig field to Cloud Tasks v2 API.\n\nPiperOrigin-RevId: 296256388\n\ne4117d5e9ed8bbca28da4a60a94947ca51cb2083\nCreate a Bazel BUILD file for the google.actions.type export.\n\nPiperOrigin-RevId: 296212567\n\na9639a0a9854fd6e1be08bba1ac3897f4f16cb2f\nAdd secretmanager.googleapis.com v1 protos\n\nPiperOrigin-RevId: 295983266\n\nce4f4c21d9dd2bfab18873a80449b9d9851efde8\nasset: v1p1beta1 remove SearchResources and SearchIamPolicies\n\nPiperOrigin-RevId: 295861722\n\ncb61d6c2d070b589980c779b68ffca617f789116\nasset: v1p1beta1 remove SearchResources and SearchIamPolicies\n\nPiperOrigin-RevId: 295855449\n\nab2685d8d3a0e191dc8aef83df36773c07cb3d06\nfix: Dataproc v1 - AutoscalingPolicy annotation\n\nThis adds the second resource name pattern to the\nAutoscalingPolicy resource.\n\nCommitter: @lukesneeringer\nPiperOrigin-RevId: 295738415\n\n8a1020bf6828f6e3c84c3014f2c51cb62b739140\nUpdate cloud asset api v1p4beta1.\n\nPiperOrigin-RevId: 295286165\n\n5cfa105206e77670369e4b2225597386aba32985\nAdd service control related proto build rule.\n\nPiperOrigin-RevId: 295262088\n\nee4dddf805072004ab19ac94df2ce669046eec26\nmonitoring v3: Add prefix \"https://cloud.google.com/\" into the link for global access\ncl 295167522, get ride of synth.py hacks\n\nPiperOrigin-RevId: 295238095\n\nd9835e922ea79eed8497db270d2f9f85099a519c\nUpdate some minor docs changes about user event proto\n\nPiperOrigin-RevId: 295185610\n\n5f311e416e69c170243de722023b22f3df89ec1c\nfix: use correct PHP package name in gapic configuration\n\nPiperOrigin-RevId: 295161330\n\n6cdd74dcdb071694da6a6b5a206e3a320b62dd11\npubsub: v1 add client config annotations and retry config\n\nPiperOrigin-RevId: 295158776\n\n5169f46d9f792e2934d9fa25c36d0515b4fd0024\nAdded cloud asset api v1p4beta1.\n\nPiperOrigin-RevId: 295026522\n\n56b55aa8818cd0a532a7d779f6ef337ba809ccbd\nFix: Resource annotations for CreateTimeSeriesRequest and ListTimeSeriesRequest should refer to valid resources. TimeSeries is not a named resource.\n\nPiperOrigin-RevId: 294931650\n\n0646bc775203077226c2c34d3e4d50cc4ec53660\nRemove unnecessary languages from bigquery-related artman configuration files.\n\nPiperOrigin-RevId: 294809380\n\n8b78aa04382e3d4147112ad6d344666771bb1909\nUpdate backend.proto for schemes and protocol\n\nPiperOrigin-RevId: 294788800\n\n80b8f8b3de2359831295e24e5238641a38d8488f\nAdds artman config files for bigquerystorage endpoints v1beta2, v1alpha2, v1\n\nPiperOrigin-RevId: 294763931\n\n2c17ac33b226194041155bb5340c3f34733f1b3a\nAdd parameter to sample generated for UpdateInstance. Related to https://github.com/googleapis/python-redis/issues/4\n\nPiperOrigin-RevId: 294734008\n\nd5e8a8953f2acdfe96fb15e85eb2f33739623957\nMove bigquery datatransfer to gapic v2.\n\nPiperOrigin-RevId: 294703703\n\nefd36705972cfcd7d00ab4c6dfa1135bafacd4ae\nfix: Add two annotations that we missed.\n\nPiperOrigin-RevId: 294664231\n\n8a36b928873ff9c05b43859b9d4ea14cd205df57\nFix: Define the \"bigquery.googleapis.com/Table\" resource in the BigQuery Storage API (v1beta2).\n\nPiperOrigin-RevId: 294459768\n\nc7a3caa2c40c49f034a3c11079dd90eb24987047\nFix: Define the \"bigquery.googleapis.com/Table\" resource in the BigQuery Storage API (v1).\n\nPiperOrigin-RevId: 294456889\n\n5006247aa157e59118833658084345ee59af7c09\nFix: Make deprecated fields optional\nFix: Deprecate SetLoggingServiceRequest.zone in line with the comments\nFeature: Add resource name method signatures where appropriate\n\nPiperOrigin-RevId: 294383128\n\neabba40dac05c5cbe0fca3a35761b17e372036c4\nFix: C# and PHP package/namespace capitalization for BigQuery Storage v1.\n\nPiperOrigin-RevId: 294382444\n\nf8d9a858a7a55eba8009a23aa3f5cc5fe5e88dde\nfix: artman configuration file for bigtable-admin\n\nPiperOrigin-RevId: 294322616\n\n0f29555d1cfcf96add5c0b16b089235afbe9b1a9\nAPI definition for (not-yet-launched) GCS gRPC.\n\nPiperOrigin-RevId: 294321472\n\nfcc86bee0e84dc11e9abbff8d7c3529c0626f390\nfix: Bigtable Admin v2\n\nChange LRO metadata from PartialUpdateInstanceMetadata\nto UpdateInstanceMetadata. (Otherwise, it will not build.)\n\nPiperOrigin-RevId: 294264582\n\n6d9361eae2ebb3f42d8c7ce5baf4bab966fee7c0\nrefactor: Add annotations to Bigtable Admin v2.\n\nPiperOrigin-RevId: 294243406\n\nad7616f3fc8e123451c8b3a7987bc91cea9e6913\nFix: Resource type in CreateLogMetricRequest should use logging.googleapis.com.\nFix: ListLogEntries should have a method signature for convenience of calling it.\n\nPiperOrigin-RevId: 294222165\n\n63796fcbb08712676069e20a3e455c9f7aa21026\nFix: Remove extraneous resource definition for cloudkms.googleapis.com/CryptoKey.\n\nPiperOrigin-RevId: 294176658\n\ne7d8a694f4559201e6913f6610069cb08b39274e\nDepend on the latest gapic-generator and resource names plugin.\n\nThis fixes the very old an very annoying bug: https://github.com/googleapis/gapic-generator/pull/3087\n\nPiperOrigin-RevId: 293903652\n\n806b2854a966d55374ee26bb0cef4e30eda17b58\nfix: correct capitalization of Ruby namespaces in SecurityCenter V1p1beta1\n\nPiperOrigin-RevId: 293903613\n\n1b83c92462b14d67a7644e2980f723112472e03a\nPublish annotations and grpc service config for Logging API.\n\nPiperOrigin-RevId: 293893514\n\ne46f761cd6ec15a9e3d5ed4ff321a4bcba8e8585\nGenerate the Bazel build file for recommendengine public api\n\nPiperOrigin-RevId: 293710856\n\n68477017c4173c98addac0373950c6aa9d7b375f\nMake `language_code` optional for UpdateIntentRequest and BatchUpdateIntentsRequest.\n\nThe comments and proto annotations describe this parameter as optional.\n\nPiperOrigin-RevId: 293703548\n\n16f823f578bca4e845a19b88bb9bc5870ea71ab2\nAdd BUILD.bazel files for managedidentities API\n\nPiperOrigin-RevId: 293698246\n\n2f53fd8178c9a9de4ad10fae8dd17a7ba36133f2\nAdd v1p1beta1 config file\n\nPiperOrigin-RevId: 293696729\n\n052b274138fce2be80f97b6dcb83ab343c7c8812\nAdd source field for user event and add field behavior annotations\n\nPiperOrigin-RevId: 293693115\n\n1e89732b2d69151b1b3418fff3d4cc0434f0dded\ndatacatalog: v1beta1 add three new RPCs to gapic v1beta1 config\n\nPiperOrigin-RevId: 293692823\n\n9c8bd09bbdc7c4160a44f1fbab279b73cd7a2337\nchange the name of AccessApproval service to AccessApprovalAdmin\n\nPiperOrigin-RevId: 293690934\n\n2e23b8fbc45f5d9e200572ca662fe1271bcd6760\nAdd ListEntryGroups method, add http bindings to support entry group tagging, and update some comments.\n\nPiperOrigin-RevId: 293666452\n\n0275e38a4ca03a13d3f47a9613aac8c8b0d3f1f2\nAdd proto_package field to managedidentities API. It is needed for APIs that still depend on artman generation.\n\nPiperOrigin-RevId: 293643323\n\n4cdfe8278cb6f308106580d70648001c9146e759\nRegenerating public protos for Data Catalog to add new Custom Type Entry feature.\n\nPiperOrigin-RevId: 293614782\n\n45d2a569ab526a1fad3720f95eefb1c7330eaada\nEnable client generation for v1 ManagedIdentities API.\n\nPiperOrigin-RevId: 293515675\n\n2c17086b77e6f3bcf04a1f65758dfb0c3da1568f\nAdd the Actions on Google common types (//google/actions/type/*).\n\nPiperOrigin-RevId: 293478245\n\n781aadb932e64a12fb6ead7cd842698d99588433\nDialogflow weekly v2/v2beta1 library update:\n- Documentation updates\nImportant updates are also posted at\nhttps://cloud.google.com/dialogflow/docs/release-notes\n\nPiperOrigin-RevId: 293443396\n\ne2602608c9138c2fca24162720e67f9307c30b95\nDialogflow weekly v2/v2beta1 library update:\n- Documentation updates\nImportant updates are also posted at\nhttps://cloud.google.com/dialogflow/docs/release-notes\n\nPiperOrigin-RevId: 293442964\n\nc8aef82028d06b7992278fa9294c18570dc86c3d\nAdd cc_proto_library and cc_grpc_library targets for Bigtable protos.\n\nAlso fix indentation of cc_grpc_library targets in Spanner and IAM protos.\n\nPiperOrigin-RevId: 293440538\n\ne2faab04f4cb7f9755072330866689b1943a16e9\ncloudtasks: v2 replace non-standard retry params in gapic config v2\n\nPiperOrigin-RevId: 293424055\n\ndfb4097ea628a8470292c6590a4313aee0c675bd\nerrorreporting: v1beta1 add legacy artman config for php\n\nPiperOrigin-RevId: 293423790\n\nb18aed55b45bfe5b62476292c72759e6c3e573c6\nasset: v1p1beta1 updated comment for `page_size` limit.\n\nPiperOrigin-RevId: 293421386\n\nc9ef36b7956d9859a2fc86ad35fcaa16958ab44f\nbazel: Refactor CI build scripts\n\nPiperOrigin-RevId: 293387911\n\na8ed9d921fdddc61d8467bfd7c1668f0ad90435c\nfix: set Ruby module name for OrgPolicy\n\nPiperOrigin-RevId: 293257997\n\n6c7d28509bd8315de8af0889688ee20099594269\nredis: v1beta1 add UpgradeInstance and connect_mode field to Instance\n\nPiperOrigin-RevId: 293242878\n\nae0abed4fcb4c21f5cb67a82349a049524c4ef68\nredis: v1 add connect_mode field to Instance\n\nPiperOrigin-RevId: 293241914\n\n3f7a0d29b28ee9365771da2b66edf7fa2b4e9c56\nAdds service config definition for bigqueryreservation v1beta1\n\nPiperOrigin-RevId: 293234418\n\n0c88168d5ed6fe353a8cf8cbdc6bf084f6bb66a5\naddition of BUILD & configuration for accessapproval v1\n\nPiperOrigin-RevId: 293219198\n\n39bedc2e30f4778ce81193f6ba1fec56107bcfc4\naccessapproval: v1 publish protos\n\nPiperOrigin-RevId: 293167048\n\n69d9945330a5721cd679f17331a78850e2618226\nAdd file-level `Session` resource definition\n\nPiperOrigin-RevId: 293080182\n\nf6a1a6b417f39694275ca286110bc3c1ca4db0dc\nAdd file-level `Session` resource definition\n\nPiperOrigin-RevId: 293080178\n\n29d40b78e3dc1579b0b209463fbcb76e5767f72a\nExpose managedidentities/v1beta1/ API for client library usage.\n\nPiperOrigin-RevId: 292979741\n\na22129a1fb6e18056d576dfb7717aef74b63734a\nExpose managedidentities/v1/ API for client library usage.\n\nPiperOrigin-RevId: 292968186\n\n" + } + }, + { + "git": { + "name": "synthtool", + "remote": "rpc://devrel/cloud/libraries/tools/autosynth", + "sha": "706a38c26db42299845396cdae55db635c38794a" } }, { "template": { "name": "ruby_library", "origin": "synthtool.gcp", - "version": "2019.10.17" + "version": "2020.2.4" } } ],