Update Dataproc v1beta2 client.

PiperOrigin-RevId: 296451205
This commit is contained in:
Google APIs 2020-02-21 09:37:03 -08:00 committed by Copybara-Service
parent de28752440
commit 7f910bcc4f
6 changed files with 164 additions and 178 deletions

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";
@ -30,7 +29,7 @@ option java_outer_classname = "AutoscalingPoliciesProto";
option java_package = "com.google.cloud.dataproc.v1beta2";
// The API interface for managing autoscaling policies in the
// Google Cloud Dataproc API.
// Cloud Dataproc API.
service AutoscalingPolicyService {
option (google.api.default_host) = "dataproc.googleapis.com";
option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform";

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";
@ -20,7 +19,7 @@ package google.cloud.dataproc.v1beta2;
import "google/api/annotations.proto";
import "google/api/client.proto";
import "google/api/field_behavior.proto";
import "google/cloud/dataproc/v1beta2/operations.proto";
import "google/api/resource.proto";
import "google/cloud/dataproc/v1beta2/shared.proto";
import "google/longrunning/operations.proto";
import "google/protobuf/duration.proto";
@ -40,7 +39,7 @@ service ClusterController {
// Creates a cluster in a project. The returned
// [Operation.metadata][google.longrunning.Operation.metadata] will be
// [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
rpc CreateCluster(CreateClusterRequest) returns (google.longrunning.Operation) {
option (google.api.http) = {
post: "/v1beta2/projects/{project_id}/regions/{region}/clusters"
@ -55,7 +54,7 @@ service ClusterController {
// Updates a cluster in a project. The returned
// [Operation.metadata][google.longrunning.Operation.metadata] will be
// [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
rpc UpdateCluster(UpdateClusterRequest) returns (google.longrunning.Operation) {
option (google.api.http) = {
patch: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}"
@ -70,7 +69,7 @@ service ClusterController {
// Deletes a cluster in a project. The returned
// [Operation.metadata][google.longrunning.Operation.metadata] will be
// [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
rpc DeleteCluster(DeleteClusterRequest) returns (google.longrunning.Operation) {
option (google.api.http) = {
delete: "/v1beta2/projects/{project_id}/regions/{region}/clusters/{cluster_name}"
@ -101,7 +100,7 @@ service ClusterController {
// Gets cluster diagnostic information. The returned
// [Operation.metadata][google.longrunning.Operation.metadata] will be
// [ClusterOperationMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
// [ClusterOperationMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#clusteroperationmetadata).
// After the operation completes,
// [Operation.response][google.longrunning.Operation.response]
// contains
@ -129,7 +128,7 @@ message Cluster {
// unique. Names of deleted clusters can be reused.
string cluster_name = 2 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster config. Note that Cloud Dataproc may set
// Required. The cluster config. Note that Dataproc may set
// default values, and values may change when clusters are updated.
ClusterConfig config = 3 [(google.api.field_behavior) = REQUIRED];
@ -148,7 +147,7 @@ message Cluster {
// Output only. The previous cluster status.
repeated ClusterStatus status_history = 7 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. A cluster UUID (Unique Universal Identifier). Cloud Dataproc
// Output only. A cluster UUID (Unique Universal Identifier). Dataproc
// generates this value when it creates the cluster.
string cluster_uuid = 6 [(google.api.field_behavior) = OUTPUT_ONLY];
@ -161,14 +160,14 @@ message Cluster {
// The cluster config.
message ClusterConfig {
// Optional. A Google Cloud Storage bucket used to stage job
// Optional. A Cloud Storage bucket used to stage job
// dependencies, config files, and job driver console output.
// If you do not specify a staging bucket, Cloud
// Dataproc will determine a Cloud Storage location (US,
// ASIA, or EU) for your cluster's staging bucket according to the Google
// ASIA, or EU) for your cluster's staging bucket according to the
// Compute Engine zone where your cluster is deployed, and then create
// and manage this project-level, per-location bucket (see
// [Cloud Dataproc staging
// [Dataproc staging
// bucket](/dataproc/docs/concepts/configuring-clusters/staging-bucket)).
string config_bucket = 1 [(google.api.field_behavior) = OPTIONAL];
@ -244,7 +243,7 @@ message AutoscalingConfig {
// * `https://www.googleapis.com/compute/v1/projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
// * `projects/[project_id]/locations/[dataproc_region]/autoscalingPolicies/[policy_id]`
//
// Note that the policy must be in the same project and Cloud Dataproc region.
// Note that the policy must be in the same project and Dataproc region.
string policy_uri = 1 [(google.api.field_behavior) = OPTIONAL];
}
@ -260,7 +259,7 @@ message EncryptionConfig {
message GceClusterConfig {
// Optional. The zone where the Compute Engine cluster will be located.
// On a create request, it is required in the "global" region. If omitted
// in a non-global Cloud Dataproc region, the service will pick a zone in the
// in a non-global Dataproc region, the service will pick a zone in the
// corresponding Compute Engine region. On a get request, zone will always be
// present.
//
@ -302,17 +301,17 @@ message GceClusterConfig {
// configured to be accessible without external IP addresses.
bool internal_ip_only = 7 [(google.api.field_behavior) = OPTIONAL];
// Optional. The service account of the instances. Defaults to the default
// Compute Engine service account. Custom service accounts need
// permissions equivalent to the following IAM roles:
// Optional. The [Dataproc service
// account](/dataproc/docs/concepts/configuring-clusters/service-accounts#service_accounts_in_cloud_dataproc)
// (also see [VM Data Plane
// identity](/dataproc/docs/concepts/iam/dataproc-principals#vm_service_account_data_plane_identity))
// used by Dataproc cluster VM instances to access Google Cloud Platform
// services.
//
// * roles/logging.logWriter
// * roles/storage.objectAdmin
//
// (see
// https://cloud.google.com/compute/docs/access/service-accounts#custom_service_accounts
// for more information).
// Example: `[account_id]@[project_id].iam.gserviceaccount.com`
// If not specified, the
// [Compute Engine default service
// account](/compute/docs/access/service-accounts#default_service_account)
// is used.
string service_account = 8 [(google.api.field_behavior) = OPTIONAL];
// Optional. The URIs of service account scopes to be included in
@ -351,7 +350,7 @@ message InstanceGroupConfig {
// For master instance groups, must be set to 1.
int32 num_instances = 1 [(google.api.field_behavior) = OPTIONAL];
// Output only. The list of instance names. Cloud Dataproc derives the names
// Output only. The list of instance names. Dataproc derives the names
// from `cluster_name`, `num_instances`, and the instance group.
repeated string instance_names = 2 [(google.api.field_behavior) = OUTPUT_ONLY];
@ -368,7 +367,7 @@ message InstanceGroupConfig {
// * `projects/[project_id]/zones/us-east1-a/machineTypes/n1-standard-2`
// * `n1-standard-2`
//
// **Auto Zone Exception**: If you are using the Cloud Dataproc
// **Auto Zone Exception**: If you are using the Dataproc
// [Auto Zone
// Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
// feature, you must use the short name of the machine type
@ -392,7 +391,7 @@ message InstanceGroupConfig {
repeated AcceleratorConfig accelerators = 8 [(google.api.field_behavior) = OPTIONAL];
// Specifies the minimum cpu platform for the Instance Group.
// See [Cloud Dataproc→Minimum CPU Platform]
// See [Dataproc→Minimum CPU Platform]
// (/dataproc/docs/concepts/compute/dataproc-min-cpu).
string min_cpu_platform = 9;
}
@ -420,7 +419,7 @@ message AcceleratorConfig {
// * `projects/[project_id]/zones/us-east1-a/acceleratorTypes/nvidia-tesla-k80`
// * `nvidia-tesla-k80`
//
// **Auto Zone Exception**: If you are using the Cloud Dataproc
// **Auto Zone Exception**: If you are using the Dataproc
// [Auto Zone
// Placement](/dataproc/docs/concepts/configuring-clusters/auto-zone#using_auto_zone_placement)
// feature, you must use the short name of the accelerator type
@ -452,29 +451,31 @@ message DiskConfig {
// Specifies the cluster auto-delete schedule configuration.
message LifecycleConfig {
// Optional. The duration to keep the cluster alive while idling.
// Passing this threshold will cause the cluster to be
// deleted. Valid range: **[10m, 14d]**.
//
// Example: **"10m"**, the minimum value, to delete the
// cluster when it has had no jobs running for 10 minutes.
// Optional. The duration to keep the cluster alive while idling (when no jobs
// are running). Passing this threshold will cause the cluster to be
// deleted. Minimum value is 10 minutes; maximum value is 14 days (see JSON
// representation of
// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json).
google.protobuf.Duration idle_delete_ttl = 1 [(google.api.field_behavior) = OPTIONAL];
// Either the exact time the cluster should be deleted at or
// the cluster maximum age.
oneof ttl {
// Optional. The time when cluster will be auto-deleted.
google.protobuf.Timestamp auto_delete_time = 2;
// Optional. The time when cluster will be auto-deleted. (see JSON representation of
// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
google.protobuf.Timestamp auto_delete_time = 2 [(google.api.field_behavior) = OPTIONAL];
// Optional. The lifetime duration of cluster. The cluster will be
// auto-deleted at the end of this period. Valid range: **[10m, 14d]**.
//
// Example: **"1d"**, to delete the cluster 1 day after its creation..
google.protobuf.Duration auto_delete_ttl = 3;
// auto-deleted at the end of this period. Minimum value is 10 minutes;
// maximum value is 14 days (see JSON representation of
// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
google.protobuf.Duration auto_delete_ttl = 3 [(google.api.field_behavior) = OPTIONAL];
}
// Output only. The time when cluster became idle (most recent job finished)
// and became eligible for deletion due to idleness.
// and became eligible for deletion due to idleness (see JSON representation
// of
// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
google.protobuf.Timestamp idle_start_time = 4 [(google.api.field_behavior) = OUTPUT_ONLY];
}
@ -560,7 +561,10 @@ message NodeInitializationAction {
string executable_file = 1 [(google.api.field_behavior) = REQUIRED];
// Optional. Amount of time executable has to complete. Default is
// 10 minutes. Cluster creation fails with an explanatory error message (the
// 10 minutes (see JSON representation of
// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
//
// Cluster creation fails with an explanatory error message (the
// name of the executable that caused the error and the exceeded timeout
// period) if the executable is not completed at end of the timeout period.
google.protobuf.Duration execution_timeout = 2 [(google.api.field_behavior) = OPTIONAL];
@ -602,7 +606,7 @@ message ClusterStatus {
UNHEALTHY = 1;
// The agent-reported status is out of date (may occur if
// Cloud Dataproc loses communication with Agent).
// Dataproc loses communication with Agent).
//
// Applies to RUNNING state.
STALE_STATUS = 2;
@ -614,7 +618,8 @@ message ClusterStatus {
// Output only. Optional details of cluster's state.
string detail = 2 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. Time when this state was entered.
// Output only. Time when this state was entered (see JSON representation of
// [Timestamp](https://developers.google.com/protocol-buffers/docs/proto3#json)).
google.protobuf.Timestamp state_start_time = 3 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. Additional state information that includes
@ -625,7 +630,7 @@ message ClusterStatus {
// Specifies the selection and config of software inside the cluster.
message SoftwareConfig {
// Optional. The version of software inside the cluster. It must be one of the
// supported [Cloud Dataproc
// supported [Dataproc
// Versions](/dataproc/docs/concepts/versioning/dataproc-versions#supported_cloud_dataproc_versions),
// such as "1.2" (including a subminor version, such as "1.2.29"), or the
// ["preview"
@ -675,7 +680,7 @@ message CreateClusterRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster to create.
@ -701,7 +706,7 @@ message UpdateClusterRequest {
// cluster belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 5 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster name.
@ -715,7 +720,8 @@ message UpdateClusterRequest {
// interrupting jobs in progress. Timeout specifies how long to wait for jobs
// in progress to finish before forcefully removing nodes (and potentially
// interrupting jobs). Default timeout is 0 (for forceful decommission), and
// the maximum allowed timeout is 1 day.
// the maximum allowed timeout is 1 day (see JSON representation of
// [Duration](https://developers.google.com/protocol-buffers/docs/proto3#json)).
//
// Only supported on Dataproc image versions 1.2 and higher.
google.protobuf.Duration graceful_decommission_timeout = 6 [(google.api.field_behavior) = OPTIONAL];
@ -802,7 +808,7 @@ message DeleteClusterRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster name.
@ -832,7 +838,7 @@ message GetClusterRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster name.
@ -845,7 +851,7 @@ message ListClustersRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 4 [(google.api.field_behavior) = REQUIRED];
// Optional. A filter constraining the clusters to list. Filters are
@ -893,7 +899,7 @@ message DiagnoseClusterRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The cluster name.

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";
@ -20,6 +19,7 @@ package google.cloud.dataproc.v1beta2;
import "google/api/annotations.proto";
import "google/api/client.proto";
import "google/api/field_behavior.proto";
import "google/longrunning/operations.proto";
import "google/protobuf/empty.proto";
import "google/protobuf/field_mask.proto";
import "google/protobuf/timestamp.proto";
@ -70,9 +70,9 @@ service JobController {
// Starts a job cancellation request. To access the job resource
// after cancellation, call
// [regions/{region}/jobs.list](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
// [regions/{region}/jobs.list](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/list)
// or
// [regions/{region}/jobs.get](/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
// [regions/{region}/jobs.get](https://cloud.google.com/dataproc/docs/reference/rest/v1beta2/projects.regions.jobs/get).
rpc CancelJob(CancelJobRequest) returns (Job) {
option (google.api.http) = {
post: "/v1beta2/projects/{project_id}/regions/{region}/jobs/{job_id}:cancel"
@ -132,7 +132,7 @@ message LoggingConfig {
map<string, Level> driver_log_levels = 2;
}
// A Cloud Dataproc job for running
// A Dataproc job for running
// [Apache Hadoop
// MapReduce](https://hadoop.apache.org/docs/current/hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html)
// jobs on [Apache Hadoop
@ -159,33 +159,33 @@ message HadoopJob {
// include arguments, such as `-libjars` or `-Dfoo=bar`, that can be set as
// job properties, since a collision may occur that causes an incorrect job
// submission.
repeated string args = 3;
repeated string args = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. Jar file URIs to add to the CLASSPATHs of the
// Hadoop driver and tasks.
repeated string jar_file_uris = 4;
repeated string jar_file_uris = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS (Hadoop Compatible Filesystem) URIs of files to be copied
// to the working directory of Hadoop drivers and distributed tasks. Useful
// for naively parallel tasks.
repeated string file_uris = 5;
repeated string file_uris = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of archives to be extracted in the working directory of
// Hadoop drivers and tasks. Supported file types:
// .jar, .tar, .tar.gz, .tgz, or .zip.
repeated string archive_uris = 6;
repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure Hadoop.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in /etc/hadoop/conf/*-site and
// classes in user code.
map<string, string> properties = 7;
map<string, string> properties = 7 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 8;
LoggingConfig logging_config = 8 [(google.api.field_behavior) = OPTIONAL];
}
// A Cloud Dataproc job for running [Apache Spark](http://spark.apache.org/)
// A Dataproc job for running [Apache Spark](http://spark.apache.org/)
// applications on YARN.
// The specification of the main method to call to drive the job.
// Specify either the jar file that contains the main class or the main class
@ -205,32 +205,32 @@ message SparkJob {
// Optional. The arguments to pass to the driver. Do not include arguments,
// such as `--conf`, that can be set as job properties, since a collision may
// occur that causes an incorrect job submission.
repeated string args = 3;
repeated string args = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of jar files to add to the CLASSPATHs of the
// Spark driver and tasks.
repeated string jar_file_uris = 4;
repeated string jar_file_uris = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of files to be copied to the working directory of
// Spark drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
repeated string file_uris = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of archives to be extracted in the working directory
// of Spark drivers and tasks. Supported file types:
// .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure Spark.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in
// /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
map<string, string> properties = 7 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 8;
LoggingConfig logging_config = 8 [(google.api.field_behavior) = OPTIONAL];
}
// A Cloud Dataproc job for running
// A Dataproc job for running
// [Apache
// PySpark](https://spark.apache.org/docs/0.9.0/python-programming-guide.html)
// applications on YARN.
@ -242,32 +242,32 @@ message PySparkJob {
// Optional. The arguments to pass to the driver. Do not include arguments,
// such as `--conf`, that can be set as job properties, since a collision may
// occur that causes an incorrect job submission.
repeated string args = 2;
repeated string args = 2 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS file URIs of Python files to pass to the PySpark
// framework. Supported file types: .py, .egg, and .zip.
repeated string python_file_uris = 3;
repeated string python_file_uris = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of jar files to add to the CLASSPATHs of the
// Python driver and tasks.
repeated string jar_file_uris = 4;
repeated string jar_file_uris = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of files to be copied to the working directory of
// Python drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 5;
repeated string file_uris = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of archives to be extracted in the working directory of
// .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 6;
repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure PySpark.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in
// /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 7;
map<string, string> properties = 7 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 8;
LoggingConfig logging_config = 8 [(google.api.field_behavior) = OPTIONAL];
}
// A list of queries to run on a cluster.
@ -289,7 +289,7 @@ message QueryList {
repeated string queries = 1 [(google.api.field_behavior) = REQUIRED];
}
// A Cloud Dataproc job for running [Apache Hive](https://hive.apache.org/)
// A Dataproc job for running [Apache Hive](https://hive.apache.org/)
// queries on YARN.
message HiveJob {
// Required. The sequence of Hive queries to execute, specified as either
@ -305,25 +305,25 @@ message HiveJob {
// Optional. Whether to continue executing queries if a query fails.
// The default value is `false`. Setting to `true` can be useful when
// executing independent parallel queries.
bool continue_on_failure = 3;
bool continue_on_failure = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. Mapping of query variable names to values (equivalent to the
// Hive command: `SET name="value";`).
map<string, string> script_variables = 4;
map<string, string> script_variables = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names and values, used to configure Hive.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
// /etc/hive/conf/hive-site.xml, and classes in user code.
map<string, string> properties = 5;
map<string, string> properties = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of jar files to add to the CLASSPATH of the
// Hive server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes
// and UDFs.
repeated string jar_file_uris = 6;
repeated string jar_file_uris = 6 [(google.api.field_behavior) = OPTIONAL];
}
// A Cloud Dataproc job for running [Apache Spark
// A Dataproc job for running [Apache Spark
// SQL](http://spark.apache.org/sql/) queries.
message SparkSqlJob {
// Required. The sequence of Spark SQL queries to execute, specified as
@ -338,21 +338,21 @@ message SparkSqlJob {
// Optional. Mapping of query variable names to values (equivalent to the
// Spark SQL command: SET `name="value";`).
map<string, string> script_variables = 3;
map<string, string> script_variables = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure
// Spark SQL's SparkConf. Properties that conflict with values set by the
// Cloud Dataproc API may be overwritten.
map<string, string> properties = 4;
// Dataproc API may be overwritten.
map<string, string> properties = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH.
repeated string jar_file_uris = 56;
repeated string jar_file_uris = 56 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 6;
LoggingConfig logging_config = 6 [(google.api.field_behavior) = OPTIONAL];
}
// A Cloud Dataproc job for running [Apache Pig](https://pig.apache.org/)
// A Dataproc job for running [Apache Pig](https://pig.apache.org/)
// queries on YARN.
message PigJob {
// Required. The sequence of Pig queries to execute, specified as an HCFS
@ -368,27 +368,27 @@ message PigJob {
// Optional. Whether to continue executing queries if a query fails.
// The default value is `false`. Setting to `true` can be useful when
// executing independent parallel queries.
bool continue_on_failure = 3;
bool continue_on_failure = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. Mapping of query variable names to values (equivalent to the Pig
// command: `name=[value]`).
map<string, string> script_variables = 4;
map<string, string> script_variables = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure Pig.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in /etc/hadoop/conf/*-site.xml,
// /etc/pig/conf/pig.properties, and classes in user code.
map<string, string> properties = 5;
map<string, string> properties = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of jar files to add to the CLASSPATH of
// the Pig Client and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs.
repeated string jar_file_uris = 6;
repeated string jar_file_uris = 6 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 7;
LoggingConfig logging_config = 7 [(google.api.field_behavior) = OPTIONAL];
}
// A Cloud Dataproc job for running
// A Dataproc job for running
// [Apache SparkR](https://spark.apache.org/docs/latest/sparkr.html)
// applications on YARN.
message SparkRJob {
@ -399,38 +399,38 @@ message SparkRJob {
// Optional. The arguments to pass to the driver. Do not include arguments,
// such as `--conf`, that can be set as job properties, since a collision may
// occur that causes an incorrect job submission.
repeated string args = 2;
repeated string args = 2 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of files to be copied to the working directory of
// R drivers and distributed tasks. Useful for naively parallel tasks.
repeated string file_uris = 3;
repeated string file_uris = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. HCFS URIs of archives to be extracted in the working directory of
// Spark drivers and tasks. Supported file types:
// .jar, .tar, .tar.gz, .tgz, and .zip.
repeated string archive_uris = 4;
repeated string archive_uris = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. A mapping of property names to values, used to configure SparkR.
// Properties that conflict with values set by the Cloud Dataproc API may be
// Properties that conflict with values set by the Dataproc API may be
// overwritten. Can include properties set in
// /etc/spark/conf/spark-defaults.conf and classes in user code.
map<string, string> properties = 5;
map<string, string> properties = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. The runtime log config for job execution.
LoggingConfig logging_config = 6;
LoggingConfig logging_config = 6 [(google.api.field_behavior) = OPTIONAL];
}
// Cloud Dataproc job config.
// Dataproc job config.
message JobPlacement {
// Required. The name of the cluster where the job will be submitted.
string cluster_name = 1 [(google.api.field_behavior) = REQUIRED];
// Output only. A cluster UUID generated by the Cloud Dataproc service when
// Output only. A cluster UUID generated by the Dataproc service when
// the job is submitted.
string cluster_uuid = 2;
string cluster_uuid = 2 [(google.api.field_behavior) = OUTPUT_ONLY];
}
// Cloud Dataproc job status.
// Dataproc job status.
message JobStatus {
// The job state.
enum State {
@ -488,7 +488,7 @@ message JobStatus {
QUEUED = 2;
// The agent-reported status is out of date, which may be caused by a
// loss of communication between the agent and Cloud Dataproc. If the
// loss of communication between the agent and Dataproc. If the
// agent does not send a timely update, the job will fail.
//
// Applies to RUNNING state.
@ -496,18 +496,18 @@ message JobStatus {
}
// Output only. A state message specifying the overall job state.
State state = 1;
State state = 1 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. Optional job state details, such as an error
// Output only. Optional Job state details, such as an error
// description if the state is <code>ERROR</code>.
string details = 2;
string details = 2 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. The time when this state was entered.
google.protobuf.Timestamp state_start_time = 6;
google.protobuf.Timestamp state_start_time = 6 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. Additional state information, which includes
// status reported by the agent.
Substate substate = 7;
Substate substate = 7 [(google.api.field_behavior) = OUTPUT_ONLY];
}
// Encapsulates the full scoping used to reference a job.
@ -517,12 +517,11 @@ message JobReference {
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Optional. The job ID, which must be unique within the project.
//
// The ID must contain only letters (a-z, A-Z), numbers (0-9),
// underscores (_), or hyphens (-). The maximum length is 100 characters.
//
// If not specified by the caller, the job ID will be provided by the server.
string job_id = 2;
string job_id = 2 [(google.api.field_behavior) = OPTIONAL];
}
// A YARN application created by a job. Application information is a subset of
@ -571,20 +570,20 @@ message YarnApplication {
// Output only. The numerical progress of the application, from 1 to 100.
float progress = 3 [(google.api.field_behavior) = OUTPUT_ONLY];
// Optional. Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
// Output only. The HTTP URL of the ApplicationMaster, HistoryServer, or
// TimelineServer that provides application-specific information. The URL uses
// the internal hostname, and requires a proxy server for resolution and,
// possibly, access.
string tracking_url = 4 [(google.api.field_behavior) = OUTPUT_ONLY];
}
// A Cloud Dataproc job resource.
// A Dataproc job resource.
message Job {
// Optional. The fully qualified reference to the job, which can be used to
// obtain the equivalent REST path of the job resource. If this property
// is not specified when a job is created, the server generates a
// <code>job_id</code>.
JobReference reference = 1;
JobReference reference = 1 [(google.api.field_behavior) = OPTIONAL];
// Required. Job information, including how, when, and where to
// run the job.
@ -592,54 +591,47 @@ message Job {
// Required. The application/framework-specific portion of the job.
oneof type_job {
// Job is a Hadoop job.
HadoopJob hadoop_job = 3;
// Job is a Spark job.
SparkJob spark_job = 4;
// Job is a Pyspark job.
PySparkJob pyspark_job = 5;
// Job is a Hive job.
HiveJob hive_job = 6;
// Job is a Pig job.
PigJob pig_job = 7;
// Job is a SparkR job.
SparkRJob spark_r_job = 21;
// Job is a SparkSql job.
SparkSqlJob spark_sql_job = 12;
}
// Output only. The job status. Additional application-specific
// status information may be contained in the <code>type_job</code>
// and <code>yarn_applications</code> fields.
JobStatus status = 8;
JobStatus status = 8 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. The previous job status.
repeated JobStatus status_history = 13;
repeated JobStatus status_history = 13 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. The collection of YARN applications spun up by this job.
//
// **Beta** Feature: This report is available for testing purposes only. It
// may be changed before final release.
repeated YarnApplication yarn_applications = 9;
repeated YarnApplication yarn_applications = 9 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. The email address of the user submitting the job. For jobs
// submitted on the cluster, the address is <code>username@hostname</code>.
string submitted_by = 10;
string submitted_by = 10 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. A URI pointing to the location of the stdout of the job's
// driver program.
string driver_output_resource_uri = 17;
string driver_output_resource_uri = 17 [(google.api.field_behavior) = OUTPUT_ONLY];
// Output only. If present, the location of miscellaneous control files
// which may be used as part of job setup and handling. If not present,
// control files may be placed in the same location as `driver_output_uri`.
string driver_control_files_uri = 15;
string driver_control_files_uri = 15 [(google.api.field_behavior) = OUTPUT_ONLY];
// Optional. The labels to associate with this job.
// Label **keys** must contain 1 to 63 characters, and must conform to
@ -648,15 +640,15 @@ message Job {
// characters, and must conform to [RFC
// 1035](https://www.ietf.org/rfc/rfc1035.txt). No more than 32 labels can be
// associated with a job.
map<string, string> labels = 18;
map<string, string> labels = 18 [(google.api.field_behavior) = OPTIONAL];
// Optional. Job scheduling configuration.
JobScheduling scheduling = 20;
JobScheduling scheduling = 20 [(google.api.field_behavior) = OPTIONAL];
// Output only. A UUID that uniquely identifies a job within the project
// over time. This is in contrast to a user-settable reference.job_id that
// may be reused over time.
string job_uuid = 22;
string job_uuid = 22 [(google.api.field_behavior) = OUTPUT_ONLY];
}
// Job scheduling options.
@ -669,7 +661,7 @@ message JobScheduling {
// 4 times within 10 minute window.
//
// Maximum value is 10.
int32 max_failures_per_hour = 1;
int32 max_failures_per_hour = 1 [(google.api.field_behavior) = OPTIONAL];
}
// A request to submit a job.
@ -678,7 +670,7 @@ message SubmitJobRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The job resource.
@ -695,7 +687,7 @@ message SubmitJobRequest {
//
// The id must contain only letters (a-z, A-Z), numbers (0-9),
// underscores (_), and hyphens (-). The maximum length is 40 characters.
string request_id = 4;
string request_id = 4 [(google.api.field_behavior) = OPTIONAL];
}
// A request to get the resource representation for a job in a project.
@ -704,7 +696,7 @@ message GetJobRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The job ID.
@ -730,25 +722,25 @@ message ListJobsRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 6 [(google.api.field_behavior) = REQUIRED];
// Optional. The number of results to return in each response.
int32 page_size = 2;
int32 page_size = 2 [(google.api.field_behavior) = OPTIONAL];
// Optional. The page token, returned by a previous call, to request the
// next page of results.
string page_token = 3;
string page_token = 3 [(google.api.field_behavior) = OPTIONAL];
// Optional. If set, the returned jobs list includes only jobs that were
// submitted to the named cluster.
string cluster_name = 4;
string cluster_name = 4 [(google.api.field_behavior) = OPTIONAL];
// Optional. Specifies enumerated categories of jobs to list.
// (default = match ALL jobs).
//
// If `filter` is provided, `jobStateMatcher` will be ignored.
JobStateMatcher job_state_matcher = 5;
JobStateMatcher job_state_matcher = 5 [(google.api.field_behavior) = OPTIONAL];
// Optional. A filter constraining the jobs to list. Filters are
// case-sensitive and have the following syntax:
@ -764,7 +756,7 @@ message ListJobsRequest {
// Example filter:
//
// status.state = ACTIVE AND labels.env = staging AND labels.starred = *
string filter = 7;
string filter = 7 [(google.api.field_behavior) = OPTIONAL];
}
// A request to update a job.
@ -773,7 +765,7 @@ message UpdateJobRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 2 [(google.api.field_behavior) = REQUIRED];
// Required. The job ID.
@ -794,12 +786,12 @@ message UpdateJobRequest {
// A list of jobs in a project.
message ListJobsResponse {
// Output only. Jobs list.
repeated Job jobs = 1;
repeated Job jobs = 1 [(google.api.field_behavior) = OUTPUT_ONLY];
// Optional. This token is included in the response if there are more results
// to fetch. To fetch additional results, provide this value as the
// `page_token` in a subsequent <code>ListJobsRequest</code>.
string next_page_token = 2;
string next_page_token = 2 [(google.api.field_behavior) = OPTIONAL];
}
// A request to cancel a job.
@ -808,7 +800,7 @@ message CancelJobRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The job ID.
@ -821,7 +813,7 @@ message DeleteJobRequest {
// belongs to.
string project_id = 1 [(google.api.field_behavior) = REQUIRED];
// Required. The Cloud Dataproc region in which to handle the request.
// Required. The Dataproc region in which to handle the request.
string region = 3 [(google.api.field_behavior) = REQUIRED];
// Required. The job ID.

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";

View File

@ -1,4 +1,4 @@
// Copyright 2019 Google LLC.
// Copyright 2020 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@ -11,7 +11,6 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
//
syntax = "proto3";
@ -33,7 +32,7 @@ option java_outer_classname = "WorkflowTemplatesProto";
option java_package = "com.google.cloud.dataproc.v1beta2";
// The API interface for managing Workflow Templates in the
// Cloud Dataproc API.
// Dataproc API.
service WorkflowTemplateService {
option (google.api.default_host) = "dataproc.googleapis.com";
option (google.api.oauth_scopes) = "https://www.googleapis.com/auth/cloud-platform";
@ -78,9 +77,9 @@ service WorkflowTemplateService {
// clusters to be deleted.
//
// The [Operation.metadata][google.longrunning.Operation.metadata] will be
// [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1beta2#workflowmetadata).
// Also see [Using
// WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
//
// On successful completion,
// [Operation.response][google.longrunning.Operation.response] will be
@ -119,9 +118,9 @@ service WorkflowTemplateService {
// clusters to be deleted.
//
// The [Operation.metadata][google.longrunning.Operation.metadata] will be
// [WorkflowMetadata](/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
// [WorkflowMetadata](https://cloud.google.com/dataproc/docs/reference/rpc/google.cloud.dataproc.v1#workflowmetadata).
// Also see [Using
// WorkflowMetadata](/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
// WorkflowMetadata](https://cloud.google.com/dataproc/docs/concepts/workflows/debugging#using_workflowmetadata).
//
// On successful completion,
// [Operation.response][google.longrunning.Operation.response] will be
@ -179,7 +178,7 @@ service WorkflowTemplateService {
}
}
// A Cloud Dataproc workflow template resource.
// A Dataproc workflow template resource.
message WorkflowTemplate {
option (google.api.resource) = {
type: "dataproc.googleapis.com/WorkflowTemplate"
@ -327,22 +326,16 @@ message OrderedJob {
// Required. The job definition.
oneof job_type {
// Job is a Hadoop job.
HadoopJob hadoop_job = 2;
// Job is a Spark job.
SparkJob spark_job = 3;
// Job is a Pyspark job.
PySparkJob pyspark_job = 4;
// Job is a Hive job.
HiveJob hive_job = 5;
// Job is a Pig job.
PigJob pig_job = 6;
// Job is a SparkSql job.
SparkSqlJob spark_sql_job = 7;
}
@ -465,7 +458,7 @@ message ValueValidation {
repeated string values = 1;
}
// A Cloud Dataproc workflow template resource.
// A Dataproc workflow template resource.
message WorkflowMetadata {
// The operation state.
enum State {
@ -721,9 +714,7 @@ message UpdateWorkflowTemplateRequest {
// Required. The updated workflow template.
//
// The `template.version` field must match the current version.
WorkflowTemplate template = 1 [
(google.api.field_behavior) = REQUIRED
];
WorkflowTemplate template = 1 [(google.api.field_behavior) = REQUIRED];
}
// A request to list workflow templates in a project.