diff --git a/.github/.OwlBot.lock.yaml b/.github/.OwlBot.lock.yaml index 7d98291cc35..ff5126c188d 100644 --- a/.github/.OwlBot.lock.yaml +++ b/.github/.OwlBot.lock.yaml @@ -1,3 +1,3 @@ docker: image: gcr.io/cloud-devrel-public-resources/owlbot-python:latest - digest: sha256:58f73ba196b5414782605236dd0712a73541b44ff2ff4d3a36ec41092dd6fa5b + digest: sha256:dfa9b663b32de8b5b327e32c1da665a80de48876558dd58091d8160c60ad7355 diff --git a/.github/release-please.yml b/.github/release-please.yml index 4507ad0598a..466597e5b19 100644 --- a/.github/release-please.yml +++ b/.github/release-please.yml @@ -1 +1,2 @@ releaseType: python +handleGHRelease: true diff --git a/.github/release-trigger.yml b/.github/release-trigger.yml new file mode 100644 index 00000000000..d4ca94189e1 --- /dev/null +++ b/.github/release-trigger.yml @@ -0,0 +1 @@ +enabled: true diff --git a/CHANGELOG.md b/CHANGELOG.md index 34469bd30f9..efe50a35a64 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,37 @@ # Changelog +## [2.35.0](https://github.com/googleapis/google-api-python-client/compare/v2.34.0...v2.35.0) (2022-01-13) + + +### Features + +* **analyticsadmin:** update the api https://github.com/googleapis/google-api-python-client/commit/0dec51cdb8227155731e9b63a4092da556c02d1e ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **androidmanagement:** update the api https://github.com/googleapis/google-api-python-client/commit/975ea480b8050a519a344cdaecb297e0cde188f3 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **bigtableadmin:** update the api https://github.com/googleapis/google-api-python-client/commit/6266ac01c6507d12bfd65ae811b4f4edd9b6954a ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **cloudbuild:** update the api https://github.com/googleapis/google-api-python-client/commit/06362dce43356f7b6dd4253f31f3533f1465885e ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **compute:** update the api https://github.com/googleapis/google-api-python-client/commit/ce5edd17bfc87c22eddfdc697b16f0b6401d4ca9 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **connectors:** update the api https://github.com/googleapis/google-api-python-client/commit/bc9062b3d55b5b4fbb54e547b8399fbcf09a596b ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **contactcenterinsights:** update the api https://github.com/googleapis/google-api-python-client/commit/5fedb046b952461db978978802f4227f1611c333 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **container:** update the api https://github.com/googleapis/google-api-python-client/commit/a9bdbde1f8c095ecc8693f7cb3ad4c6390d2ed68 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **datacatalog:** update the api https://github.com/googleapis/google-api-python-client/commit/135360cca2be8f746940df79eefd576b07af125c ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **dataproc:** update the api https://github.com/googleapis/google-api-python-client/commit/38c6ad15c360531c96008b81563ca0245e859b0d ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **dialogflow:** update the api https://github.com/googleapis/google-api-python-client/commit/f4a3a0f31de62c1823c4168df67385deb8e8e2c8 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **dns:** update the api https://github.com/googleapis/google-api-python-client/commit/d039cf7eadc91759236bbfd6de2d123ae3eae948 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **eventarc:** update the api https://github.com/googleapis/google-api-python-client/commit/a8766f583f266fbe7b2d299c3992cc86b758ea31 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **networkservices:** update the api https://github.com/googleapis/google-api-python-client/commit/288618ff2301e4e6a49c5211ed7ae8cef978fc60 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **notebooks:** update the api https://github.com/googleapis/google-api-python-client/commit/66584093cbed2d24f3184def26e164026607e5ed ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **orgpolicy:** update the api https://github.com/googleapis/google-api-python-client/commit/35d0c20e2c3f7d418703d7b3eea7c353094985e5 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **prod_tt_sasportal:** update the api https://github.com/googleapis/google-api-python-client/commit/bbbf1b21186421b529edd01eb99b2a18bf812f9d ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **retail:** update the api https://github.com/googleapis/google-api-python-client/commit/a2a9eb8c0d104d3320b2928e568bef5e6b439541 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **run:** update the api https://github.com/googleapis/google-api-python-client/commit/af2b78b469d050c5eeca0996f7eaaea65930c1e5 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **script:** update the api https://github.com/googleapis/google-api-python-client/commit/79b1292ea77bfaebc837aacffac320e780452a0a ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) +* **texttospeech:** update the api https://github.com/googleapis/google-api-python-client/commit/46901569d6e918d86f2d9dc175ecdd2f8719bb93 ([4896942](https://github.com/googleapis/google-api-python-client/commit/48969422d118e0a30a166de6bd178d45573fe9f3)) + + +### Bug Fixes + +* remove adsense.v1.4.json ([#1616](https://github.com/googleapis/google-api-python-client/issues/1616)) ([5d3e588](https://github.com/googleapis/google-api-python-client/commit/5d3e5883b63d6038aa18e48397ce2eab50cb5e2c)) + ## [2.34.0](https://www.github.com/googleapis/google-api-python-client/compare/v2.33.0...v2.34.0) (2022-01-05) diff --git a/docs/dyn/analyticsadmin_v1alpha.accounts.html b/docs/dyn/analyticsadmin_v1alpha.accounts.html index 04257e281b9..1dd40d318d5 100644 --- a/docs/dyn/analyticsadmin_v1alpha.accounts.html +++ b/docs/dyn/analyticsadmin_v1alpha.accounts.html @@ -348,14 +348,6 @@
checkValidCreds(name, body=None, x__xgafv=None)
Returns true if valid credentials exist for the given data source and requesting user. Some data sources doesn't support service account, so we need to talk to them on behalf of the end user. This API just checks whether we have OAuth token for the particular user, which is a pre-requisite before user can create a transfer config.
+Returns true if valid credentials exist for the given data source and requesting user.
Close httplib2 connections.
-Retrieves a supported data source and returns its settings, which can be used for UI rendering.
+Retrieves a supported data source and returns its settings.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists supported data sources and returns their settings, which can be used for UI rendering.
+Lists supported data sources and returns their settings.
list_next(previous_request, previous_response)
Retrieves the next page of results.
checkValidCreds(name, body=None, x__xgafv=None)
- Returns true if valid credentials exist for the given data source and requesting user. Some data sources doesn't support service account, so we need to talk to them on behalf of the end user. This API just checks whether we have OAuth token for the particular user, which is a pre-requisite before user can create a transfer config.
+ Returns true if valid credentials exist for the given data source and requesting user.
Args:
name: string, Required. The data source in the form: `projects/{project_id}/dataSources/{data_source_id}` or `projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}`. (required)
@@ -122,7 +122,7 @@ Method Details
get(name, x__xgafv=None)
- Retrieves a supported data source and returns its settings, which can be used for UI rendering.
+ Retrieves a supported data source and returns its settings.
Args:
name: string, Required. The field will contain name of the resource requested, for example: `projects/{project_id}/dataSources/{data_source_id}` or `projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}` (required)
@@ -134,7 +134,7 @@ Method Details
Returns:
An object of the form:
- { # Represents data source metadata. Metadata is sufficient to render UI and request proper OAuth tokens.
+ { # Defines the properties and custom parameters for a data source.
"authorizationType": "A String", # Indicates the type of authorization.
"clientId": "A String", # Data source client id which should be used to receive refresh token.
"dataRefreshType": "A String", # Specifies whether the data source supports automatic data refresh for the past few days, and how it's supported. For some data sources, data might not be complete until a few days later, so it's useful to refresh data automatically.
@@ -148,7 +148,7 @@ Method Details
"minimumScheduleInterval": "A String", # The minimum interval for scheduler to schedule runs.
"name": "A String", # Output only. Data source resource name.
"parameters": [ # Data source parameters.
- { # Represents a data source parameter with validation rules, so that parameters can be rendered in the UI. These parameters are given to us by supported data sources, and include all needed information for rendering and validation. Thus, whoever uses this api can decide to generate either generic ui, or custom data source specific forms.
+ { # A parameter used to define custom fields in a data source definition.
"allowedValues": [ # All possible values for the parameter.
"A String",
],
@@ -183,7 +183,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists supported data sources and returns their settings, which can be used for UI rendering.
+ Lists supported data sources and returns their settings.
Args:
parent: string, Required. The BigQuery project id for which data sources should be returned. Must be in the form: `projects/{project_id}` or `projects/{project_id}/locations/{location_id} (required)
@@ -199,7 +199,7 @@ Method Details
{ # Returns list of supported data sources and their metadata.
"dataSources": [ # List of supported data sources and their transfer settings.
- { # Represents data source metadata. Metadata is sufficient to render UI and request proper OAuth tokens.
+ { # Defines the properties and custom parameters for a data source.
"authorizationType": "A String", # Indicates the type of authorization.
"clientId": "A String", # Data source client id which should be used to receive refresh token.
"dataRefreshType": "A String", # Specifies whether the data source supports automatic data refresh for the past few days, and how it's supported. For some data sources, data might not be complete until a few days later, so it's useful to refresh data automatically.
@@ -213,7 +213,7 @@ Method Details
"minimumScheduleInterval": "A String", # The minimum interval for scheduler to schedule runs.
"name": "A String", # Output only. Data source resource name.
"parameters": [ # Data source parameters.
- { # Represents a data source parameter with validation rules, so that parameters can be rendered in the UI. These parameters are given to us by supported data sources, and include all needed information for rendering and validation. Thus, whoever uses this api can decide to generate either generic ui, or custom data source specific forms.
+ { # A parameter used to define custom fields in a data source definition.
"allowedValues": [ # All possible values for the parameter.
"A String",
],
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.locations.dataSources.html b/docs/dyn/bigquerydatatransfer_v1.projects.locations.dataSources.html
index bebfab1b90d..3f74b3194ee 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.locations.dataSources.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.locations.dataSources.html
@@ -76,23 +76,23 @@ BigQuery Data Transfer API . Instance Methods
checkValidCreds(name, body=None, x__xgafv=None)
-Returns true if valid credentials exist for the given data source and requesting user. Some data sources doesn't support service account, so we need to talk to them on behalf of the end user. This API just checks whether we have OAuth token for the particular user, which is a pre-requisite before user can create a transfer config.
+Returns true if valid credentials exist for the given data source and requesting user.
Close httplib2 connections.
-Retrieves a supported data source and returns its settings, which can be used for UI rendering.
+Retrieves a supported data source and returns its settings.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists supported data sources and returns their settings, which can be used for UI rendering.
+Lists supported data sources and returns their settings.
list_next(previous_request, previous_response)
Retrieves the next page of results.
Method Details
checkValidCreds(name, body=None, x__xgafv=None)
- Returns true if valid credentials exist for the given data source and requesting user. Some data sources doesn't support service account, so we need to talk to them on behalf of the end user. This API just checks whether we have OAuth token for the particular user, which is a pre-requisite before user can create a transfer config.
+ Returns true if valid credentials exist for the given data source and requesting user.
Args:
name: string, Required. The data source in the form: `projects/{project_id}/dataSources/{data_source_id}` or `projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}`. (required)
@@ -122,7 +122,7 @@ Method Details
get(name, x__xgafv=None)
- Retrieves a supported data source and returns its settings, which can be used for UI rendering.
+ Retrieves a supported data source and returns its settings.
Args:
name: string, Required. The field will contain name of the resource requested, for example: `projects/{project_id}/dataSources/{data_source_id}` or `projects/{project_id}/locations/{location_id}/dataSources/{data_source_id}` (required)
@@ -134,7 +134,7 @@ Method Details
Returns:
An object of the form:
- { # Represents data source metadata. Metadata is sufficient to render UI and request proper OAuth tokens.
+ { # Defines the properties and custom parameters for a data source.
"authorizationType": "A String", # Indicates the type of authorization.
"clientId": "A String", # Data source client id which should be used to receive refresh token.
"dataRefreshType": "A String", # Specifies whether the data source supports automatic data refresh for the past few days, and how it's supported. For some data sources, data might not be complete until a few days later, so it's useful to refresh data automatically.
@@ -148,7 +148,7 @@ Method Details
"minimumScheduleInterval": "A String", # The minimum interval for scheduler to schedule runs.
"name": "A String", # Output only. Data source resource name.
"parameters": [ # Data source parameters.
- { # Represents a data source parameter with validation rules, so that parameters can be rendered in the UI. These parameters are given to us by supported data sources, and include all needed information for rendering and validation. Thus, whoever uses this api can decide to generate either generic ui, or custom data source specific forms.
+ { # A parameter used to define custom fields in a data source definition.
"allowedValues": [ # All possible values for the parameter.
"A String",
],
@@ -183,7 +183,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists supported data sources and returns their settings, which can be used for UI rendering.
+ Lists supported data sources and returns their settings.
Args:
parent: string, Required. The BigQuery project id for which data sources should be returned. Must be in the form: `projects/{project_id}` or `projects/{project_id}/locations/{location_id} (required)
@@ -199,7 +199,7 @@ Method Details
{ # Returns list of supported data sources and their metadata.
"dataSources": [ # List of supported data sources and their transfer settings.
- { # Represents data source metadata. Metadata is sufficient to render UI and request proper OAuth tokens.
+ { # Defines the properties and custom parameters for a data source.
"authorizationType": "A String", # Indicates the type of authorization.
"clientId": "A String", # Data source client id which should be used to receive refresh token.
"dataRefreshType": "A String", # Specifies whether the data source supports automatic data refresh for the past few days, and how it's supported. For some data sources, data might not be complete until a few days later, so it's useful to refresh data automatically.
@@ -213,7 +213,7 @@ Method Details
"minimumScheduleInterval": "A String", # The minimum interval for scheduler to schedule runs.
"name": "A String", # Output only. Data source resource name.
"parameters": [ # Data source parameters.
- { # Represents a data source parameter with validation rules, so that parameters can be rendered in the UI. These parameters are given to us by supported data sources, and include all needed information for rendering and validation. Thus, whoever uses this api can decide to generate either generic ui, or custom data source specific forms.
+ { # A parameter used to define custom fields in a data source definition.
"allowedValues": [ # All possible values for the parameter.
"A String",
],
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.html b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.html
index f0287f93104..0a07c664fef 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.html
@@ -90,7 +90,7 @@ Instance Methods
Returns information about the particular transfer run.
list(parent, pageSize=None, pageToken=None, runAttempt=None, states=None, x__xgafv=None)
-Returns information about running and completed jobs.
+Returns information about running and completed transfer runs.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -165,7 +165,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, runAttempt=None, states=None, x__xgafv=None)
- Returns information about running and completed jobs.
+ Returns information about running and completed transfer runs.
Args:
parent: string, Required. Name of transfer configuration for which transfer runs should be retrieved. Format of transfer configuration resource name is: `projects/{project_id}/transferConfigs/{config_id}` or `projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}`. (required)
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.transferLogs.html b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.transferLogs.html
index 430c3522414..89c1435d42f 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.transferLogs.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.locations.transferConfigs.runs.transferLogs.html
@@ -79,7 +79,7 @@ Instance Methods
Close httplib2 connections.
list(parent, messageTypes=None, pageSize=None, pageToken=None, x__xgafv=None)
-Returns user facing log messages for the data transfer run.
+Returns log messages for the transfer run.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -91,7 +91,7 @@ Method Details
list(parent, messageTypes=None, pageSize=None, pageToken=None, x__xgafv=None)
- Returns user facing log messages for the data transfer run.
+ Returns log messages for the transfer run.
Args:
parent: string, Required. Transfer run name in the form: `projects/{project_id}/transferConfigs/{config_id}/runs/{run_id}` or `projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}/runs/{run_id}` (required)
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.html b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.html
index 8e4fc26510c..1d5c568ab15 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.html
@@ -90,7 +90,7 @@ Instance Methods
Returns information about the particular transfer run.
list(parent, pageSize=None, pageToken=None, runAttempt=None, states=None, x__xgafv=None)
-Returns information about running and completed jobs.
+Returns information about running and completed transfer runs.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -165,7 +165,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, runAttempt=None, states=None, x__xgafv=None)
- Returns information about running and completed jobs.
+ Returns information about running and completed transfer runs.
Args:
parent: string, Required. Name of transfer configuration for which transfer runs should be retrieved. Format of transfer configuration resource name is: `projects/{project_id}/transferConfigs/{config_id}` or `projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}`. (required)
diff --git a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.transferLogs.html b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.transferLogs.html
index b84460c27b7..e35e08f985c 100644
--- a/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.transferLogs.html
+++ b/docs/dyn/bigquerydatatransfer_v1.projects.transferConfigs.runs.transferLogs.html
@@ -79,7 +79,7 @@ Instance Methods
Close httplib2 connections.
list(parent, messageTypes=None, pageSize=None, pageToken=None, x__xgafv=None)
-Returns user facing log messages for the data transfer run.
+Returns log messages for the transfer run.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -91,7 +91,7 @@ Method Details
list(parent, messageTypes=None, pageSize=None, pageToken=None, x__xgafv=None)
- Returns user facing log messages for the data transfer run.
+ Returns log messages for the transfer run.
Args:
parent: string, Required. Transfer run name in the form: `projects/{project_id}/transferConfigs/{config_id}/runs/{run_id}` or `projects/{project_id}/locations/{location_id}/transferConfigs/{config_id}/runs/{run_id}` (required)
diff --git a/docs/dyn/calendar_v3.events.html b/docs/dyn/calendar_v3.events.html
index aa4ace9f4d9..e56d9b6c6a9 100644
--- a/docs/dyn/calendar_v3.events.html
+++ b/docs/dyn/calendar_v3.events.html
@@ -288,7 +288,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -554,7 +554,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -816,7 +816,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -1082,7 +1082,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -1353,7 +1353,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -1650,7 +1650,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -1992,7 +1992,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -2291,7 +2291,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -2558,7 +2558,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -2830,7 +2830,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -3106,7 +3106,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -3373,7 +3373,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
@@ -3645,7 +3645,7 @@ Method Details
},
},
"signature": "A String", # The signature of the conference data.
- # Generated on server side. Must be preserved while copying the conference data between events, otherwise the conference data will not be copied.
+ # Generated on server side.
# Unset for a conference with a failed create request.
# Optional for a conference with a pending create request.
},
diff --git a/docs/dyn/cloudbuild_v1.html b/docs/dyn/cloudbuild_v1.html
index ed9c61f05c8..da0ff4a7033 100644
--- a/docs/dyn/cloudbuild_v1.html
+++ b/docs/dyn/cloudbuild_v1.html
@@ -74,6 +74,11 @@
Cloud Build API
Instance Methods
+
+ locations()
+
+Returns the locations Resource.
+
diff --git a/docs/dyn/cloudbuild_v1.locations.html b/docs/dyn/cloudbuild_v1.locations.html
new file mode 100644
index 00000000000..047d9310059
--- /dev/null
+++ b/docs/dyn/cloudbuild_v1.locations.html
@@ -0,0 +1,121 @@
+
+
+
+Cloud Build API . locations
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ regionalWebhook(location, body=None, webhookKey=None, x__xgafv=None)
+ReceiveRegionalWebhook is called when the API receives a regional GitHub webhook.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ regionalWebhook(location, body=None, webhookKey=None, x__xgafv=None)
+ ReceiveRegionalWebhook is called when the API receives a regional GitHub webhook.
+
+Args:
+ location: string, Required. The location where the webhook should be sent. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Message that represents an arbitrary HTTP body. It should only be used for payload formats that can't be represented as JSON, such as raw binary or an HTML page. This message can be used both in streaming and non-streaming API methods in the request as well as the response. It can be used as a top-level request field, which is convenient if one wants to extract parameters from either the URL or HTTP template into the request fields and also want access to the raw HTTP body. Example: message GetResourceRequest { // A unique request id. string request_id = 1; // The raw HTTP body is bound to this field. google.api.HttpBody http_body = 2; } service ResourceService { rpc GetResource(GetResourceRequest) returns (google.api.HttpBody); rpc UpdateResource(google.api.HttpBody) returns (google.protobuf.Empty); } Example with streaming methods: service CaldavService { rpc GetCalendar(stream google.api.HttpBody) returns (stream google.api.HttpBody); rpc UpdateCalendar(stream google.api.HttpBody) returns (stream google.api.HttpBody); } Use of this type only changes how the request and response bodies are handled, all other features will continue to work unchanged.
+ "contentType": "A String", # The HTTP Content-Type header value specifying the content type of the body.
+ "data": "A String", # The HTTP request/response body as raw binary.
+ "extensions": [ # Application specific response metadata. Must be set in the first response for streaming APIs.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+}
+
+ webhookKey: string, For GitHub Enterprise webhooks, this key is used to associate the webhook request with the GitHubEnterpriseConfig to use for validation.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.connectedRepositories.html b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.connectedRepositories.html
new file mode 100644
index 00000000000..e2f78463e83
--- /dev/null
+++ b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.connectedRepositories.html
@@ -0,0 +1,152 @@
+
+
+
+Cloud Build API . projects . locations . bitbucketServerConfigs . connectedRepositories
+Instance Methods
+
+ batchCreate(parent, body=None, x__xgafv=None)
+Batch connecting Bitbucket Server repositories to Cloud Build.
+
+ close()
+Close httplib2 connections.
+Method Details
+
+ batchCreate(parent, body=None, x__xgafv=None)
+ Batch connecting Bitbucket Server repositories to Cloud Build.
+
+Args:
+ parent: string, The name of the `BitbucketServerConfig` that added connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}` (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # RPC request object accepted by BatchCreateBitbucketServerConnectedRepositories RPC method.
+ "requests": [ # Required. Requests to connect Bitbucket Server repositories.
+ { # Request to connect a repository from a connected Bitbucket Server host.
+ "bitbucketServerConnectedRepository": { # / BitbucketServerConnectedRepository represents a connected Bitbucket Server / repository. # Required. The Bitbucket Server repository to connect.
+ "parent": "A String", # The name of the `BitbucketServerConfig` that added connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}`
+ "repo": { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server. # The Bitbucket Server repositories to connect.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ "status": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. The status of the repo connection request.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ },
+ "parent": "A String", # Required. The name of the `BitbucketServerConfig` that added connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}`
+ },
+ ],
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ close()
+ Close httplib2 connections.
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.html b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.html
new file mode 100644
index 00000000000..84823c2a951
--- /dev/null
+++ b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.html
@@ -0,0 +1,440 @@
+
+
+
+Cloud Build API . projects . locations . bitbucketServerConfigs
+Instance Methods
+
+Returns the connectedRepositories Resource.
+
+
+ repos()
+
+Returns the repos Resource.
+
+
+ addBitbucketServerConnectedRepository(config, body=None, x__xgafv=None)
+Add a Bitbucket Server repository to a given BitbucketServerConfig's connected repositories. This API is experimental.
+
+ close()
+Close httplib2 connections.
+
+ create(parent, bitbucketServerConfigId=None, body=None, x__xgafv=None)
+Creates a new `BitbucketServerConfig`. This API is experimental.
+
+Delete a `BitbucketServerConfig`. This API is experimental.
+
+Retrieve a `BitbucketServerConfig`. This API is experimental.
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+List all `BitbucketServerConfigs` for a given project. This API is experimental.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates an existing `BitbucketServerConfig`. This API is experimental.
+
+ removeBitbucketServerConnectedRepository(config, body=None, x__xgafv=None)
+Remove a Bitbucket Server repository from an given BitbucketServerConfig’s connected repositories. This API is experimental.
+Method Details
+
+ addBitbucketServerConnectedRepository(config, body=None, x__xgafv=None)
+ Add a Bitbucket Server repository to a given BitbucketServerConfig's connected repositories. This API is experimental.
+
+Args:
+ config: string, Required. The name of the `BitbucketServerConfig` to add a connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}` (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # RPC request object accepted by the AddBitbucketServerConnectedRepository RPC method.
+ "connectedRepository": { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server. # The connected repository to add.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # RPC request object returned by the AddBitbucketServerConnectedRepository RPC method.
+ "config": "A String", # The name of the `BitbucketServerConfig` that added connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}`
+ "connectedRepository": { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server. # The connected repository.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+}
+
+
+
+ close()
+ Close httplib2 connections.
+
+
+
+ create(parent, bitbucketServerConfigId=None, body=None, x__xgafv=None)
+ Creates a new `BitbucketServerConfig`. This API is experimental.
+
+Args:
+ parent: string, Required. Name of the parent resource. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # BitbucketServerConfig represents the configuration for a Bitbucket Server.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+}
+
+ bitbucketServerConfigId: string, Optional. The ID to use for the BitbucketServerConfig, which will become the final component of the BitbucketServerConfig's resource name. bitbucket_server_config_id must meet the following requirements: + They must contain only alphanumeric characters and dashes. + They can be 1-64 characters long. + They must begin and end with an alphanumeric character.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ delete(name, x__xgafv=None)
+ Delete a `BitbucketServerConfig`. This API is experimental.
+
+Args:
+ name: string, Required. The config resource name. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ get(name, x__xgafv=None)
+ Retrieve a `BitbucketServerConfig`. This API is experimental.
+
+Args:
+ name: string, Required. The config resource name. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # BitbucketServerConfig represents the configuration for a Bitbucket Server.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+}
+
+
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ List all `BitbucketServerConfigs` for a given project. This API is experimental.
+
+Args:
+ parent: string, Required. Name of the parent resource. (required)
+ pageSize: integer, The maximum number of configs to return. The service may return fewer than this value. If unspecified, at most 50 configs will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000.
+ pageToken: string, A page token, received from a previous `ListBitbucketServerConfigsRequest` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListBitbucketServerConfigsRequest` must match the call that provided the page token.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # RPC response object returned by ListBitbucketServerConfigs RPC method.
+ "bitbucketServerConfigs": [ # A list of BitbucketServerConfigs
+ { # BitbucketServerConfig represents the configuration for a Bitbucket Server.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ ],
+ "nextPageToken": "A String", # A token that can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates an existing `BitbucketServerConfig`. This API is experimental.
+
+Args:
+ name: string, The resource name for the config. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # BitbucketServerConfig represents the configuration for a Bitbucket Server.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+}
+
+ updateMask: string, Update mask for the resource. If this is set, the server will only update the fields specified in the field mask. Otherwise, a full update of the mutable resource fields will be performed.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # This resource represents a long-running operation that is the result of a network API call.
+ "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+ "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+ },
+ "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+ "response": { # The normal response of the operation in case of success. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+}
+
+
+
+ removeBitbucketServerConnectedRepository(config, body=None, x__xgafv=None)
+ Remove a Bitbucket Server repository from an given BitbucketServerConfig’s connected repositories. This API is experimental.
+
+Args:
+ config: string, Required. The name of the `BitbucketServerConfig` to remove a connected repository. Format: `projects/{project}/locations/{location}/bitbucketServerConfigs/{config}` (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # RPC request object accepted by RemoveBitbucketServerConnectedRepository RPC method.
+ "connectedRepository": { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server. # The connected repository to remove.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.repos.html b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.repos.html
new file mode 100644
index 00000000000..6c7eac622c3
--- /dev/null
+++ b/docs/dyn/cloudbuild_v1.projects.locations.bitbucketServerConfigs.repos.html
@@ -0,0 +1,140 @@
+
+
+
+Cloud Build API . projects . locations . bitbucketServerConfigs . repos
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+List all repositories for a given `BitbucketServerConfig`. This API is experimental.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ List all repositories for a given `BitbucketServerConfig`. This API is experimental.
+
+Args:
+ parent: string, Required. Name of the parent resource. (required)
+ pageSize: integer, The maximum number of configs to return. The service may return fewer than this value. If unspecified, at most 50 configs will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000.
+ pageToken: string, A page token, received from a previous `ListBitbucketServerRepositoriesRequest` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListBitbucketServerConfigsRequest` must match the call that provided the page token.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # RPC response object returned by the ListBitbucketServerRepositories RPC method.
+ "bitbucketServerRepositories": [ # List of Bitbucket Server repositories.
+ { # BitbucketServerRepository represents a repository hosted on a Bitbucket Server.
+ "browseUri": "A String", # Link to the browse repo page on the Bitbucket Server instance.
+ "description": "A String", # Description of the repository.
+ "displayName": "A String", # Display name of the repository.
+ "name": "A String", # The resource name of the repository.
+ "repoId": { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server. # Identifier for a repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ },
+ ],
+ "nextPageToken": "A String", # A token that can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.html b/docs/dyn/cloudbuild_v1.projects.locations.html
index a977d5c0407..44a4d690402 100644
--- a/docs/dyn/cloudbuild_v1.projects.locations.html
+++ b/docs/dyn/cloudbuild_v1.projects.locations.html
@@ -74,6 +74,11 @@
Cloud Build API . projects . locations
Instance Methods
+
+Returns the bitbucketServerConfigs Resource.
+
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.triggers.html b/docs/dyn/cloudbuild_v1.projects.locations.triggers.html
index b7ea054de8d..feaf58ee245 100644
--- a/docs/dyn/cloudbuild_v1.projects.locations.triggers.html
+++ b/docs/dyn/cloudbuild_v1.projects.locations.triggers.html
@@ -121,6 +121,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -364,7 +401,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -448,6 +485,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -691,7 +765,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -803,6 +877,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1046,7 +1157,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1142,6 +1253,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1385,7 +1533,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1486,6 +1634,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1729,7 +1914,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1814,6 +1999,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -2057,7 +2279,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -2139,7 +2361,7 @@ Method Details
{ # Specifies a build trigger to run and the source to use.
"projectId": "A String", # Required. ID of the project.
- "source": { # Location of the source in a Google Cloud Source Repository. # Source to build against this trigger.
+ "source": { # Location of the source in a Google Cloud Source Repository. # Source to build against this trigger. Branch and tag names cannot consist of regular expressions.
"branchName": "A String", # Regex matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
"commitSha": "A String", # Explicit commit SHA to build.
"dir": "A String", # Directory, relative to the source root, in which to run the build. This must be a relative path. If a step's `dir` is specified and is an absolute path, this value is ignored for that step's execution.
diff --git a/docs/dyn/cloudbuild_v1.projects.locations.workerPools.html b/docs/dyn/cloudbuild_v1.projects.locations.workerPools.html
index 568621b8cee..5c31c2fb652 100644
--- a/docs/dyn/cloudbuild_v1.projects.locations.workerPools.html
+++ b/docs/dyn/cloudbuild_v1.projects.locations.workerPools.html
@@ -119,7 +119,7 @@ Method Details
"displayName": "A String", # A user-specified, human-readable name for the `WorkerPool`. If provided, this value must be 1-63 characters.
"etag": "A String", # Output only. Checksum computed by the server. May be sent on update and delete requests to ensure that the client has an up-to-date value before proceeding.
"name": "A String", # Output only. The resource name of the `WorkerPool`, with format `projects/{project}/locations/{location}/workerPools/{worker_pool}`. The value of `{worker_pool}` is provided by `worker_pool_id` in `CreateWorkerPool` request and the value of `{location}` is determined by the endpoint accessed.
- "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Private Pool using a v1 configuration.
+ "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Legacy Private Pool configuration.
"networkConfig": { # Defines the network configuration for the pool. # Network configuration for the pool.
"egressOption": "A String", # Option to configure network egress for the workers.
"peeredNetwork": "A String", # Required. Immutable. The network definition that the workers are peered to. If this section is left empty, the workers will be peered to `WorkerPool.project_id` on the service producer network. Must be in the format `projects/{project}/global/networks/{network}`, where `{project}` is a project number, such as `12345`, and `{network}` is the name of a VPC network in the project. See [Understanding network configuration options](https://cloud.google.com/build/docs/private-pools/set-up-private-pool-environment)
@@ -226,7 +226,7 @@ Method Details
"displayName": "A String", # A user-specified, human-readable name for the `WorkerPool`. If provided, this value must be 1-63 characters.
"etag": "A String", # Output only. Checksum computed by the server. May be sent on update and delete requests to ensure that the client has an up-to-date value before proceeding.
"name": "A String", # Output only. The resource name of the `WorkerPool`, with format `projects/{project}/locations/{location}/workerPools/{worker_pool}`. The value of `{worker_pool}` is provided by `worker_pool_id` in `CreateWorkerPool` request and the value of `{location}` is determined by the endpoint accessed.
- "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Private Pool using a v1 configuration.
+ "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Legacy Private Pool configuration.
"networkConfig": { # Defines the network configuration for the pool. # Network configuration for the pool.
"egressOption": "A String", # Option to configure network egress for the workers.
"peeredNetwork": "A String", # Required. Immutable. The network definition that the workers are peered to. If this section is left empty, the workers will be peered to `WorkerPool.project_id` on the service producer network. Must be in the format `projects/{project}/global/networks/{network}`, where `{project}` is a project number, such as `12345`, and `{network}` is the name of a VPC network in the project. See [Understanding network configuration options](https://cloud.google.com/build/docs/private-pools/set-up-private-pool-environment)
@@ -270,7 +270,7 @@ Method Details
"displayName": "A String", # A user-specified, human-readable name for the `WorkerPool`. If provided, this value must be 1-63 characters.
"etag": "A String", # Output only. Checksum computed by the server. May be sent on update and delete requests to ensure that the client has an up-to-date value before proceeding.
"name": "A String", # Output only. The resource name of the `WorkerPool`, with format `projects/{project}/locations/{location}/workerPools/{worker_pool}`. The value of `{worker_pool}` is provided by `worker_pool_id` in `CreateWorkerPool` request and the value of `{location}` is determined by the endpoint accessed.
- "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Private Pool using a v1 configuration.
+ "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Legacy Private Pool configuration.
"networkConfig": { # Defines the network configuration for the pool. # Network configuration for the pool.
"egressOption": "A String", # Option to configure network egress for the workers.
"peeredNetwork": "A String", # Required. Immutable. The network definition that the workers are peered to. If this section is left empty, the workers will be peered to `WorkerPool.project_id` on the service producer network. Must be in the format `projects/{project}/global/networks/{network}`, where `{project}` is a project number, such as `12345`, and `{network}` is the name of a VPC network in the project. See [Understanding network configuration options](https://cloud.google.com/build/docs/private-pools/set-up-private-pool-environment)
@@ -320,7 +320,7 @@ Method Details
"displayName": "A String", # A user-specified, human-readable name for the `WorkerPool`. If provided, this value must be 1-63 characters.
"etag": "A String", # Output only. Checksum computed by the server. May be sent on update and delete requests to ensure that the client has an up-to-date value before proceeding.
"name": "A String", # Output only. The resource name of the `WorkerPool`, with format `projects/{project}/locations/{location}/workerPools/{worker_pool}`. The value of `{worker_pool}` is provided by `worker_pool_id` in `CreateWorkerPool` request and the value of `{location}` is determined by the endpoint accessed.
- "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Private Pool using a v1 configuration.
+ "privatePoolV1Config": { # Configuration for a V1 `PrivatePool`. # Legacy Private Pool configuration.
"networkConfig": { # Defines the network configuration for the pool. # Network configuration for the pool.
"egressOption": "A String", # Option to configure network egress for the workers.
"peeredNetwork": "A String", # Required. Immutable. The network definition that the workers are peered to. If this section is left empty, the workers will be peered to `WorkerPool.project_id` on the service producer network. Must be in the format `projects/{project}/global/networks/{network}`, where `{project}` is a project number, such as `12345`, and `{network}` is the name of a VPC network in the project. See [Understanding network configuration options](https://cloud.google.com/build/docs/private-pools/set-up-private-pool-environment)
diff --git a/docs/dyn/cloudbuild_v1.projects.triggers.html b/docs/dyn/cloudbuild_v1.projects.triggers.html
index 4420bc2b330..5b42eed3359 100644
--- a/docs/dyn/cloudbuild_v1.projects.triggers.html
+++ b/docs/dyn/cloudbuild_v1.projects.triggers.html
@@ -121,6 +121,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -364,7 +401,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -448,6 +485,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -691,7 +765,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -803,6 +877,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1046,7 +1157,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1142,6 +1253,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1385,7 +1533,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1487,6 +1635,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -1730,7 +1915,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
@@ -1813,6 +1998,43 @@ Method Details
"approvalRequired": True or False, # Whether or not approval is needed. If this is set on a build, it will become pending when created, and will need to be explicitly approved to start.
},
"autodetect": True or False, # Autodetect build configuration. The following precedence is used (case insensitive): 1. cloudbuild.yaml 2. cloudbuild.yml 3. cloudbuild.json 4. Dockerfile Currently only available for GitHub App Triggers.
+ "bitbucketServerTriggerConfig": { # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received. # BitbucketServerTriggerConfig describes the configuration of a trigger that creates a build whenever a Bitbucket Server event is received.
+ "bitbucketServerConfig": { # BitbucketServerConfig represents the configuration for a Bitbucket Server. # Output only. The BitbucketServerConfig specified in the bitbucket_server_config_resource field.
+ "apiKey": "A String", # Required. Immutable. API Key that will be attached to webhook. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "connectedRepositories": [ # Output only. Connected Bitbucket Server repositories for this config.
+ { # BitbucketServerRepositoryId identifies a specific repository hosted on a Bitbucket Server.
+ "projectKey": "A String", # Required. Identifier for the project storing the repository.
+ "repoSlug": "A String", # Required. Identifier for the repository.
+ "webhookId": 42, # Output only. The ID of the webhook that was created for receiving events from this repo. We only create and manage a single webhook for each repo.
+ },
+ ],
+ "createTime": "A String", # Time when the config was created.
+ "hostUri": "A String", # Required. Immutable. The URI of the Bitbucket Server host. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ "name": "A String", # The resource name for the config.
+ "peeredNetwork": "A String", # Optional. The network to be used when reaching out to the Bitbucket Server instance. The VPC network must be enabled for private service connection. This should be set if the Bitbucket Server instance is hosted on-premises and not reachable by public internet. If this field is left empty, no network peering will occur and calls to the Bitbucket Server instance will be made over the public internet. Must be in the format `projects/{project}/global/networks/{network}`, where {project} is a project number or id and {network} is the name of a VPC network in the project.
+ "secrets": { # BitbucketServerSecrets represents the secrets in Secret Manager for a Bitbucket Server. # Required. Secret Manager secrets needed by the config.
+ "adminAccessTokenVersionName": "A String", # Required. The resource name for the admin access token's secret version.
+ "readAccessTokenVersionName": "A String", # Required. The resource name for the read access token's secret version.
+ "webhookSecretVersionName": "A String", # Required. Immutable. The resource name for the webhook secret's secret version. Once this field has been set, it cannot be changed. If you need to change it, please create another BitbucketServerConfig.
+ },
+ "sslCa": "A String", # Optional. SSL certificate to use for requests to Bitbucket Server. The format should be PEM format but the extension can be one of .pem, .cer, or .crt.
+ "username": "A String", # Username of the account Cloud Build will use on Bitbucket Server.
+ "webhookKey": "A String", # Output only. UUID included in webhook requests. The UUID is used to look up the corresponding config.
+ },
+ "bitbucketServerConfigResource": "A String", # Required. The Bitbucket server config resource that this trigger config maps to.
+ "projectKey": "A String", # Required. Key of the project that the repo is in. For example: The key for http://mybitbucket.server/projects/TEST/repos/test-repo is "TEST".
+ "pullRequest": { # PullRequestFilter contains filter properties for matching GitHub Pull Requests. # Filter to match changes in pull requests.
+ "branch": "A String", # Regex of branches to match. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "commentControl": "A String", # Configure builds to run whether a repository owner or collaborator need to comment `/gcbrun`.
+ "invertRegex": True or False, # If true, branches that do NOT match the git_ref will trigger a build.
+ },
+ "push": { # Push contains filter properties for matching GitHub git pushes. # Filter to match changes in refs like branches, tags.
+ "branch": "A String", # Regexes matching branches to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ "invertRegex": True or False, # When true, only trigger a build if the revision regex does NOT match the git_ref regex.
+ "tag": "A String", # Regexes matching tags to build. The syntax of the regular expressions accepted is the syntax accepted by RE2 and described at https://github.com/google/re2/wiki/Syntax
+ },
+ "repoSlug": "A String", # Required. Slug of the repository. A repository slug is a URL-friendly version of a repository name, automatically generated by Bitbucket for use in the URL. For example, if the repository name is 'test repo', in the URL it would become 'test-repo' as in http://mybitbucket.server/projects/TEST/repos/test-repo.
+ },
"build": { # A build resource in the Cloud Build API. At a high level, a `Build` describes where to find source code, how to build it (for example, the builder image to run on the source), and where to store the built artifacts. Fields can include the following variables, which will be expanded when the build is created: - $PROJECT_ID: the project ID of the build. - $PROJECT_NUMBER: the project number of the build. - $LOCATION: the location/region of the build. - $BUILD_ID: the autogenerated ID of the build. - $REPO_NAME: the source repository name specified by RepoSource. - $BRANCH_NAME: the branch name specified by RepoSource. - $TAG_NAME: the tag name specified by RepoSource. - $REVISION_ID or $COMMIT_SHA: the commit SHA specified by RepoSource or resolved from the specified branch or tag. - $SHORT_SHA: first 7 characters of $REVISION_ID or $COMMIT_SHA. # Contents of the build template.
"approval": { # BuildApproval describes a build's approval configuration, state, and result. # Output only. Describes this build's approval configuration, status, and result.
"config": { # ApprovalConfig describes configuration for manual approval of a build. # Output only. Configuration for manual approval of this build.
@@ -2056,7 +2278,7 @@ Method Details
"createTime": "A String", # Output only. Time when the trigger was created.
"description": "A String", # Human-readable description of this trigger.
"disabled": True or False, # If true, the trigger will never automatically execute a build.
- "eventType": "A String", # Optional. EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field is optional but will be validated against the rest of the configuration if it is set.
+ "eventType": "A String", # EventType allows the user to explicitly set the type of event to which this BuildTrigger should respond. This field will be validated against the rest of the configuration if it is set.
"filename": "A String", # Path, from the source root, to the build configuration file (i.e. cloudbuild.yaml).
"filter": "A String", # A Common Expression Language string.
"gitFileSource": { # GitFileSource describes a file within a (possibly remote) code repository. # The file source describing the local or remote Build template.
diff --git a/docs/dyn/cloudkms_v1.projects.locations.html b/docs/dyn/cloudkms_v1.projects.locations.html
index 819e470d3c2..0ba05e26b37 100644
--- a/docs/dyn/cloudkms_v1.projects.locations.html
+++ b/docs/dyn/cloudkms_v1.projects.locations.html
@@ -116,7 +116,7 @@ Method Details
{ # Request message for KeyManagementService.GenerateRandomBytes.
"lengthBytes": 42, # The length in bytes of the amount of randomness to retrieve. Minimum 8 bytes, maximum 1024 bytes.
- "protectionLevel": "A String", # The ProtectionLevel to use when generating the random data. Defaults to SOFTWARE.
+ "protectionLevel": "A String", # The ProtectionLevel to use when generating the random data. Currently, only HSM protection level is supported.
}
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html
index be8decef84a..081433ba968 100644
--- a/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html
+++ b/docs/dyn/cloudkms_v1.projects.locations.keyRings.cryptoKeys.cryptoKeyVersions.html
@@ -412,7 +412,7 @@ Method Details
"algorithm": "A String", # Required. The algorithm of the key being imported. This does not need to match the version_template of the CryptoKey this version imports into.
"cryptoKeyVersion": "A String", # Optional. The optional name of an existing CryptoKeyVersion to target for an import operation. If this field is not present, a new CryptoKeyVersion containing the supplied key material is created. If this field is present, the supplied key material is imported into the existing CryptoKeyVersion. To import into an existing CryptoKeyVersion, the CryptoKeyVersion must be a child of ImportCryptoKeyVersionRequest.parent, have been previously created via ImportCryptoKeyVersion, and be in DESTROYED or IMPORT_FAILED state. The key material and algorithm must match the previous CryptoKeyVersion exactly if the CryptoKeyVersion has ever contained key material.
"importJob": "A String", # Required. The name of the ImportJob that was used to wrap this key material.
- "rsaAesWrappedKey": "A String", # Wrapped key material produced with RSA_OAEP_3072_SHA1_AES_256 or RSA_OAEP_4096_SHA1_AES_256. This field contains the concatenation of two wrapped keys: 1. An ephemeral AES-256 wrapping key wrapped with the public_key using RSAES-OAEP with SHA-1, MGF1 with SHA-1, and an empty label. 2. The key to be imported, wrapped with the ephemeral AES-256 key using AES-KWP (RFC 5649). If importing symmetric key material, it is expected that the unwrapped key contains plain bytes. If importing asymmetric key material, it is expected that the unwrapped key is in PKCS#8-encoded DER format (the PrivateKeyInfo structure from RFC 5208). This format is the same as the format produced by PKCS#11 mechanism CKM_RSA_AES_KEY_WRAP.
+ "rsaAesWrappedKey": "A String", # Wrapped key material produced with RSA_OAEP_3072_SHA1_AES_256 or RSA_OAEP_4096_SHA1_AES_256. This field contains the concatenation of two wrapped keys: 1. An ephemeral AES-256 wrapping key wrapped with the public_key using RSAES-OAEP with SHA-1/SHA-256, MGF1 with SHA-1/SHA-256, and an empty label. 2. The key to be imported, wrapped with the ephemeral AES-256 key using AES-KWP (RFC 5649). If importing symmetric key material, it is expected that the unwrapped key contains plain bytes. If importing asymmetric key material, it is expected that the unwrapped key is in PKCS#8-encoded DER format (the PrivateKeyInfo structure from RFC 5208). This format is the same as the format produced by PKCS#11 mechanism CKM_RSA_AES_KEY_WRAP.
}
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/cloudresourcemanager_v1.projects.html b/docs/dyn/cloudresourcemanager_v1.projects.html
index 277a2cc0a57..f8befa9a5e4 100644
--- a/docs/dyn/cloudresourcemanager_v1.projects.html
+++ b/docs/dyn/cloudresourcemanager_v1.projects.html
@@ -121,7 +121,7 @@ Instance Methods
Retrieves the next page of results.
setIamPolicy(resource, body=None, x__xgafv=None)
-Sets the IAM access control policy for the specified Project. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. NOTE: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. For additional information about `resource` (e.g. my-project-id) structure and identification, see [Resource Names](https://cloud.google.com/apis/design/resource_names). The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited via Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + You can only grant ownership of a project to a member by using the GCP Console. Inviting a member will deliver an invitation email that they must accept. An invitation email is not generated if you are granting a role other than owner, or if both the member you are inviting and the project are part of your organization. + Membership changes that leave the project without any owners that have accepted the Terms of Service (ToS) will be rejected. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. Authorization requires the Google IAM permission `resourcemanager.projects.setIamPolicy` on the project
+Sets the IAM access control policy for the specified Project. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. NOTE: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. For additional information about `resource` (e.g. my-project-id) structure and identification, see [Resource Names](https://cloud.google.com/apis/design/resource_names). The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited via Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + You can only grant ownership of a project to a member by using the GCP Console. Inviting a member will deliver an invitation email that they must accept. An invitation email is not generated if you are granting a role other than owner, or if both the member you are inviting and the project are part of your organization. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. If the project is part of an organization, you can remove all owners, potentially making the organization inaccessible. Authorization requires the Google IAM permission `resourcemanager.projects.setIamPolicy` on the project
setOrgPolicy(resource, body=None, x__xgafv=None)
Updates the specified `Policy` on the resource. Creates a new `Policy` for that `Constraint` on the resource if one does not exist. Not supplying an `etag` on the request `Policy` results in an unconditional write of the `Policy`.
@@ -618,7 +618,7 @@ Method Details
setIamPolicy(resource, body=None, x__xgafv=None)
- Sets the IAM access control policy for the specified Project. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. NOTE: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. For additional information about `resource` (e.g. my-project-id) structure and identification, see [Resource Names](https://cloud.google.com/apis/design/resource_names). The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited via Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + You can only grant ownership of a project to a member by using the GCP Console. Inviting a member will deliver an invitation email that they must accept. An invitation email is not generated if you are granting a role other than owner, or if both the member you are inviting and the project are part of your organization. + Membership changes that leave the project without any owners that have accepted the Terms of Service (ToS) will be rejected. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. Authorization requires the Google IAM permission `resourcemanager.projects.setIamPolicy` on the project
+ Sets the IAM access control policy for the specified Project. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. NOTE: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. For additional information about `resource` (e.g. my-project-id) structure and identification, see [Resource Names](https://cloud.google.com/apis/design/resource_names). The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited via Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + You can only grant ownership of a project to a member by using the GCP Console. Inviting a member will deliver an invitation email that they must accept. An invitation email is not generated if you are granting a role other than owner, or if both the member you are inviting and the project are part of your organization. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. If the project is part of an organization, you can remove all owners, potentially making the organization inaccessible. Authorization requires the Google IAM permission `resourcemanager.projects.setIamPolicy` on the project
Args:
resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
diff --git a/docs/dyn/cloudresourcemanager_v3.projects.html b/docs/dyn/cloudresourcemanager_v3.projects.html
index 4e79297cf91..c38af07aa00 100644
--- a/docs/dyn/cloudresourcemanager_v3.projects.html
+++ b/docs/dyn/cloudresourcemanager_v3.projects.html
@@ -109,7 +109,7 @@ Instance Methods
Retrieves the next page of results.
setIamPolicy(resource, body=None, x__xgafv=None)
-Sets the IAM access control policy for the specified project, in the format `projects/{ProjectIdOrNumber}` e.g. projects/123. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. Note: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited using the Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + Invitations to grant the owner role cannot be sent using `setIamPolicy()`; they must be sent only using the Cloud Platform Console. + Membership changes that leave the project without any owners that have accepted the Terms of Service (ToS) will be rejected. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. + Calling this method requires enabling the App Engine Admin API.
+Sets the IAM access control policy for the specified project, in the format `projects/{ProjectIdOrNumber}` e.g. projects/123. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. Note: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited using the Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + Invitations to grant the owner role cannot be sent using `setIamPolicy()`; they must be sent only using the Cloud Platform Console. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. If the project is part of an organization, you can remove all owners, potentially making the organization inaccessible. + Calling this method requires enabling the App Engine Admin API.
testIamPermissions(resource, body=None, x__xgafv=None)
Returns permissions that a caller has on the specified project, in the format `projects/{ProjectIdOrNumber}` e.g. projects/123..
@@ -496,7 +496,7 @@ Method Details
setIamPolicy(resource, body=None, x__xgafv=None)
- Sets the IAM access control policy for the specified project, in the format `projects/{ProjectIdOrNumber}` e.g. projects/123. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. Note: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited using the Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + Invitations to grant the owner role cannot be sent using `setIamPolicy()`; they must be sent only using the Cloud Platform Console. + Membership changes that leave the project without any owners that have accepted the Terms of Service (ToS) will be rejected. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. + Calling this method requires enabling the App Engine Admin API.
+ Sets the IAM access control policy for the specified project, in the format `projects/{ProjectIdOrNumber}` e.g. projects/123. CAUTION: This method will replace the existing policy, and cannot be used to append additional IAM settings. Note: Removing service accounts from policies or changing their roles can render services completely inoperable. It is important to understand how the service account is being used before removing or updating its roles. The following constraints apply when using `setIamPolicy()`: + Project does not support `allUsers` and `allAuthenticatedUsers` as `members` in a `Binding` of a `Policy`. + The owner role can be granted to a `user`, `serviceAccount`, or a group that is part of an organization. For example, group@myownpersonaldomain.com could be added as an owner to a project in the myownpersonaldomain.com organization, but not the examplepetstore.com organization. + Service accounts can be made owners of a project directly without any restrictions. However, to be added as an owner, a user must be invited using the Cloud Platform console and must accept the invitation. + A user cannot be granted the owner role using `setIamPolicy()`. The user must be granted the owner role using the Cloud Platform Console and must explicitly accept the invitation. + Invitations to grant the owner role cannot be sent using `setIamPolicy()`; they must be sent only using the Cloud Platform Console. + If the project is not part of an organization, there must be at least one owner who has accepted the Terms of Service (ToS) agreement in the policy. Calling `setIamPolicy()` to remove the last ToS-accepted owner from the policy will fail. This restriction also applies to legacy projects that no longer have owners who have accepted the ToS. Edits to IAM policies will be rejected until the lack of a ToS-accepting owner is rectified. If the project is part of an organization, you can remove all owners, potentially making the organization inaccessible. + Calling this method requires enabling the App Engine Admin API.
Args:
resource: string, REQUIRED: The resource for which the policy is being specified. See the operation documentation for the appropriate value for this field. (required)
diff --git a/docs/dyn/compute_alpha.instances.html b/docs/dyn/compute_alpha.instances.html
index 6bfc9bc52dd..d9b280d28bd 100644
--- a/docs/dyn/compute_alpha.instances.html
+++ b/docs/dyn/compute_alpha.instances.html
@@ -1634,7 +1634,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
"targetShape": "A String", # Strategy for distributing VMs across zones in a region.
diff --git a/docs/dyn/compute_alpha.regionInstances.html b/docs/dyn/compute_alpha.regionInstances.html
index bcb0297593d..6b6293076f0 100644
--- a/docs/dyn/compute_alpha.regionInstances.html
+++ b/docs/dyn/compute_alpha.regionInstances.html
@@ -722,7 +722,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
"targetShape": "A String", # Strategy for distributing VMs across zones in a region.
diff --git a/docs/dyn/compute_alpha.routers.html b/docs/dyn/compute_alpha.routers.html
index e51af5a7b2d..05111150c4e 100644
--- a/docs/dyn/compute_alpha.routers.html
+++ b/docs/dyn/compute_alpha.routers.html
@@ -679,6 +679,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
@@ -727,6 +728,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
@@ -777,6 +779,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
diff --git a/docs/dyn/compute_alpha.routes.html b/docs/dyn/compute_alpha.routes.html
index 5d60b937855..839c10e36f6 100644
--- a/docs/dyn/compute_alpha.routes.html
+++ b/docs/dyn/compute_alpha.routes.html
@@ -210,6 +210,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
@@ -267,6 +268,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
@@ -396,6 +398,7 @@ Method Details
"nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.
"nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets.
"priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive.
+ "routeStatus": "A String", # [Output only] The status of the route.
"routeType": "A String", # [Output Only] The type of this route, which can be one of the following values: - 'TRANSIT' for a transit route that this router learned from another Cloud Router and will readvertise to one of its BGP peers - 'SUBNET' for a route from a subnet of the VPC - 'BGP' for a route learned from a BGP peer of this router - 'STATIC' for a static route
"selfLink": "A String", # [Output Only] Server-defined fully-qualified URL for this resource.
"selfLinkWithId": "A String", # [Output Only] Server-defined URL for this resource with the resource id.
diff --git a/docs/dyn/compute_beta.instanceGroupManagers.html b/docs/dyn/compute_beta.instanceGroupManagers.html
index c3d23f91256..356ae82778a 100644
--- a/docs/dyn/compute_beta.instanceGroupManagers.html
+++ b/docs/dyn/compute_beta.instanceGroupManagers.html
@@ -281,6 +281,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -853,6 +857,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -978,6 +986,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1171,6 +1183,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1599,6 +1615,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -2350,6 +2370,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
diff --git a/docs/dyn/compute_beta.instanceTemplates.html b/docs/dyn/compute_beta.instanceTemplates.html
index 772716394c4..e893b1cb4c2 100644
--- a/docs/dyn/compute_beta.instanceTemplates.html
+++ b/docs/dyn/compute_beta.instanceTemplates.html
@@ -194,6 +194,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -568,6 +569,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -907,6 +909,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
diff --git a/docs/dyn/compute_beta.instances.html b/docs/dyn/compute_beta.instances.html
index 4d52bf34bdb..339e7a70f68 100644
--- a/docs/dyn/compute_beta.instances.html
+++ b/docs/dyn/compute_beta.instances.html
@@ -415,6 +415,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -912,6 +913,7 @@ Method Details
"instanceProperties": { # The instance properties defining the VM instances to be created. Required if sourceInstanceTemplate is not provided.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -1153,7 +1155,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
"targetShape": "A String", # Strategy for distributing VMs across zones in a region.
@@ -1443,6 +1445,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -2174,6 +2177,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -2532,6 +2536,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -4684,6 +4689,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
diff --git a/docs/dyn/compute_beta.machineImages.html b/docs/dyn/compute_beta.machineImages.html
index 5ee9bdcb39b..4d748ba1a05 100644
--- a/docs/dyn/compute_beta.machineImages.html
+++ b/docs/dyn/compute_beta.machineImages.html
@@ -193,6 +193,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -732,6 +733,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -1237,6 +1239,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
diff --git a/docs/dyn/compute_beta.regionCommitments.html b/docs/dyn/compute_beta.regionCommitments.html
index 8bc35bcf098..e1fdaebeb01 100644
--- a/docs/dyn/compute_beta.regionCommitments.html
+++ b/docs/dyn/compute_beta.regionCommitments.html
@@ -168,6 +168,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -313,6 +314,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -403,6 +405,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -561,6 +564,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -713,6 +717,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -846,6 +851,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
diff --git a/docs/dyn/compute_beta.regionInstanceGroupManagers.html b/docs/dyn/compute_beta.regionInstanceGroupManagers.html
index a22b24c4593..b2f7791d9c3 100644
--- a/docs/dyn/compute_beta.regionInstanceGroupManagers.html
+++ b/docs/dyn/compute_beta.regionInstanceGroupManagers.html
@@ -663,6 +663,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -788,6 +792,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -981,6 +989,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1409,6 +1421,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -2160,6 +2176,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
diff --git a/docs/dyn/compute_beta.regionInstances.html b/docs/dyn/compute_beta.regionInstances.html
index 654492ea5d8..4581882d9e2 100644
--- a/docs/dyn/compute_beta.regionInstances.html
+++ b/docs/dyn/compute_beta.regionInstances.html
@@ -96,6 +96,7 @@ Method Details
"instanceProperties": { # The instance properties defining the VM instances to be created. Required if sourceInstanceTemplate is not provided.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -337,7 +338,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
"targetShape": "A String", # Strategy for distributing VMs across zones in a region.
diff --git a/docs/dyn/compute_beta.reservations.html b/docs/dyn/compute_beta.reservations.html
index 095e9743aab..3625bf6f25c 100644
--- a/docs/dyn/compute_beta.reservations.html
+++ b/docs/dyn/compute_beta.reservations.html
@@ -160,6 +160,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -338,6 +339,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -501,6 +503,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -628,6 +631,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
@@ -1030,6 +1034,7 @@ Method Details
"shareType": "A String", # Type of sharing for this shared-reservation
},
"specificReservation": { # This reservation type allows to pre allocate specific instance configuration. Next ID: 5 # Reservation for instances with specific machine shapes.
+ "assuredCount": "A String", # [Output Only] Indicates how many instances are actually usable currently.
"count": "A String", # Specifies the number of resources that are allocated.
"inUseCount": "A String", # [Output Only] Indicates how many instances are in use.
"instanceProperties": { # Properties of the SKU instances being reserved. Next ID: 9 # The instance properties for the reservation.
diff --git a/docs/dyn/compute_v1.instanceGroupManagers.html b/docs/dyn/compute_v1.instanceGroupManagers.html
index f7628196762..9d9749e1bac 100644
--- a/docs/dyn/compute_v1.instanceGroupManagers.html
+++ b/docs/dyn/compute_v1.instanceGroupManagers.html
@@ -269,6 +269,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -809,6 +813,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -920,6 +928,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1099,6 +1111,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1459,6 +1475,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
diff --git a/docs/dyn/compute_v1.instanceTemplates.html b/docs/dyn/compute_v1.instanceTemplates.html
index 3db451f5656..fe76fea15ef 100644
--- a/docs/dyn/compute_v1.instanceTemplates.html
+++ b/docs/dyn/compute_v1.instanceTemplates.html
@@ -194,6 +194,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -366,6 +367,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -543,6 +547,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -715,6 +720,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -857,6 +865,7 @@ Method Details
"properties": { # The instance properties for this instance template.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -1029,6 +1038,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
diff --git a/docs/dyn/compute_v1.instances.html b/docs/dyn/compute_v1.instances.html
index 3fe18151ad4..8f215125e4d 100644
--- a/docs/dyn/compute_v1.instances.html
+++ b/docs/dyn/compute_v1.instances.html
@@ -397,6 +397,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -856,6 +857,7 @@ Method Details
"instanceProperties": { # The instance properties defining the VM instances to be created. Required if sourceInstanceTemplate is not provided.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -1028,6 +1030,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -1072,7 +1077,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
},
@@ -1361,6 +1366,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -1954,6 +1960,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -2284,6 +2291,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
@@ -4108,6 +4116,7 @@ Method Details
{ # Represents an Instance resource. An instance is a virtual machine that is hosted on Google Cloud Platform. For more information, read Virtual Machine Instances.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Allows this instance to send and receive packets with non-matching destination or source IPs. This is required if you plan to use this instance to forward routes. For more information, see Enabling IP Forwarding .
diff --git a/docs/dyn/compute_v1.machineImages.html b/docs/dyn/compute_v1.machineImages.html
index c9d604fd194..15d85b8e3ba 100644
--- a/docs/dyn/compute_v1.machineImages.html
+++ b/docs/dyn/compute_v1.machineImages.html
@@ -193,6 +193,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -365,6 +366,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -703,6 +707,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -875,6 +880,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -1179,6 +1187,7 @@ Method Details
"instanceProperties": { # [Output Only] Properties of source instance
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -1351,6 +1360,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
diff --git a/docs/dyn/compute_v1.regionInstanceGroupManagers.html b/docs/dyn/compute_v1.regionInstanceGroupManagers.html
index fb170e8c3fb..4d029b306dc 100644
--- a/docs/dyn/compute_v1.regionInstanceGroupManagers.html
+++ b/docs/dyn/compute_v1.regionInstanceGroupManagers.html
@@ -633,6 +633,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -744,6 +748,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -923,6 +931,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
@@ -1283,6 +1295,10 @@ Method Details
"recreating": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be recreated or are currently being being recreated. Recreating an instance deletes the existing root persistent disk and creates a new disk from the image that is defined in the instance template.
"refreshing": 42, # [Output Only] The number of instances in the managed instance group that are being reconfigured with properties that do not require a restart or a recreate action. For example, setting or removing target pools for the instance.
"restarting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be restarted or are currently being restarted.
+ "resuming": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be resumed or are currently being resumed.
+ "starting": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be started or are currently being started.
+ "stopping": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be stopped or are currently being stopped.
+ "suspending": 42, # [Output Only] The number of instances in the managed instance group that are scheduled to be suspended or are currently being suspended.
"verifying": 42, # [Output Only] The number of instances in the managed instance group that are being verified. See the managedInstances[].currentAction property in the listManagedInstances method documentation.
},
"description": "A String", # An optional description of this resource.
diff --git a/docs/dyn/compute_v1.regionInstances.html b/docs/dyn/compute_v1.regionInstances.html
index 46fd26b62d9..e8dc55ed57b 100644
--- a/docs/dyn/compute_v1.regionInstances.html
+++ b/docs/dyn/compute_v1.regionInstances.html
@@ -96,6 +96,7 @@ Method Details
"instanceProperties": { # The instance properties defining the VM instances to be created. Required if sourceInstanceTemplate is not provided.
"advancedMachineFeatures": { # Specifies options for controlling advanced machine features. Options that would traditionally be configured in a BIOS belong here. Features that require operating system support may have corresponding entries in the GuestOsFeatures of an Image (e.g., whether or not the OS in the Image supports nested virtualization being enabled or disabled). # Controls for advanced machine-related behavior features. Note that for MachineImage, this is not supported yet.
"enableNestedVirtualization": True or False, # Whether to enable nested virtualization or not (default is false).
+ "enableUefiNetworking": True or False, # Whether to enable UEFI networking for instance creation.
"threadsPerCore": 42, # The number of threads per physical core. To disable simultaneous multithreading (SMT) set this to 1. If unset, the maximum number of threads supported per core by the underlying processor is assumed.
},
"canIpForward": True or False, # Enables instances created based on these properties to send packets with source IP addresses other than their own and receive packets with destination IP addresses other than their own. If these instances will be used as an IP gateway or it will be set as the next-hop in a Route resource, specify true. If unsure, leave this set to false. See the Enable IP forwarding documentation for more information.
@@ -268,6 +269,9 @@ Method Details
"A String",
],
},
+ "resourceManagerTags": { # Resource manager tags to be bound to the instance. Tag keys and values have the same definition as resource manager tags. Keys must be in the format `tagKeys/{tag_key_id}`, and values are in the format `tagValues/456`. The field is ignored (both PUT & PATCH) when empty.
+ "a_key": "A String",
+ },
"resourcePolicies": [ # Resource policies (names, not URLs) applied to instances created from these properties. Note that for MachineImage, this is not supported yet.
"A String",
],
@@ -312,7 +316,7 @@ Method Details
"locationPolicy": { # Configuration for location policy among multiple possible locations (e.g. preferences for zone selection among zones in a single region). # Policy for chosing target zone.
"locations": { # Location configurations mapped by location name. Currently only zone names are supported and must be represented as valid internal URLs, such as zones/us-central1-a.
"a_key": {
- "preference": "A String", # Preference for a given location: ALLOW or DENY.
+ "preference": "A String", # Preference for a given location.
},
},
},
diff --git a/docs/dyn/connectors_v1.projects.locations.global_.providers.connectors.html b/docs/dyn/connectors_v1.projects.locations.global_.providers.connectors.html
index c967b72f3ae..b89358cb83a 100644
--- a/docs/dyn/connectors_v1.projects.locations.global_.providers.connectors.html
+++ b/docs/dyn/connectors_v1.projects.locations.global_.providers.connectors.html
@@ -120,6 +120,7 @@ Method Details
"labels": { # Output only. Resource labels to represent user-provided metadata. Refer to cloud documentation on labels for more details. https://cloud.google.com/compute/docs/labeling-resources
"a_key": "A String",
},
+ "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage.
"name": "A String", # Output only. Resource name of the Connector. Format: projects/{project}/locations/{location}/providers/{provider}/connectors/{connector}
"updateTime": "A String", # Output only. Updated time.
"webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI.
@@ -153,6 +154,7 @@ Method Details
"labels": { # Output only. Resource labels to represent user-provided metadata. Refer to cloud documentation on labels for more details. https://cloud.google.com/compute/docs/labeling-resources
"a_key": "A String",
},
+ "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage.
"name": "A String", # Output only. Resource name of the Connector. Format: projects/{project}/locations/{location}/providers/{provider}/connectors/{connector}
"updateTime": "A String", # Output only. Updated time.
"webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI.
diff --git a/docs/dyn/connectors_v1.projects.locations.global_.providers.html b/docs/dyn/connectors_v1.projects.locations.global_.providers.html
index 4511566fb7f..da308c8a53d 100644
--- a/docs/dyn/connectors_v1.projects.locations.global_.providers.html
+++ b/docs/dyn/connectors_v1.projects.locations.global_.providers.html
@@ -120,6 +120,7 @@ Method Details
"labels": { # Output only. Resource labels to represent user-provided metadata. Refer to cloud documentation on labels for more details. https://cloud.google.com/compute/docs/labeling-resources
"a_key": "A String",
},
+ "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage.
"name": "A String", # Output only. Resource name of the Provider. Format: projects/{project}/locations/{location}/providers/{provider}
"updateTime": "A String", # Output only. Updated time.
"webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI.
@@ -154,6 +155,7 @@ Method Details
"labels": { # Output only. Resource labels to represent user-provided metadata. Refer to cloud documentation on labels for more details. https://cloud.google.com/compute/docs/labeling-resources
"a_key": "A String",
},
+ "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage.
"name": "A String", # Output only. Resource name of the Provider. Format: projects/{project}/locations/{location}/providers/{provider}
"updateTime": "A String", # Output only. Updated time.
"webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI.
diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html
index e628ee01a76..89e51bf009d 100644
--- a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html
+++ b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.analyses.html
@@ -169,7 +169,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -324,7 +324,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -438,7 +438,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html
index 77313b68d32..41335d3bf15 100644
--- a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html
+++ b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html
@@ -255,7 +255,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -486,7 +486,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -747,7 +747,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -994,7 +994,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -1243,7 +1243,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
@@ -1474,7 +1474,7 @@ Method Details
},
},
"issueModelResult": { # Issue Modeling result on a conversation. # Overall conversation-level issue modeling result.
- "issueModel": "A String", # Issue model that generates the result.
+ "issueModel": "A String", # Issue model that generates the result. Format: projects/{project}/locations/{location}/issueModels/{issue_model}
"issues": [ # All the matched issues.
{ # Information about the issue.
"displayName": "A String", # Immutable. Display name of the assigned issue. This field is set at time of analyis and immutable since then.
diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.html
index 385bdf145a3..0551fc6aca7 100644
--- a/docs/dyn/contactcenterinsights_v1.projects.locations.html
+++ b/docs/dyn/contactcenterinsights_v1.projects.locations.html
@@ -99,6 +99,11 @@ Instance Methods
Returns the phraseMatchers Resource.
+
+ views()
+
+Returns the views Resource.
+
Close httplib2 connections.
diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.views.html b/docs/dyn/contactcenterinsights_v1.projects.locations.views.html
new file mode 100644
index 00000000000..ccc59373182
--- /dev/null
+++ b/docs/dyn/contactcenterinsights_v1.projects.locations.views.html
@@ -0,0 +1,258 @@
+
+
+
+Contact Center AI Insights API . projects . locations . views
+Instance Methods
+
+ close()
+Close httplib2 connections.
+
+ create(parent, body=None, x__xgafv=None)
+Creates a view.
+
+Deletes a view.
+
+Gets a view.
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists views.
+
+ list_next(previous_request, previous_response)
+Retrieves the next page of results.
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates a view.
+Method Details
+
+ close()
+ Close httplib2 connections.
+
+
+
+ create(parent, body=None, x__xgafv=None)
+ Creates a view.
+
+Args:
+ parent: string, Required. The parent resource of the view. Required. The location to create a view for. Format: `projects//locations/` or `projects//locations/` (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+}
+
+
+
+ delete(name, x__xgafv=None)
+ Deletes a view.
+
+Args:
+ name: string, Required. The name of the view to delete. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON representation for `Empty` is empty JSON object `{}`.
+}
+
+
+
+ get(name, x__xgafv=None)
+ Gets a view.
+
+Args:
+ name: string, Required. The name of the view to get. (required)
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+}
+
+
+
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ Lists views.
+
+Args:
+ parent: string, Required. The parent resource of the views. (required)
+ pageSize: integer, The maximum number of views to return in the response. If this value is zero, the service will select a default size. A call may return fewer objects than requested. A non-empty `next_page_token` in the response indicates that more data is available.
+ pageToken: string, The value returned by the last `ListViewsResponse`; indicates that this is a continuation of a prior `ListViews` call and the system should return the next page of data.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response of listing views.
+ "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
+ "views": [ # The views that match the request.
+ { # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+ },
+ ],
+}
+
+
+
+ list_next(previous_request, previous_response)
+ Retrieves the next page of results.
+
+Args:
+ previous_request: The request for the previous page. (required)
+ previous_response: The response from the request for the previous page. (required)
+
+Returns:
+ A request object that you can call 'execute()' on to request the next
+ page. Returns None if there are no more items in the collection.
+
+
+
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates a view.
+
+Args:
+ name: string, Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view} (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+}
+
+ updateMask: string, The list of fields to be updated.
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The View resource.
+ "createTime": "A String", # Output only. The time at which this view was created.
+ "displayName": "A String", # The human-readable display name of the view.
+ "name": "A String", # Immutable. The resource name of the view. Format: projects/{project}/locations/{location}/views/{view}
+ "updateTime": "A String", # Output only. The most recent time at which the view was updated.
+ "value": "A String", # String with specific view properties.
+}
+
+
+
\ No newline at end of file
diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.html b/docs/dyn/container_v1beta1.projects.locations.clusters.html
index 3e1417ba8e1..dd6ee389318 100644
--- a/docs/dyn/container_v1beta1.projects.locations.clusters.html
+++ b/docs/dyn/container_v1beta1.projects.locations.clusters.html
@@ -252,6 +252,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -690,6 +693,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -940,6 +946,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -1378,6 +1387,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -1531,6 +1543,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -1969,6 +1984,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -2082,6 +2100,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -2947,6 +2968,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html b/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html
index c5a9815f7d3..0bd2aabf82d 100644
--- a/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html
+++ b/docs/dyn/container_v1beta1.projects.locations.clusters.nodePools.html
@@ -239,6 +239,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -522,6 +525,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -671,6 +677,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.html b/docs/dyn/container_v1beta1.projects.zones.clusters.html
index 9992b652008..1db80adc666 100644
--- a/docs/dyn/container_v1beta1.projects.zones.clusters.html
+++ b/docs/dyn/container_v1beta1.projects.zones.clusters.html
@@ -163,6 +163,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -363,6 +366,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -801,6 +807,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -1051,6 +1060,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -1489,6 +1501,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -1686,6 +1701,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
@@ -2124,6 +2142,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -3008,6 +3029,9 @@ Method Details
"gcpFilestoreCsiDriverConfig": { # Configuration for the GCP Filestore CSI driver. # Configuration for the GCP Filestore CSI driver.
"enabled": True or False, # Whether the GCP Filestore CSI driver is enabled for this cluster.
},
+ "gkeBackupAgentConfig": { # Configuration for the Backup for GKE Agent. # Configuration for the Backup for GKE agent addon.
+ "enabled": True or False, # Whether the Backup for GKE agent is enabled for this cluster.
+ },
"horizontalPodAutoscaling": { # Configuration options for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods. # Configuration for the horizontal pod autoscaling feature, which increases or decreases the number of replica pods a replication controller has based on the resource usage of the existing pods.
"disabled": True or False, # Whether the Horizontal Pod Autoscaling feature is enabled in the cluster. When enabled, it ensures that metrics are collected into Stackdriver Monitoring.
},
diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html b/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html
index 3a412a8b505..5a7d2fa7499 100644
--- a/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html
+++ b/docs/dyn/container_v1beta1.projects.zones.clusters.nodePools.html
@@ -328,6 +328,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -611,6 +614,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
@@ -760,6 +766,9 @@ Method Details
"podIpv4CidrBlock": "A String", # The IP address range for pod IPs in this node pool. Only applicable if `create_pod_range` is true. Set to blank to have a range chosen with the default size. Set to /netmask (e.g. `/14`) to have a range chosen with a specific netmask. Set to a [CIDR](https://en.wikipedia.org/wiki/Classless_Inter-Domain_Routing) notation (e.g. `10.96.0.0/14`) to pick a specific range to use. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
"podRange": "A String", # The ID of the secondary range for pod IPs. If `create_pod_range` is true, this ID is used for the new range. If `create_pod_range` is false, uses an existing secondary range with this ID. Only applicable if `ip_allocation_policy.use_ip_aliases` is true. This field cannot be changed after the node pool has been created.
},
+ "placementPolicy": { # PlacementPolicy defines the placement policy used by the node pool. # Specifies the node placement policy.
+ "type": "A String", # The type of placement.
+ },
"podIpv4CidrSize": 42, # [Output only] The pod CIDR block size per node in this node pool.
"selfLink": "A String", # [Output only] Server-defined URL for the resource.
"status": "A String", # [Output only] The status of the nodes in this pool instance.
diff --git a/docs/dyn/datacatalog_v1.entries.html b/docs/dyn/datacatalog_v1.entries.html
index 2fcc4ae2f16..d7922433425 100644
--- a/docs/dyn/datacatalog_v1.entries.html
+++ b/docs/dyn/datacatalog_v1.entries.html
@@ -118,6 +118,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
diff --git a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.html b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.html
index 6d273cdb6ab..726ef811dc8 100644
--- a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.html
+++ b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.html
@@ -100,6 +100,12 @@ Instance Methods
list_next(previous_request, previous_response)
Retrieves the next page of results.
+
+ modifyEntryContacts(name, body=None, x__xgafv=None)
+Modifies contacts, part of the business context of an Entry. To call this method, you must have the `datacatalog.entries.updateContacts` IAM permission on the corresponding project.
+
+ modifyEntryOverview(name, body=None, x__xgafv=None)
+Modifies entry overview, part of the business context of an Entry. To call this method, you must have the `datacatalog.entries.updateOverview` IAM permission on the corresponding project.
patch(name, body=None, updateMask=None, x__xgafv=None)
Updates an existing entry. You must enable the Data Catalog API in the project identified by the `entry.name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
@@ -143,6 +149,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
@@ -267,6 +286,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
@@ -415,6 +447,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
@@ -593,6 +638,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
@@ -710,6 +768,72 @@ Method Details
+
+ modifyEntryContacts(name, body=None, x__xgafv=None)
+ Modifies contacts, part of the business context of an Entry. To call this method, you must have the `datacatalog.entries.updateContacts` IAM permission on the corresponding project.
+
+Args:
+ name: string, Required. The full resource name of the entry. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for ModifyEntryContacts.
+ "contacts": { # Contact people for the entry. # Required. The new value for the Contacts.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+}
+
+
+
+ modifyEntryOverview(name, body=None, x__xgafv=None)
+ Modifies entry overview, part of the business context of an Entry. To call this method, you must have the `datacatalog.entries.updateOverview` IAM permission on the corresponding project.
+
+Args:
+ name: string, Required. The full resource name of the entry. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # Request message for ModifyEntryOverview.
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Required. The new value for the Entry Overview.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+}
+
+
patch(name, body=None, updateMask=None, x__xgafv=None)
Updates an existing entry. You must enable the Data Catalog API in the project identified by the `entry.name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
@@ -735,6 +859,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
@@ -859,6 +996,19 @@ Method Details
"viewQuery": "A String", # Output only. The query that defines the table view.
},
},
+ "businessContext": { # Business Context of the entry. # Business Context of the entry.
+ "contacts": { # Contact people for the entry. # Contact people for the entry.
+ "people": [ # The list of contact people for the entry.
+ { # A contact person for the entry.
+ "designation": "A String", # Designation of the person, for example, Data Steward.
+ "email": "A String", # Email of the person in the format of `john.doe@xyz`, ``, or `John Doe`.
+ },
+ ],
+ },
+ "entryOverview": { # Entry overview fields for rich text descriptions of entries. # Entry overview fields for rich text descriptions of entries.
+ "overview": "A String", # Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
+ },
+ },
"dataSource": { # Physical location of an entry. # Output only. Physical location of the entry.
"resource": "A String", # Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
"service": "A String", # Service that physically stores the data.
diff --git a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.tags.html b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.tags.html
index 11ae5dd03ae..d1511c17019 100644
--- a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.tags.html
+++ b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.entries.tags.html
@@ -85,7 +85,7 @@ Instance Methods
Deletes a tag.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists tags assigned to an Entry.
+Lists tags assigned to an Entry. The columns in the response are lowercased.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -178,7 +178,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists tags assigned to an Entry.
+ Lists tags assigned to an Entry. The columns in the response are lowercased.
Args:
parent: string, Required. The name of the Data Catalog resource to list the tags of. The resource can be an Entry or an EntryGroup (without `/entries/{entries}` at the end). (required)
diff --git a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.tags.html b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.tags.html
index a6406fa6824..d8d99c35f89 100644
--- a/docs/dyn/datacatalog_v1.projects.locations.entryGroups.tags.html
+++ b/docs/dyn/datacatalog_v1.projects.locations.entryGroups.tags.html
@@ -85,7 +85,7 @@ Instance Methods
Deletes a tag.
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
-Lists tags assigned to an Entry.
+Lists tags assigned to an Entry. The columns in the response are lowercased.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -178,7 +178,7 @@ Method Details
list(parent, pageSize=None, pageToken=None, x__xgafv=None)
- Lists tags assigned to an Entry.
+ Lists tags assigned to an Entry. The columns in the response are lowercased.
Args:
parent: string, Required. The name of the Data Catalog resource to list the tags of. The resource can be an Entry or an EntryGroup (without `/entries/{entries}` at the end). (required)
diff --git a/docs/dyn/datamigration_v1.projects.locations.connectionProfiles.html b/docs/dyn/datamigration_v1.projects.locations.connectionProfiles.html
index 2dbf23a11d5..16340cd9bca 100644
--- a/docs/dyn/datamigration_v1.projects.locations.connectionProfiles.html
+++ b/docs/dyn/datamigration_v1.projects.locations.connectionProfiles.html
@@ -91,7 +91,7 @@ Instance Methods
Gets the access control policy for a resource. Returns an empty policy if the resource exists and does not have a policy set.
list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
-Retrieve a list of all connection profiles in a given project and location.
+Retrieves a list of all connection profiles in a given project and location.
list_next(previous_request, previous_response)
Retrieves the next page of results.
@@ -428,12 +428,12 @@ Method Details
list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)
- Retrieve a list of all connection profiles in a given project and location.
+ Retrieves a list of all connection profiles in a given project and location.
Args:
parent: string, Required. The parent, which owns this collection of connection profiles. (required)
filter: string, A filter expression that filters connection profiles listed in the response. The expression must specify the field name, a comparison operator, and the value that you want to use for filtering. The value must be a string, a number, or a boolean. The comparison operator must be either =, !=, >, or <. For example, list connection profiles created this year by specifying **createTime %gt; 2020-01-01T00:00:00.000000000Z**. You can also filter nested fields. For example, you could specify **mySql.username = %lt;my_username%gt;** to list all connection profiles configured to connect with a specific username.
- orderBy: string, the order by fields for the result.
+ orderBy: string, A comma-separated list of fields to order results according to.
pageSize: integer, The maximum number of connection profiles to return. The service may return fewer than this value. If unspecified, at most 50 connection profiles will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000.
pageToken: string, A page token, received from a previous `ListConnectionProfiles` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListConnectionProfiles` must match the call that provided the page token.
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html b/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
index 9b737511751..324bb3de1a1 100644
--- a/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
+++ b/docs/dyn/dataproc_v1.projects.locations.workflowTemplates.html
@@ -121,7 +121,7 @@ Method Details
Creates new workflow template.
Args:
- parent: string, Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. For projects.regions.workflowTemplates,create, the resource name of the region has the following format: projects/{project_id}/regions/{region} For projects.locations.workflowTemplates.create, the resource name of the location has the following format: projects/{project_id}/locations/{location} (required)
+ parent: string, Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. For projects.regions.workflowTemplates.create, the resource name of the region has the following format: projects/{project_id}/regions/{region} For projects.locations.workflowTemplates.create, the resource name of the location has the following format: projects/{project_id}/locations/{location} (required)
body: object, The request body.
The object takes the form of:
@@ -250,8 +250,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -433,6 +433,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -470,6 +471,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -538,6 +540,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -705,8 +708,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -888,6 +891,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -925,6 +929,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -993,6 +998,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1187,8 +1193,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1370,6 +1376,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1407,6 +1414,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1475,6 +1483,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1733,8 +1742,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1916,6 +1925,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1953,6 +1963,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2021,6 +2032,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2229,8 +2241,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -2412,6 +2424,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2449,6 +2462,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2517,6 +2531,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2790,8 +2805,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -2973,6 +2988,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3010,6 +3026,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3078,6 +3095,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3245,8 +3263,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -3428,6 +3446,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3465,6 +3484,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3533,6 +3553,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
diff --git a/docs/dyn/dataproc_v1.projects.regions.clusters.html b/docs/dyn/dataproc_v1.projects.regions.clusters.html
index 8304fec014c..466b9b319de 100644
--- a/docs/dyn/dataproc_v1.projects.regions.clusters.html
+++ b/docs/dyn/dataproc_v1.projects.regions.clusters.html
@@ -215,6 +215,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -252,6 +253,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -320,6 +322,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -586,6 +589,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -623,6 +627,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -691,6 +696,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -933,6 +939,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -970,6 +977,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1038,6 +1046,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1200,6 +1209,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1237,6 +1247,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1305,6 +1316,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
diff --git a/docs/dyn/dataproc_v1.projects.regions.jobs.html b/docs/dyn/dataproc_v1.projects.regions.jobs.html
index 487c553fd1e..1ae95117d34 100644
--- a/docs/dyn/dataproc_v1.projects.regions.jobs.html
+++ b/docs/dyn/dataproc_v1.projects.regions.jobs.html
@@ -265,8 +265,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -532,8 +532,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -827,8 +827,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1081,8 +1081,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1315,8 +1315,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1610,8 +1610,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1845,8 +1845,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -2082,8 +2082,8 @@ Method Details
"projectId": "A String", # Optional. The ID of the Google Cloud Platform project that the job belongs to. If specified, must match the request project ID.
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
diff --git a/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html b/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html
index c536740031e..4d87eb5d304 100644
--- a/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html
+++ b/docs/dyn/dataproc_v1.projects.regions.workflowTemplates.html
@@ -121,7 +121,7 @@ Method Details
Creates new workflow template.
Args:
- parent: string, Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. For projects.regions.workflowTemplates,create, the resource name of the region has the following format: projects/{project_id}/regions/{region} For projects.locations.workflowTemplates.create, the resource name of the location has the following format: projects/{project_id}/locations/{location} (required)
+ parent: string, Required. The resource name of the region or location, as described in https://cloud.google.com/apis/design/resource_names. For projects.regions.workflowTemplates.create, the resource name of the region has the following format: projects/{project_id}/regions/{region} For projects.locations.workflowTemplates.create, the resource name of the location has the following format: projects/{project_id}/locations/{location} (required)
body: object, The request body.
The object takes the form of:
@@ -250,8 +250,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -433,6 +433,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -470,6 +471,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -538,6 +540,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -705,8 +708,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -888,6 +891,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -925,6 +929,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -993,6 +998,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1187,8 +1193,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1370,6 +1376,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1407,6 +1414,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1475,6 +1483,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1733,8 +1742,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -1916,6 +1925,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -1953,6 +1963,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2021,6 +2032,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2229,8 +2241,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -2412,6 +2424,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2449,6 +2462,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2517,6 +2531,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -2790,8 +2805,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -2973,6 +2988,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3010,6 +3026,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3078,6 +3095,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3245,8 +3263,8 @@ Method Details
],
},
"scheduling": { # Job scheduling options. # Optional. Job scheduling configuration.
- "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.
- "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.
+ "maxFailuresPerHour": 42, # Optional. Maximum number of times per hour a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed.A job may be reported as thrashing if driver exits with non-zero code 4 times within 10 minute window.Maximum value is 10.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
+ "maxFailuresTotal": 42, # Optional. Maximum number of times in total a driver may be restarted as a result of driver exiting with non-zero code before job is reported failed. Maximum value is 240.Note: Currently, this restartable job option is not supported in Dataproc workflow template (https://cloud.google.com/dataproc/docs/concepts/workflows/using-workflows#adding_jobs_to_a_template) jobs.
},
"sparkJob": { # A Dataproc job for running Apache Spark (http://spark.apache.org/) applications on YARN. # Optional. Job is a Spark job.
"archiveUris": [ # Optional. HCFS URIs of archives to be extracted into the working directory of each executor. Supported file types: .jar, .tar, .tar.gz, .tgz, and .zip.
@@ -3428,6 +3446,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3465,6 +3484,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
@@ -3533,6 +3553,7 @@ Method Details
"diskConfig": { # Specifies the config of disk options for a group of VM instances. # Optional. Disk option config settings.
"bootDiskSizeGb": 42, # Optional. Size in GB of the boot disk (default is 500GB).
"bootDiskType": "A String", # Optional. Type of the boot disk (default is "pd-standard"). Valid values: "pd-balanced" (Persistent Disk Balanced Solid State Drive), "pd-ssd" (Persistent Disk Solid State Drive), or "pd-standard" (Persistent Disk Hard Disk Drive). See Disk types (https://cloud.google.com/compute/docs/disks#disk-types).
+ "localSsdInterface": "A String", # Optional. Interface type of local SSDs (default is "scsi"). Valid values: "scsi" (Small Computer System Interface), "nvme" (Non-Volatile Memory Express). See SSD Interface types (https://cloud.google.com/compute/docs/disks/local-ssd#performance).
"numLocalSsds": 42, # Optional. Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are not attached, the boot disk is used to store runtime logs and HDFS (https://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) data. If one or more SSDs are attached, this runtime bulk data is spread across them, and the boot disk contains only basic config and installed binaries.
},
"imageUri": "A String", # Optional. The Compute Engine image resource used for cluster instances.The URI can represent an image or image family.Image examples: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/[image-id] projects/[project_id]/global/images/[image-id] image-idImage family examples. Dataproc will use the most recent image from the family: https://www.googleapis.com/compute/beta/projects/[project_id]/global/images/family/[custom-image-family-name] projects/[project_id]/global/images/family/[custom-image-family-name]If the URI is unspecified, it will be inferred from SoftwareConfig.image_version or the system default.
diff --git a/docs/dyn/datastore_v1beta3.projects.html b/docs/dyn/datastore_v1beta3.projects.html
index 5bd9b3632f1..68130019c47 100644
--- a/docs/dyn/datastore_v1beta3.projects.html
+++ b/docs/dyn/datastore_v1beta3.projects.html
@@ -232,7 +232,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
"update": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # The entity to update. The entity must already exist. Must have a complete key path.
@@ -250,7 +283,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
"upsert": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # The entity to upsert. The entity may or may not already exist. The entity key's final path element may be incomplete.
@@ -268,7 +334,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
},
@@ -381,7 +480,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
"version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads.
@@ -405,7 +537,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
"version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads.
@@ -503,24 +668,7 @@ Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
"booleanValue": True or False, # A boolean value.
"doubleValue": 3.14, # A double value.
- "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
- "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key.
- "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
- "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
- "projectId": "A String", # The ID of the project to which the entities belong.
- },
- "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
- { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
- "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
- "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- },
- ],
- },
- "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
- },
- },
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
"excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
"geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
@@ -559,24 +707,7 @@ Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
"booleanValue": True or False, # A boolean value.
"doubleValue": 3.14, # A double value.
- "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
- "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key.
- "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
- "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
- "projectId": "A String", # The ID of the project to which the entities belong.
- },
- "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
- { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
- "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
- "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- },
- ],
- },
- "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
- },
- },
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
"excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
"geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
@@ -637,24 +768,7 @@ Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
"booleanValue": True or False, # A boolean value.
"doubleValue": 3.14, # A double value.
- "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
- "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key.
- "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
- "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
- "projectId": "A String", # The ID of the project to which the entities belong.
- },
- "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
- { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
- "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
- "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- },
- ],
- },
- "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
- },
- },
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
"excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
"geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
@@ -741,7 +855,40 @@ Method Details
],
},
"properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
+ "a_key": { # A message that can hold any of the supported value types and associated metadata.
+ "arrayValue": { # An array value. # An array value. Cannot contain another array value. A `Value` instance that sets field `array_value` must not set fields `meaning` or `exclude_from_indexes`.
+ "values": [ # Values in the array. The order of values in an array is preserved as long as all values have identical settings for 'exclude_from_indexes'.
+ # Object with schema name: Value
+ ],
+ },
+ "blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
+ "booleanValue": True or False, # A boolean value.
+ "doubleValue": 3.14, # A double value.
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
+ "excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
+ "geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
+ "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
+ "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
+ },
+ "integerValue": "A String", # An integer value.
+ "keyValue": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # A key value.
+ "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
+ "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
+ "projectId": "A String", # The ID of the project to which the entities belong.
+ },
+ "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
+ { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
+ "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
+ "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
+ },
+ ],
+ },
+ "meaning": 42, # The `meaning` field should only be populated for backwards compatibility.
+ "nullValue": "A String", # A null value.
+ "stringValue": "A String", # A UTF-8 encoded string value. When `exclude_from_indexes` is false (it is indexed) , may have at most 1500 bytes. Otherwise, may be set to at most 1,000,000 bytes.
+ "timestampValue": "A String", # A timestamp value. When stored in the Datastore, precise only to microseconds; any additional precision is rounded down.
+ },
},
},
"version": "A String", # The version of the entity, a strictly positive number that monotonically increases with changes to the entity. This field is set for `FULL` entity results. For missing entities in `LookupResponse`, this is the version of the snapshot that was used to look up the entity, and it is always set except for eventually consistent reads.
@@ -780,24 +927,7 @@ Method Details
"blobValue": "A String", # A blob value. May have at most 1,000,000 bytes. When `exclude_from_indexes` is false, may have at most 1500 bytes. In JSON requests, must be base64-encoded.
"booleanValue": True or False, # A boolean value.
"doubleValue": 3.14, # A double value.
- "entityValue": { # A Datastore data object. An entity is limited to 1 megabyte when stored. That _roughly_ corresponds to a limit of 1 megabyte for the serialized form of this message. # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
- "key": { # A unique identifier for an entity. If a key's partition ID or any of its path kinds or names are reserved/read-only, the key is reserved/read-only. A reserved/read-only key is forbidden in certain documented contexts. # The entity's key. An entity must have a key, unless otherwise documented (for example, an entity in `Value.entity_value` may have no key). An entity's kind is its key path's last element's kind, or null if it has no key.
- "partitionId": { # A partition ID identifies a grouping of entities. The grouping is always by project and namespace, however the namespace ID may be empty. A partition ID contains several dimensions: project ID and namespace ID. Partition dimensions: - May be `""`. - Must be valid UTF-8 bytes. - Must have values that match regex `[A-Za-z\d\.\-_]{1,100}` If the value of any dimension matches regex `__.*__`, the partition is reserved/read-only. A reserved/read-only partition ID is forbidden in certain documented contexts. Foreign partition IDs (in which the project ID does not match the context project ID ) are discouraged. Reads and writes of foreign partition IDs may fail if the project is not in an active state. # Entities are partitioned into subsets, currently identified by a project ID and namespace ID. Queries are scoped to a single partition.
- "namespaceId": "A String", # If not empty, the ID of the namespace to which the entities belong.
- "projectId": "A String", # The ID of the project to which the entities belong.
- },
- "path": [ # The entity path. An entity path consists of one or more elements composed of a kind and a string or numerical identifier, which identify entities. The first element identifies a _root entity_, the second element identifies a _child_ of the root entity, the third element identifies a child of the second entity, and so forth. The entities identified by all prefixes of the path are called the element's _ancestors_. An entity path is always fully complete: *all* of the entity's ancestors are required to be in the path along with the entity identifier itself. The only exception is that in some documented cases, the identifier in the last path element (for the entity) itself may be omitted. For example, the last path element of the key of `Mutation.insert` may have no identifier. A path can never be empty, and a path can have at most 100 elements.
- { # A (kind, ID/name) pair used to construct a key path. If either name or ID is set, the element is complete. If neither is set, the element is incomplete.
- "id": "A String", # The auto-allocated ID of the entity. Never equal to zero. Values less than zero are discouraged and may not be supported in the future.
- "kind": "A String", # The kind of the entity. A kind matching regex `__.*__` is reserved/read-only. A kind must not contain more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- "name": "A String", # The name of the entity. A name matching regex `__.*__` is reserved/read-only. A name must not be more than 1500 bytes when UTF-8 encoded. Cannot be `""`.
- },
- ],
- },
- "properties": { # The entity's properties. The map's keys are property names. A property name matching regex `__.*__` is reserved. A reserved property name is forbidden in certain documented contexts. The name must not contain more than 500 characters. The name cannot be `""`.
- "a_key": # Object with schema name: Value
- },
- },
+ "entityValue": # Object with schema name: Entity # An entity value. - May have no key. - May have a key with an incomplete key path. - May have a reserved/read-only key.
"excludeFromIndexes": True or False, # If the value should be excluded from all indexes including those defined explicitly.
"geoPointValue": { # An object that represents a latitude/longitude pair. This is expressed as a pair of doubles to represent degrees latitude and degrees longitude. Unless specified otherwise, this object must conform to the WGS84 standard. Values must be within normalized ranges. # A geo point value representing a point on the surface of Earth.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
diff --git a/docs/dyn/dfareporting_v3_3.conversions.html b/docs/dyn/dfareporting_v3_3.conversions.html
index 6eb22b5f4d8..2c5757e1ae5 100644
--- a/docs/dyn/dfareporting_v3_3.conversions.html
+++ b/docs/dyn/dfareporting_v3_3.conversions.html
@@ -97,8 +97,8 @@ Method Details
"conversions": [ # The set of conversions to insert.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -147,8 +147,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -198,8 +198,8 @@ Method Details
"conversions": [ # The set of conversions to update.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -248,8 +248,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
diff --git a/docs/dyn/dfareporting_v3_3.placements.html b/docs/dyn/dfareporting_v3_3.placements.html
index 3f1861e7d53..d544e1f4e74 100644
--- a/docs/dyn/dfareporting_v3_3.placements.html
+++ b/docs/dyn/dfareporting_v3_3.placements.html
@@ -133,6 +133,7 @@ Method Details
PLACEMENT_TAG_INTERSTITIAL_IFRAME_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INTERSTITIAL_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INSTREAM_VIDEO_PREFETCH_VAST_4 -
+ PLACEMENT_TAG_TRACKING_THIRD_PARTY_MEASUREMENT -
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
diff --git a/docs/dyn/dfareporting_v3_4.conversions.html b/docs/dyn/dfareporting_v3_4.conversions.html
index 87c8a6c64e3..e72d682ce5f 100644
--- a/docs/dyn/dfareporting_v3_4.conversions.html
+++ b/docs/dyn/dfareporting_v3_4.conversions.html
@@ -97,8 +97,8 @@ Method Details
"conversions": [ # The set of conversions to insert.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -148,8 +148,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -200,8 +200,8 @@ Method Details
"conversions": [ # The set of conversions to update.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -251,8 +251,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
diff --git a/docs/dyn/dfareporting_v3_4.placements.html b/docs/dyn/dfareporting_v3_4.placements.html
index fe308a46902..e1ce9c867b8 100644
--- a/docs/dyn/dfareporting_v3_4.placements.html
+++ b/docs/dyn/dfareporting_v3_4.placements.html
@@ -133,6 +133,7 @@ Method Details
PLACEMENT_TAG_INTERSTITIAL_IFRAME_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INTERSTITIAL_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INSTREAM_VIDEO_PREFETCH_VAST_4 -
+ PLACEMENT_TAG_TRACKING_THIRD_PARTY_MEASUREMENT -
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
diff --git a/docs/dyn/dfareporting_v3_5.conversions.html b/docs/dyn/dfareporting_v3_5.conversions.html
index f099a2b5c19..05e11beb3e4 100644
--- a/docs/dyn/dfareporting_v3_5.conversions.html
+++ b/docs/dyn/dfareporting_v3_5.conversions.html
@@ -97,8 +97,8 @@ Method Details
"conversions": [ # The set of conversions to insert.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -148,8 +148,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -200,8 +200,8 @@ Method Details
"conversions": [ # The set of conversions to update.
{ # A Conversion represents when a user successfully performs a desired action after seeing an ad.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
@@ -251,8 +251,8 @@ Method Details
{ # The original conversion that was inserted or updated and whether there were any errors.
"conversion": { # A Conversion represents when a user successfully performs a desired action after seeing an ad. # The original conversion that was inserted or updated.
"childDirectedTreatment": True or False, # Whether this particular request may come from a user under the age of 13, under COPPA compliance.
- "customVariables": [ # Custom floodlight variables.
- { # A custom floodlight variable.
+ "customVariables": [ # Custom floodlight variables. This field may only be used when calling batchinsert; it is not supported by batchupdate.
+ { # A custom floodlight variable. This field may only be used when calling batchinsert; it is not supported by batchupdate.
"kind": "A String", # Identifies what kind of resource this is. Value: the fixed string "dfareporting#customFloodlightVariable".
"type": "A String", # The type of custom floodlight variable to supply a value for. These map to the "u[1-20]=" in the tags.
"value": "A String", # The value of the custom floodlight variable. The length of string must not exceed 100 characters.
diff --git a/docs/dyn/dfareporting_v3_5.placements.html b/docs/dyn/dfareporting_v3_5.placements.html
index 1a99f62cf70..ff56f352619 100644
--- a/docs/dyn/dfareporting_v3_5.placements.html
+++ b/docs/dyn/dfareporting_v3_5.placements.html
@@ -133,6 +133,7 @@ Method Details
PLACEMENT_TAG_INTERSTITIAL_IFRAME_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INTERSTITIAL_JAVASCRIPT_LEGACY -
PLACEMENT_TAG_INSTREAM_VIDEO_PREFETCH_VAST_4 -
+ PLACEMENT_TAG_TRACKING_THIRD_PARTY_MEASUREMENT -
x__xgafv: string, V1 error format.
Allowed values
1 - v1 error format
diff --git a/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.documents.html b/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.documents.html
index 674e0fd5374..e2a2352eab9 100644
--- a/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.documents.html
+++ b/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.documents.html
@@ -87,7 +87,7 @@ Instance Methods
get(name, x__xgafv=None)
Retrieves the specified document.
- list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Returns the list of all documents of the knowledge base.
list_next(previous_request, previous_response)
@@ -247,11 +247,12 @@ Method Details
- list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Returns the list of all documents of the knowledge base.
Args:
parent: string, Required. The knowledge base to list all documents for. Format: `projects//locations//knowledgeBases/`. (required)
+ filter: string, The filter expression used to filter documents returned by the list method. The expression has the following syntax: [AND ] ... The following fields and operators are supported: * knowledge_types with has(:) operator * display_name with has(:) operator * state with equals(=) operator Examples: * "knowledge_types:FAQ" matches documents with FAQ knowledge type. * "display_name:customer" matches documents whose display name contains "customer". * "state=ACTIVE" matches documents with ACTIVE state. * "knowledge_types:FAQ AND state=ACTIVE" matches all active FAQ documents. For more information about filtering, see [API Filtering](https://aip.dev/160).
pageSize: integer, The maximum number of items to return in a single page. By default 10 and at most 100.
pageToken: string, The next_page_token value returned from a previous list request.
x__xgafv: string, V1 error format.
@@ -386,6 +387,8 @@ Method Details
{ # Request message for Documents.ReloadDocument.
"contentUri": "A String", # Optional. The path of gcs source file for reloading document content. For now, only gcs uri is supported. For documents stored in Google Cloud Storage, these URIs must have the form `gs:///`.
+ "importGcsCustomMetadata": True or False, # Optional. Whether to import custom metadata from Google Cloud Storage. Only valid when the document source is Google Cloud Storage URI.
+ "smartMessagingPartialUpdate": True or False, # Optional. When enabled, the reload request is to apply partial update to the smart messaging allowlist.
}
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.html b/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.html
index c80cf437f4d..185bfea4a9b 100644
--- a/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.html
+++ b/docs/dyn/dialogflow_v2.projects.agent.knowledgeBases.html
@@ -92,7 +92,7 @@ Instance Methods
get(name, x__xgafv=None)
Retrieves the specified knowledge base.
- list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Returns the list of all knowledge bases of the specified agent.
list_next(previous_request, previous_response)
@@ -177,11 +177,12 @@ Method Details
- list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+ list(parent, filter=None, pageSize=None, pageToken=None, x__xgafv=None)
Returns the list of all knowledge bases of the specified agent.
Args:
parent: string, Required. The project to list of knowledge bases for. Format: `projects//locations/`. (required)
+ filter: string, The filter expression used to filter knowledge bases returned by the list method. The expression has the following syntax: [AND ] ... The following fields and operators are supported: * display_name with has(:) operator * language_code with equals(=) operator Examples: * 'language_code=en-us' matches knowledge bases with en-us language code. * 'display_name:articles' matches knowledge bases whose display name contains "articles". * 'display_name:"Best Articles"' matches knowledge bases whose display name contains "Best Articles". * 'language_code=en-gb AND display_name=articles' matches all knowledge bases whose display name contains "articles" and whose language code is "en-gb". Note: An empty filter string (i.e. "") is a no-op and will result in no filtering. For more information about filtering, see [API Filtering](https://aip.dev/160).
pageSize: integer, The maximum number of items to return in a single page. By default 10 and at most 100.
pageToken: string, The next_page_token value returned from a previous list request.
x__xgafv: string, V1 error format.
diff --git a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html
index 7ddd9546dd9..bcfb9a1f63c 100644
--- a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html
+++ b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html
@@ -123,6 +123,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -163,6 +166,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -259,6 +265,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -299,6 +308,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -420,6 +432,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -460,6 +475,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -567,6 +585,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -607,6 +628,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -722,6 +746,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -762,6 +789,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -859,6 +889,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
@@ -899,6 +932,9 @@ Method Details
"conversationModelConfig": { # Custom conversation models used in agent assist feature. Supported feature: ARTICLE_SUGGESTION, SMART_COMPOSE, SMART_REPLY. # Configs of custom conversation model.
"model": "A String", # Conversation model resource name. Format: `projects//conversationModels/`.
},
+ "conversationProcessConfig": { # Config to process conversation. # Configs for processing conversation.
+ "recentSentencesCount": 42, # Number of recent non-small-talk sentences to use as context for article and FAQ suggestion
+ },
"enableEventBasedSuggestion": True or False, # Automatically iterates all participants and tries to compile suggestions. Supported features: ARTICLE_SUGGESTION, FAQ, DIALOGFLOW_ASSIST.
"queryConfig": { # Config for suggestion query. # Configs of query.
"confidenceThreshold": 3.14, # Confidence threshold of query result. Agent Assist gives each suggestion a score in the range [0.0, 1.0], based on the relevance between the suggestion and the current conversation context. A score of 0.0 has no relevance, while a score of 1.0 has high relevance. Only suggestions with a score greater than or equal to the value of this field are included in the results. For a baseline model (the default), the recommended value is in the range [0.05, 0.1]. For a custom model, there is no recommended value. Tune this value by starting from a very low value and slowly increasing until you have desired results. If this field is not set, it defaults to 0.0, which means that all suggestions are returned. Supported features: ARTICLE_SUGGESTION, FAQ, SMART_REPLY, SMART_COMPOSE.
diff --git a/docs/dyn/dialogflow_v2.projects.conversations.messages.html b/docs/dyn/dialogflow_v2.projects.conversations.messages.html
index 3a469f86bc8..945e55436c5 100644
--- a/docs/dyn/dialogflow_v2.projects.conversations.messages.html
+++ b/docs/dyn/dialogflow_v2.projects.conversations.messages.html
@@ -110,7 +110,7 @@ Method Details
"messages": [ # The list of messages. There will be a maximum number of items returned based on the page_size field in the request. `messages` is sorted by `create_time` in descending order.
{ # Represents a message posted into a conversation.
"content": "A String", # Required. The message content.
- "createTime": "A String", # Output only. The time when the message was created.
+ "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
"languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
"messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
"containEntities": True or False, # Indicates whether the text message contains entities.
@@ -122,9 +122,16 @@ Method Details
},
],
},
- "name": "A String", # The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+ "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
"participant": "A String", # Output only. The participant that sends this message.
"participantRole": "A String", # Output only. The role of the participant.
+ "sendTime": "A String", # Optional. The time when the message was sent.
+ "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+ "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. # The sentiment analysis result for `query_text`.
+ "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+ "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+ },
+ },
},
],
"nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list.
diff --git a/docs/dyn/dialogflow_v2.projects.conversations.participants.html b/docs/dyn/dialogflow_v2.projects.conversations.participants.html
index 9b473b8b442..f15310af1e5 100644
--- a/docs/dyn/dialogflow_v2.projects.conversations.participants.html
+++ b/docs/dyn/dialogflow_v2.projects.conversations.participants.html
@@ -732,6 +732,17 @@ Method Details
],
"latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`.
},
+ "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY.
+ "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation.
+ "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`.
+ "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig.
+ { # Represents a smart reply answer.
+ "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/"
+ "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain).
+ "reply": "A String", # The content of the reply.
+ },
+ ],
+ },
},
],
"humanAgentSuggestionResults": [ # The suggestions for most recent human agent. The order is the same as HumanAgentAssistantConfig.SuggestionConfig.feature_configs of HumanAgentAssistantConfig.human_agent_suggestion_config.
@@ -779,11 +790,22 @@ Method Details
],
"latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`.
},
+ "suggestSmartRepliesResponse": { # The response message for Participants.SuggestSmartReplies. # SuggestSmartRepliesResponse if request is for SMART_REPLY.
+ "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation.
+ "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`.
+ "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig.
+ { # Represents a smart reply answer.
+ "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/"
+ "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain).
+ "reply": "A String", # The content of the reply.
+ },
+ ],
+ },
},
],
"message": { # Represents a message posted into a conversation. # Message analyzed by CCAI.
"content": "A String", # Required. The message content.
- "createTime": "A String", # Output only. The time when the message was created.
+ "createTime": "A String", # Output only. The time when the message was created in Contact Center AI.
"languageCode": "A String", # Optional. The message language. This should be a [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US".
"messageAnnotation": { # Represents the result of annotation for the message. # Output only. The annotation for the message.
"containEntities": True or False, # Indicates whether the text message contains entities.
@@ -795,9 +817,16 @@ Method Details
},
],
},
- "name": "A String", # The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
+ "name": "A String", # Optional. The unique identifier of the message. Format: `projects//locations//conversations//messages/`.
"participant": "A String", # Output only. The participant that sends this message.
"participantRole": "A String", # Output only. The role of the participant.
+ "sendTime": "A String", # Optional. The time when the message was sent.
+ "sentimentAnalysis": { # The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For Participants.DetectIntent, it needs to be configured in DetectIntentRequest.query_params. For Participants.StreamingDetectIntent, it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config # Output only. The sentiment analysis result for the message.
+ "queryTextSentiment": { # The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text. # The sentiment analysis result for `query_text`.
+ "magnitude": 3.14, # A non-negative number in the [0, +inf) range, which represents the absolute magnitude of sentiment, regardless of score (positive or negative).
+ "score": 3.14, # Sentiment score between -1.0 (negative sentiment) and 1.0 (positive sentiment).
+ },
+ },
},
"replyAudio": { # Represents the natural language speech audio to be played to the end user. # The audio data bytes encoded as specified in the request. This field is set if: - `reply_audio_config` was specified in the request, or - The automated agent responded with audio to play to the user. In such case, `reply_audio.config` contains settings used to synthesize the speech. In some scenarios, multiple output audio fields may be present in the response structure. In these cases, only the top-most-level audio output has content.
"audio": "A String", # The natural language speech audio.
diff --git a/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html b/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html
index e4d831092ec..be399ac9440 100644
--- a/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html
+++ b/docs/dyn/dialogflow_v2.projects.conversations.participants.suggestions.html
@@ -83,6 +83,9 @@ Instance Methods
suggestFaqAnswers(parent, body=None, x__xgafv=None)
Gets suggested faq answers for a participant based on specific historical messages.
+
+ suggestSmartReplies(parent, body=None, x__xgafv=None)
+Gets smart replies for a participant based on specific historical messages.
Method Details
close()
@@ -104,8 +107,8 @@ Method Details
"a_key": "A String",
},
},
- "contextSize": 42, # Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50.
- "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`.
+ "contextSize": 42, # Optional. Max number of messages prior to and including latest_message to use as context when compiling the suggestion. By default 20 and at most 50.
+ "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`.
}
x__xgafv: string, V1 error format.
@@ -151,8 +154,8 @@ Method Details
"a_key": "A String",
},
},
- "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50.
- "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`.
+ "contextSize": 42, # Optional. Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50.
+ "latestMessage": "A String", # Optional. The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`.
}
x__xgafv: string, V1 error format.
@@ -181,4 +184,43 @@ Method Details
}
+
+ suggestSmartReplies(parent, body=None, x__xgafv=None)
+ Gets smart replies for a participant based on specific historical messages.
+
+Args:
+ parent: string, Required. The name of the participant to fetch suggestion for. Format: `projects//locations//conversations//participants/`. (required)
+ body: object, The request body.
+ The object takes the form of:
+
+{ # The request message for Participants.SuggestSmartReplies.
+ "contextSize": 42, # Max number of messages prior to and including [latest_message] to use as context when compiling the suggestion. By default 20 and at most 50.
+ "currentTextInput": { # Represents the natural language text to be processed. # The current natural language text segment to compile suggestion for. This provides a way for user to get follow up smart reply suggestion after a smart reply selection, without sending a text message.
+ "languageCode": "A String", # Required. The language of this conversational query. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. Note that queries in the same session do not necessarily need to specify the same language.
+ "text": "A String", # Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters.
+ },
+ "latestMessage": "A String", # The name of the latest conversation message to compile suggestion for. If empty, it will be the latest message of the conversation. Format: `projects//locations//conversations//messages/`.
+}
+
+ x__xgafv: string, V1 error format.
+ Allowed values
+ 1 - v1 error format
+ 2 - v2 error format
+
+Returns:
+ An object of the form:
+
+ { # The response message for Participants.SuggestSmartReplies.
+ "contextSize": 42, # Number of messages prior to and including latest_message to compile the suggestion. It may be smaller than the SuggestSmartRepliesRequest.context_size field in the request if there aren't that many messages in the conversation.
+ "latestMessage": "A String", # The name of the latest conversation message used to compile suggestion for. Format: `projects//locations//conversations//messages/`.
+ "smartReplyAnswers": [ # Output only. Multiple reply options provided by smart reply service. The order is based on the rank of the model prediction. The maximum number of the returned replies is set in SmartReplyConfig.
+ { # Represents a smart reply answer.
+ "answerRecord": "A String", # The name of answer record, in the format of "projects//locations//answerRecords/"
+ "confidence": 3.14, # Smart reply confidence. The system's confidence score that this reply is a good match for this conversation, as a value from 0.0 (completely uncertain) to 1.0 (completely certain).
+ "reply": "A String", # The content of the reply.
+ },
+ ],
+}
+
+